~bzr-pqm/bzr/bzr.dev

« back to all changes in this revision

Viewing changes to bzrlib/tests/per_versionedfile.py

  • Committer: Patch Queue Manager
  • Date: 2011-10-14 16:54:26 UTC
  • mfrom: (6216.1.1 remove-this-file)
  • Revision ID: pqm@pqm.ubuntu.com-20111014165426-tjix4e6idryf1r2z
(jelmer) Remove an accidentally committed .THIS file. (Jelmer Vernooij)

Show diffs side-by-side

added added

removed removed

Lines of Context:
1
 
# Copyright (C) 2005, 2009 Canonical Ltd
 
1
# Copyright (C) 2006-2011 Canonical Ltd
2
2
#
3
3
# Authors:
4
4
#   Johan Rydberg <jrydberg@gnu.org>
21
21
# TODO: might be nice to create a versionedfile with some type of corruption
22
22
# considered typical and check that it can be detected/corrected.
23
23
 
 
24
from gzip import GzipFile
24
25
from itertools import chain, izip
25
26
from StringIO import StringIO
26
27
 
27
28
from bzrlib import (
28
29
    errors,
 
30
    graph as _mod_graph,
29
31
    groupcompress,
30
32
    knit as _mod_knit,
31
33
    osutils,
32
34
    progress,
 
35
    transport,
 
36
    ui,
33
37
    )
34
38
from bzrlib.errors import (
35
39
                           RevisionNotPresent,
36
40
                           RevisionAlreadyPresent,
37
 
                           WeaveParentMismatch
38
41
                           )
39
42
from bzrlib.knit import (
40
43
    cleanup_pack_knit,
41
44
    make_file_factory,
42
45
    make_pack_factory,
43
 
    KnitAnnotateFactory,
44
 
    KnitPlainFactory,
45
46
    )
46
47
from bzrlib.tests import (
47
48
    TestCase,
48
49
    TestCaseWithMemoryTransport,
49
50
    TestNotApplicable,
50
51
    TestSkipped,
51
 
    condition_isinstance,
52
 
    split_suite_by_condition,
53
 
    multiply_tests,
54
52
    )
55
53
from bzrlib.tests.http_utils import TestCaseWithWebserver
56
 
from bzrlib.trace import mutter
57
 
from bzrlib.transport import get_transport
58
54
from bzrlib.transport.memory import MemoryTransport
59
 
from bzrlib.tsort import topo_sort
60
 
from bzrlib.tuned_gzip import GzipFile
61
55
import bzrlib.versionedfile as versionedfile
62
56
from bzrlib.versionedfile import (
63
57
    ConstantMapper,
67
61
    make_versioned_files_factory,
68
62
    )
69
63
from bzrlib.weave import WeaveFile
70
 
from bzrlib.weavefile import read_weave, write_weave
71
 
 
72
 
 
73
 
def load_tests(standard_tests, module, loader):
74
 
    """Parameterize VersionedFiles tests for different implementations."""
75
 
    to_adapt, result = split_suite_by_condition(
76
 
        standard_tests, condition_isinstance(TestVersionedFiles))
77
 
    # We want to be sure of behaviour for:
78
 
    # weaves prefix layout (weave texts)
79
 
    # individually named weaves (weave inventories)
80
 
    # annotated knits - prefix|hash|hash-escape layout, we test the third only
81
 
    #                   as it is the most complex mapper.
82
 
    # individually named knits
83
 
    # individual no-graph knits in packs (signatures)
84
 
    # individual graph knits in packs (inventories)
85
 
    # individual graph nocompression knits in packs (revisions)
86
 
    # plain text knits in packs (texts)
87
 
    len_one_scenarios = [
88
 
        ('weave-named', {
89
 
            'cleanup':None,
90
 
            'factory':make_versioned_files_factory(WeaveFile,
91
 
                ConstantMapper('inventory')),
92
 
            'graph':True,
93
 
            'key_length':1,
94
 
            'support_partial_insertion': False,
95
 
            }),
96
 
        ('named-knit', {
97
 
            'cleanup':None,
98
 
            'factory':make_file_factory(False, ConstantMapper('revisions')),
99
 
            'graph':True,
100
 
            'key_length':1,
101
 
            'support_partial_insertion': False,
102
 
            }),
103
 
        ('named-nograph-nodelta-knit-pack', {
104
 
            'cleanup':cleanup_pack_knit,
105
 
            'factory':make_pack_factory(False, False, 1),
106
 
            'graph':False,
107
 
            'key_length':1,
108
 
            'support_partial_insertion': False,
109
 
            }),
110
 
        ('named-graph-knit-pack', {
111
 
            'cleanup':cleanup_pack_knit,
112
 
            'factory':make_pack_factory(True, True, 1),
113
 
            'graph':True,
114
 
            'key_length':1,
115
 
            'support_partial_insertion': True,
116
 
            }),
117
 
        ('named-graph-nodelta-knit-pack', {
118
 
            'cleanup':cleanup_pack_knit,
119
 
            'factory':make_pack_factory(True, False, 1),
120
 
            'graph':True,
121
 
            'key_length':1,
122
 
            'support_partial_insertion': False,
123
 
            }),
124
 
        ('groupcompress-nograph', {
125
 
            'cleanup':groupcompress.cleanup_pack_group,
126
 
            'factory':groupcompress.make_pack_factory(False, False, 1),
127
 
            'graph': False,
128
 
            'key_length':1,
129
 
            'support_partial_insertion':False,
130
 
            }),
131
 
        ]
132
 
    len_two_scenarios = [
133
 
        ('weave-prefix', {
134
 
            'cleanup':None,
135
 
            'factory':make_versioned_files_factory(WeaveFile,
136
 
                PrefixMapper()),
137
 
            'graph':True,
138
 
            'key_length':2,
139
 
            'support_partial_insertion': False,
140
 
            }),
141
 
        ('annotated-knit-escape', {
142
 
            'cleanup':None,
143
 
            'factory':make_file_factory(True, HashEscapedPrefixMapper()),
144
 
            'graph':True,
145
 
            'key_length':2,
146
 
            'support_partial_insertion': False,
147
 
            }),
148
 
        ('plain-knit-pack', {
149
 
            'cleanup':cleanup_pack_knit,
150
 
            'factory':make_pack_factory(True, True, 2),
151
 
            'graph':True,
152
 
            'key_length':2,
153
 
            'support_partial_insertion': True,
154
 
            }),
155
 
        ('groupcompress', {
156
 
            'cleanup':groupcompress.cleanup_pack_group,
157
 
            'factory':groupcompress.make_pack_factory(True, False, 1),
158
 
            'graph': True,
159
 
            'key_length':1,
160
 
            'support_partial_insertion':False,
161
 
            }),
162
 
        ]
163
 
    scenarios = len_one_scenarios + len_two_scenarios
164
 
    return multiply_tests(to_adapt, scenarios, result)
 
64
from bzrlib.weavefile import write_weave
 
65
from bzrlib.tests.scenarios import load_tests_apply_scenarios
 
66
 
 
67
 
 
68
load_tests = load_tests_apply_scenarios
165
69
 
166
70
 
167
71
def get_diamond_vf(f, trailing_eol=True, left_only=False):
732
636
        # the ordering here is to make a tree so that dumb searches have
733
637
        # more changes to muck up.
734
638
 
735
 
        class InstrumentedProgress(progress.DummyProgress):
 
639
        class InstrumentedProgress(progress.ProgressTask):
736
640
 
737
641
            def __init__(self):
738
 
 
739
 
                progress.DummyProgress.__init__(self)
 
642
                progress.ProgressTask.__init__(self)
740
643
                self.updates = []
741
644
 
742
645
            def update(self, msg=None, current=None, total=None):
848
751
        self.assertEquals(('references_ghost', 'line_c\n'), origins[2])
849
752
 
850
753
    def test_readonly_mode(self):
851
 
        transport = get_transport(self.get_url('.'))
 
754
        t = self.get_transport()
852
755
        factory = self.get_factory()
853
 
        vf = factory('id', transport, 0777, create=True, access_mode='w')
854
 
        vf = factory('id', transport, access_mode='r')
 
756
        vf = factory('id', t, 0777, create=True, access_mode='w')
 
757
        vf = factory('id', t, access_mode='r')
855
758
        self.assertRaises(errors.ReadOnlyError, vf.add_lines, 'base', [], [])
856
759
        self.assertRaises(errors.ReadOnlyError,
857
760
                          vf.add_lines_with_ghosts,
879
782
class TestWeave(TestCaseWithMemoryTransport, VersionedFileTestMixIn):
880
783
 
881
784
    def get_file(self, name='foo'):
882
 
        return WeaveFile(name, get_transport(self.get_url('.')), create=True,
883
 
            get_scope=self.get_transaction)
 
785
        return WeaveFile(name, self.get_transport(),
 
786
                         create=True,
 
787
                         get_scope=self.get_transaction)
884
788
 
885
789
    def get_file_corrupted_text(self):
886
 
        w = WeaveFile('foo', get_transport(self.get_url('.')), create=True,
887
 
            get_scope=self.get_transaction)
 
790
        w = WeaveFile('foo', self.get_transport(),
 
791
                      create=True,
 
792
                      get_scope=self.get_transaction)
888
793
        w.add_lines('v1', [], ['hello\n'])
889
794
        w.add_lines('v2', ['v1'], ['hello\n', 'there\n'])
890
795
 
918
823
        return w
919
824
 
920
825
    def reopen_file(self, name='foo', create=False):
921
 
        return WeaveFile(name, get_transport(self.get_url('.')), create=create,
922
 
            get_scope=self.get_transaction)
 
826
        return WeaveFile(name, self.get_transport(),
 
827
                         create=create,
 
828
                         get_scope=self.get_transaction)
923
829
 
924
830
    def test_no_implicit_create(self):
925
831
        self.assertRaises(errors.NoSuchFile,
926
832
                          WeaveFile,
927
833
                          'foo',
928
 
                          get_transport(self.get_url('.')),
 
834
                          self.get_transport(),
929
835
                          get_scope=self.get_transaction)
930
836
 
931
837
    def get_factory(self):
998
904
        # we should be able to read from http with a versioned file.
999
905
        vf = self.get_file()
1000
906
        # try an empty file access
1001
 
        readonly_vf = self.get_factory()('foo', get_transport(self.get_readonly_url('.')))
 
907
        readonly_vf = self.get_factory()('foo',
 
908
            transport.get_transport_from_url(self.get_readonly_url('.')))
1002
909
        self.assertEqual([], readonly_vf.versions())
 
910
 
 
911
    def test_readonly_http_works_with_feeling(self):
 
912
        # we should be able to read from http with a versioned file.
 
913
        vf = self.get_file()
1003
914
        # now with feeling.
1004
915
        vf.add_lines('1', [], ['a\n'])
1005
916
        vf.add_lines('2', ['1'], ['b\n', 'a\n'])
1006
 
        readonly_vf = self.get_factory()('foo', get_transport(self.get_readonly_url('.')))
 
917
        readonly_vf = self.get_factory()('foo',
 
918
            transport.get_transport_from_url(self.get_readonly_url('.')))
1007
919
        self.assertEqual(['1', '2'], vf.versions())
 
920
        self.assertEqual(['1', '2'], readonly_vf.versions())
1008
921
        for version in readonly_vf.versions():
1009
922
            readonly_vf.get_lines(version)
1010
923
 
1012
925
class TestWeaveHTTP(TestCaseWithWebserver, TestReadonlyHttpMixin):
1013
926
 
1014
927
    def get_file(self):
1015
 
        return WeaveFile('foo', get_transport(self.get_url('.')), create=True,
1016
 
            get_scope=self.get_transaction)
 
928
        return WeaveFile('foo', self.get_transport(),
 
929
                         create=True,
 
930
                         get_scope=self.get_transaction)
1017
931
 
1018
932
    def get_factory(self):
1019
933
        return WeaveFile
1263
1177
class TestWeaveMerge(TestCaseWithMemoryTransport, MergeCasesMixin):
1264
1178
 
1265
1179
    def get_file(self, name='foo'):
1266
 
        return WeaveFile(name, get_transport(self.get_url('.')), create=True)
 
1180
        return WeaveFile(name, self.get_transport(),
 
1181
                         create=True)
1267
1182
 
1268
1183
    def log_contents(self, w):
1269
1184
        self.log('weave is:')
1462
1377
class TestVersionedFiles(TestCaseWithMemoryTransport):
1463
1378
    """Tests for the multiple-file variant of VersionedFile."""
1464
1379
 
 
1380
    # We want to be sure of behaviour for:
 
1381
    # weaves prefix layout (weave texts)
 
1382
    # individually named weaves (weave inventories)
 
1383
    # annotated knits - prefix|hash|hash-escape layout, we test the third only
 
1384
    #                   as it is the most complex mapper.
 
1385
    # individually named knits
 
1386
    # individual no-graph knits in packs (signatures)
 
1387
    # individual graph knits in packs (inventories)
 
1388
    # individual graph nocompression knits in packs (revisions)
 
1389
    # plain text knits in packs (texts)
 
1390
    len_one_scenarios = [
 
1391
        ('weave-named', {
 
1392
            'cleanup':None,
 
1393
            'factory':make_versioned_files_factory(WeaveFile,
 
1394
                ConstantMapper('inventory')),
 
1395
            'graph':True,
 
1396
            'key_length':1,
 
1397
            'support_partial_insertion': False,
 
1398
            }),
 
1399
        ('named-knit', {
 
1400
            'cleanup':None,
 
1401
            'factory':make_file_factory(False, ConstantMapper('revisions')),
 
1402
            'graph':True,
 
1403
            'key_length':1,
 
1404
            'support_partial_insertion': False,
 
1405
            }),
 
1406
        ('named-nograph-nodelta-knit-pack', {
 
1407
            'cleanup':cleanup_pack_knit,
 
1408
            'factory':make_pack_factory(False, False, 1),
 
1409
            'graph':False,
 
1410
            'key_length':1,
 
1411
            'support_partial_insertion': False,
 
1412
            }),
 
1413
        ('named-graph-knit-pack', {
 
1414
            'cleanup':cleanup_pack_knit,
 
1415
            'factory':make_pack_factory(True, True, 1),
 
1416
            'graph':True,
 
1417
            'key_length':1,
 
1418
            'support_partial_insertion': True,
 
1419
            }),
 
1420
        ('named-graph-nodelta-knit-pack', {
 
1421
            'cleanup':cleanup_pack_knit,
 
1422
            'factory':make_pack_factory(True, False, 1),
 
1423
            'graph':True,
 
1424
            'key_length':1,
 
1425
            'support_partial_insertion': False,
 
1426
            }),
 
1427
        ('groupcompress-nograph', {
 
1428
            'cleanup':groupcompress.cleanup_pack_group,
 
1429
            'factory':groupcompress.make_pack_factory(False, False, 1),
 
1430
            'graph': False,
 
1431
            'key_length':1,
 
1432
            'support_partial_insertion':False,
 
1433
            }),
 
1434
        ]
 
1435
    len_two_scenarios = [
 
1436
        ('weave-prefix', {
 
1437
            'cleanup':None,
 
1438
            'factory':make_versioned_files_factory(WeaveFile,
 
1439
                PrefixMapper()),
 
1440
            'graph':True,
 
1441
            'key_length':2,
 
1442
            'support_partial_insertion': False,
 
1443
            }),
 
1444
        ('annotated-knit-escape', {
 
1445
            'cleanup':None,
 
1446
            'factory':make_file_factory(True, HashEscapedPrefixMapper()),
 
1447
            'graph':True,
 
1448
            'key_length':2,
 
1449
            'support_partial_insertion': False,
 
1450
            }),
 
1451
        ('plain-knit-pack', {
 
1452
            'cleanup':cleanup_pack_knit,
 
1453
            'factory':make_pack_factory(True, True, 2),
 
1454
            'graph':True,
 
1455
            'key_length':2,
 
1456
            'support_partial_insertion': True,
 
1457
            }),
 
1458
        ('groupcompress', {
 
1459
            'cleanup':groupcompress.cleanup_pack_group,
 
1460
            'factory':groupcompress.make_pack_factory(True, False, 1),
 
1461
            'graph': True,
 
1462
            'key_length':1,
 
1463
            'support_partial_insertion':False,
 
1464
            }),
 
1465
        ]
 
1466
 
 
1467
    scenarios = len_one_scenarios + len_two_scenarios
 
1468
 
1465
1469
    def get_versionedfiles(self, relpath='files'):
1466
1470
        transport = self.get_transport(relpath)
1467
1471
        if relpath != '.':
1468
1472
            transport.mkdir('.')
1469
1473
        files = self.factory(transport)
1470
1474
        if self.cleanup is not None:
1471
 
            self.addCleanup(lambda:self.cleanup(files))
 
1475
            self.addCleanup(self.cleanup, files)
1472
1476
        return files
1473
1477
 
1474
1478
    def get_simple_key(self, suffix):
1478
1482
        else:
1479
1483
            return ('FileA',) + (suffix,)
1480
1484
 
 
1485
    def test_add_fallback_implies_without_fallbacks(self):
 
1486
        f = self.get_versionedfiles('files')
 
1487
        if getattr(f, 'add_fallback_versioned_files', None) is None:
 
1488
            raise TestNotApplicable("%s doesn't support fallbacks"
 
1489
                                    % (f.__class__.__name__,))
 
1490
        g = self.get_versionedfiles('fallback')
 
1491
        key_a = self.get_simple_key('a')
 
1492
        g.add_lines(key_a, [], ['\n'])
 
1493
        f.add_fallback_versioned_files(g)
 
1494
        self.assertTrue(key_a in f.get_parent_map([key_a]))
 
1495
        self.assertFalse(key_a in f.without_fallbacks().get_parent_map([key_a]))
 
1496
 
1481
1497
    def test_add_lines(self):
1482
1498
        f = self.get_versionedfiles()
1483
1499
        key0 = self.get_simple_key('r0')
1557
1573
        self.assertRaises(RevisionNotPresent,
1558
1574
            files.annotate, prefix + ('missing-key',))
1559
1575
 
1560
 
    def test_get_annotator(self):
 
1576
    def test_check_no_parameters(self):
 
1577
        files = self.get_versionedfiles()
 
1578
 
 
1579
    def test_check_progressbar_parameter(self):
 
1580
        """A progress bar can be supplied because check can be a generator."""
 
1581
        pb = ui.ui_factory.nested_progress_bar()
 
1582
        self.addCleanup(pb.finished)
 
1583
        files = self.get_versionedfiles()
 
1584
        files.check(progress_bar=pb)
 
1585
 
 
1586
    def test_check_with_keys_becomes_generator(self):
1561
1587
        files = self.get_versionedfiles()
1562
1588
        self.get_diamond_files(files)
1563
 
        origin_key = self.get_simple_key('origin')
1564
 
        base_key = self.get_simple_key('base')
1565
 
        left_key = self.get_simple_key('left')
1566
 
        right_key = self.get_simple_key('right')
1567
 
        merged_key = self.get_simple_key('merged')
1568
 
        # annotator = files.get_annotator()
1569
 
        # introduced full text
1570
 
        origins, lines = files.get_annotator().annotate(origin_key)
1571
 
        self.assertEqual([(origin_key,)], origins)
1572
 
        self.assertEqual(['origin\n'], lines)
1573
 
        # a delta
1574
 
        origins, lines = files.get_annotator().annotate(base_key)
1575
 
        self.assertEqual([(base_key,)], origins)
1576
 
        # a merge
1577
 
        origins, lines = files.get_annotator().annotate(merged_key)
1578
 
        if self.graph:
1579
 
            self.assertEqual([
1580
 
                (base_key,),
1581
 
                (left_key,),
1582
 
                (right_key,),
1583
 
                (merged_key,),
1584
 
                ], origins)
1585
 
        else:
1586
 
            # Without a graph everything is new.
1587
 
            self.assertEqual([
1588
 
                (merged_key,),
1589
 
                (merged_key,),
1590
 
                (merged_key,),
1591
 
                (merged_key,),
1592
 
                ], origins)
1593
 
        self.assertRaises(RevisionNotPresent,
1594
 
            files.get_annotator().annotate, self.get_simple_key('missing-key'))
 
1589
        keys = files.keys()
 
1590
        entries = files.check(keys=keys)
 
1591
        seen = set()
 
1592
        # Texts output should be fulltexts.
 
1593
        self.capture_stream(files, entries, seen.add,
 
1594
            files.get_parent_map(keys), require_fulltext=True)
 
1595
        # All texts should be output.
 
1596
        self.assertEqual(set(keys), seen)
 
1597
 
 
1598
    def test_clear_cache(self):
 
1599
        files = self.get_versionedfiles()
 
1600
        files.clear_cache()
1595
1601
 
1596
1602
    def test_construct(self):
1597
1603
        """Each parameterised test can be constructed on a transport."""
1750
1756
            f.get_record_stream([key_b], 'unordered', True
1751
1757
                ).next().get_bytes_as('fulltext'))
1752
1758
 
 
1759
    def test_get_known_graph_ancestry(self):
 
1760
        f = self.get_versionedfiles()
 
1761
        if not self.graph:
 
1762
            raise TestNotApplicable('ancestry info only relevant with graph.')
 
1763
        key_a = self.get_simple_key('a')
 
1764
        key_b = self.get_simple_key('b')
 
1765
        key_c = self.get_simple_key('c')
 
1766
        # A
 
1767
        # |\
 
1768
        # | B
 
1769
        # |/
 
1770
        # C
 
1771
        f.add_lines(key_a, [], ['\n'])
 
1772
        f.add_lines(key_b, [key_a], ['\n'])
 
1773
        f.add_lines(key_c, [key_a, key_b], ['\n'])
 
1774
        kg = f.get_known_graph_ancestry([key_c])
 
1775
        self.assertIsInstance(kg, _mod_graph.KnownGraph)
 
1776
        self.assertEqual([key_a, key_b, key_c], list(kg.topo_sort()))
 
1777
 
 
1778
    def test_known_graph_with_fallbacks(self):
 
1779
        f = self.get_versionedfiles('files')
 
1780
        if not self.graph:
 
1781
            raise TestNotApplicable('ancestry info only relevant with graph.')
 
1782
        if getattr(f, 'add_fallback_versioned_files', None) is None:
 
1783
            raise TestNotApplicable("%s doesn't support fallbacks"
 
1784
                                    % (f.__class__.__name__,))
 
1785
        key_a = self.get_simple_key('a')
 
1786
        key_b = self.get_simple_key('b')
 
1787
        key_c = self.get_simple_key('c')
 
1788
        # A     only in fallback
 
1789
        # |\
 
1790
        # | B
 
1791
        # |/
 
1792
        # C
 
1793
        g = self.get_versionedfiles('fallback')
 
1794
        g.add_lines(key_a, [], ['\n'])
 
1795
        f.add_fallback_versioned_files(g)
 
1796
        f.add_lines(key_b, [key_a], ['\n'])
 
1797
        f.add_lines(key_c, [key_a, key_b], ['\n'])
 
1798
        kg = f.get_known_graph_ancestry([key_c])
 
1799
        self.assertEqual([key_a, key_b, key_c], list(kg.topo_sort()))
 
1800
 
1753
1801
    def test_get_record_stream_empty(self):
1754
1802
        """An empty stream can be requested without error."""
1755
1803
        f = self.get_versionedfiles()
1766
1814
             'knit-delta-closure', 'knit-delta-closure-ref',
1767
1815
             'groupcompress-block', 'groupcompress-block-ref'])
1768
1816
 
1769
 
    def capture_stream(self, f, entries, on_seen, parents):
 
1817
    def capture_stream(self, f, entries, on_seen, parents,
 
1818
        require_fulltext=False):
1770
1819
        """Capture a stream for testing."""
1771
1820
        for factory in entries:
1772
1821
            on_seen(factory.key)
1777
1826
            self.assertEqual(parents[factory.key], factory.parents)
1778
1827
            self.assertIsInstance(factory.get_bytes_as(factory.storage_kind),
1779
1828
                str)
 
1829
            if require_fulltext:
 
1830
                factory.get_bytes_as('fulltext')
1780
1831
 
1781
1832
    def test_get_record_stream_interface(self):
1782
1833
        """each item in a stream has to provide a regular interface."""
2157
2208
        else:
2158
2209
            return None
2159
2210
 
 
2211
    def test_get_annotator(self):
 
2212
        files = self.get_versionedfiles()
 
2213
        self.get_diamond_files(files)
 
2214
        origin_key = self.get_simple_key('origin')
 
2215
        base_key = self.get_simple_key('base')
 
2216
        left_key = self.get_simple_key('left')
 
2217
        right_key = self.get_simple_key('right')
 
2218
        merged_key = self.get_simple_key('merged')
 
2219
        # annotator = files.get_annotator()
 
2220
        # introduced full text
 
2221
        origins, lines = files.get_annotator().annotate(origin_key)
 
2222
        self.assertEqual([(origin_key,)], origins)
 
2223
        self.assertEqual(['origin\n'], lines)
 
2224
        # a delta
 
2225
        origins, lines = files.get_annotator().annotate(base_key)
 
2226
        self.assertEqual([(base_key,)], origins)
 
2227
        # a merge
 
2228
        origins, lines = files.get_annotator().annotate(merged_key)
 
2229
        if self.graph:
 
2230
            self.assertEqual([
 
2231
                (base_key,),
 
2232
                (left_key,),
 
2233
                (right_key,),
 
2234
                (merged_key,),
 
2235
                ], origins)
 
2236
        else:
 
2237
            # Without a graph everything is new.
 
2238
            self.assertEqual([
 
2239
                (merged_key,),
 
2240
                (merged_key,),
 
2241
                (merged_key,),
 
2242
                (merged_key,),
 
2243
                ], origins)
 
2244
        self.assertRaises(RevisionNotPresent,
 
2245
            files.get_annotator().annotate, self.get_simple_key('missing-key'))
 
2246
 
2160
2247
    def test_get_parent_map(self):
2161
2248
        files = self.get_versionedfiles()
2162
2249
        if self.key_length == 1:
2365
2452
        else:
2366
2453
            self.assertIdenticalVersionedFile(source, files)
2367
2454
 
 
2455
    def test_insert_record_stream_long_parent_chain_out_of_order(self):
 
2456
        """An out of order stream can either error or work."""
 
2457
        if not self.graph:
 
2458
            raise TestNotApplicable('ancestry info only relevant with graph.')
 
2459
        # Create a reasonably long chain of records based on each other, where
 
2460
        # most will be deltas.
 
2461
        source = self.get_versionedfiles('source')
 
2462
        parents = ()
 
2463
        keys = []
 
2464
        content = [('same same %d\n' % n) for n in range(500)]
 
2465
        for letter in 'abcdefghijklmnopqrstuvwxyz':
 
2466
            key = ('key-' + letter,)
 
2467
            if self.key_length == 2:
 
2468
                key = ('prefix',) + key
 
2469
            content.append('content for ' + letter + '\n')
 
2470
            source.add_lines(key, parents, content)
 
2471
            keys.append(key)
 
2472
            parents = (key,)
 
2473
        # Create a stream of these records, excluding the first record that the
 
2474
        # rest ultimately depend upon, and insert it into a new vf.
 
2475
        streams = []
 
2476
        for key in reversed(keys):
 
2477
            streams.append(source.get_record_stream([key], 'unordered', False))
 
2478
        deltas = chain(*streams[:-1])
 
2479
        files = self.get_versionedfiles()
 
2480
        try:
 
2481
            files.insert_record_stream(deltas)
 
2482
        except RevisionNotPresent:
 
2483
            # Must not have corrupted the file.
 
2484
            files.check()
 
2485
        else:
 
2486
            # Must only report either just the first key as a missing parent,
 
2487
            # no key as missing (for nodelta scenarios).
 
2488
            missing = set(files.get_missing_compression_parent_keys())
 
2489
            missing.discard(keys[0])
 
2490
            self.assertEqual(set(), missing)
 
2491
 
2368
2492
    def get_knit_delta_source(self):
2369
2493
        """Get a source that can produce a stream with knit delta records,
2370
2494
        regardless of this test's scenario.
2438
2562
        # the ordering here is to make a tree so that dumb searches have
2439
2563
        # more changes to muck up.
2440
2564
 
2441
 
        class InstrumentedProgress(progress.DummyProgress):
 
2565
        class InstrumentedProgress(progress.ProgressTask):
2442
2566
 
2443
2567
            def __init__(self):
2444
 
 
2445
 
                progress.DummyProgress.__init__(self)
 
2568
                progress.ProgressTask.__init__(self)
2446
2569
                self.updates = []
2447
2570
 
2448
2571
            def update(self, msg=None, current=None, total=None):
2637
2760
        self.assertRaises(NotImplementedError,
2638
2761
                self.texts.add_mpdiffs, [])
2639
2762
 
2640
 
    def test_check(self):
2641
 
        self.assertTrue(self.texts.check())
 
2763
    def test_check_noerrors(self):
 
2764
        self.texts.check()
2642
2765
 
2643
2766
    def test_insert_record_stream(self):
2644
2767
        self.assertRaises(NotImplementedError, self.texts.insert_record_stream,