~bzr-pqm/bzr/bzr.dev

« back to all changes in this revision

Viewing changes to bzrlib/tests/test_repository.py

  • Committer: Martin Pool
  • Date: 2009-01-13 03:11:04 UTC
  • mto: This revision was merged to the branch mainline in revision 3937.
  • Revision ID: mbp@sourcefrog.net-20090113031104-03my054s02i9l2pe
Bump version to 1.12 and add news template

Show diffs side-by-side

added added

removed removed

Lines of Context:
1
 
# Copyright (C) 2006-2010 Canonical Ltd
 
1
# Copyright (C) 2006, 2007, 2008 Canonical Ltd
2
2
#
3
3
# This program is free software; you can redistribute it and/or modify
4
4
# it under the terms of the GNU General Public License as published by
12
12
#
13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
 
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
16
16
 
17
17
"""Tests for the Repository facility that are not interface tests.
18
18
 
23
23
"""
24
24
 
25
25
from stat import S_ISDIR
26
 
import sys
 
26
from StringIO import StringIO
27
27
 
28
28
import bzrlib
29
 
from bzrlib.errors import (NoSuchFile,
 
29
from bzrlib.errors import (NotBranchError,
 
30
                           NoSuchFile,
30
31
                           UnknownFormatError,
31
32
                           UnsupportedFormatError,
32
33
                           )
33
 
from bzrlib import (
34
 
    graph,
35
 
    tests,
36
 
    )
 
34
from bzrlib import graph
37
35
from bzrlib.btree_index import BTreeBuilder, BTreeGraphIndex
38
 
from bzrlib.index import GraphIndex
 
36
from bzrlib.index import GraphIndex, InMemoryGraphIndex
39
37
from bzrlib.repository import RepositoryFormat
 
38
from bzrlib.smart import server
40
39
from bzrlib.tests import (
41
40
    TestCase,
42
41
    TestCaseWithTransport,
 
42
    TestSkipped,
 
43
    test_knit,
43
44
    )
44
45
from bzrlib.transport import (
 
46
    fakenfs,
45
47
    get_transport,
46
48
    )
 
49
from bzrlib.transport.memory import MemoryServer
 
50
from bzrlib.util import bencode
47
51
from bzrlib import (
48
52
    bzrdir,
49
53
    errors,
50
54
    inventory,
51
55
    osutils,
 
56
    progress,
52
57
    repository,
53
58
    revision as _mod_revision,
 
59
    symbol_versioning,
54
60
    upgrade,
55
 
    versionedfile,
56
61
    workingtree,
57
62
    )
58
 
from bzrlib.repofmt import (
59
 
    groupcompress_repo,
60
 
    knitrepo,
61
 
    pack_repo,
62
 
    weaverepo,
63
 
    )
 
63
from bzrlib.repofmt import knitrepo, weaverepo, pack_repo
64
64
 
65
65
 
66
66
class TestDefaultFormat(TestCase):
95
95
class SampleRepositoryFormat(repository.RepositoryFormat):
96
96
    """A sample format
97
97
 
98
 
    this format is initializable, unsupported to aid in testing the
 
98
    this format is initializable, unsupported to aid in testing the 
99
99
    open and open(unsupported=True) routines.
100
100
    """
101
101
 
122
122
    def test_find_format(self):
123
123
        # is the right format object found for a repository?
124
124
        # create a branch with a few known format objects.
125
 
        # this is not quite the same as
 
125
        # this is not quite the same as 
126
126
        self.build_tree(["foo/", "bar/"])
127
127
        def check_format(format, url):
128
128
            dir = format._matchingbzrdir.initialize(url)
131
131
            found_format = repository.RepositoryFormat.find_format(dir)
132
132
            self.failUnless(isinstance(found_format, format.__class__))
133
133
        check_format(weaverepo.RepositoryFormat7(), "bar")
134
 
 
 
134
        
135
135
    def test_find_format_no_repository(self):
136
136
        dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
137
137
        self.assertRaises(errors.NoRepositoryPresent,
167
167
        """Weaves need topological data insertion."""
168
168
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
169
169
        repo = weaverepo.RepositoryFormat6().initialize(control)
170
 
        self.assertEqual('topological', repo._format._fetch_order)
 
170
        self.assertEqual('topological', repo._fetch_order)
171
171
 
172
172
    def test_attribute__fetch_uses_deltas(self):
173
173
        """Weaves do not reuse deltas."""
174
174
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
175
175
        repo = weaverepo.RepositoryFormat6().initialize(control)
176
 
        self.assertEqual(False, repo._format._fetch_uses_deltas)
 
176
        self.assertEqual(False, repo._fetch_uses_deltas)
177
177
 
178
178
    def test_attribute__fetch_reconcile(self):
179
179
        """Weave repositories need a reconcile after fetch."""
180
180
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
181
181
        repo = weaverepo.RepositoryFormat6().initialize(control)
182
 
        self.assertEqual(True, repo._format._fetch_reconcile)
 
182
        self.assertEqual(True, repo._fetch_reconcile)
183
183
 
184
184
    def test_no_ancestry_weave(self):
185
185
        control = bzrdir.BzrDirFormat6().initialize(self.get_url())
202
202
        """Weaves need topological data insertion."""
203
203
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
204
204
        repo = weaverepo.RepositoryFormat7().initialize(control)
205
 
        self.assertEqual('topological', repo._format._fetch_order)
 
205
        self.assertEqual('topological', repo._fetch_order)
206
206
 
207
207
    def test_attribute__fetch_uses_deltas(self):
208
208
        """Weaves do not reuse deltas."""
209
209
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
210
210
        repo = weaverepo.RepositoryFormat7().initialize(control)
211
 
        self.assertEqual(False, repo._format._fetch_uses_deltas)
 
211
        self.assertEqual(False, repo._fetch_uses_deltas)
212
212
 
213
213
    def test_attribute__fetch_reconcile(self):
214
214
        """Weave repositories need a reconcile after fetch."""
215
215
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
216
216
        repo = weaverepo.RepositoryFormat7().initialize(control)
217
 
        self.assertEqual(True, repo._format._fetch_reconcile)
 
217
        self.assertEqual(True, repo._fetch_reconcile)
218
218
 
219
219
    def test_disk_layout(self):
220
220
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
243
243
        tree = control.create_workingtree()
244
244
        tree.add(['foo'], ['Foo:Bar'], ['file'])
245
245
        tree.put_file_bytes_non_atomic('Foo:Bar', 'content\n')
246
 
        try:
247
 
            tree.commit('first post', rev_id='first')
248
 
        except errors.IllegalPath:
249
 
            if sys.platform != 'win32':
250
 
                raise
251
 
            self.knownFailure('Foo:Bar cannot be used as a file-id on windows'
252
 
                              ' in repo format 7')
253
 
            return
 
246
        tree.commit('first post', rev_id='first')
254
247
        self.assertEqualDiff(
255
248
            '# bzr weave file v5\n'
256
249
            'i\n'
291
284
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
292
285
        repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
293
286
        t = control.get_repository_transport(None)
294
 
        # TODO: Should check there is a 'lock' toplevel directory,
 
287
        # TODO: Should check there is a 'lock' toplevel directory, 
295
288
        # regardless of contents
296
289
        self.assertFalse(t.has('lock/held/info'))
297
290
        repo.lock_write()
350
343
 
351
344
 
352
345
class TestFormatKnit1(TestCaseWithTransport):
353
 
 
 
346
    
354
347
    def test_attribute__fetch_order(self):
355
348
        """Knits need topological data insertion."""
356
349
        repo = self.make_repository('.',
357
350
                format=bzrdir.format_registry.get('knit')())
358
 
        self.assertEqual('topological', repo._format._fetch_order)
 
351
        self.assertEqual('topological', repo._fetch_order)
359
352
 
360
353
    def test_attribute__fetch_uses_deltas(self):
361
354
        """Knits reuse deltas."""
362
355
        repo = self.make_repository('.',
363
356
                format=bzrdir.format_registry.get('knit')())
364
 
        self.assertEqual(True, repo._format._fetch_uses_deltas)
 
357
        self.assertEqual(True, repo._fetch_uses_deltas)
365
358
 
366
359
    def test_disk_layout(self):
367
360
        control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
454
447
        repo = self.make_repository('.',
455
448
                format=bzrdir.format_registry.get('knit')())
456
449
        inv_xml = '<inventory format="5">\n</inventory>\n'
457
 
        inv = repo._deserialise_inventory('test-rev-id', inv_xml)
 
450
        inv = repo.deserialise_inventory('test-rev-id', inv_xml)
458
451
        self.assertEqual('test-rev-id', inv.root.revision)
459
452
 
460
453
    def test_deserialise_uses_global_revision_id(self):
466
459
        # Arguably, the deserialise_inventory should detect a mismatch, and
467
460
        # raise an error, rather than silently using one revision_id over the
468
461
        # other.
469
 
        self.assertRaises(AssertionError, repo._deserialise_inventory,
 
462
        self.assertRaises(AssertionError, repo.deserialise_inventory,
470
463
            'test-rev-id', inv_xml)
471
 
        inv = repo._deserialise_inventory('other-rev-id', inv_xml)
 
464
        inv = repo.deserialise_inventory('other-rev-id', inv_xml)
472
465
        self.assertEqual('other-rev-id', inv.root.revision)
473
466
 
474
467
    def test_supports_external_lookups(self):
484
477
    _serializer = None
485
478
 
486
479
    def supports_rich_root(self):
487
 
        if self._format is not None:
488
 
            return self._format.rich_root_data
489
480
        return False
490
481
 
491
482
    def get_graph(self):
506
497
    @staticmethod
507
498
    def is_compatible(repo_source, repo_target):
508
499
        """InterDummy is compatible with DummyRepository."""
509
 
        return (isinstance(repo_source, DummyRepository) and
 
500
        return (isinstance(repo_source, DummyRepository) and 
510
501
            isinstance(repo_target, DummyRepository))
511
502
 
512
503
 
525
516
 
526
517
    def assertGetsDefaultInterRepository(self, repo_a, repo_b):
527
518
        """Asserts that InterRepository.get(repo_a, repo_b) -> the default.
528
 
 
 
519
        
529
520
        The effective default is now InterSameDataRepository because there is
530
521
        no actual sane default in the presence of incompatible data models.
531
522
        """
542
533
        # pair that it returns true on for the is_compatible static method
543
534
        # check
544
535
        dummy_a = DummyRepository()
545
 
        dummy_a._format = RepositoryFormat()
546
536
        dummy_b = DummyRepository()
547
 
        dummy_b._format = RepositoryFormat()
548
537
        repo = self.make_repository('.')
549
538
        # hack dummies to look like repo somewhat.
550
539
        dummy_a._serializer = repo._serializer
551
 
        dummy_a._format.supports_tree_reference = repo._format.supports_tree_reference
552
 
        dummy_a._format.rich_root_data = repo._format.rich_root_data
553
540
        dummy_b._serializer = repo._serializer
554
 
        dummy_b._format.supports_tree_reference = repo._format.supports_tree_reference
555
 
        dummy_b._format.rich_root_data = repo._format.rich_root_data
556
541
        repository.InterRepository.register_optimiser(InterDummy)
557
542
        try:
558
543
            # we should get the default for something InterDummy returns False
621
606
 
622
607
 
623
608
class TestMisc(TestCase):
624
 
 
 
609
    
625
610
    def test_unescape_xml(self):
626
611
        """We get some kind of error when malformed entities are passed"""
627
 
        self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;')
 
612
        self.assertRaises(KeyError, repository._unescape_xml, 'foo&bar;') 
628
613
 
629
614
 
630
615
class TestRepositoryFormatKnit3(TestCaseWithTransport):
634
619
        format = bzrdir.BzrDirMetaFormat1()
635
620
        format.repository_format = knitrepo.RepositoryFormatKnit3()
636
621
        repo = self.make_repository('.', format=format)
637
 
        self.assertEqual('topological', repo._format._fetch_order)
 
622
        self.assertEqual('topological', repo._fetch_order)
638
623
 
639
624
    def test_attribute__fetch_uses_deltas(self):
640
625
        """Knits reuse deltas."""
641
626
        format = bzrdir.BzrDirMetaFormat1()
642
627
        format.repository_format = knitrepo.RepositoryFormatKnit3()
643
628
        repo = self.make_repository('.', format=format)
644
 
        self.assertEqual(True, repo._format._fetch_uses_deltas)
 
629
        self.assertEqual(True, repo._fetch_uses_deltas)
645
630
 
646
631
    def test_convert(self):
647
632
        """Ensure the upgrade adds weaves for roots"""
679
664
        self.assertFalse(repo._format.supports_external_lookups)
680
665
 
681
666
 
682
 
class Test2a(tests.TestCaseWithMemoryTransport):
683
 
 
684
 
    def test_fetch_combines_groups(self):
685
 
        builder = self.make_branch_builder('source', format='2a')
686
 
        builder.start_series()
687
 
        builder.build_snapshot('1', None, [
688
 
            ('add', ('', 'root-id', 'directory', '')),
689
 
            ('add', ('file', 'file-id', 'file', 'content\n'))])
690
 
        builder.build_snapshot('2', ['1'], [
691
 
            ('modify', ('file-id', 'content-2\n'))])
692
 
        builder.finish_series()
693
 
        source = builder.get_branch()
694
 
        target = self.make_repository('target', format='2a')
695
 
        target.fetch(source.repository)
696
 
        target.lock_read()
697
 
        self.addCleanup(target.unlock)
698
 
        details = target.texts._index.get_build_details(
699
 
            [('file-id', '1',), ('file-id', '2',)])
700
 
        file_1_details = details[('file-id', '1')]
701
 
        file_2_details = details[('file-id', '2')]
702
 
        # The index, and what to read off disk, should be the same for both
703
 
        # versions of the file.
704
 
        self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
705
 
 
706
 
    def test_fetch_combines_groups(self):
707
 
        builder = self.make_branch_builder('source', format='2a')
708
 
        builder.start_series()
709
 
        builder.build_snapshot('1', None, [
710
 
            ('add', ('', 'root-id', 'directory', '')),
711
 
            ('add', ('file', 'file-id', 'file', 'content\n'))])
712
 
        builder.build_snapshot('2', ['1'], [
713
 
            ('modify', ('file-id', 'content-2\n'))])
714
 
        builder.finish_series()
715
 
        source = builder.get_branch()
716
 
        target = self.make_repository('target', format='2a')
717
 
        target.fetch(source.repository)
718
 
        target.lock_read()
719
 
        self.addCleanup(target.unlock)
720
 
        details = target.texts._index.get_build_details(
721
 
            [('file-id', '1',), ('file-id', '2',)])
722
 
        file_1_details = details[('file-id', '1')]
723
 
        file_2_details = details[('file-id', '2')]
724
 
        # The index, and what to read off disk, should be the same for both
725
 
        # versions of the file.
726
 
        self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
727
 
 
728
 
    def test_fetch_combines_groups(self):
729
 
        builder = self.make_branch_builder('source', format='2a')
730
 
        builder.start_series()
731
 
        builder.build_snapshot('1', None, [
732
 
            ('add', ('', 'root-id', 'directory', '')),
733
 
            ('add', ('file', 'file-id', 'file', 'content\n'))])
734
 
        builder.build_snapshot('2', ['1'], [
735
 
            ('modify', ('file-id', 'content-2\n'))])
736
 
        builder.finish_series()
737
 
        source = builder.get_branch()
738
 
        target = self.make_repository('target', format='2a')
739
 
        target.fetch(source.repository)
740
 
        target.lock_read()
741
 
        self.addCleanup(target.unlock)
742
 
        details = target.texts._index.get_build_details(
743
 
            [('file-id', '1',), ('file-id', '2',)])
744
 
        file_1_details = details[('file-id', '1')]
745
 
        file_2_details = details[('file-id', '2')]
746
 
        # The index, and what to read off disk, should be the same for both
747
 
        # versions of the file.
748
 
        self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
749
 
 
750
 
    def test_format_pack_compresses_True(self):
751
 
        repo = self.make_repository('repo', format='2a')
752
 
        self.assertTrue(repo._format.pack_compresses)
753
 
 
754
 
    def test_inventories_use_chk_map_with_parent_base_dict(self):
755
 
        tree = self.make_branch_and_memory_tree('repo', format="2a")
756
 
        tree.lock_write()
757
 
        tree.add([''], ['TREE_ROOT'])
758
 
        revid = tree.commit("foo")
759
 
        tree.unlock()
760
 
        tree.lock_read()
761
 
        self.addCleanup(tree.unlock)
762
 
        inv = tree.branch.repository.get_inventory(revid)
763
 
        self.assertNotEqual(None, inv.parent_id_basename_to_file_id)
764
 
        inv.parent_id_basename_to_file_id._ensure_root()
765
 
        inv.id_to_entry._ensure_root()
766
 
        self.assertEqual(65536, inv.id_to_entry._root_node.maximum_size)
767
 
        self.assertEqual(65536,
768
 
            inv.parent_id_basename_to_file_id._root_node.maximum_size)
769
 
 
770
 
    def test_autopack_unchanged_chk_nodes(self):
771
 
        # at 20 unchanged commits, chk pages are packed that are split into
772
 
        # two groups such that the new pack being made doesn't have all its
773
 
        # pages in the source packs (though they are in the repository).
774
 
        # Use a memory backed repository, we don't need to hit disk for this
775
 
        tree = self.make_branch_and_memory_tree('tree', format='2a')
776
 
        tree.lock_write()
777
 
        self.addCleanup(tree.unlock)
778
 
        tree.add([''], ['TREE_ROOT'])
779
 
        for pos in range(20):
780
 
            tree.commit(str(pos))
781
 
 
782
 
    def test_pack_with_hint(self):
783
 
        tree = self.make_branch_and_memory_tree('tree', format='2a')
784
 
        tree.lock_write()
785
 
        self.addCleanup(tree.unlock)
786
 
        tree.add([''], ['TREE_ROOT'])
787
 
        # 1 commit to leave untouched
788
 
        tree.commit('1')
789
 
        to_keep = tree.branch.repository._pack_collection.names()
790
 
        # 2 to combine
791
 
        tree.commit('2')
792
 
        tree.commit('3')
793
 
        all = tree.branch.repository._pack_collection.names()
794
 
        combine = list(set(all) - set(to_keep))
795
 
        self.assertLength(3, all)
796
 
        self.assertLength(2, combine)
797
 
        tree.branch.repository.pack(hint=combine)
798
 
        final = tree.branch.repository._pack_collection.names()
799
 
        self.assertLength(2, final)
800
 
        self.assertFalse(combine[0] in final)
801
 
        self.assertFalse(combine[1] in final)
802
 
        self.assertSubset(to_keep, final)
803
 
 
804
 
    def test_stream_source_to_gc(self):
805
 
        source = self.make_repository('source', format='2a')
806
 
        target = self.make_repository('target', format='2a')
807
 
        stream = source._get_source(target._format)
808
 
        self.assertIsInstance(stream, groupcompress_repo.GroupCHKStreamSource)
809
 
 
810
 
    def test_stream_source_to_non_gc(self):
811
 
        source = self.make_repository('source', format='2a')
812
 
        target = self.make_repository('target', format='rich-root-pack')
813
 
        stream = source._get_source(target._format)
814
 
        # We don't want the child GroupCHKStreamSource
815
 
        self.assertIs(type(stream), repository.StreamSource)
816
 
 
817
 
    def test_get_stream_for_missing_keys_includes_all_chk_refs(self):
818
 
        source_builder = self.make_branch_builder('source',
819
 
                            format='2a')
820
 
        # We have to build a fairly large tree, so that we are sure the chk
821
 
        # pages will have split into multiple pages.
822
 
        entries = [('add', ('', 'a-root-id', 'directory', None))]
823
 
        for i in 'abcdefghijklmnopqrstuvwxyz123456789':
824
 
            for j in 'abcdefghijklmnopqrstuvwxyz123456789':
825
 
                fname = i + j
826
 
                fid = fname + '-id'
827
 
                content = 'content for %s\n' % (fname,)
828
 
                entries.append(('add', (fname, fid, 'file', content)))
829
 
        source_builder.start_series()
830
 
        source_builder.build_snapshot('rev-1', None, entries)
831
 
        # Now change a few of them, so we get a few new pages for the second
832
 
        # revision
833
 
        source_builder.build_snapshot('rev-2', ['rev-1'], [
834
 
            ('modify', ('aa-id', 'new content for aa-id\n')),
835
 
            ('modify', ('cc-id', 'new content for cc-id\n')),
836
 
            ('modify', ('zz-id', 'new content for zz-id\n')),
837
 
            ])
838
 
        source_builder.finish_series()
839
 
        source_branch = source_builder.get_branch()
840
 
        source_branch.lock_read()
841
 
        self.addCleanup(source_branch.unlock)
842
 
        target = self.make_repository('target', format='2a')
843
 
        source = source_branch.repository._get_source(target._format)
844
 
        self.assertIsInstance(source, groupcompress_repo.GroupCHKStreamSource)
845
 
 
846
 
        # On a regular pass, getting the inventories and chk pages for rev-2
847
 
        # would only get the newly created chk pages
848
 
        search = graph.SearchResult(set(['rev-2']), set(['rev-1']), 1,
849
 
                                    set(['rev-2']))
850
 
        simple_chk_records = []
851
 
        for vf_name, substream in source.get_stream(search):
852
 
            if vf_name == 'chk_bytes':
853
 
                for record in substream:
854
 
                    simple_chk_records.append(record.key)
855
 
            else:
856
 
                for _ in substream:
857
 
                    continue
858
 
        # 3 pages, the root (InternalNode), + 2 pages which actually changed
859
 
        self.assertEqual([('sha1:91481f539e802c76542ea5e4c83ad416bf219f73',),
860
 
                          ('sha1:4ff91971043668583985aec83f4f0ab10a907d3f',),
861
 
                          ('sha1:81e7324507c5ca132eedaf2d8414ee4bb2226187',),
862
 
                          ('sha1:b101b7da280596c71a4540e9a1eeba8045985ee0',)],
863
 
                         simple_chk_records)
864
 
        # Now, when we do a similar call using 'get_stream_for_missing_keys'
865
 
        # we should get a much larger set of pages.
866
 
        missing = [('inventories', 'rev-2')]
867
 
        full_chk_records = []
868
 
        for vf_name, substream in source.get_stream_for_missing_keys(missing):
869
 
            if vf_name == 'inventories':
870
 
                for record in substream:
871
 
                    self.assertEqual(('rev-2',), record.key)
872
 
            elif vf_name == 'chk_bytes':
873
 
                for record in substream:
874
 
                    full_chk_records.append(record.key)
875
 
            else:
876
 
                self.fail('Should not be getting a stream of %s' % (vf_name,))
877
 
        # We have 257 records now. This is because we have 1 root page, and 256
878
 
        # leaf pages in a complete listing.
879
 
        self.assertEqual(257, len(full_chk_records))
880
 
        self.assertSubset(simple_chk_records, full_chk_records)
881
 
 
882
 
    def test_inconsistency_fatal(self):
883
 
        repo = self.make_repository('repo', format='2a')
884
 
        self.assertTrue(repo.revisions._index._inconsistency_fatal)
885
 
        self.assertFalse(repo.texts._index._inconsistency_fatal)
886
 
        self.assertFalse(repo.inventories._index._inconsistency_fatal)
887
 
        self.assertFalse(repo.signatures._index._inconsistency_fatal)
888
 
        self.assertFalse(repo.chk_bytes._index._inconsistency_fatal)
889
 
 
890
 
 
891
 
class TestKnitPackStreamSource(tests.TestCaseWithMemoryTransport):
892
 
 
893
 
    def test_source_to_exact_pack_092(self):
894
 
        source = self.make_repository('source', format='pack-0.92')
895
 
        target = self.make_repository('target', format='pack-0.92')
896
 
        stream_source = source._get_source(target._format)
897
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
898
 
 
899
 
    def test_source_to_exact_pack_rich_root_pack(self):
900
 
        source = self.make_repository('source', format='rich-root-pack')
901
 
        target = self.make_repository('target', format='rich-root-pack')
902
 
        stream_source = source._get_source(target._format)
903
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
904
 
 
905
 
    def test_source_to_exact_pack_19(self):
906
 
        source = self.make_repository('source', format='1.9')
907
 
        target = self.make_repository('target', format='1.9')
908
 
        stream_source = source._get_source(target._format)
909
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
910
 
 
911
 
    def test_source_to_exact_pack_19_rich_root(self):
912
 
        source = self.make_repository('source', format='1.9-rich-root')
913
 
        target = self.make_repository('target', format='1.9-rich-root')
914
 
        stream_source = source._get_source(target._format)
915
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
916
 
 
917
 
    def test_source_to_remote_exact_pack_19(self):
918
 
        trans = self.make_smart_server('target')
919
 
        trans.ensure_base()
920
 
        source = self.make_repository('source', format='1.9')
921
 
        target = self.make_repository('target', format='1.9')
922
 
        target = repository.Repository.open(trans.base)
923
 
        stream_source = source._get_source(target._format)
924
 
        self.assertIsInstance(stream_source, pack_repo.KnitPackStreamSource)
925
 
 
926
 
    def test_stream_source_to_non_exact(self):
927
 
        source = self.make_repository('source', format='pack-0.92')
928
 
        target = self.make_repository('target', format='1.9')
929
 
        stream = source._get_source(target._format)
930
 
        self.assertIs(type(stream), repository.StreamSource)
931
 
 
932
 
    def test_stream_source_to_non_exact_rich_root(self):
933
 
        source = self.make_repository('source', format='1.9')
934
 
        target = self.make_repository('target', format='1.9-rich-root')
935
 
        stream = source._get_source(target._format)
936
 
        self.assertIs(type(stream), repository.StreamSource)
937
 
 
938
 
    def test_source_to_remote_non_exact_pack_19(self):
939
 
        trans = self.make_smart_server('target')
940
 
        trans.ensure_base()
941
 
        source = self.make_repository('source', format='1.9')
942
 
        target = self.make_repository('target', format='1.6')
943
 
        target = repository.Repository.open(trans.base)
944
 
        stream_source = source._get_source(target._format)
945
 
        self.assertIs(type(stream_source), repository.StreamSource)
946
 
 
947
 
    def test_stream_source_to_knit(self):
948
 
        source = self.make_repository('source', format='pack-0.92')
949
 
        target = self.make_repository('target', format='dirstate')
950
 
        stream = source._get_source(target._format)
951
 
        self.assertIs(type(stream), repository.StreamSource)
952
 
 
953
 
 
954
 
class TestDevelopment6FindParentIdsOfRevisions(TestCaseWithTransport):
955
 
    """Tests for _find_parent_ids_of_revisions."""
956
 
 
957
 
    def setUp(self):
958
 
        super(TestDevelopment6FindParentIdsOfRevisions, self).setUp()
959
 
        self.builder = self.make_branch_builder('source',
960
 
            format='development6-rich-root')
961
 
        self.builder.start_series()
962
 
        self.builder.build_snapshot('initial', None,
963
 
            [('add', ('', 'tree-root', 'directory', None))])
964
 
        self.repo = self.builder.get_branch().repository
965
 
        self.addCleanup(self.builder.finish_series)
966
 
 
967
 
    def assertParentIds(self, expected_result, rev_set):
968
 
        self.assertEqual(sorted(expected_result),
969
 
            sorted(self.repo._find_parent_ids_of_revisions(rev_set)))
970
 
 
971
 
    def test_simple(self):
972
 
        self.builder.build_snapshot('revid1', None, [])
973
 
        self.builder.build_snapshot('revid2', ['revid1'], [])
974
 
        rev_set = ['revid2']
975
 
        self.assertParentIds(['revid1'], rev_set)
976
 
 
977
 
    def test_not_first_parent(self):
978
 
        self.builder.build_snapshot('revid1', None, [])
979
 
        self.builder.build_snapshot('revid2', ['revid1'], [])
980
 
        self.builder.build_snapshot('revid3', ['revid2'], [])
981
 
        rev_set = ['revid3', 'revid2']
982
 
        self.assertParentIds(['revid1'], rev_set)
983
 
 
984
 
    def test_not_null(self):
985
 
        rev_set = ['initial']
986
 
        self.assertParentIds([], rev_set)
987
 
 
988
 
    def test_not_null_set(self):
989
 
        self.builder.build_snapshot('revid1', None, [])
990
 
        rev_set = [_mod_revision.NULL_REVISION]
991
 
        self.assertParentIds([], rev_set)
992
 
 
993
 
    def test_ghost(self):
994
 
        self.builder.build_snapshot('revid1', None, [])
995
 
        rev_set = ['ghost', 'revid1']
996
 
        self.assertParentIds(['initial'], rev_set)
997
 
 
998
 
    def test_ghost_parent(self):
999
 
        self.builder.build_snapshot('revid1', None, [])
1000
 
        self.builder.build_snapshot('revid2', ['revid1', 'ghost'], [])
1001
 
        rev_set = ['revid2', 'revid1']
1002
 
        self.assertParentIds(['ghost', 'initial'], rev_set)
1003
 
 
1004
 
    def test_righthand_parent(self):
1005
 
        self.builder.build_snapshot('revid1', None, [])
1006
 
        self.builder.build_snapshot('revid2a', ['revid1'], [])
1007
 
        self.builder.build_snapshot('revid2b', ['revid1'], [])
1008
 
        self.builder.build_snapshot('revid3', ['revid2a', 'revid2b'], [])
1009
 
        rev_set = ['revid3', 'revid2a']
1010
 
        self.assertParentIds(['revid1', 'revid2b'], rev_set)
1011
 
 
1012
 
 
1013
667
class TestWithBrokenRepo(TestCaseWithTransport):
1014
668
    """These tests seem to be more appropriate as interface tests?"""
1015
669
 
1028
682
            inv = inventory.Inventory(revision_id='rev1a')
1029
683
            inv.root.revision = 'rev1a'
1030
684
            self.add_file(repo, inv, 'file1', 'rev1a', [])
1031
 
            repo.texts.add_lines((inv.root.file_id, 'rev1a'), [], [])
1032
685
            repo.add_inventory('rev1a', inv, [])
1033
686
            revision = _mod_revision.Revision('rev1a',
1034
687
                committer='jrandom@example.com', timestamp=0,
1069
722
    def add_revision(self, repo, revision_id, inv, parent_ids):
1070
723
        inv.revision_id = revision_id
1071
724
        inv.root.revision = revision_id
1072
 
        repo.texts.add_lines((inv.root.file_id, revision_id), [], [])
1073
725
        repo.add_inventory(revision_id, inv, parent_ids)
1074
726
        revision = _mod_revision.Revision(revision_id,
1075
727
            committer='jrandom@example.com', timestamp=0, inventory_sha1='',
1092
744
        """
1093
745
        broken_repo = self.make_broken_repository()
1094
746
        empty_repo = self.make_repository('empty-repo')
1095
 
        try:
1096
 
            empty_repo.fetch(broken_repo)
1097
 
        except (errors.RevisionNotPresent, errors.BzrCheckError):
1098
 
            # Test successful: compression parent not being copied leads to
1099
 
            # error.
1100
 
            return
1101
 
        empty_repo.lock_read()
1102
 
        self.addCleanup(empty_repo.unlock)
1103
 
        text = empty_repo.texts.get_record_stream(
1104
 
            [('file2-id', 'rev3')], 'topological', True).next()
1105
 
        self.assertEqual('line\n', text.get_bytes_as('fulltext'))
 
747
        self.assertRaises((errors.RevisionNotPresent, errors.BzrCheckError),
 
748
                          empty_repo.fetch, broken_repo)
1106
749
 
1107
750
 
1108
751
class TestRepositoryPackCollection(TestCaseWithTransport):
1117
760
 
1118
761
    def make_packs_and_alt_repo(self, write_lock=False):
1119
762
        """Create a pack repo with 3 packs, and access it via a second repo."""
1120
 
        tree = self.make_branch_and_tree('.', format=self.get_format())
 
763
        tree = self.make_branch_and_tree('.')
1121
764
        tree.lock_write()
1122
765
        self.addCleanup(tree.unlock)
1123
766
        rev1 = tree.commit('one')
1133
776
        packs.ensure_loaded()
1134
777
        return tree, r, packs, [rev1, rev2, rev3]
1135
778
 
1136
 
    def test__clear_obsolete_packs(self):
1137
 
        packs = self.get_packs()
1138
 
        obsolete_pack_trans = packs.transport.clone('obsolete_packs')
1139
 
        obsolete_pack_trans.put_bytes('a-pack.pack', 'content\n')
1140
 
        obsolete_pack_trans.put_bytes('a-pack.rix', 'content\n')
1141
 
        obsolete_pack_trans.put_bytes('a-pack.iix', 'content\n')
1142
 
        obsolete_pack_trans.put_bytes('another-pack.pack', 'foo\n')
1143
 
        obsolete_pack_trans.put_bytes('not-a-pack.rix', 'foo\n')
1144
 
        res = packs._clear_obsolete_packs()
1145
 
        self.assertEqual(['a-pack', 'another-pack'], sorted(res))
1146
 
        self.assertEqual([], obsolete_pack_trans.list_dir('.'))
1147
 
 
1148
 
    def test__clear_obsolete_packs_preserve(self):
1149
 
        packs = self.get_packs()
1150
 
        obsolete_pack_trans = packs.transport.clone('obsolete_packs')
1151
 
        obsolete_pack_trans.put_bytes('a-pack.pack', 'content\n')
1152
 
        obsolete_pack_trans.put_bytes('a-pack.rix', 'content\n')
1153
 
        obsolete_pack_trans.put_bytes('a-pack.iix', 'content\n')
1154
 
        obsolete_pack_trans.put_bytes('another-pack.pack', 'foo\n')
1155
 
        obsolete_pack_trans.put_bytes('not-a-pack.rix', 'foo\n')
1156
 
        res = packs._clear_obsolete_packs(preserve=set(['a-pack']))
1157
 
        self.assertEqual(['a-pack', 'another-pack'], sorted(res))
1158
 
        self.assertEqual(['a-pack.iix', 'a-pack.pack', 'a-pack.rix'],
1159
 
                         sorted(obsolete_pack_trans.list_dir('.')))
1160
 
 
1161
779
    def test__max_pack_count(self):
1162
780
        """The maximum pack count is a function of the number of revisions."""
1163
781
        # no revisions - one pack, so that we can have a revision free repo
1183
801
        # check some arbitrary big numbers
1184
802
        self.assertEqual(25, packs._max_pack_count(112894))
1185
803
 
1186
 
    def test_repr(self):
1187
 
        packs = self.get_packs()
1188
 
        self.assertContainsRe(repr(packs),
1189
 
            'RepositoryPackCollection(.*Repository(.*))')
1190
 
 
1191
 
    def test__obsolete_packs(self):
1192
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1193
 
        names = packs.names()
1194
 
        pack = packs.get_pack_by_name(names[0])
1195
 
        # Schedule this one for removal
1196
 
        packs._remove_pack_from_memory(pack)
1197
 
        # Simulate a concurrent update by renaming the .pack file and one of
1198
 
        # the indices
1199
 
        packs.transport.rename('packs/%s.pack' % (names[0],),
1200
 
                               'obsolete_packs/%s.pack' % (names[0],))
1201
 
        packs.transport.rename('indices/%s.iix' % (names[0],),
1202
 
                               'obsolete_packs/%s.iix' % (names[0],))
1203
 
        # Now trigger the obsoletion, and ensure that all the remaining files
1204
 
        # are still renamed
1205
 
        packs._obsolete_packs([pack])
1206
 
        self.assertEqual([n + '.pack' for n in names[1:]],
1207
 
                         sorted(packs._pack_transport.list_dir('.')))
1208
 
        # names[0] should not be present in the index anymore
1209
 
        self.assertEqual(names[1:],
1210
 
            sorted(set([osutils.splitext(n)[0] for n in
1211
 
                        packs._index_transport.list_dir('.')])))
1212
 
 
1213
804
    def test_pack_distribution_zero(self):
1214
805
        packs = self.get_packs()
1215
806
        self.assertEqual([0], packs.pack_distribution(0))
1338
929
        tree.lock_read()
1339
930
        self.addCleanup(tree.unlock)
1340
931
        packs = tree.branch.repository._pack_collection
1341
 
        packs.reset()
1342
932
        packs.ensure_loaded()
1343
933
        name = packs.names()[0]
1344
934
        pack_1 = packs.get_pack_by_name(name)
1383
973
        self.assertEqual({revs[-1]:(revs[-2],)}, r.get_parent_map([revs[-1]]))
1384
974
        self.assertFalse(packs.reload_pack_names())
1385
975
 
1386
 
    def test_reload_pack_names_preserves_pending(self):
1387
 
        # TODO: Update this to also test for pending-deleted names
1388
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1389
 
        # We will add one pack (via start_write_group + insert_record_stream),
1390
 
        # and remove another pack (via _remove_pack_from_memory)
1391
 
        orig_names = packs.names()
1392
 
        orig_at_load = packs._packs_at_load
1393
 
        to_remove_name = iter(orig_names).next()
1394
 
        r.start_write_group()
1395
 
        self.addCleanup(r.abort_write_group)
1396
 
        r.texts.insert_record_stream([versionedfile.FulltextContentFactory(
1397
 
            ('text', 'rev'), (), None, 'content\n')])
1398
 
        new_pack = packs._new_pack
1399
 
        self.assertTrue(new_pack.data_inserted())
1400
 
        new_pack.finish()
1401
 
        packs.allocate(new_pack)
1402
 
        packs._new_pack = None
1403
 
        removed_pack = packs.get_pack_by_name(to_remove_name)
1404
 
        packs._remove_pack_from_memory(removed_pack)
1405
 
        names = packs.names()
1406
 
        all_nodes, deleted_nodes, new_nodes, _ = packs._diff_pack_names()
1407
 
        new_names = set([x[0][0] for x in new_nodes])
1408
 
        self.assertEqual(names, sorted([x[0][0] for x in all_nodes]))
1409
 
        self.assertEqual(set(names) - set(orig_names), new_names)
1410
 
        self.assertEqual(set([new_pack.name]), new_names)
1411
 
        self.assertEqual([to_remove_name],
1412
 
                         sorted([x[0][0] for x in deleted_nodes]))
1413
 
        packs.reload_pack_names()
1414
 
        reloaded_names = packs.names()
1415
 
        self.assertEqual(orig_at_load, packs._packs_at_load)
1416
 
        self.assertEqual(names, reloaded_names)
1417
 
        all_nodes, deleted_nodes, new_nodes, _ = packs._diff_pack_names()
1418
 
        new_names = set([x[0][0] for x in new_nodes])
1419
 
        self.assertEqual(names, sorted([x[0][0] for x in all_nodes]))
1420
 
        self.assertEqual(set(names) - set(orig_names), new_names)
1421
 
        self.assertEqual(set([new_pack.name]), new_names)
1422
 
        self.assertEqual([to_remove_name],
1423
 
                         sorted([x[0][0] for x in deleted_nodes]))
1424
 
 
1425
 
    def test_autopack_obsoletes_new_pack(self):
1426
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1427
 
        packs._max_pack_count = lambda x: 1
1428
 
        packs.pack_distribution = lambda x: [10]
1429
 
        r.start_write_group()
1430
 
        r.revisions.insert_record_stream([versionedfile.FulltextContentFactory(
1431
 
            ('bogus-rev',), (), None, 'bogus-content\n')])
1432
 
        # This should trigger an autopack, which will combine everything into a
1433
 
        # single pack file.
1434
 
        new_names = r.commit_write_group()
1435
 
        names = packs.names()
1436
 
        self.assertEqual(1, len(names))
1437
 
        self.assertEqual([names[0] + '.pack'],
1438
 
                         packs._pack_transport.list_dir('.'))
1439
 
 
1440
976
    def test_autopack_reloads_and_stops(self):
1441
977
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1442
978
        # After we have determined what needs to be autopacked, trigger a
1454
990
        self.assertEqual(tree.branch.repository._pack_collection.names(),
1455
991
                         packs.names())
1456
992
 
1457
 
    def test__save_pack_names(self):
1458
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1459
 
        names = packs.names()
1460
 
        pack = packs.get_pack_by_name(names[0])
1461
 
        packs._remove_pack_from_memory(pack)
1462
 
        packs._save_pack_names(obsolete_packs=[pack])
1463
 
        cur_packs = packs._pack_transport.list_dir('.')
1464
 
        self.assertEqual([n + '.pack' for n in names[1:]], sorted(cur_packs))
1465
 
        # obsolete_packs will also have stuff like .rix and .iix present.
1466
 
        obsolete_packs = packs.transport.list_dir('obsolete_packs')
1467
 
        obsolete_names = set([osutils.splitext(n)[0] for n in obsolete_packs])
1468
 
        self.assertEqual([pack.name], sorted(obsolete_names))
1469
 
 
1470
 
    def test__save_pack_names_already_obsoleted(self):
1471
 
        tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1472
 
        names = packs.names()
1473
 
        pack = packs.get_pack_by_name(names[0])
1474
 
        packs._remove_pack_from_memory(pack)
1475
 
        # We are going to simulate a concurrent autopack by manually obsoleting
1476
 
        # the pack directly.
1477
 
        packs._obsolete_packs([pack])
1478
 
        packs._save_pack_names(clear_obsolete_packs=True,
1479
 
                               obsolete_packs=[pack])
1480
 
        cur_packs = packs._pack_transport.list_dir('.')
1481
 
        self.assertEqual([n + '.pack' for n in names[1:]], sorted(cur_packs))
1482
 
        # Note that while we set clear_obsolete_packs=True, it should not
1483
 
        # delete a pack file that we have also scheduled for obsoletion.
1484
 
        obsolete_packs = packs.transport.list_dir('obsolete_packs')
1485
 
        obsolete_names = set([osutils.splitext(n)[0] for n in obsolete_packs])
1486
 
        self.assertEqual([pack.name], sorted(obsolete_names))
1487
 
 
1488
 
 
1489
993
 
1490
994
class TestPack(TestCaseWithTransport):
1491
995
    """Tests for the Pack object."""
1545
1049
        pack_transport = self.get_transport('pack')
1546
1050
        index_transport = self.get_transport('index')
1547
1051
        upload_transport.mkdir('.')
1548
 
        collection = pack_repo.RepositoryPackCollection(
1549
 
            repo=None,
 
1052
        collection = pack_repo.RepositoryPackCollection(repo=None,
1550
1053
            transport=self.get_transport('.'),
1551
1054
            index_transport=index_transport,
1552
1055
            upload_transport=upload_transport,
1553
1056
            pack_transport=pack_transport,
1554
1057
            index_builder_class=BTreeBuilder,
1555
 
            index_class=BTreeGraphIndex,
1556
 
            use_chk_index=False)
 
1058
            index_class=BTreeGraphIndex)
1557
1059
        pack = pack_repo.NewPack(collection)
1558
 
        self.addCleanup(pack.abort) # Make sure the write stream gets closed
1559
1060
        self.assertIsInstance(pack.revision_index, BTreeBuilder)
1560
1061
        self.assertIsInstance(pack.inventory_index, BTreeBuilder)
1561
1062
        self.assertIsInstance(pack._hash, type(osutils.md5()))
1572
1073
    """Tests for the packs repository Packer class."""
1573
1074
 
1574
1075
    def test_pack_optimizes_pack_order(self):
1575
 
        builder = self.make_branch_builder('.', format="1.9")
 
1076
        builder = self.make_branch_builder('.')
1576
1077
        builder.start_series()
1577
1078
        builder.build_snapshot('A', None, [
1578
1079
            ('add', ('', 'root-id', 'directory', None)),
1614
1115
        packer = pack_repo.OptimisingPacker(self.get_pack_collection(),
1615
1116
                                            [], '.test')
1616
1117
        new_pack = packer.open_pack()
1617
 
        self.addCleanup(new_pack.abort) # ensure cleanup
1618
1118
        self.assertIsInstance(new_pack, pack_repo.NewPack)
1619
1119
        self.assertTrue(new_pack.revision_index._optimize_for_size)
1620
1120
        self.assertTrue(new_pack.inventory_index._optimize_for_size)
1622
1122
        self.assertTrue(new_pack.signature_index._optimize_for_size)
1623
1123
 
1624
1124
 
1625
 
class TestCrossFormatPacks(TestCaseWithTransport):
1626
 
 
1627
 
    def log_pack(self, hint=None):
1628
 
        self.calls.append(('pack', hint))
1629
 
        self.orig_pack(hint=hint)
1630
 
        if self.expect_hint:
1631
 
            self.assertTrue(hint)
1632
 
 
1633
 
    def run_stream(self, src_fmt, target_fmt, expect_pack_called):
1634
 
        self.expect_hint = expect_pack_called
1635
 
        self.calls = []
1636
 
        source_tree = self.make_branch_and_tree('src', format=src_fmt)
1637
 
        source_tree.lock_write()
1638
 
        self.addCleanup(source_tree.unlock)
1639
 
        tip = source_tree.commit('foo')
1640
 
        target = self.make_repository('target', format=target_fmt)
1641
 
        target.lock_write()
1642
 
        self.addCleanup(target.unlock)
1643
 
        source = source_tree.branch.repository._get_source(target._format)
1644
 
        self.orig_pack = target.pack
1645
 
        target.pack = self.log_pack
1646
 
        search = target.search_missing_revision_ids(
1647
 
            source_tree.branch.repository, tip)
1648
 
        stream = source.get_stream(search)
1649
 
        from_format = source_tree.branch.repository._format
1650
 
        sink = target._get_sink()
1651
 
        sink.insert_stream(stream, from_format, [])
1652
 
        if expect_pack_called:
1653
 
            self.assertLength(1, self.calls)
1654
 
        else:
1655
 
            self.assertLength(0, self.calls)
1656
 
 
1657
 
    def run_fetch(self, src_fmt, target_fmt, expect_pack_called):
1658
 
        self.expect_hint = expect_pack_called
1659
 
        self.calls = []
1660
 
        source_tree = self.make_branch_and_tree('src', format=src_fmt)
1661
 
        source_tree.lock_write()
1662
 
        self.addCleanup(source_tree.unlock)
1663
 
        tip = source_tree.commit('foo')
1664
 
        target = self.make_repository('target', format=target_fmt)
1665
 
        target.lock_write()
1666
 
        self.addCleanup(target.unlock)
1667
 
        source = source_tree.branch.repository
1668
 
        self.orig_pack = target.pack
1669
 
        target.pack = self.log_pack
1670
 
        target.fetch(source)
1671
 
        if expect_pack_called:
1672
 
            self.assertLength(1, self.calls)
1673
 
        else:
1674
 
            self.assertLength(0, self.calls)
1675
 
 
1676
 
    def test_sink_format_hint_no(self):
1677
 
        # When the target format says packing makes no difference, pack is not
1678
 
        # called.
1679
 
        self.run_stream('1.9', 'rich-root-pack', False)
1680
 
 
1681
 
    def test_sink_format_hint_yes(self):
1682
 
        # When the target format says packing makes a difference, pack is
1683
 
        # called.
1684
 
        self.run_stream('1.9', '2a', True)
1685
 
 
1686
 
    def test_sink_format_same_no(self):
1687
 
        # When the formats are the same, pack is not called.
1688
 
        self.run_stream('2a', '2a', False)
1689
 
 
1690
 
    def test_IDS_format_hint_no(self):
1691
 
        # When the target format says packing makes no difference, pack is not
1692
 
        # called.
1693
 
        self.run_fetch('1.9', 'rich-root-pack', False)
1694
 
 
1695
 
    def test_IDS_format_hint_yes(self):
1696
 
        # When the target format says packing makes a difference, pack is
1697
 
        # called.
1698
 
        self.run_fetch('1.9', '2a', True)
1699
 
 
1700
 
    def test_IDS_format_same_no(self):
1701
 
        # When the formats are the same, pack is not called.
1702
 
        self.run_fetch('2a', '2a', False)
 
1125
class TestInterDifferingSerializer(TestCaseWithTransport):
 
1126
 
 
1127
    def test_progress_bar(self):
 
1128
        tree = self.make_branch_and_tree('tree')
 
1129
        tree.commit('rev1', rev_id='rev-1')
 
1130
        tree.commit('rev2', rev_id='rev-2')
 
1131
        tree.commit('rev3', rev_id='rev-3')
 
1132
        repo = self.make_repository('repo')
 
1133
        inter_repo = repository.InterDifferingSerializer(
 
1134
            tree.branch.repository, repo)
 
1135
        pb = progress.InstrumentedProgress(to_file=StringIO())
 
1136
        pb.never_throttle = True
 
1137
        inter_repo.fetch('rev-1', pb)
 
1138
        self.assertEqual('Transferring revisions', pb.last_msg)
 
1139
        self.assertEqual(1, pb.last_cnt)
 
1140
        self.assertEqual(1, pb.last_total)
 
1141
        inter_repo.fetch('rev-3', pb)
 
1142
        self.assertEqual(2, pb.last_cnt)
 
1143
        self.assertEqual(2, pb.last_total)