136
138
def check_format(format, url):
137
139
dir = format._matchingbzrdir.initialize(url)
138
140
format.initialize(dir)
139
t = get_transport(url)
140
found_format = repository.RepositoryFormat.find_format(dir)
141
self.failUnless(isinstance(found_format, format.__class__))
142
check_format(weaverepo.RepositoryFormat7(), "bar")
141
t = transport.get_transport_from_path(url)
142
found_format = repository.RepositoryFormatMetaDir.find_format(dir)
143
self.assertIsInstance(found_format, format.__class__)
144
check_format(repository.format_registry.get_default(), "bar")
144
146
def test_find_format_no_repository(self):
145
147
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
146
148
self.assertRaises(errors.NoRepositoryPresent,
147
repository.RepositoryFormat.find_format,
149
repository.RepositoryFormatMetaDir.find_format,
152
def test_from_string(self):
153
self.assertIsInstance(
154
SampleRepositoryFormat.from_string(
155
"Sample .bzr repository format."),
156
SampleRepositoryFormat)
157
self.assertRaises(AssertionError,
158
SampleRepositoryFormat.from_string,
159
"Different .bzr repository format.")
150
161
def test_find_format_unknown_format(self):
151
162
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
152
163
SampleRepositoryFormat().initialize(dir)
153
164
self.assertRaises(UnknownFormatError,
154
repository.RepositoryFormat.find_format,
165
repository.RepositoryFormatMetaDir.find_format,
168
def test_find_format_with_features(self):
169
tree = self.make_branch_and_tree('.', format='2a')
170
tree.branch.repository.update_feature_flags({"name": "necessity"})
171
found_format = repository.RepositoryFormatMetaDir.find_format(tree.bzrdir)
172
self.assertIsInstance(found_format, repository.RepositoryFormatMetaDir)
173
self.assertEquals(found_format.features.get("name"), "necessity")
174
self.assertRaises(errors.MissingFeature, found_format.check_support_status,
176
self.addCleanup(repository.RepositoryFormatMetaDir.unregister_feature,
178
repository.RepositoryFormatMetaDir.register_feature("name")
179
found_format.check_support_status(True)
182
class TestRepositoryFormatRegistry(TestCase):
185
super(TestRepositoryFormatRegistry, self).setUp()
186
self.registry = repository.RepositoryFormatRegistry()
157
188
def test_register_unregister_format(self):
158
189
format = SampleRepositoryFormat()
160
dir = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
162
format.initialize(dir)
163
# register a format for it.
164
repository.RepositoryFormat.register_format(format)
165
# which repository.Open will refuse (not supported)
166
self.assertRaises(UnsupportedFormatError, repository.Repository.open, self.get_url())
167
# but open(unsupported) will work
168
self.assertEqual(format.open(dir), "opened repository.")
169
# unregister the format
170
repository.RepositoryFormat.unregister_format(format)
173
class TestFormat6(TestCaseWithTransport):
175
def test_attribute__fetch_order(self):
176
"""Weaves need topological data insertion."""
177
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
178
repo = weaverepo.RepositoryFormat6().initialize(control)
179
self.assertEqual('topological', repo._format._fetch_order)
181
def test_attribute__fetch_uses_deltas(self):
182
"""Weaves do not reuse deltas."""
183
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
184
repo = weaverepo.RepositoryFormat6().initialize(control)
185
self.assertEqual(False, repo._format._fetch_uses_deltas)
187
def test_attribute__fetch_reconcile(self):
188
"""Weave repositories need a reconcile after fetch."""
189
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
190
repo = weaverepo.RepositoryFormat6().initialize(control)
191
self.assertEqual(True, repo._format._fetch_reconcile)
193
def test_no_ancestry_weave(self):
194
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
195
repo = weaverepo.RepositoryFormat6().initialize(control)
196
# We no longer need to create the ancestry.weave file
197
# since it is *never* used.
198
self.assertRaises(NoSuchFile,
199
control.transport.get,
202
def test_supports_external_lookups(self):
203
control = bzrdir.BzrDirFormat6().initialize(self.get_url())
204
repo = weaverepo.RepositoryFormat6().initialize(control)
205
self.assertFalse(repo._format.supports_external_lookups)
208
class TestFormat7(TestCaseWithTransport):
210
def test_attribute__fetch_order(self):
211
"""Weaves need topological data insertion."""
212
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
213
repo = weaverepo.RepositoryFormat7().initialize(control)
214
self.assertEqual('topological', repo._format._fetch_order)
216
def test_attribute__fetch_uses_deltas(self):
217
"""Weaves do not reuse deltas."""
218
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
219
repo = weaverepo.RepositoryFormat7().initialize(control)
220
self.assertEqual(False, repo._format._fetch_uses_deltas)
222
def test_attribute__fetch_reconcile(self):
223
"""Weave repositories need a reconcile after fetch."""
224
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
225
repo = weaverepo.RepositoryFormat7().initialize(control)
226
self.assertEqual(True, repo._format._fetch_reconcile)
228
def test_disk_layout(self):
229
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
230
repo = weaverepo.RepositoryFormat7().initialize(control)
231
# in case of side effects of locking.
235
# format 'Bazaar-NG Repository format 7'
237
# inventory.weave == empty_weave
238
# empty revision-store directory
239
# empty weaves directory
240
t = control.get_repository_transport(None)
241
self.assertEqualDiff('Bazaar-NG Repository format 7',
242
t.get('format').read())
243
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
244
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
245
self.assertEqualDiff('# bzr weave file v5\n'
248
t.get('inventory.weave').read())
249
# Creating a file with id Foo:Bar results in a non-escaped file name on
251
control.create_branch()
252
tree = control.create_workingtree()
253
tree.add(['foo'], ['Foo:Bar'], ['file'])
254
tree.put_file_bytes_non_atomic('Foo:Bar', 'content\n')
255
tree.commit('first post', rev_id='first')
256
self.assertEqualDiff(
257
'# bzr weave file v5\n'
259
'1 7fe70820e08a1aac0ef224d9c66ab66831cc4ab1\n'
267
t.get('weaves/74/Foo%3ABar.weave').read())
269
def test_shared_disk_layout(self):
270
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
271
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
273
# format 'Bazaar-NG Repository format 7'
274
# inventory.weave == empty_weave
275
# empty revision-store directory
276
# empty weaves directory
277
# a 'shared-storage' marker file.
278
# lock is not present when unlocked
279
t = control.get_repository_transport(None)
280
self.assertEqualDiff('Bazaar-NG Repository format 7',
281
t.get('format').read())
282
self.assertEqualDiff('', t.get('shared-storage').read())
283
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
284
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
285
self.assertEqualDiff('# bzr weave file v5\n'
288
t.get('inventory.weave').read())
289
self.assertFalse(t.has('branch-lock'))
291
def test_creates_lockdir(self):
292
"""Make sure it appears to be controlled by a LockDir existence"""
293
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
294
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
295
t = control.get_repository_transport(None)
296
# TODO: Should check there is a 'lock' toplevel directory,
297
# regardless of contents
298
self.assertFalse(t.has('lock/held/info'))
301
self.assertTrue(t.has('lock/held/info'))
303
# unlock so we don't get a warning about failing to do so
306
def test_uses_lockdir(self):
307
"""repo format 7 actually locks on lockdir"""
308
base_url = self.get_url()
309
control = bzrdir.BzrDirMetaFormat1().initialize(base_url)
310
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
311
t = control.get_repository_transport(None)
315
# make sure the same lock is created by opening it
316
repo = repository.Repository.open(base_url)
318
self.assertTrue(t.has('lock/held/info'))
320
self.assertFalse(t.has('lock/held/info'))
322
def test_shared_no_tree_disk_layout(self):
323
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
324
repo = weaverepo.RepositoryFormat7().initialize(control, shared=True)
325
repo.set_make_working_trees(False)
327
# format 'Bazaar-NG Repository format 7'
329
# inventory.weave == empty_weave
330
# empty revision-store directory
331
# empty weaves directory
332
# a 'shared-storage' marker file.
333
t = control.get_repository_transport(None)
334
self.assertEqualDiff('Bazaar-NG Repository format 7',
335
t.get('format').read())
336
## self.assertEqualDiff('', t.get('lock').read())
337
self.assertEqualDiff('', t.get('shared-storage').read())
338
self.assertEqualDiff('', t.get('no-working-trees').read())
339
repo.set_make_working_trees(True)
340
self.assertFalse(t.has('no-working-trees'))
341
self.assertTrue(S_ISDIR(t.stat('revision-store').st_mode))
342
self.assertTrue(S_ISDIR(t.stat('weaves').st_mode))
343
self.assertEqualDiff('# bzr weave file v5\n'
346
t.get('inventory.weave').read())
348
def test_supports_external_lookups(self):
349
control = bzrdir.BzrDirMetaFormat1().initialize(self.get_url())
350
repo = weaverepo.RepositoryFormat7().initialize(control)
351
self.assertFalse(repo._format.supports_external_lookups)
190
self.registry.register(format)
191
self.assertEquals(format, self.registry.get("Sample .bzr repository format."))
192
self.registry.remove(format)
193
self.assertRaises(KeyError, self.registry.get, "Sample .bzr repository format.")
195
def test_get_all(self):
196
format = SampleRepositoryFormat()
197
self.assertEquals([], self.registry._get_all())
198
self.registry.register(format)
199
self.assertEquals([format], self.registry._get_all())
201
def test_register_extra(self):
202
format = SampleExtraRepositoryFormat()
203
self.assertEquals([], self.registry._get_all())
204
self.registry.register_extra(format)
205
self.assertEquals([format], self.registry._get_all())
207
def test_register_extra_lazy(self):
208
self.assertEquals([], self.registry._get_all())
209
self.registry.register_extra_lazy("bzrlib.tests.test_repository",
210
"SampleExtraRepositoryFormat")
211
formats = self.registry._get_all()
212
self.assertEquals(1, len(formats))
213
self.assertIsInstance(formats[0], SampleExtraRepositoryFormat)
354
216
class TestFormatKnit1(TestCaseWithTransport):
573
441
self.assertGetsDefaultInterRepository(dummy_a, dummy_b)
576
class TestInterWeaveRepo(TestCaseWithTransport):
578
def test_is_compatible_and_registered(self):
579
# InterWeaveRepo is compatible when either side
580
# is a format 5/6/7 branch
581
from bzrlib.repofmt import knitrepo, weaverepo
582
formats = [weaverepo.RepositoryFormat5(),
583
weaverepo.RepositoryFormat6(),
584
weaverepo.RepositoryFormat7()]
585
incompatible_formats = [weaverepo.RepositoryFormat4(),
586
knitrepo.RepositoryFormatKnit1(),
588
repo_a = self.make_repository('a')
589
repo_b = self.make_repository('b')
590
is_compatible = repository.InterWeaveRepo.is_compatible
591
for source in incompatible_formats:
592
# force incompatible left then right
593
repo_a._format = source
594
repo_b._format = formats[0]
595
self.assertFalse(is_compatible(repo_a, repo_b))
596
self.assertFalse(is_compatible(repo_b, repo_a))
597
for source in formats:
598
repo_a._format = source
599
for target in formats:
600
repo_b._format = target
601
self.assertTrue(is_compatible(repo_a, repo_b))
602
self.assertEqual(repository.InterWeaveRepo,
603
repository.InterRepository.get(repo_a,
444
class TestRepositoryFormat1(knitrepo.RepositoryFormatKnit1):
447
def get_format_string(cls):
448
return "Test Format 1"
451
class TestRepositoryFormat2(knitrepo.RepositoryFormatKnit1):
454
def get_format_string(cls):
455
return "Test Format 2"
607
458
class TestRepositoryConverter(TestCaseWithTransport):
609
460
def test_convert_empty(self):
610
t = get_transport(self.get_url('.'))
461
source_format = TestRepositoryFormat1()
462
target_format = TestRepositoryFormat2()
463
repository.format_registry.register(source_format)
464
self.addCleanup(repository.format_registry.remove,
466
repository.format_registry.register(target_format)
467
self.addCleanup(repository.format_registry.remove,
469
t = self.get_transport()
611
470
t.mkdir('repository')
612
471
repo_dir = bzrdir.BzrDirMetaFormat1().initialize('repository')
613
repo = weaverepo.RepositoryFormat7().initialize(repo_dir)
614
target_format = knitrepo.RepositoryFormatKnit1()
472
repo = TestRepositoryFormat1().initialize(repo_dir)
615
473
converter = repository.CopyConverter(target_format)
616
474
pb = bzrlib.ui.ui_factory.nested_progress_bar()
681
533
self.assertFalse(repo._format.supports_external_lookups)
684
class Test2a(TestCaseWithTransport):
536
class Test2a(tests.TestCaseWithMemoryTransport):
538
def test_chk_bytes_uses_custom_btree_parser(self):
539
mt = self.make_branch_and_memory_tree('test', format='2a')
541
self.addCleanup(mt.unlock)
542
mt.add([''], ['root-id'])
544
index = mt.branch.repository.chk_bytes._index._graph_index._indices[0]
545
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
546
# It should also work if we re-open the repo
547
repo = mt.branch.repository.bzrdir.open_repository()
549
self.addCleanup(repo.unlock)
550
index = repo.chk_bytes._index._graph_index._indices[0]
551
self.assertEqual(btree_index._gcchk_factory, index._leaf_factory)
553
def test_fetch_combines_groups(self):
554
builder = self.make_branch_builder('source', format='2a')
555
builder.start_series()
556
builder.build_snapshot('1', None, [
557
('add', ('', 'root-id', 'directory', '')),
558
('add', ('file', 'file-id', 'file', 'content\n'))])
559
builder.build_snapshot('2', ['1'], [
560
('modify', ('file-id', 'content-2\n'))])
561
builder.finish_series()
562
source = builder.get_branch()
563
target = self.make_repository('target', format='2a')
564
target.fetch(source.repository)
566
self.addCleanup(target.unlock)
567
details = target.texts._index.get_build_details(
568
[('file-id', '1',), ('file-id', '2',)])
569
file_1_details = details[('file-id', '1')]
570
file_2_details = details[('file-id', '2')]
571
# The index, and what to read off disk, should be the same for both
572
# versions of the file.
573
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
575
def test_fetch_combines_groups(self):
576
builder = self.make_branch_builder('source', format='2a')
577
builder.start_series()
578
builder.build_snapshot('1', None, [
579
('add', ('', 'root-id', 'directory', '')),
580
('add', ('file', 'file-id', 'file', 'content\n'))])
581
builder.build_snapshot('2', ['1'], [
582
('modify', ('file-id', 'content-2\n'))])
583
builder.finish_series()
584
source = builder.get_branch()
585
target = self.make_repository('target', format='2a')
586
target.fetch(source.repository)
588
self.addCleanup(target.unlock)
589
details = target.texts._index.get_build_details(
590
[('file-id', '1',), ('file-id', '2',)])
591
file_1_details = details[('file-id', '1')]
592
file_2_details = details[('file-id', '2')]
593
# The index, and what to read off disk, should be the same for both
594
# versions of the file.
595
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
597
def test_fetch_combines_groups(self):
598
builder = self.make_branch_builder('source', format='2a')
599
builder.start_series()
600
builder.build_snapshot('1', None, [
601
('add', ('', 'root-id', 'directory', '')),
602
('add', ('file', 'file-id', 'file', 'content\n'))])
603
builder.build_snapshot('2', ['1'], [
604
('modify', ('file-id', 'content-2\n'))])
605
builder.finish_series()
606
source = builder.get_branch()
607
target = self.make_repository('target', format='2a')
608
target.fetch(source.repository)
610
self.addCleanup(target.unlock)
611
details = target.texts._index.get_build_details(
612
[('file-id', '1',), ('file-id', '2',)])
613
file_1_details = details[('file-id', '1')]
614
file_2_details = details[('file-id', '2')]
615
# The index, and what to read off disk, should be the same for both
616
# versions of the file.
617
self.assertEqual(file_1_details[0][:3], file_2_details[0][:3])
686
619
def test_format_pack_compresses_True(self):
687
620
repo = self.make_repository('repo', format='2a')
688
621
self.assertTrue(repo._format.pack_compresses)
690
623
def test_inventories_use_chk_map_with_parent_base_dict(self):
691
tree = self.make_branch_and_tree('repo', format="2a")
624
tree = self.make_branch_and_memory_tree('repo', format="2a")
626
tree.add([''], ['TREE_ROOT'])
692
627
revid = tree.commit("foo")
694
630
self.addCleanup(tree.unlock)
695
631
inv = tree.branch.repository.get_inventory(revid)
1082
1051
# check some arbitrary big numbers
1083
1052
self.assertEqual(25, packs._max_pack_count(112894))
1054
def test_repr(self):
1055
packs = self.get_packs()
1056
self.assertContainsRe(repr(packs),
1057
'RepositoryPackCollection(.*Repository(.*))')
1059
def test__obsolete_packs(self):
1060
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1061
names = packs.names()
1062
pack = packs.get_pack_by_name(names[0])
1063
# Schedule this one for removal
1064
packs._remove_pack_from_memory(pack)
1065
# Simulate a concurrent update by renaming the .pack file and one of
1067
packs.transport.rename('packs/%s.pack' % (names[0],),
1068
'obsolete_packs/%s.pack' % (names[0],))
1069
packs.transport.rename('indices/%s.iix' % (names[0],),
1070
'obsolete_packs/%s.iix' % (names[0],))
1071
# Now trigger the obsoletion, and ensure that all the remaining files
1073
packs._obsolete_packs([pack])
1074
self.assertEqual([n + '.pack' for n in names[1:]],
1075
sorted(packs._pack_transport.list_dir('.')))
1076
# names[0] should not be present in the index anymore
1077
self.assertEqual(names[1:],
1078
sorted(set([osutils.splitext(n)[0] for n in
1079
packs._index_transport.list_dir('.')])))
1081
def test__obsolete_packs_missing_directory(self):
1082
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1083
r.control_transport.rmdir('obsolete_packs')
1084
names = packs.names()
1085
pack = packs.get_pack_by_name(names[0])
1086
# Schedule this one for removal
1087
packs._remove_pack_from_memory(pack)
1088
# Now trigger the obsoletion, and ensure that all the remaining files
1090
packs._obsolete_packs([pack])
1091
self.assertEqual([n + '.pack' for n in names[1:]],
1092
sorted(packs._pack_transport.list_dir('.')))
1093
# names[0] should not be present in the index anymore
1094
self.assertEqual(names[1:],
1095
sorted(set([osutils.splitext(n)[0] for n in
1096
packs._index_transport.list_dir('.')])))
1085
1098
def test_pack_distribution_zero(self):
1086
1099
packs = self.get_packs()
1087
1100
self.assertEqual([0], packs.pack_distribution(0))
1255
1268
self.assertEqual({revs[-1]:(revs[-2],)}, r.get_parent_map([revs[-1]]))
1256
1269
self.assertFalse(packs.reload_pack_names())
1271
def test_reload_pack_names_preserves_pending(self):
1272
# TODO: Update this to also test for pending-deleted names
1273
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1274
# We will add one pack (via start_write_group + insert_record_stream),
1275
# and remove another pack (via _remove_pack_from_memory)
1276
orig_names = packs.names()
1277
orig_at_load = packs._packs_at_load
1278
to_remove_name = iter(orig_names).next()
1279
r.start_write_group()
1280
self.addCleanup(r.abort_write_group)
1281
r.texts.insert_record_stream([versionedfile.FulltextContentFactory(
1282
('text', 'rev'), (), None, 'content\n')])
1283
new_pack = packs._new_pack
1284
self.assertTrue(new_pack.data_inserted())
1286
packs.allocate(new_pack)
1287
packs._new_pack = None
1288
removed_pack = packs.get_pack_by_name(to_remove_name)
1289
packs._remove_pack_from_memory(removed_pack)
1290
names = packs.names()
1291
all_nodes, deleted_nodes, new_nodes, _ = packs._diff_pack_names()
1292
new_names = set([x[0][0] for x in new_nodes])
1293
self.assertEqual(names, sorted([x[0][0] for x in all_nodes]))
1294
self.assertEqual(set(names) - set(orig_names), new_names)
1295
self.assertEqual(set([new_pack.name]), new_names)
1296
self.assertEqual([to_remove_name],
1297
sorted([x[0][0] for x in deleted_nodes]))
1298
packs.reload_pack_names()
1299
reloaded_names = packs.names()
1300
self.assertEqual(orig_at_load, packs._packs_at_load)
1301
self.assertEqual(names, reloaded_names)
1302
all_nodes, deleted_nodes, new_nodes, _ = packs._diff_pack_names()
1303
new_names = set([x[0][0] for x in new_nodes])
1304
self.assertEqual(names, sorted([x[0][0] for x in all_nodes]))
1305
self.assertEqual(set(names) - set(orig_names), new_names)
1306
self.assertEqual(set([new_pack.name]), new_names)
1307
self.assertEqual([to_remove_name],
1308
sorted([x[0][0] for x in deleted_nodes]))
1310
def test_autopack_obsoletes_new_pack(self):
1311
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1312
packs._max_pack_count = lambda x: 1
1313
packs.pack_distribution = lambda x: [10]
1314
r.start_write_group()
1315
r.revisions.insert_record_stream([versionedfile.FulltextContentFactory(
1316
('bogus-rev',), (), None, 'bogus-content\n')])
1317
# This should trigger an autopack, which will combine everything into a
1319
new_names = r.commit_write_group()
1320
names = packs.names()
1321
self.assertEqual(1, len(names))
1322
self.assertEqual([names[0] + '.pack'],
1323
packs._pack_transport.list_dir('.'))
1258
1325
def test_autopack_reloads_and_stops(self):
1259
1326
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1260
1327
# After we have determined what needs to be autopacked, trigger a
1272
1339
self.assertEqual(tree.branch.repository._pack_collection.names(),
1342
def test__save_pack_names(self):
1343
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1344
names = packs.names()
1345
pack = packs.get_pack_by_name(names[0])
1346
packs._remove_pack_from_memory(pack)
1347
packs._save_pack_names(obsolete_packs=[pack])
1348
cur_packs = packs._pack_transport.list_dir('.')
1349
self.assertEqual([n + '.pack' for n in names[1:]], sorted(cur_packs))
1350
# obsolete_packs will also have stuff like .rix and .iix present.
1351
obsolete_packs = packs.transport.list_dir('obsolete_packs')
1352
obsolete_names = set([osutils.splitext(n)[0] for n in obsolete_packs])
1353
self.assertEqual([pack.name], sorted(obsolete_names))
1355
def test__save_pack_names_already_obsoleted(self):
1356
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1357
names = packs.names()
1358
pack = packs.get_pack_by_name(names[0])
1359
packs._remove_pack_from_memory(pack)
1360
# We are going to simulate a concurrent autopack by manually obsoleting
1361
# the pack directly.
1362
packs._obsolete_packs([pack])
1363
packs._save_pack_names(clear_obsolete_packs=True,
1364
obsolete_packs=[pack])
1365
cur_packs = packs._pack_transport.list_dir('.')
1366
self.assertEqual([n + '.pack' for n in names[1:]], sorted(cur_packs))
1367
# Note that while we set clear_obsolete_packs=True, it should not
1368
# delete a pack file that we have also scheduled for obsoletion.
1369
obsolete_packs = packs.transport.list_dir('obsolete_packs')
1370
obsolete_names = set([osutils.splitext(n)[0] for n in obsolete_packs])
1371
self.assertEqual([pack.name], sorted(obsolete_names))
1373
def test_pack_no_obsolete_packs_directory(self):
1374
"""Bug #314314, don't fail if obsolete_packs directory does
1376
tree, r, packs, revs = self.make_packs_and_alt_repo(write_lock=True)
1377
r.control_transport.rmdir('obsolete_packs')
1378
packs._clear_obsolete_packs()
1276
1381
class TestPack(TestCaseWithTransport):
1277
1382
"""Tests for the Pack object."""
1406
1513
self.assertTrue(new_pack.signature_index._optimize_for_size)
1516
class TestGCCHKPacker(TestCaseWithTransport):
1518
def make_abc_branch(self):
1519
builder = self.make_branch_builder('source')
1520
builder.start_series()
1521
builder.build_snapshot('A', None, [
1522
('add', ('', 'root-id', 'directory', None)),
1523
('add', ('file', 'file-id', 'file', 'content\n')),
1525
builder.build_snapshot('B', ['A'], [
1526
('add', ('dir', 'dir-id', 'directory', None))])
1527
builder.build_snapshot('C', ['B'], [
1528
('modify', ('file-id', 'new content\n'))])
1529
builder.finish_series()
1530
return builder.get_branch()
1532
def make_branch_with_disjoint_inventory_and_revision(self):
1533
"""a repo with separate packs for a revisions Revision and Inventory.
1535
There will be one pack file that holds the Revision content, and one
1536
for the Inventory content.
1538
:return: (repository,
1539
pack_name_with_rev_A_Revision,
1540
pack_name_with_rev_A_Inventory,
1541
pack_name_with_rev_C_content)
1543
b_source = self.make_abc_branch()
1544
b_base = b_source.bzrdir.sprout('base', revision_id='A').open_branch()
1545
b_stacked = b_base.bzrdir.sprout('stacked', stacked=True).open_branch()
1546
b_stacked.lock_write()
1547
self.addCleanup(b_stacked.unlock)
1548
b_stacked.fetch(b_source, 'B')
1549
# Now re-open the stacked repo directly (no fallbacks) so that we can
1550
# fill in the A rev.
1551
repo_not_stacked = b_stacked.bzrdir.open_repository()
1552
repo_not_stacked.lock_write()
1553
self.addCleanup(repo_not_stacked.unlock)
1554
# Now we should have a pack file with A's inventory, but not its
1556
self.assertEqual([('A',), ('B',)],
1557
sorted(repo_not_stacked.inventories.keys()))
1558
self.assertEqual([('B',)],
1559
sorted(repo_not_stacked.revisions.keys()))
1560
stacked_pack_names = repo_not_stacked._pack_collection.names()
1561
# We have a couple names here, figure out which has A's inventory
1562
for name in stacked_pack_names:
1563
pack = repo_not_stacked._pack_collection.get_pack_by_name(name)
1564
keys = [n[1] for n in pack.inventory_index.iter_all_entries()]
1566
inv_a_pack_name = name
1569
self.fail('Could not find pack containing A\'s inventory')
1570
repo_not_stacked.fetch(b_source.repository, 'A')
1571
self.assertEqual([('A',), ('B',)],
1572
sorted(repo_not_stacked.revisions.keys()))
1573
new_pack_names = set(repo_not_stacked._pack_collection.names())
1574
rev_a_pack_names = new_pack_names.difference(stacked_pack_names)
1575
self.assertEqual(1, len(rev_a_pack_names))
1576
rev_a_pack_name = list(rev_a_pack_names)[0]
1577
# Now fetch 'C', so we have a couple pack files to join
1578
repo_not_stacked.fetch(b_source.repository, 'C')
1579
rev_c_pack_names = set(repo_not_stacked._pack_collection.names())
1580
rev_c_pack_names = rev_c_pack_names.difference(new_pack_names)
1581
self.assertEqual(1, len(rev_c_pack_names))
1582
rev_c_pack_name = list(rev_c_pack_names)[0]
1583
return (repo_not_stacked, rev_a_pack_name, inv_a_pack_name,
1586
def test_pack_with_distant_inventories(self):
1587
# See https://bugs.launchpad.net/bzr/+bug/437003
1588
# When repacking, it is possible to have an inventory in a different
1589
# pack file than the associated revision. An autopack can then come
1590
# along, and miss that inventory, and complain.
1591
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1592
) = self.make_branch_with_disjoint_inventory_and_revision()
1593
a_pack = repo._pack_collection.get_pack_by_name(rev_a_pack_name)
1594
c_pack = repo._pack_collection.get_pack_by_name(rev_c_pack_name)
1595
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1596
[a_pack, c_pack], '.test-pack')
1597
# This would raise ValueError in bug #437003, but should not raise an
1601
def test_pack_with_missing_inventory(self):
1602
# Similar to test_pack_with_missing_inventory, but this time, we force
1603
# the A inventory to actually be gone from the repository.
1604
(repo, rev_a_pack_name, inv_a_pack_name, rev_c_pack_name
1605
) = self.make_branch_with_disjoint_inventory_and_revision()
1606
inv_a_pack = repo._pack_collection.get_pack_by_name(inv_a_pack_name)
1607
repo._pack_collection._remove_pack_from_memory(inv_a_pack)
1608
packer = groupcompress_repo.GCCHKPacker(repo._pack_collection,
1609
repo._pack_collection.all_packs(), '.test-pack')
1610
e = self.assertRaises(ValueError, packer.pack)
1611
packer.new_pack.abort()
1612
self.assertContainsRe(str(e),
1613
r"We are missing inventories for revisions: .*'A'")
1409
1616
class TestCrossFormatPacks(TestCaseWithTransport):
1411
1618
def log_pack(self, hint=None):