47
48
def fetch_steps(self, br_a, br_b, writable_a):
48
49
"""A foreign test method for testing fetch locally and remotely."""
50
51
# TODO RBC 20060201 make this a repository test.
51
52
repo_b = br_b.repository
52
53
self.assertFalse(repo_b.has_revision(br_a.revision_history()[3]))
53
54
self.assertTrue(repo_b.has_revision(br_a.revision_history()[2]))
54
55
self.assertEquals(len(br_b.revision_history()), 7)
55
self.assertEquals(br_b.fetch(br_a, br_a.revision_history()[2])[0], 0)
56
br_b.fetch(br_a, br_a.revision_history()[2])
56
57
# branch.fetch is not supposed to alter the revision history
57
58
self.assertEquals(len(br_b.revision_history()), 7)
58
59
self.assertFalse(repo_b.has_revision(br_a.revision_history()[3]))
60
61
# fetching the next revision up in sample data copies one revision
61
self.assertEquals(br_b.fetch(br_a, br_a.revision_history()[3])[0], 1)
62
br_b.fetch(br_a, br_a.revision_history()[3])
62
63
self.assertTrue(repo_b.has_revision(br_a.revision_history()[3]))
63
64
self.assertFalse(has_revision(br_a, br_b.revision_history()[6]))
64
65
self.assertTrue(br_a.repository.has_revision(br_b.revision_history()[5]))
66
67
# When a non-branch ancestor is missing, it should be unlisted...
67
68
# as its not reference from the inventory weave.
68
69
br_b4 = self.make_branch('br_4')
69
count, failures = br_b4.fetch(br_b)
70
self.assertEqual(count, 7)
71
self.assertEqual(failures, [])
73
self.assertEqual(writable_a.fetch(br_b)[0], 1)
72
writable_a.fetch(br_b)
74
73
self.assertTrue(has_revision(br_a, br_b.revision_history()[3]))
75
74
self.assertTrue(has_revision(br_a, br_b.revision_history()[4]))
77
76
br_b2 = self.make_branch('br_b2')
78
self.assertEquals(br_b2.fetch(br_b)[0], 7)
79
78
self.assertTrue(has_revision(br_b2, br_b.revision_history()[4]))
80
79
self.assertTrue(has_revision(br_b2, br_a.revision_history()[2]))
81
80
self.assertFalse(has_revision(br_b2, br_a.revision_history()[3]))
83
82
br_a2 = self.make_branch('br_a2')
84
self.assertEquals(br_a2.fetch(br_a)[0], 9)
85
84
self.assertTrue(has_revision(br_a2, br_b.revision_history()[4]))
86
85
self.assertTrue(has_revision(br_a2, br_a.revision_history()[3]))
87
86
self.assertTrue(has_revision(br_a2, br_a.revision_history()[2]))
89
88
br_a3 = self.make_branch('br_a3')
90
# pulling a branch with no revisions grabs nothing, regardless of
89
# pulling a branch with no revisions grabs nothing, regardless of
91
90
# whats in the inventory.
92
self.assertEquals(br_a3.fetch(br_a2)[0], 0)
93
92
for revno in range(4):
95
94
br_a3.repository.has_revision(br_a.revision_history()[revno]))
96
self.assertEqual(br_a3.fetch(br_a2, br_a.revision_history()[2])[0], 3)
95
br_a3.fetch(br_a2, br_a.revision_history()[2])
97
96
# pull the 3 revisions introduced by a@u-0-3
98
fetched = br_a3.fetch(br_a2, br_a.revision_history()[3])[0]
99
self.assertEquals(fetched, 3, "fetched %d instead of 3" % fetched)
100
# InstallFailed should be raised if the branch is missing the revision
97
br_a3.fetch(br_a2, br_a.revision_history()[3])
98
# NoSuchRevision should be raised if the branch is missing the revision
101
99
# that was requested.
102
self.assertRaises(errors.InstallFailed, br_a3.fetch, br_a2, 'pizza')
100
self.assertRaises(errors.NoSuchRevision, br_a3.fetch, br_a2, 'pizza')
104
102
# TODO: Test trying to fetch from a branch that points to a revision not
105
103
# actually present in its repository. Not every branch format allows you
283
281
wt.commit("changed file")
284
282
target = BzrDir.create_branch_and_repo("target/")
285
283
source = Branch.open(self.get_readonly_url("source/"))
286
self.assertEqual(target.fetch(source), (2, []))
287
# this is the path to the literal file. As format changes
285
# this is the path to the literal file. As format changes
288
286
# occur it needs to be updated. FIXME: ask the store for the
290
288
self.log("web server logs are:")
291
289
http_logs = self.get_readonly_server().logs
292
290
self.log('\n'.join(http_logs))
293
# unfortunately this log entry is branch format specific. We could
294
# factor out the 'what files does this format use' to a method on the
291
# unfortunately this log entry is branch format specific. We could
292
# factor out the 'what files does this format use' to a method on the
295
293
# repository, which would let us to this generically. RBC 20060419
296
294
# RBC 20080408: Or perhaps we can assert that no files are fully read
298
296
self.assertEqual(1, self._count_log_matches('/ce/id.kndx', http_logs))
299
297
self.assertEqual(1, self._count_log_matches('/ce/id.knit', http_logs))
300
298
self.assertEqual(1, self._count_log_matches('inventory.kndx', http_logs))
301
# this r-h check test will prevent regressions, but it currently already
299
# this r-h check test will prevent regressions, but it currently already
302
300
# passes, before the patch to cache-rh is applied :[
303
301
self.assertTrue(1 >= self._count_log_matches('revision-history',
323
321
self.assertEqual(1, self._count_log_matches('branch/format', http_logs))
324
322
self.assertEqual(1, self._count_log_matches('repository/format',
324
self.assertEqual(1, self._count_log_matches('revisions.kndx',
326
326
self.assertTrue(1 >= self._count_log_matches('revision-history',
328
328
self.assertTrue(1 >= self._count_log_matches('last-revision',
330
self.assertEqual(4, len(http_logs))
330
self.assertLength(5, http_logs)
333
333
class TestKnitToPackFetch(TestCaseWithTransport):
335
def find_get_record_stream(self, calls):
336
"""In a list of calls, find 'get_record_stream' calls.
335
def find_get_record_stream(self, calls, expected_count=1):
336
"""In a list of calls, find the last 'get_record_stream'.
338
This also ensures that there is only one get_record_stream call.
338
:param expected_count: The number of calls we should exepect to find.
339
If a different number is found, an assertion is raised.
340
341
get_record_call = None
341
343
for call in calls:
342
344
if call[0] == 'get_record_stream':
343
self.assertIs(None, get_record_call,
344
"there should only be one call to"
345
" get_record_stream")
346
346
get_record_call = call
347
self.assertIsNot(None, get_record_call,
348
"there should be exactly one call to "
349
" get_record_stream")
347
self.assertEqual(expected_count, call_count)
350
348
return get_record_call
352
350
def test_fetch_with_deltas_no_delta_closure(self):
366
364
source.inventories = versionedfile.RecordingVersionedFilesDecorator(
367
365
source.inventories)
369
self.assertTrue(target._fetch_uses_deltas)
367
self.assertTrue(target._format._fetch_uses_deltas)
370
368
target.fetch(source, revision_id='rev-one')
371
369
self.assertEqual(('get_record_stream', [('file-id', 'rev-one')],
372
target._fetch_order, False),
370
target._format._fetch_order, False),
373
371
self.find_get_record_stream(source.texts.calls))
374
372
self.assertEqual(('get_record_stream', [('rev-one',)],
375
target._fetch_order, False),
376
self.find_get_record_stream(source.inventories.calls))
377
# Because of bugs in the old fetch code, revisions could accidentally
378
# have deltas present in knits. However, it was never intended, so we
379
# always for include_delta_closure=True, to make sure we get fulltexts.
373
target._format._fetch_order, False),
374
self.find_get_record_stream(source.inventories.calls, 2))
381
375
self.assertEqual(('get_record_stream', [('rev-one',)],
382
target._fetch_order, True),
376
target._format._fetch_order, False),
383
377
self.find_get_record_stream(source.revisions.calls))
384
378
# XXX: Signatures is special, and slightly broken. The
385
379
# standard item_keys_introduced_by actually does a lookup for every
409
403
source.revisions)
410
404
source.inventories = versionedfile.RecordingVersionedFilesDecorator(
411
405
source.inventories)
412
target._fetch_uses_deltas = False
406
# XXX: This won't work in general, but for the dirstate format it does.
407
old_fetch_uses_deltas_setting = target._format._fetch_uses_deltas
409
target._format._fetch_uses_deltas = old_fetch_uses_deltas_setting
410
self.addCleanup(restore)
411
target._format._fetch_uses_deltas = False
413
412
target.fetch(source, revision_id='rev-one')
414
413
self.assertEqual(('get_record_stream', [('file-id', 'rev-one')],
415
target._fetch_order, True),
414
target._format._fetch_order, True),
416
415
self.find_get_record_stream(source.texts.calls))
417
416
self.assertEqual(('get_record_stream', [('rev-one',)],
418
target._fetch_order, True),
419
self.find_get_record_stream(source.inventories.calls))
417
target._format._fetch_order, True),
418
self.find_get_record_stream(source.inventories.calls, 2))
420
419
self.assertEqual(('get_record_stream', [('rev-one',)],
421
target._fetch_order, True),
420
target._format._fetch_order, True),
422
421
self.find_get_record_stream(source.revisions.calls))
423
422
# XXX: Signatures is special, and slightly broken. The
424
423
# standard item_keys_introduced_by actually does a lookup for every
430
429
signature_calls = source.signatures.calls[-1:]
431
430
self.assertEqual(('get_record_stream', [('rev-one',)],
432
target._fetch_order, True),
431
target._format._fetch_order, True),
433
432
self.find_get_record_stream(signature_calls))
434
def test_fetch_revisions_with_deltas_into_pack(self):
435
# See BUG #261339, dev versions of bzr could accidentally create deltas
436
# in revision texts in knit branches (when fetching from packs). So we
437
# ensure that *if* a knit repository has a delta in revisions, that it
438
# gets properly expanded back into a fulltext when stored in the pack
440
tree = self.make_branch_and_tree('source', format='dirstate')
441
target = self.make_repository('target', format='pack-0.92')
442
self.build_tree(['source/file'])
443
tree.set_root_id('root-id')
444
tree.add('file', 'file-id')
445
tree.commit('one', rev_id='rev-one')
446
# Hack the KVF for revisions so that it "accidentally" allows a delta
447
tree.branch.repository.revisions._max_delta_chain = 200
448
tree.commit('two', rev_id='rev-two')
449
source = tree.branch.repository
450
# Ensure that we stored a delta
452
self.addCleanup(source.unlock)
453
record = source.revisions.get_record_stream([('rev-two',)],
454
'unordered', False).next()
455
self.assertEqual('knit-delta-gz', record.storage_kind)
456
target.fetch(tree.branch.repository, revision_id='rev-two')
457
# The record should get expanded back to a fulltext
459
self.addCleanup(target.unlock)
460
record = target.revisions.get_record_stream([('rev-two',)],
461
'unordered', False).next()
462
self.assertEqual('knit-ft-gz', record.storage_kind)
464
def test_fetch_with_fallback_and_merge(self):
465
builder = self.make_branch_builder('source', format='pack-0.92')
466
builder.start_series()
477
# A & B are present in the base (stacked-on) repository, A-E are
478
# present in the source.
479
# This reproduces bug #304841
480
# We need a large enough inventory that total size of compressed deltas
481
# is shorter than the size of a compressed fulltext. We have to use
482
# random ids because otherwise the inventory fulltext compresses too
483
# well and the deltas get bigger.
485
('add', ('', 'TREE_ROOT', 'directory', None))]
487
fname = 'file%03d' % (i,)
488
fileid = '%s-%s' % (fname, osutils.rand_chars(64))
489
to_add.append(('add', (fname, fileid, 'file', 'content\n')))
490
builder.build_snapshot('A', None, to_add)
491
builder.build_snapshot('B', ['A'], [])
492
builder.build_snapshot('C', ['A'], [])
493
builder.build_snapshot('D', ['C'], [])
494
builder.build_snapshot('E', ['D'], [])
495
builder.build_snapshot('F', ['E', 'B'], [])
496
builder.finish_series()
497
source_branch = builder.get_branch()
498
source_branch.bzrdir.sprout('base', revision_id='B')
499
target_branch = self.make_branch('target', format='1.6')
500
target_branch.set_stacked_on_url('../base')
501
source = source_branch.repository
503
self.addCleanup(source.unlock)
504
source.inventories = versionedfile.OrderingVersionedFilesDecorator(
506
key_priority={('E',): 1, ('D',): 2, ('C',): 4,
508
# Ensure that the content is yielded in the proper order, and given as
510
records = [(record.key, record.storage_kind)
511
for record in source.inventories.get_record_stream(
512
[('D',), ('C',), ('E',), ('F',)], 'unordered', False)]
513
self.assertEqual([(('E',), 'knit-delta-gz'), (('D',), 'knit-delta-gz'),
514
(('F',), 'knit-delta-gz'), (('C',), 'knit-delta-gz')],
517
target_branch.lock_write()
518
self.addCleanup(target_branch.unlock)
519
target = target_branch.repository
520
target.fetch(source, revision_id='F')
521
# 'C' should be expanded to a fulltext, but D and E should still be
523
stream = target.inventories.get_record_stream(
524
[('C',), ('D',), ('E',), ('F',)],
526
kinds = dict((record.key, record.storage_kind) for record in stream)
527
self.assertEqual({('C',): 'knit-ft-gz', ('D',): 'knit-delta-gz',
528
('E',): 'knit-delta-gz', ('F',): 'knit-delta-gz'},
436
532
class Test1To2Fetch(TestCaseWithTransport):
437
533
"""Tests for Model1To2 failure modes"""