57
53
# When a non-branch ancestor is missing, it should be unlisted...
58
54
# as its not reference from the inventory weave.
59
55
br_b4 = self.make_branch('br_4')
56
count, failures = br_b4.fetch(br_b)
57
self.assertEqual(count, 7)
58
self.assertEqual(failures, [])
62
writable_a.fetch(br_b)
60
self.assertEqual(writable_a.fetch(br_b)[0], 1)
63
61
self.assertTrue(has_revision(br_a, br_b.revision_history()[3]))
64
62
self.assertTrue(has_revision(br_a, br_b.revision_history()[4]))
66
64
br_b2 = self.make_branch('br_b2')
65
self.assertEquals(br_b2.fetch(br_b)[0], 7)
68
66
self.assertTrue(has_revision(br_b2, br_b.revision_history()[4]))
69
67
self.assertTrue(has_revision(br_b2, br_a.revision_history()[2]))
70
68
self.assertFalse(has_revision(br_b2, br_a.revision_history()[3]))
72
70
br_a2 = self.make_branch('br_a2')
71
self.assertEquals(br_a2.fetch(br_a)[0], 9)
74
72
self.assertTrue(has_revision(br_a2, br_b.revision_history()[4]))
75
73
self.assertTrue(has_revision(br_a2, br_a.revision_history()[3]))
76
74
self.assertTrue(has_revision(br_a2, br_a.revision_history()[2]))
78
76
br_a3 = self.make_branch('br_a3')
79
# pulling a branch with no revisions grabs nothing, regardless of
77
# pulling a branch with no revisions grabs nothing, regardless of
80
78
# whats in the inventory.
79
self.assertEquals(br_a3.fetch(br_a2)[0], 0)
82
80
for revno in range(4):
84
82
br_a3.repository.has_revision(br_a.revision_history()[revno]))
85
br_a3.fetch(br_a2, br_a.revision_history()[2])
83
self.assertEqual(br_a3.fetch(br_a2, br_a.revision_history()[2])[0], 3)
86
84
# pull the 3 revisions introduced by a@u-0-3
87
br_a3.fetch(br_a2, br_a.revision_history()[3])
88
# NoSuchRevision should be raised if the branch is missing the revision
85
fetched = br_a3.fetch(br_a2, br_a.revision_history()[3])[0]
86
self.assertEquals(fetched, 3, "fetched %d instead of 3" % fetched)
87
# InstallFailed should be raised if the branch is missing the revision
89
88
# that was requested.
90
self.assertRaises(errors.NoSuchRevision, br_a3.fetch, br_a2, 'pizza')
89
self.assertRaises(bzrlib.errors.InstallFailed, br_a3.fetch, br_a2, 'pizza')
90
# InstallFailed should be raised if the branch is missing a revision
91
# from its own revision history
92
br_a2.append_revision('a-b-c')
93
self.assertRaises(bzrlib.errors.InstallFailed, br_a3.fetch, br_a2)
92
# TODO: Test trying to fetch from a branch that points to a revision not
93
# actually present in its repository. Not every branch format allows you
94
# to directly point to such revisions, so it's a bit complicated to
95
# construct. One way would be to uncommit and gc the revision, but not
96
# every branch supports that. -- mbp 20070814
95
# TODO: jam 20051218 Branch should no longer allow append_revision for revisions
96
# which don't exist. So this test needs to be rewritten
97
# RBC 20060403 the way to do this is to uncommit the revision from the
98
# repository after the commit
98
100
#TODO: test that fetch correctly does reweaving when needed. RBC 20051008
99
# Note that this means - updating the weave when ghosts are filled in to
101
# Note that this means - updating the weave when ghosts are filled in to
100
102
# add the right parents.
238
185
rev_id).get_file_text('this-file-id'), text)
241
class TestKnitToPackFetch(TestCaseWithTransport):
243
def find_get_record_stream(self, calls, expected_count=1):
244
"""In a list of calls, find the last 'get_record_stream'.
246
:param expected_count: The number of calls we should exepect to find.
247
If a different number is found, an assertion is raised.
249
get_record_call = None
252
if call[0] == 'get_record_stream':
254
get_record_call = call
255
self.assertEqual(expected_count, call_count)
256
return get_record_call
258
def test_fetch_with_deltas_no_delta_closure(self):
259
tree = self.make_branch_and_tree('source', format='dirstate')
260
target = self.make_repository('target', format='pack-0.92')
261
self.build_tree(['source/file'])
262
tree.set_root_id('root-id')
263
tree.add('file', 'file-id')
264
tree.commit('one', rev_id='rev-one')
265
source = tree.branch.repository
266
source.texts = versionedfile.RecordingVersionedFilesDecorator(
268
source.signatures = versionedfile.RecordingVersionedFilesDecorator(
270
source.revisions = versionedfile.RecordingVersionedFilesDecorator(
272
source.inventories = versionedfile.RecordingVersionedFilesDecorator(
275
self.assertTrue(target._format._fetch_uses_deltas)
276
target.fetch(source, revision_id='rev-one')
277
self.assertEqual(('get_record_stream', [('file-id', 'rev-one')],
278
target._format._fetch_order, False),
279
self.find_get_record_stream(source.texts.calls))
280
self.assertEqual(('get_record_stream', [('rev-one',)],
281
target._format._fetch_order, False),
282
self.find_get_record_stream(source.inventories.calls, 2))
283
self.assertEqual(('get_record_stream', [('rev-one',)],
284
target._format._fetch_order, False),
285
self.find_get_record_stream(source.revisions.calls))
286
# XXX: Signatures is special, and slightly broken. The
287
# standard item_keys_introduced_by actually does a lookup for every
288
# signature to see if it exists, rather than waiting to do them all at
289
# once at the end. The fetch code then does an all-at-once and just
290
# allows for some of them to be missing.
291
# So we know there will be extra calls, but the *last* one is the one
293
signature_calls = source.signatures.calls[-1:]
294
self.assertEqual(('get_record_stream', [('rev-one',)],
295
target._format._fetch_order, False),
296
self.find_get_record_stream(signature_calls))
298
def test_fetch_no_deltas_with_delta_closure(self):
299
tree = self.make_branch_and_tree('source', format='dirstate')
300
target = self.make_repository('target', format='pack-0.92')
301
self.build_tree(['source/file'])
302
tree.set_root_id('root-id')
303
tree.add('file', 'file-id')
304
tree.commit('one', rev_id='rev-one')
305
source = tree.branch.repository
306
source.texts = versionedfile.RecordingVersionedFilesDecorator(
308
source.signatures = versionedfile.RecordingVersionedFilesDecorator(
310
source.revisions = versionedfile.RecordingVersionedFilesDecorator(
312
source.inventories = versionedfile.RecordingVersionedFilesDecorator(
314
# XXX: This won't work in general, but for the dirstate format it does.
315
self.overrideAttr(target._format, '_fetch_uses_deltas', False)
316
target.fetch(source, revision_id='rev-one')
317
self.assertEqual(('get_record_stream', [('file-id', 'rev-one')],
318
target._format._fetch_order, True),
319
self.find_get_record_stream(source.texts.calls))
320
self.assertEqual(('get_record_stream', [('rev-one',)],
321
target._format._fetch_order, True),
322
self.find_get_record_stream(source.inventories.calls, 2))
323
self.assertEqual(('get_record_stream', [('rev-one',)],
324
target._format._fetch_order, True),
325
self.find_get_record_stream(source.revisions.calls))
326
# XXX: Signatures is special, and slightly broken. The
327
# standard item_keys_introduced_by actually does a lookup for every
328
# signature to see if it exists, rather than waiting to do them all at
329
# once at the end. The fetch code then does an all-at-once and just
330
# allows for some of them to be missing.
331
# So we know there will be extra calls, but the *last* one is the one
333
signature_calls = source.signatures.calls[-1:]
334
self.assertEqual(('get_record_stream', [('rev-one',)],
335
target._format._fetch_order, True),
336
self.find_get_record_stream(signature_calls))
338
def test_fetch_revisions_with_deltas_into_pack(self):
339
# See BUG #261339, dev versions of bzr could accidentally create deltas
340
# in revision texts in knit branches (when fetching from packs). So we
341
# ensure that *if* a knit repository has a delta in revisions, that it
342
# gets properly expanded back into a fulltext when stored in the pack
344
tree = self.make_branch_and_tree('source', format='dirstate')
345
target = self.make_repository('target', format='pack-0.92')
346
self.build_tree(['source/file'])
347
tree.set_root_id('root-id')
348
tree.add('file', 'file-id')
349
tree.commit('one', rev_id='rev-one')
350
# Hack the KVF for revisions so that it "accidentally" allows a delta
351
tree.branch.repository.revisions._max_delta_chain = 200
352
tree.commit('two', rev_id='rev-two')
353
source = tree.branch.repository
354
# Ensure that we stored a delta
356
self.addCleanup(source.unlock)
357
record = source.revisions.get_record_stream([('rev-two',)],
358
'unordered', False).next()
359
self.assertEqual('knit-delta-gz', record.storage_kind)
360
target.fetch(tree.branch.repository, revision_id='rev-two')
361
# The record should get expanded back to a fulltext
363
self.addCleanup(target.unlock)
364
record = target.revisions.get_record_stream([('rev-two',)],
365
'unordered', False).next()
366
self.assertEqual('knit-ft-gz', record.storage_kind)
368
def test_fetch_with_fallback_and_merge(self):
369
builder = self.make_branch_builder('source', format='pack-0.92')
370
builder.start_series()
381
# A & B are present in the base (stacked-on) repository, A-E are
382
# present in the source.
383
# This reproduces bug #304841
384
# We need a large enough inventory that total size of compressed deltas
385
# is shorter than the size of a compressed fulltext. We have to use
386
# random ids because otherwise the inventory fulltext compresses too
387
# well and the deltas get bigger.
389
('add', ('', 'TREE_ROOT', 'directory', None))]
391
fname = 'file%03d' % (i,)
392
fileid = '%s-%s' % (fname, osutils.rand_chars(64))
393
to_add.append(('add', (fname, fileid, 'file', 'content\n')))
394
builder.build_snapshot('A', None, to_add)
395
builder.build_snapshot('B', ['A'], [])
396
builder.build_snapshot('C', ['A'], [])
397
builder.build_snapshot('D', ['C'], [])
398
builder.build_snapshot('E', ['D'], [])
399
builder.build_snapshot('F', ['E', 'B'], [])
400
builder.finish_series()
401
source_branch = builder.get_branch()
402
source_branch.bzrdir.sprout('base', revision_id='B')
403
target_branch = self.make_branch('target', format='1.6')
404
target_branch.set_stacked_on_url('../base')
405
source = source_branch.repository
407
self.addCleanup(source.unlock)
408
source.inventories = versionedfile.OrderingVersionedFilesDecorator(
410
key_priority={('E',): 1, ('D',): 2, ('C',): 4,
412
# Ensure that the content is yielded in the proper order, and given as
414
records = [(record.key, record.storage_kind)
415
for record in source.inventories.get_record_stream(
416
[('D',), ('C',), ('E',), ('F',)], 'unordered', False)]
417
self.assertEqual([(('E',), 'knit-delta-gz'), (('D',), 'knit-delta-gz'),
418
(('F',), 'knit-delta-gz'), (('C',), 'knit-delta-gz')],
421
target_branch.lock_write()
422
self.addCleanup(target_branch.unlock)
423
target = target_branch.repository
424
target.fetch(source, revision_id='F')
425
# 'C' should be expanded to a fulltext, but D and E should still be
427
stream = target.inventories.get_record_stream(
428
[('C',), ('D',), ('E',), ('F',)],
430
kinds = dict((record.key, record.storage_kind) for record in stream)
431
self.assertEqual({('C',): 'knit-ft-gz', ('D',): 'knit-delta-gz',
432
('E',): 'knit-delta-gz', ('F',): 'knit-delta-gz'},
436
class Test1To2Fetch(TestCaseWithTransport):
437
"""Tests for Model1To2 failure modes"""
439
def make_tree_and_repo(self):
440
self.tree = self.make_branch_and_tree('tree', format='pack-0.92')
441
self.repo = self.make_repository('rich-repo', format='rich-root-pack')
442
self.repo.lock_write()
443
self.addCleanup(self.repo.unlock)
445
def do_fetch_order_test(self, first, second):
446
"""Test that fetch works no matter what the set order of revision is.
448
This test depends on the order of items in a set, which is
449
implementation-dependant, so we test A, B and then B, A.
451
self.make_tree_and_repo()
452
self.tree.commit('Commit 1', rev_id=first)
453
self.tree.commit('Commit 2', rev_id=second)
454
self.repo.fetch(self.tree.branch.repository, second)
456
def test_fetch_order_AB(self):
457
"""See do_fetch_order_test"""
458
self.do_fetch_order_test('A', 'B')
460
def test_fetch_order_BA(self):
461
"""See do_fetch_order_test"""
462
self.do_fetch_order_test('B', 'A')
464
def get_parents(self, file_id, revision_id):
465
self.repo.lock_read()
467
parent_map = self.repo.texts.get_parent_map([(file_id, revision_id)])
468
return parent_map[(file_id, revision_id)]
472
def test_fetch_ghosts(self):
473
self.make_tree_and_repo()
474
self.tree.commit('first commit', rev_id='left-parent')
475
self.tree.add_parent_tree_id('ghost-parent')
476
fork = self.tree.bzrdir.sprout('fork', 'null:').open_workingtree()
477
fork.commit('not a ghost', rev_id='not-ghost-parent')
478
self.tree.branch.repository.fetch(fork.branch.repository,
480
self.tree.add_parent_tree_id('not-ghost-parent')
481
self.tree.commit('second commit', rev_id='second-id')
482
self.repo.fetch(self.tree.branch.repository, 'second-id')
483
root_id = self.tree.get_root_id()
485
((root_id, 'left-parent'), (root_id, 'not-ghost-parent')),
486
self.get_parents(root_id, 'second-id'))
488
def make_two_commits(self, change_root, fetch_twice):
489
self.make_tree_and_repo()
490
self.tree.commit('first commit', rev_id='first-id')
492
self.tree.set_root_id('unique-id')
493
self.tree.commit('second commit', rev_id='second-id')
495
self.repo.fetch(self.tree.branch.repository, 'first-id')
496
self.repo.fetch(self.tree.branch.repository, 'second-id')
498
def test_fetch_changed_root(self):
499
self.make_two_commits(change_root=True, fetch_twice=False)
500
self.assertEqual((), self.get_parents('unique-id', 'second-id'))
502
def test_two_fetch_changed_root(self):
503
self.make_two_commits(change_root=True, fetch_twice=True)
504
self.assertEqual((), self.get_parents('unique-id', 'second-id'))
506
def test_two_fetches(self):
507
self.make_two_commits(change_root=False, fetch_twice=True)
508
self.assertEqual((('TREE_ROOT', 'first-id'),),
509
self.get_parents('TREE_ROOT', 'second-id'))
188
class TestHttpFetch(TestCaseWithWebserver):
189
# FIXME RBC 20060124 this really isn't web specific, perhaps an
190
# instrumented readonly transport? Can we do an instrumented
191
# adapter and use self.get_readonly_url ?
193
def test_fetch(self):
194
#highest indices a: 5, b: 7
195
br_a, br_b = make_branches(self)
196
br_rem_a = Branch.open(self.get_readonly_url('branch1'))
197
fetch_steps(self, br_rem_a, br_b, br_a)
199
def _count_log_matches(self, target, logs):
200
"""Count the number of times the target file pattern was fetched in an http log"""
201
log_pattern = '%s HTTP/1.1" 200 - "-" "bzr/%s' % \
202
(target, bzrlib.__version__)
205
# TODO: perhaps use a regexp instead so we can match more
207
if line.find(log_pattern) > -1:
211
def test_weaves_are_retrieved_once(self):
212
self.build_tree(("source/", "source/file", "target/"))
213
wt = self.make_branch_and_tree('source')
215
wt.add(["file"], ["id"])
216
wt.commit("added file")
217
print >>open("source/file", 'w'), "blah"
218
wt.commit("changed file")
219
target = BzrDir.create_branch_and_repo("target/")
220
source = Branch.open(self.get_readonly_url("source/"))
221
self.assertEqual(target.fetch(source), (2, []))
222
log_pattern = '%%s HTTP/1.1" 200 - "-" "bzr/%s' % bzrlib.__version__
223
# this is the path to the literal file. As format changes
224
# occur it needs to be updated. FIXME: ask the store for the
226
self.log("web server logs are:")
227
http_logs = self.get_readonly_server().logs
228
self.log('\n'.join(http_logs))
229
# unfortunately this log entry is branch format specific. We could
230
# factor out the 'what files does this format use' to a method on the
231
# repository, which would let us to this generically. RBC 20060419
232
self.assertEqual(1, self._count_log_matches('/ce/id.kndx', http_logs))
233
self.assertEqual(1, self._count_log_matches('/ce/id.knit', http_logs))
234
self.assertEqual(1, self._count_log_matches('inventory.kndx', http_logs))
235
self.assertEqual(1, self._count_log_matches('inventory.knit', http_logs))
236
# this r-h check test will prevent regressions, but it currently already
237
# passes, before the patch to cache-rh is applied :[
238
self.assertEqual(1, self._count_log_matches('revision-history', http_logs))
239
# FIXME naughty poking in there.
240
self.get_readonly_server().logs = []
241
# check there is nothing more to fetch
242
source = Branch.open(self.get_readonly_url("source/"))
243
self.assertEqual(target.fetch(source), (0, []))
244
# should make just two requests
245
http_logs = self.get_readonly_server().logs
246
self.log("web server logs are:")
247
self.log('\n'.join(http_logs))
248
self.assertEqual(1, self._count_log_matches('branch-format', http_logs))
249
self.assertEqual(1, self._count_log_matches('branch/format', http_logs))
250
self.assertEqual(1, self._count_log_matches('repository/format', http_logs))
251
self.assertEqual(1, self._count_log_matches('revision-history', http_logs))
252
self.assertEqual(4, len(http_logs))