86
71
class Merger(object):
87
def __init__(self, this_branch, other_tree=None, base_tree=None,
88
this_tree=None, pb=DummyProgress()):
72
def __init__(self, this_branch, other_tree=None, base_tree=None,
73
this_tree=None, pb=None, change_reporter=None,
74
recurse='down', revision_graph=None):
89
75
object.__init__(self)
90
assert this_tree is not None, "this_tree is required"
91
76
self.this_branch = this_branch
92
self.this_basis = this_branch.last_revision()
77
self.this_basis = _mod_revision.ensure_null(
78
this_branch.last_revision())
93
79
self.this_rev_id = None
94
80
self.this_tree = this_tree
95
81
self.this_revision_tree = None
96
82
self.this_basis_tree = None
97
83
self.other_tree = other_tree
84
self.other_branch = None
98
85
self.base_tree = base_tree
99
86
self.ignore_zero = False
100
87
self.backup_files = False
101
88
self.interesting_ids = None
89
self.interesting_files = None
102
90
self.show_base = False
103
91
self.reprocess = False
108
def revision_tree(self, revision_id):
109
return self.this_branch.repository.revision_tree(revision_id)
96
self.recurse = recurse
97
self.change_reporter = change_reporter
98
self._cached_trees = {}
99
self._revision_graph = revision_graph
100
self._base_is_ancestor = None
101
self._base_is_other_ancestor = None
102
self._is_criss_cross = None
103
self._lca_trees = None
106
def revision_graph(self):
107
if self._revision_graph is None:
108
self._revision_graph = self.this_branch.repository.get_graph()
109
return self._revision_graph
111
def _set_base_is_ancestor(self, value):
112
self._base_is_ancestor = value
114
def _get_base_is_ancestor(self):
115
if self._base_is_ancestor is None:
116
self._base_is_ancestor = self.revision_graph.is_ancestor(
117
self.base_rev_id, self.this_basis)
118
return self._base_is_ancestor
120
base_is_ancestor = property(_get_base_is_ancestor, _set_base_is_ancestor)
122
def _set_base_is_other_ancestor(self, value):
123
self._base_is_other_ancestor = value
125
def _get_base_is_other_ancestor(self):
126
if self._base_is_other_ancestor is None:
127
if self.other_basis is None:
129
self._base_is_other_ancestor = self.revision_graph.is_ancestor(
130
self.base_rev_id, self.other_basis)
131
return self._base_is_other_ancestor
133
base_is_other_ancestor = property(_get_base_is_other_ancestor,
134
_set_base_is_other_ancestor)
137
def from_uncommitted(tree, other_tree, pb=None, base_tree=None):
138
"""Return a Merger for uncommitted changes in other_tree.
140
:param tree: The tree to merge into
141
:param other_tree: The tree to get uncommitted changes from
142
:param pb: A progress indicator
143
:param base_tree: The basis to use for the merge. If unspecified,
144
other_tree.basis_tree() will be used.
146
if base_tree is None:
147
base_tree = other_tree.basis_tree()
148
merger = Merger(tree.branch, other_tree, base_tree, tree, pb)
149
merger.base_rev_id = merger.base_tree.get_revision_id()
150
merger.other_rev_id = None
151
merger.other_basis = merger.base_rev_id
155
def from_mergeable(klass, tree, mergeable, pb):
156
"""Return a Merger for a bundle or merge directive.
158
:param tree: The tree to merge changes into
159
:param mergeable: A merge directive or bundle
160
:param pb: A progress indicator
162
mergeable.install_revisions(tree.branch.repository)
163
base_revision_id, other_revision_id, verified =\
164
mergeable.get_merge_request(tree.branch.repository)
165
revision_graph = tree.branch.repository.get_graph()
166
if base_revision_id is not None:
167
if (base_revision_id != _mod_revision.NULL_REVISION and
168
revision_graph.is_ancestor(
169
base_revision_id, tree.branch.last_revision())):
170
base_revision_id = None
172
warning('Performing cherrypick')
173
merger = klass.from_revision_ids(pb, tree, other_revision_id,
174
base_revision_id, revision_graph=
176
return merger, verified
179
def from_revision_ids(pb, tree, other, base=None, other_branch=None,
180
base_branch=None, revision_graph=None,
182
"""Return a Merger for revision-ids.
184
:param pb: A progress indicator
185
:param tree: The tree to merge changes into
186
:param other: The revision-id to use as OTHER
187
:param base: The revision-id to use as BASE. If not specified, will
189
:param other_branch: A branch containing the other revision-id. If
190
not supplied, tree.branch is used.
191
:param base_branch: A branch containing the base revision-id. If
192
not supplied, other_branch or tree.branch will be used.
193
:param revision_graph: If you have a revision_graph precomputed, pass
194
it in, otherwise it will be created for you.
195
:param tree_branch: The branch associated with tree. If not supplied,
196
tree.branch will be used.
198
if tree_branch is None:
199
tree_branch = tree.branch
200
merger = Merger(tree_branch, this_tree=tree, pb=pb,
201
revision_graph=revision_graph)
202
if other_branch is None:
203
other_branch = tree.branch
204
merger.set_other_revision(other, other_branch)
208
if base_branch is None:
209
base_branch = other_branch
210
merger.set_base_revision(base, base_branch)
213
def revision_tree(self, revision_id, branch=None):
214
if revision_id not in self._cached_trees:
216
branch = self.this_branch
218
tree = self.this_tree.revision_tree(revision_id)
219
except errors.NoSuchRevisionInTree:
220
tree = branch.repository.revision_tree(revision_id)
221
self._cached_trees[revision_id] = tree
222
return self._cached_trees[revision_id]
224
def _get_tree(self, treespec, possible_transports=None):
225
from bzrlib import workingtree
226
location, revno = treespec
228
tree = workingtree.WorkingTree.open_containing(location)[0]
229
return tree.branch, tree
230
branch = Branch.open_containing(location, possible_transports)[0]
232
revision_id = branch.last_revision()
234
revision_id = branch.get_rev_id(revno)
235
revision_id = ensure_null(revision_id)
236
return branch, self.revision_tree(revision_id, branch)
111
238
def ensure_revision_trees(self):
112
239
if self.this_revision_tree is None:
113
self.this_basis_tree = self.this_branch.repository.revision_tree(
240
self.this_basis_tree = self.revision_tree(self.this_basis)
115
241
if self.this_basis == self.this_rev_id:
116
242
self.this_revision_tree = self.this_basis_tree
118
244
if self.other_rev_id is None:
119
245
other_basis_tree = self.revision_tree(self.other_basis)
120
changes = other_basis_tree.changes_from(self.other_tree)
121
if changes.has_changed():
246
if other_basis_tree.has_changes(self.other_tree):
122
247
raise WorkingTreeNotRevision(self.this_tree)
123
248
other_rev_id = self.other_basis
124
249
self.other_tree = other_basis_tree
140
264
def check_basis(self, check_clean, require_commits=True):
141
265
if self.this_basis is None and require_commits is True:
142
raise BzrCommandError("This branch has no commits")
266
raise BzrCommandError("This branch has no commits."
267
" (perhaps you would prefer 'bzr pull')")
144
269
self.compare_basis()
145
270
if self.this_basis != self.this_rev_id:
146
raise BzrCommandError("Working tree has uncommitted changes.")
271
raise errors.UncommittedChanges(self.this_tree)
148
273
def compare_basis(self):
149
changes = self.this_tree.changes_from(self.this_tree.basis_tree())
150
if not changes.has_changed():
275
basis_tree = self.revision_tree(self.this_tree.last_revision())
276
except errors.NoSuchRevision:
277
basis_tree = self.this_tree.basis_tree()
278
if not self.this_tree.has_changes(basis_tree):
151
279
self.this_rev_id = self.this_basis
153
281
def set_interesting_files(self, file_list):
155
self._set_interesting_files(file_list)
156
except NotVersionedError, e:
157
raise BzrCommandError("%s is not a source file in any"
160
def _set_interesting_files(self, file_list):
161
"""Set the list of interesting ids from a list of files."""
162
if file_list is None:
163
self.interesting_ids = None
166
interesting_ids = set()
167
for path in file_list:
169
for tree in (self.this_tree, self.base_tree, self.other_tree):
170
file_id = tree.inventory.path2id(path)
171
if file_id is not None:
172
interesting_ids.add(file_id)
175
raise NotVersionedError(path=path)
176
self.interesting_ids = interesting_ids
282
self.interesting_files = file_list
178
284
def set_pending(self):
179
if not self.base_is_ancestor:
181
if self.other_rev_id is None:
183
ancestry = self.this_branch.repository.get_ancestry(self.this_basis)
184
if self.other_rev_id in ancestry:
186
self.this_tree.add_pending_merge(self.other_rev_id)
188
def set_other(self, other_revision):
189
other_branch, self.other_tree = _get_tree(other_revision,
285
if not self.base_is_ancestor or not self.base_is_other_ancestor or self.other_rev_id is None:
289
def _add_parent(self):
290
new_parents = self.this_tree.get_parent_ids() + [self.other_rev_id]
291
new_parent_trees = []
292
for revision_id in new_parents:
294
tree = self.revision_tree(revision_id)
295
except errors.NoSuchRevision:
299
new_parent_trees.append((revision_id, tree))
301
self.this_tree.set_parent_trees(new_parent_trees,
302
allow_leftmost_as_ghost=True)
304
for _revision_id, tree in new_parent_trees:
308
def set_other(self, other_revision, possible_transports=None):
309
"""Set the revision and tree to merge from.
311
This sets the other_tree, other_rev_id, other_basis attributes.
313
:param other_revision: The [path, revision] list to merge from.
315
self.other_branch, self.other_tree = self._get_tree(other_revision,
191
317
if other_revision[1] == -1:
192
self.other_rev_id = other_branch.last_revision()
193
if self.other_rev_id is None:
194
raise NoCommits(other_branch)
318
self.other_rev_id = _mod_revision.ensure_null(
319
self.other_branch.last_revision())
320
if _mod_revision.is_null(self.other_rev_id):
321
raise NoCommits(self.other_branch)
195
322
self.other_basis = self.other_rev_id
196
323
elif other_revision[1] is not None:
197
self.other_rev_id = other_branch.get_rev_id(other_revision[1])
324
self.other_rev_id = self.other_branch.get_rev_id(other_revision[1])
198
325
self.other_basis = self.other_rev_id
200
327
self.other_rev_id = None
201
self.other_basis = other_branch.last_revision()
328
self.other_basis = self.other_branch.last_revision()
202
329
if self.other_basis is None:
203
raise NoCommits(other_branch)
204
if other_branch.base != self.this_branch.base:
205
self.this_branch.fetch(other_branch, last_revision=self.other_basis)
330
raise NoCommits(self.other_branch)
331
if self.other_rev_id is not None:
332
self._cached_trees[self.other_rev_id] = self.other_tree
333
self._maybe_fetch(self.other_branch,self.this_branch, self.other_basis)
335
def set_other_revision(self, revision_id, other_branch):
336
"""Set 'other' based on a branch and revision id
338
:param revision_id: The revision to use for a tree
339
:param other_branch: The branch containing this tree
341
self.other_rev_id = revision_id
342
self.other_branch = other_branch
343
self._maybe_fetch(other_branch, self.this_branch, self.other_rev_id)
344
self.other_tree = self.revision_tree(revision_id)
345
self.other_basis = revision_id
347
def set_base_revision(self, revision_id, branch):
348
"""Set 'base' based on a branch and revision id
350
:param revision_id: The revision to use for a tree
351
:param branch: The branch containing this tree
353
self.base_rev_id = revision_id
354
self.base_branch = branch
355
self._maybe_fetch(branch, self.this_branch, revision_id)
356
self.base_tree = self.revision_tree(revision_id)
358
def _maybe_fetch(self, source, target, revision_id):
359
if not source.repository.has_same_location(target.repository):
360
target.fetch(source, revision_id)
207
362
def find_base(self):
208
self.set_base([None, None])
363
revisions = [ensure_null(self.this_basis),
364
ensure_null(self.other_basis)]
365
if NULL_REVISION in revisions:
366
self.base_rev_id = NULL_REVISION
367
self.base_tree = self.revision_tree(self.base_rev_id)
368
self._is_criss_cross = False
370
lcas = self.revision_graph.find_lca(revisions[0], revisions[1])
371
self._is_criss_cross = False
373
self.base_rev_id = NULL_REVISION
375
self.base_rev_id = list(lcas)[0]
376
else: # len(lcas) > 1
378
# find_unique_lca can only handle 2 nodes, so we have to
379
# start back at the beginning. It is a shame to traverse
380
# the graph again, but better than re-implementing
382
self.base_rev_id = self.revision_graph.find_unique_lca(
383
revisions[0], revisions[1])
385
self.base_rev_id = self.revision_graph.find_unique_lca(
387
self._is_criss_cross = True
388
if self.base_rev_id == NULL_REVISION:
389
raise UnrelatedBranches()
390
if self._is_criss_cross:
391
warning('Warning: criss-cross merge encountered. See bzr'
392
' help criss-cross.')
393
mutter('Criss-cross lcas: %r' % lcas)
394
interesting_revision_ids = [self.base_rev_id]
395
interesting_revision_ids.extend(lcas)
396
interesting_trees = dict((t.get_revision_id(), t)
397
for t in self.this_branch.repository.revision_trees(
398
interesting_revision_ids))
399
self._cached_trees.update(interesting_trees)
400
self.base_tree = interesting_trees.pop(self.base_rev_id)
401
sorted_lca_keys = self.revision_graph.find_merge_order(
403
self._lca_trees = [interesting_trees[key]
404
for key in sorted_lca_keys]
406
self.base_tree = self.revision_tree(self.base_rev_id)
407
self.base_is_ancestor = True
408
self.base_is_other_ancestor = True
409
mutter('Base revid: %r' % self.base_rev_id)
210
411
def set_base(self, base_revision):
412
"""Set the base revision to use for the merge.
414
:param base_revision: A 2-list containing a path and revision number.
211
416
mutter("doing merge() with no base_revision specified")
212
417
if base_revision == [None, None]:
214
pb = bzrlib.ui.ui_factory.nested_progress_bar()
216
this_repo = self.this_branch.repository
217
self.base_rev_id = common_ancestor(self.this_basis,
222
except NoCommonAncestor:
223
raise UnrelatedBranches()
224
self.base_tree = _get_revid_tree(self.this_branch, self.base_rev_id,
226
self.base_is_ancestor = True
228
base_branch, self.base_tree = _get_tree(base_revision)
420
base_branch, self.base_tree = self._get_tree(base_revision)
229
421
if base_revision[1] == -1:
230
422
self.base_rev_id = base_branch.last_revision()
231
423
elif base_revision[1] is None:
232
self.base_rev_id = None
424
self.base_rev_id = _mod_revision.NULL_REVISION
234
self.base_rev_id = base_branch.get_rev_id(base_revision[1])
235
if self.this_branch.base != base_branch.base:
236
self.this_branch.fetch(base_branch)
237
self.base_is_ancestor = is_ancestor(self.this_basis,
426
self.base_rev_id = _mod_revision.ensure_null(
427
base_branch.get_rev_id(base_revision[1]))
428
self._maybe_fetch(base_branch, self.this_branch, self.base_rev_id)
242
kwargs = {'working_tree':self.this_tree, 'this_tree': self.this_tree,
243
'other_tree': self.other_tree,
430
def make_merger(self):
431
kwargs = {'working_tree':self.this_tree, 'this_tree': self.this_tree,
432
'other_tree': self.other_tree,
244
433
'interesting_ids': self.interesting_ids,
434
'interesting_files': self.interesting_files,
246
437
if self.merge_type.requires_base:
247
438
kwargs['base_tree'] = self.base_tree
248
439
if self.merge_type.supports_reprocess:
254
445
kwargs['show_base'] = self.show_base
255
446
elif self.show_base:
256
447
raise BzrError("Showing base is not supported for this"
257
" merge type. %s" % self.merge_type)
258
merge = self.merge_type(pb=self._pb, **kwargs)
448
" merge type. %s" % self.merge_type)
449
if (not getattr(self.merge_type, 'supports_reverse_cherrypick', True)
450
and not self.base_is_other_ancestor):
451
raise errors.CannotReverseCherrypick()
452
if self.merge_type.supports_cherrypick:
453
kwargs['cherrypick'] = (not self.base_is_ancestor or
454
not self.base_is_other_ancestor)
455
if self._is_criss_cross and getattr(self.merge_type,
456
'supports_lca_trees', False):
457
kwargs['lca_trees'] = self._lca_trees
458
return self.merge_type(pb=self._pb,
459
change_reporter=self.change_reporter,
462
def _do_merge_to(self, merge):
463
if self.other_branch is not None:
464
self.other_branch.update_references(self.this_branch)
466
if self.recurse == 'down':
467
for relpath, file_id in self.this_tree.iter_references():
468
sub_tree = self.this_tree.get_nested_tree(file_id, relpath)
469
other_revision = self.other_tree.get_reference_revision(
471
if other_revision == sub_tree.last_revision():
473
sub_merge = Merger(sub_tree.branch, this_tree=sub_tree)
474
sub_merge.merge_type = self.merge_type
475
other_branch = self.other_branch.reference_parent(file_id, relpath)
476
sub_merge.set_other_revision(other_revision, other_branch)
477
base_revision = self.base_tree.get_reference_revision(file_id)
478
sub_merge.base_tree = \
479
sub_tree.branch.repository.revision_tree(base_revision)
480
sub_merge.base_rev_id = base_revision
484
self.this_tree.lock_tree_write()
486
if self.base_tree is not None:
487
self.base_tree.lock_read()
489
if self.other_tree is not None:
490
self.other_tree.lock_read()
492
merge = self.make_merger()
493
self._do_merge_to(merge)
495
if self.other_tree is not None:
496
self.other_tree.unlock()
498
if self.base_tree is not None:
499
self.base_tree.unlock()
501
self.this_tree.unlock()
259
502
if len(merge.cooked_conflicts) == 0:
260
if not self.ignore_zero:
503
if not self.ignore_zero and not is_quiet():
261
504
note("All changes applied successfully.")
263
506
note("%d conflicts encountered." % len(merge.cooked_conflicts))
265
508
return len(merge.cooked_conflicts)
267
def regen_inventory(self, new_entries):
268
old_entries = self.this_tree.read_working_inventory()
272
for path, file_id in new_entries:
275
new_entries_map[file_id] = path
277
def id2path(file_id):
278
path = new_entries_map.get(file_id)
281
entry = old_entries[file_id]
282
if entry.parent_id is None:
284
return pathjoin(id2path(entry.parent_id), entry.name)
286
for file_id in old_entries:
287
entry = old_entries[file_id]
288
path = id2path(file_id)
289
if file_id in self.base_tree.inventory:
290
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
292
executable = getattr(entry, 'executable', False)
293
new_inventory[file_id] = (path, file_id, entry.parent_id,
294
entry.kind, executable)
296
by_path[path] = file_id
301
for path, file_id in new_entries:
303
del new_inventory[file_id]
306
new_path_list.append((path, file_id))
307
if file_id not in old_entries:
309
# Ensure no file is added before its parent
311
for path, file_id in new_path_list:
315
parent = by_path[os.path.dirname(path)]
316
abspath = pathjoin(self.this_tree.basedir, path)
317
kind = bzrlib.osutils.file_kind(abspath)
318
if file_id in self.base_tree.inventory:
319
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
322
new_inventory[file_id] = (path, file_id, parent, kind, executable)
323
by_path[path] = file_id
325
# Get a list in insertion order
326
new_inventory_list = new_inventory.values()
327
mutter ("""Inventory regeneration:
328
old length: %i insertions: %i deletions: %i new_length: %i"""\
329
% (len(old_entries), insertions, deletions,
330
len(new_inventory_list)))
331
assert len(new_inventory_list) == len(old_entries) + insertions\
333
new_inventory_list.sort()
334
return new_inventory_list
511
class _InventoryNoneEntry(object):
512
"""This represents an inventory entry which *isn't there*.
514
It simplifies the merging logic if we always have an InventoryEntry, even
515
if it isn't actually present
522
symlink_target = None
525
_none_entry = _InventoryNoneEntry()
337
528
class Merge3Merger(object):
353
579
self.cooked_conflicts = []
354
580
self.reprocess = reprocess
355
581
self.show_base = show_base
582
self._lca_trees = lca_trees
583
# Uncommenting this will change the default algorithm to always use
584
# _entries_lca. This can be useful for running the test suite and
585
# making sure we haven't missed any corner cases.
586
# if lca_trees is None:
587
# self._lca_trees = [self.base_tree]
590
self.change_reporter = change_reporter
591
self.cherrypick = cherrypick
358
592
if self.pp is None:
359
593
self.pp = ProgressPhase("Merge phase", 3, self.pb)
361
if interesting_ids is not None:
362
all_ids = interesting_ids
364
all_ids = set(base_tree)
365
all_ids.update(other_tree)
366
working_tree.lock_write()
367
self.tt = TreeTransform(working_tree, self.pb)
598
self.this_tree.lock_tree_write()
599
self.base_tree.lock_read()
600
self.other_tree.lock_read()
601
self.tt = TreeTransform(self.this_tree, self.pb)
369
603
self.pp.next_phase()
370
child_pb = ui.ui_factory.nested_progress_bar()
372
for num, file_id in enumerate(all_ids):
373
child_pb.update('Preparing file merge', num, len(all_ids))
374
self.merge_names(file_id)
375
file_status = self.merge_contents(file_id)
376
self.merge_executable(file_id, file_status)
381
child_pb = ui.ui_factory.nested_progress_bar()
383
fs_conflicts = resolve_conflicts(self.tt, child_pb)
386
self.cook_conflicts(fs_conflicts)
387
for conflict in self.cooked_conflicts:
390
results = self.tt.apply()
604
self._compute_transform()
606
results = self.tt.apply(no_conflicts=True)
391
607
self.write_modified(results)
393
working_tree.add_conflicts(self.cooked_conflicts)
609
self.this_tree.add_conflicts(self.cooked_conflicts)
394
610
except UnsupportedOperation:
397
613
self.tt.finalize()
398
working_tree.unlock()
614
self.other_tree.unlock()
615
self.base_tree.unlock()
616
self.this_tree.unlock()
619
def make_preview_transform(self):
620
self.base_tree.lock_read()
621
self.other_tree.lock_read()
622
self.tt = TransformPreview(self.this_tree)
625
self._compute_transform()
628
self.other_tree.unlock()
629
self.base_tree.unlock()
633
def _compute_transform(self):
634
if self._lca_trees is None:
635
entries = self._entries3()
636
resolver = self._three_way
638
entries = self._entries_lca()
639
resolver = self._lca_multi_way
640
child_pb = ui.ui_factory.nested_progress_bar()
642
for num, (file_id, changed, parents3, names3,
643
executable3) in enumerate(entries):
644
child_pb.update('Preparing file merge', num, len(entries))
645
self._merge_names(file_id, parents3, names3, resolver=resolver)
647
file_status = self.merge_contents(file_id)
649
file_status = 'unmodified'
650
self._merge_executable(file_id,
651
executable3, file_status, resolver=resolver)
656
child_pb = ui.ui_factory.nested_progress_bar()
658
fs_conflicts = resolve_conflicts(self.tt, child_pb,
659
lambda t, c: conflict_pass(t, c, self.other_tree))
662
if self.change_reporter is not None:
663
from bzrlib import delta
664
delta.report_changes(
665
self.tt.iter_changes(), self.change_reporter)
666
self.cook_conflicts(fs_conflicts)
667
for conflict in self.cooked_conflicts:
671
"""Gather data about files modified between three trees.
673
Return a list of tuples of file_id, changed, parents3, names3,
674
executable3. changed is a boolean indicating whether the file contents
675
or kind were changed. parents3 is a tuple of parent ids for base,
676
other and this. names3 is a tuple of names for base, other and this.
677
executable3 is a tuple of execute-bit values for base, other and this.
680
iterator = self.other_tree.iter_changes(self.base_tree,
681
include_unchanged=True, specific_files=self.interesting_files,
682
extra_trees=[self.this_tree])
683
this_entries = dict((e.file_id, e) for p, e in
684
self.this_tree.iter_entries_by_dir(
685
self.interesting_ids))
686
for (file_id, paths, changed, versioned, parents, names, kind,
687
executable) in iterator:
688
if (self.interesting_ids is not None and
689
file_id not in self.interesting_ids):
691
entry = this_entries.get(file_id)
692
if entry is not None:
693
this_name = entry.name
694
this_parent = entry.parent_id
695
this_executable = entry.executable
699
this_executable = None
700
parents3 = parents + (this_parent,)
701
names3 = names + (this_name,)
702
executable3 = executable + (this_executable,)
703
result.append((file_id, changed, parents3, names3, executable3))
706
def _entries_lca(self):
707
"""Gather data about files modified between multiple trees.
709
This compares OTHER versus all LCA trees, and for interesting entries,
710
it then compares with THIS and BASE.
712
For the multi-valued entries, the format will be (BASE, [lca1, lca2])
713
:return: [(file_id, changed, parents, names, executable)]
714
file_id Simple file_id of the entry
715
changed Boolean, True if the kind or contents changed
717
parents ((base, [parent_id, in, lcas]), parent_id_other,
719
names ((base, [name, in, lcas]), name_in_other, name_in_this)
720
executable ((base, [exec, in, lcas]), exec_in_other, exec_in_this)
722
if self.interesting_files is not None:
723
lookup_trees = [self.this_tree, self.base_tree]
724
lookup_trees.extend(self._lca_trees)
725
# I think we should include the lca trees as well
726
interesting_ids = self.other_tree.paths2ids(self.interesting_files,
729
interesting_ids = self.interesting_ids
731
walker = _mod_tree.MultiWalker(self.other_tree, self._lca_trees)
733
base_inventory = self.base_tree.inventory
734
this_inventory = self.this_tree.inventory
735
for path, file_id, other_ie, lca_values in walker.iter_all():
736
# Is this modified at all from any of the other trees?
738
other_ie = _none_entry
739
if interesting_ids is not None and file_id not in interesting_ids:
742
# If other_revision is found in any of the lcas, that means this
743
# node is uninteresting. This is because when merging, if there are
744
# multiple heads(), we have to create a new node. So if we didn't,
745
# we know that the ancestry is linear, and that OTHER did not
747
# See doc/developers/lca_merge_resolution.txt for details
748
other_revision = other_ie.revision
749
if other_revision is not None:
750
# We can't use this shortcut when other_revision is None,
751
# because it may be None because things are WorkingTrees, and
752
# not because it is *actually* None.
753
is_unmodified = False
754
for lca_path, ie in lca_values:
755
if ie is not None and ie.revision == other_revision:
762
for lca_path, lca_ie in lca_values:
764
lca_entries.append(_none_entry)
766
lca_entries.append(lca_ie)
768
if file_id in base_inventory:
769
base_ie = base_inventory[file_id]
771
base_ie = _none_entry
773
if file_id in this_inventory:
774
this_ie = this_inventory[file_id]
776
this_ie = _none_entry
782
for lca_ie in lca_entries:
783
lca_kinds.append(lca_ie.kind)
784
lca_parent_ids.append(lca_ie.parent_id)
785
lca_names.append(lca_ie.name)
786
lca_executable.append(lca_ie.executable)
788
kind_winner = self._lca_multi_way(
789
(base_ie.kind, lca_kinds),
790
other_ie.kind, this_ie.kind)
791
parent_id_winner = self._lca_multi_way(
792
(base_ie.parent_id, lca_parent_ids),
793
other_ie.parent_id, this_ie.parent_id)
794
name_winner = self._lca_multi_way(
795
(base_ie.name, lca_names),
796
other_ie.name, this_ie.name)
798
content_changed = True
799
if kind_winner == 'this':
800
# No kind change in OTHER, see if there are *any* changes
801
if other_ie.kind == 'directory':
802
if parent_id_winner == 'this' and name_winner == 'this':
803
# No change for this directory in OTHER, skip
805
content_changed = False
806
elif other_ie.kind is None or other_ie.kind == 'file':
807
def get_sha1(ie, tree):
808
if ie.kind != 'file':
810
return tree.get_file_sha1(file_id)
811
base_sha1 = get_sha1(base_ie, self.base_tree)
812
lca_sha1s = [get_sha1(ie, tree) for ie, tree
813
in zip(lca_entries, self._lca_trees)]
814
this_sha1 = get_sha1(this_ie, self.this_tree)
815
other_sha1 = get_sha1(other_ie, self.other_tree)
816
sha1_winner = self._lca_multi_way(
817
(base_sha1, lca_sha1s), other_sha1, this_sha1,
818
allow_overriding_lca=False)
819
exec_winner = self._lca_multi_way(
820
(base_ie.executable, lca_executable),
821
other_ie.executable, this_ie.executable)
822
if (parent_id_winner == 'this' and name_winner == 'this'
823
and sha1_winner == 'this' and exec_winner == 'this'):
824
# No kind, parent, name, exec, or content change for
825
# OTHER, so this node is not considered interesting
827
if sha1_winner == 'this':
828
content_changed = False
829
elif other_ie.kind == 'symlink':
830
def get_target(ie, tree):
831
if ie.kind != 'symlink':
833
return tree.get_symlink_target(file_id)
834
base_target = get_target(base_ie, self.base_tree)
835
lca_targets = [get_target(ie, tree) for ie, tree
836
in zip(lca_entries, self._lca_trees)]
837
this_target = get_target(this_ie, self.this_tree)
838
other_target = get_target(other_ie, self.other_tree)
839
target_winner = self._lca_multi_way(
840
(base_target, lca_targets),
841
other_target, this_target)
842
if (parent_id_winner == 'this' and name_winner == 'this'
843
and target_winner == 'this'):
844
# No kind, parent, name, or symlink target change
847
if target_winner == 'this':
848
content_changed = False
849
elif other_ie.kind == 'tree-reference':
850
# The 'changed' information seems to be handled at a higher
851
# level. At least, _entries3 returns False for content
852
# changed, even when at a new revision_id.
853
content_changed = False
854
if (parent_id_winner == 'this' and name_winner == 'this'):
855
# Nothing interesting
858
raise AssertionError('unhandled kind: %s' % other_ie.kind)
859
# XXX: We need to handle kind == 'symlink'
861
# If we have gotten this far, that means something has changed
862
result.append((file_id, content_changed,
863
((base_ie.parent_id, lca_parent_ids),
864
other_ie.parent_id, this_ie.parent_id),
865
((base_ie.name, lca_names),
866
other_ie.name, this_ie.name),
867
((base_ie.executable, lca_executable),
868
other_ie.executable, this_ie.executable)
875
self.tt.final_kind(self.tt.root)
877
self.tt.cancel_deletion(self.tt.root)
878
if self.tt.final_file_id(self.tt.root) is None:
879
self.tt.version_file(self.tt.tree_file_id(self.tt.root),
881
other_root_file_id = self.other_tree.get_root_id()
882
if other_root_file_id is None:
884
other_root = self.tt.trans_id_file_id(other_root_file_id)
885
if other_root == self.tt.root:
888
self.tt.final_kind(other_root)
891
if self.other_tree.inventory.root.file_id in self.this_tree.inventory:
892
# the other tree's root is a non-root in the current tree
894
self.reparent_children(self.other_tree.inventory.root, self.tt.root)
895
self.tt.cancel_creation(other_root)
896
self.tt.cancel_versioning(other_root)
898
def reparent_children(self, ie, target):
899
for thing, child in ie.children.iteritems():
900
trans_id = self.tt.trans_id_file_id(child.file_id)
901
self.tt.adjust_path(self.tt.final_name(trans_id), target, trans_id)
401
903
def write_modified(self, results):
402
904
modified_hashes = {}
876
1473
name = self.tt.final_name(trans_id)
877
1474
parent_id = self.tt.final_parent(trans_id)
878
1475
self._dump_conflicts(name, parent_id, file_id)
879
self._raw_conflicts.append(('text conflict', trans_id))
1476
self._raw_conflicts.append(('text conflict', trans_id))
1478
osutils.rmtree(temp_dir)
884
1481
def merge_inner(this_branch, other_tree, base_tree, ignore_zero=False,
886
merge_type=Merge3Merger,
887
interesting_ids=None,
1483
merge_type=Merge3Merger,
1484
interesting_ids=None,
890
1487
other_rev_id=None,
891
1488
interesting_files=None,
894
"""Primary interface for merging.
1491
change_reporter=None):
1492
"""Primary interface for merging.
896
typical use is probably
1494
typical use is probably
897
1495
'merge_inner(branch, branch.get_revision_tree(other_revision),
898
1496
branch.get_revision_tree(base_revision))'
900
1498
if this_tree is None:
901
warnings.warn("bzrlib.merge.merge_inner requires a this_tree parameter as of "
902
"bzrlib version 0.8.",
905
this_tree = this_branch.bzrdir.open_workingtree()
906
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1499
raise BzrError("bzrlib.merge.merge_inner requires a this_tree "
1500
"parameter as of bzrlib version 0.8.")
1501
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1502
pb=pb, change_reporter=change_reporter)
908
1503
merger.backup_files = backup_files
909
1504
merger.merge_type = merge_type
910
1505
merger.interesting_ids = interesting_ids
911
1506
merger.ignore_zero = ignore_zero
912
1507
if interesting_files:
913
assert not interesting_ids, ('Only supply interesting_ids'
914
' or interesting_files')
915
merger._set_interesting_files(interesting_files)
916
merger.show_base = show_base
1509
raise ValueError('Only supply interesting_ids'
1510
' or interesting_files')
1511
merger.interesting_files = interesting_files
1512
merger.show_base = show_base
917
1513
merger.reprocess = reprocess
918
1514
merger.other_rev_id = other_rev_id
919
1515
merger.other_basis = other_rev_id
1516
get_revision_id = getattr(base_tree, 'get_revision_id', None)
1517
if get_revision_id is None:
1518
get_revision_id = base_tree.last_revision
1519
merger.set_base_revision(get_revision_id(), this_branch)
920
1520
return merger.do_merge()
923
merge_types = { "merge3": (Merge3Merger, "Native diff3-style merge"),
924
"diff3": (Diff3Merger, "Merge using external diff3"),
925
'weave': (WeaveMerger, "Weave-based merge")
929
def merge_type_help():
930
templ = '%s%%7s: %%s' % (' '*12)
931
lines = [templ % (f[0], f[1][1]) for f in merge_types.iteritems()]
932
return '\n'.join(lines)
1522
def get_merge_type_registry():
1523
"""Merge type registry is in bzrlib.option to avoid circular imports.
1525
This method provides a sanctioned way to retrieve it.
1527
from bzrlib import option
1528
return option._merge_type_registry
1531
def _plan_annotate_merge(annotated_a, annotated_b, ancestors_a, ancestors_b):
1532
def status_a(revision, text):
1533
if revision in ancestors_b:
1534
return 'killed-b', text
1536
return 'new-a', text
1538
def status_b(revision, text):
1539
if revision in ancestors_a:
1540
return 'killed-a', text
1542
return 'new-b', text
1544
plain_a = [t for (a, t) in annotated_a]
1545
plain_b = [t for (a, t) in annotated_b]
1546
matcher = patiencediff.PatienceSequenceMatcher(None, plain_a, plain_b)
1547
blocks = matcher.get_matching_blocks()
1550
for ai, bi, l in blocks:
1551
# process all mismatched sections
1552
# (last mismatched section is handled because blocks always
1553
# includes a 0-length last block)
1554
for revision, text in annotated_a[a_cur:ai]:
1555
yield status_a(revision, text)
1556
for revision, text in annotated_b[b_cur:bi]:
1557
yield status_b(revision, text)
1558
# and now the matched section
1561
for text_a in plain_a[ai:a_cur]:
1562
yield "unchanged", text_a
1565
class _PlanMergeBase(object):
1567
def __init__(self, a_rev, b_rev, vf, key_prefix):
1570
:param a_rev: Revision-id of one revision to merge
1571
:param b_rev: Revision-id of the other revision to merge
1572
:param vf: A VersionedFiles containing both revisions
1573
:param key_prefix: A prefix for accessing keys in vf, typically
1579
self._last_lines = None
1580
self._last_lines_revision_id = None
1581
self._cached_matching_blocks = {}
1582
self._key_prefix = key_prefix
1583
self._precache_tip_lines()
1585
def _precache_tip_lines(self):
1586
lines = self.get_lines([self.a_rev, self.b_rev])
1587
self.lines_a = lines[self.a_rev]
1588
self.lines_b = lines[self.b_rev]
1590
def get_lines(self, revisions):
1591
"""Get lines for revisions from the backing VersionedFiles.
1593
:raises RevisionNotPresent: on absent texts.
1595
keys = [(self._key_prefix + (rev,)) for rev in revisions]
1597
for record in self.vf.get_record_stream(keys, 'unordered', True):
1598
if record.storage_kind == 'absent':
1599
raise errors.RevisionNotPresent(record.key, self.vf)
1600
result[record.key[-1]] = osutils.chunks_to_lines(
1601
record.get_bytes_as('chunked'))
1604
def plan_merge(self):
1605
"""Generate a 'plan' for merging the two revisions.
1607
This involves comparing their texts and determining the cause of
1608
differences. If text A has a line and text B does not, then either the
1609
line was added to text A, or it was deleted from B. Once the causes
1610
are combined, they are written out in the format described in
1611
VersionedFile.plan_merge
1613
blocks = self._get_matching_blocks(self.a_rev, self.b_rev)
1614
unique_a, unique_b = self._unique_lines(blocks)
1615
new_a, killed_b = self._determine_status(self.a_rev, unique_a)
1616
new_b, killed_a = self._determine_status(self.b_rev, unique_b)
1617
return self._iter_plan(blocks, new_a, killed_b, new_b, killed_a)
1619
def _iter_plan(self, blocks, new_a, killed_b, new_b, killed_a):
1622
for i, j, n in blocks:
1623
for a_index in range(last_i, i):
1624
if a_index in new_a:
1625
if a_index in killed_b:
1626
yield 'conflicted-a', self.lines_a[a_index]
1628
yield 'new-a', self.lines_a[a_index]
1630
yield 'killed-b', self.lines_a[a_index]
1631
for b_index in range(last_j, j):
1632
if b_index in new_b:
1633
if b_index in killed_a:
1634
yield 'conflicted-b', self.lines_b[b_index]
1636
yield 'new-b', self.lines_b[b_index]
1638
yield 'killed-a', self.lines_b[b_index]
1639
# handle common lines
1640
for a_index in range(i, i+n):
1641
yield 'unchanged', self.lines_a[a_index]
1645
def _get_matching_blocks(self, left_revision, right_revision):
1646
"""Return a description of which sections of two revisions match.
1648
See SequenceMatcher.get_matching_blocks
1650
cached = self._cached_matching_blocks.get((left_revision,
1652
if cached is not None:
1654
if self._last_lines_revision_id == left_revision:
1655
left_lines = self._last_lines
1656
right_lines = self.get_lines([right_revision])[right_revision]
1658
lines = self.get_lines([left_revision, right_revision])
1659
left_lines = lines[left_revision]
1660
right_lines = lines[right_revision]
1661
self._last_lines = right_lines
1662
self._last_lines_revision_id = right_revision
1663
matcher = patiencediff.PatienceSequenceMatcher(None, left_lines,
1665
return matcher.get_matching_blocks()
1667
def _unique_lines(self, matching_blocks):
1668
"""Analyse matching_blocks to determine which lines are unique
1670
:return: a tuple of (unique_left, unique_right), where the values are
1671
sets of line numbers of unique lines.
1677
for i, j, n in matching_blocks:
1678
unique_left.extend(range(last_i, i))
1679
unique_right.extend(range(last_j, j))
1682
return unique_left, unique_right
1685
def _subtract_plans(old_plan, new_plan):
1686
"""Remove changes from new_plan that came from old_plan.
1688
It is assumed that the difference between the old_plan and new_plan
1689
is their choice of 'b' text.
1691
All lines from new_plan that differ from old_plan are emitted
1692
verbatim. All lines from new_plan that match old_plan but are
1693
not about the 'b' revision are emitted verbatim.
1695
Lines that match and are about the 'b' revision are the lines we
1696
don't want, so we convert 'killed-b' -> 'unchanged', and 'new-b'
1697
is skipped entirely.
1699
matcher = patiencediff.PatienceSequenceMatcher(None, old_plan,
1702
for i, j, n in matcher.get_matching_blocks():
1703
for jj in range(last_j, j):
1705
for jj in range(j, j+n):
1706
plan_line = new_plan[jj]
1707
if plan_line[0] == 'new-b':
1709
elif plan_line[0] == 'killed-b':
1710
yield 'unchanged', plan_line[1]
1716
class _PlanMerge(_PlanMergeBase):
1717
"""Plan an annotate merge using on-the-fly annotation"""
1719
def __init__(self, a_rev, b_rev, vf, key_prefix):
1720
super(_PlanMerge, self).__init__(a_rev, b_rev, vf, key_prefix)
1721
self.a_key = self._key_prefix + (self.a_rev,)
1722
self.b_key = self._key_prefix + (self.b_rev,)
1723
self.graph = Graph(self.vf)
1724
heads = self.graph.heads((self.a_key, self.b_key))
1726
# one side dominates, so we can just return its values, yay for
1728
# Ideally we would know that before we get this far
1729
self._head_key = heads.pop()
1730
if self._head_key == self.a_key:
1734
mutter('found dominating revision for %s\n%s > %s', self.vf,
1735
self._head_key[-1], other)
1738
self._head_key = None
1741
def _precache_tip_lines(self):
1742
# Turn this into a no-op, because we will do this later
1745
def _find_recursive_lcas(self):
1746
"""Find all the ancestors back to a unique lca"""
1747
cur_ancestors = (self.a_key, self.b_key)
1748
# graph.find_lca(uncommon, keys) now returns plain NULL_REVISION,
1749
# rather than a key tuple. We will just map that directly to no common
1753
next_lcas = self.graph.find_lca(*cur_ancestors)
1754
# Map a plain NULL_REVISION to a simple no-ancestors
1755
if next_lcas == set([NULL_REVISION]):
1757
# Order the lca's based on when they were merged into the tip
1758
# While the actual merge portion of weave merge uses a set() of
1759
# active revisions, the order of insertion *does* effect the
1760
# implicit ordering of the texts.
1761
for rev_key in cur_ancestors:
1762
ordered_parents = tuple(self.graph.find_merge_order(rev_key,
1764
parent_map[rev_key] = ordered_parents
1765
if len(next_lcas) == 0:
1767
elif len(next_lcas) == 1:
1768
parent_map[list(next_lcas)[0]] = ()
1770
elif len(next_lcas) > 2:
1771
# More than 2 lca's, fall back to grabbing all nodes between
1772
# this and the unique lca.
1773
mutter('More than 2 LCAs, falling back to all nodes for:'
1774
' %s, %s\n=> %s', self.a_key, self.b_key, cur_ancestors)
1775
cur_lcas = next_lcas
1776
while len(cur_lcas) > 1:
1777
cur_lcas = self.graph.find_lca(*cur_lcas)
1778
if len(cur_lcas) == 0:
1779
# No common base to find, use the full ancestry
1782
unique_lca = list(cur_lcas)[0]
1783
if unique_lca == NULL_REVISION:
1784
# find_lca will return a plain 'NULL_REVISION' rather
1785
# than a key tuple when there is no common ancestor, we
1786
# prefer to just use None, because it doesn't confuse
1787
# _get_interesting_texts()
1789
parent_map.update(self._find_unique_parents(next_lcas,
1792
cur_ancestors = next_lcas
1795
def _find_unique_parents(self, tip_keys, base_key):
1796
"""Find ancestors of tip that aren't ancestors of base.
1798
:param tip_keys: Nodes that are interesting
1799
:param base_key: Cull all ancestors of this node
1800
:return: The parent map for all revisions between tip_keys and
1801
base_key. base_key will be included. References to nodes outside of
1802
the ancestor set will also be removed.
1804
# TODO: this would be simpler if find_unique_ancestors took a list
1805
# instead of a single tip, internally it supports it, but it
1806
# isn't a "backwards compatible" api change.
1807
if base_key is None:
1808
parent_map = dict(self.graph.iter_ancestry(tip_keys))
1809
# We remove NULL_REVISION because it isn't a proper tuple key, and
1810
# thus confuses things like _get_interesting_texts, and our logic
1811
# to add the texts into the memory weave.
1812
if NULL_REVISION in parent_map:
1813
parent_map.pop(NULL_REVISION)
1816
for tip in tip_keys:
1818
self.graph.find_unique_ancestors(tip, [base_key]))
1819
parent_map = self.graph.get_parent_map(interesting)
1820
parent_map[base_key] = ()
1821
culled_parent_map, child_map, tails = self._remove_external_references(
1823
# Remove all the tails but base_key
1824
if base_key is not None:
1825
tails.remove(base_key)
1826
self._prune_tails(culled_parent_map, child_map, tails)
1827
# Now remove all the uninteresting 'linear' regions
1828
simple_map = _mod_graph.collapse_linear_regions(culled_parent_map)
1832
def _remove_external_references(parent_map):
1833
"""Remove references that go outside of the parent map.
1835
:param parent_map: Something returned from Graph.get_parent_map(keys)
1836
:return: (filtered_parent_map, child_map, tails)
1837
filtered_parent_map is parent_map without external references
1838
child_map is the {parent_key: [child_keys]} mapping
1839
tails is a list of nodes that do not have any parents in the map
1841
# TODO: The basic effect of this function seems more generic than
1842
# _PlanMerge. But the specific details of building a child_map,
1843
# and computing tails seems very specific to _PlanMerge.
1844
# Still, should this be in Graph land?
1845
filtered_parent_map = {}
1848
for key, parent_keys in parent_map.iteritems():
1849
culled_parent_keys = [p for p in parent_keys if p in parent_map]
1850
if not culled_parent_keys:
1852
for parent_key in culled_parent_keys:
1853
child_map.setdefault(parent_key, []).append(key)
1854
# TODO: Do we want to do this, it adds overhead for every node,
1855
# just to say that the node has no children
1856
child_map.setdefault(key, [])
1857
filtered_parent_map[key] = culled_parent_keys
1858
return filtered_parent_map, child_map, tails
1861
def _prune_tails(parent_map, child_map, tails_to_remove):
1862
"""Remove tails from the parent map.
1864
This will remove the supplied revisions until no more children have 0
1867
:param parent_map: A dict of {child: [parents]}, this dictionary will
1868
be modified in place.
1869
:param tails_to_remove: A list of tips that should be removed,
1870
this list will be consumed
1871
:param child_map: The reverse dict of parent_map ({parent: [children]})
1872
this dict will be modified
1873
:return: None, parent_map will be modified in place.
1875
while tails_to_remove:
1876
next = tails_to_remove.pop()
1877
parent_map.pop(next)
1878
children = child_map.pop(next)
1879
for child in children:
1880
child_parents = parent_map[child]
1881
child_parents.remove(next)
1882
if len(child_parents) == 0:
1883
tails_to_remove.append(child)
1885
def _get_interesting_texts(self, parent_map):
1886
"""Return a dict of texts we are interested in.
1888
Note that the input is in key tuples, but the output is in plain
1891
:param parent_map: The output from _find_recursive_lcas
1892
:return: A dict of {'revision_id':lines} as returned by
1893
_PlanMergeBase.get_lines()
1895
all_revision_keys = set(parent_map)
1896
all_revision_keys.add(self.a_key)
1897
all_revision_keys.add(self.b_key)
1899
# Everything else is in 'keys' but get_lines is in 'revision_ids'
1900
all_texts = self.get_lines([k[-1] for k in all_revision_keys])
1903
def _build_weave(self):
1904
from bzrlib import weave
1905
self._weave = weave.Weave(weave_name='in_memory_weave',
1906
allow_reserved=True)
1907
parent_map = self._find_recursive_lcas()
1909
all_texts = self._get_interesting_texts(parent_map)
1911
# Note: Unfortunately, the order given by topo_sort will effect the
1912
# ordering resolution in the output. Specifically, if you add A then B,
1913
# then in the output text A lines will show up before B lines. And, of
1914
# course, topo_sort doesn't guarantee any real ordering.
1915
# So we use merge_sort, and add a fake node on the tip.
1916
# This ensures that left-hand parents will always be inserted into the
1917
# weave before right-hand parents.
1918
tip_key = self._key_prefix + (_mod_revision.CURRENT_REVISION,)
1919
parent_map[tip_key] = (self.a_key, self.b_key)
1921
for seq_num, key, depth, eom in reversed(tsort.merge_sort(parent_map,
1925
# for key in tsort.topo_sort(parent_map):
1926
parent_keys = parent_map[key]
1927
revision_id = key[-1]
1928
parent_ids = [k[-1] for k in parent_keys]
1929
self._weave.add_lines(revision_id, parent_ids,
1930
all_texts[revision_id])
1932
def plan_merge(self):
1933
"""Generate a 'plan' for merging the two revisions.
1935
This involves comparing their texts and determining the cause of
1936
differences. If text A has a line and text B does not, then either the
1937
line was added to text A, or it was deleted from B. Once the causes
1938
are combined, they are written out in the format described in
1939
VersionedFile.plan_merge
1941
if self._head_key is not None: # There was a single head
1942
if self._head_key == self.a_key:
1945
if self._head_key != self.b_key:
1946
raise AssertionError('There was an invalid head: %s != %s'
1947
% (self.b_key, self._head_key))
1949
head_rev = self._head_key[-1]
1950
lines = self.get_lines([head_rev])[head_rev]
1951
return ((plan, line) for line in lines)
1952
return self._weave.plan_merge(self.a_rev, self.b_rev)
1955
class _PlanLCAMerge(_PlanMergeBase):
1957
This merge algorithm differs from _PlanMerge in that:
1958
1. comparisons are done against LCAs only
1959
2. cases where a contested line is new versus one LCA but old versus
1960
another are marked as conflicts, by emitting the line as conflicted-a
1963
This is faster, and hopefully produces more useful output.
1966
def __init__(self, a_rev, b_rev, vf, key_prefix, graph):
1967
_PlanMergeBase.__init__(self, a_rev, b_rev, vf, key_prefix)
1968
lcas = graph.find_lca(key_prefix + (a_rev,), key_prefix + (b_rev,))
1971
if lca == NULL_REVISION:
1974
self.lcas.add(lca[-1])
1975
for lca in self.lcas:
1976
if _mod_revision.is_null(lca):
1979
lca_lines = self.get_lines([lca])[lca]
1980
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_a,
1982
blocks = list(matcher.get_matching_blocks())
1983
self._cached_matching_blocks[(a_rev, lca)] = blocks
1984
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_b,
1986
blocks = list(matcher.get_matching_blocks())
1987
self._cached_matching_blocks[(b_rev, lca)] = blocks
1989
def _determine_status(self, revision_id, unique_line_numbers):
1990
"""Determines the status unique lines versus all lcas.
1992
Basically, determines why the line is unique to this revision.
1994
A line may be determined new, killed, or both.
1996
If a line is determined new, that means it was not present in at least
1997
one LCA, and is not present in the other merge revision.
1999
If a line is determined killed, that means the line was present in
2002
If a line is killed and new, this indicates that the two merge
2003
revisions contain differing conflict resolutions.
2004
:param revision_id: The id of the revision in which the lines are
2006
:param unique_line_numbers: The line numbers of unique lines.
2007
:return a tuple of (new_this, killed_other):
2011
unique_line_numbers = set(unique_line_numbers)
2012
for lca in self.lcas:
2013
blocks = self._get_matching_blocks(revision_id, lca)
2014
unique_vs_lca, _ignored = self._unique_lines(blocks)
2015
new.update(unique_line_numbers.intersection(unique_vs_lca))
2016
killed.update(unique_line_numbers.difference(unique_vs_lca))