71
88
class Merger(object):
72
def __init__(self, this_branch, other_tree=None, base_tree=None,
73
this_tree=None, pb=None, change_reporter=None,
74
recurse='down', revision_graph=None):
89
def __init__(self, this_branch, other_tree=None, base_tree=None,
90
this_tree=None, pb=DummyProgress()):
75
91
object.__init__(self)
92
assert this_tree is not None, "this_tree is required"
76
93
self.this_branch = this_branch
77
self.this_basis = _mod_revision.ensure_null(
78
this_branch.last_revision())
94
self.this_basis = this_branch.last_revision()
79
95
self.this_rev_id = None
80
96
self.this_tree = this_tree
81
97
self.this_revision_tree = None
82
98
self.this_basis_tree = None
83
99
self.other_tree = other_tree
84
self.other_branch = None
85
100
self.base_tree = base_tree
86
101
self.ignore_zero = False
87
102
self.backup_files = False
88
103
self.interesting_ids = None
89
self.interesting_files = None
90
104
self.show_base = False
91
105
self.reprocess = False
96
self.recurse = recurse
97
self.change_reporter = change_reporter
98
self._cached_trees = {}
99
self._revision_graph = revision_graph
100
self._base_is_ancestor = None
101
self._base_is_other_ancestor = None
102
self._is_criss_cross = None
103
self._lca_trees = None
106
def revision_graph(self):
107
if self._revision_graph is None:
108
self._revision_graph = self.this_branch.repository.get_graph()
109
return self._revision_graph
111
def _set_base_is_ancestor(self, value):
112
self._base_is_ancestor = value
114
def _get_base_is_ancestor(self):
115
if self._base_is_ancestor is None:
116
self._base_is_ancestor = self.revision_graph.is_ancestor(
117
self.base_rev_id, self.this_basis)
118
return self._base_is_ancestor
120
base_is_ancestor = property(_get_base_is_ancestor, _set_base_is_ancestor)
122
def _set_base_is_other_ancestor(self, value):
123
self._base_is_other_ancestor = value
125
def _get_base_is_other_ancestor(self):
126
if self._base_is_other_ancestor is None:
127
if self.other_basis is None:
129
self._base_is_other_ancestor = self.revision_graph.is_ancestor(
130
self.base_rev_id, self.other_basis)
131
return self._base_is_other_ancestor
133
base_is_other_ancestor = property(_get_base_is_other_ancestor,
134
_set_base_is_other_ancestor)
137
def from_uncommitted(tree, other_tree, pb=None, base_tree=None):
138
"""Return a Merger for uncommitted changes in other_tree.
140
:param tree: The tree to merge into
141
:param other_tree: The tree to get uncommitted changes from
142
:param pb: A progress indicator
143
:param base_tree: The basis to use for the merge. If unspecified,
144
other_tree.basis_tree() will be used.
146
if base_tree is None:
147
base_tree = other_tree.basis_tree()
148
merger = Merger(tree.branch, other_tree, base_tree, tree, pb)
149
merger.base_rev_id = merger.base_tree.get_revision_id()
150
merger.other_rev_id = None
151
merger.other_basis = merger.base_rev_id
155
def from_mergeable(klass, tree, mergeable, pb):
156
"""Return a Merger for a bundle or merge directive.
158
:param tree: The tree to merge changes into
159
:param mergeable: A merge directive or bundle
160
:param pb: A progress indicator
162
mergeable.install_revisions(tree.branch.repository)
163
base_revision_id, other_revision_id, verified =\
164
mergeable.get_merge_request(tree.branch.repository)
165
revision_graph = tree.branch.repository.get_graph()
166
if base_revision_id is not None:
167
if (base_revision_id != _mod_revision.NULL_REVISION and
168
revision_graph.is_ancestor(
169
base_revision_id, tree.branch.last_revision())):
170
base_revision_id = None
172
warning('Performing cherrypick')
173
merger = klass.from_revision_ids(pb, tree, other_revision_id,
174
base_revision_id, revision_graph=
176
return merger, verified
179
def from_revision_ids(pb, tree, other, base=None, other_branch=None,
180
base_branch=None, revision_graph=None,
182
"""Return a Merger for revision-ids.
184
:param pb: A progress indicator
185
:param tree: The tree to merge changes into
186
:param other: The revision-id to use as OTHER
187
:param base: The revision-id to use as BASE. If not specified, will
189
:param other_branch: A branch containing the other revision-id. If
190
not supplied, tree.branch is used.
191
:param base_branch: A branch containing the base revision-id. If
192
not supplied, other_branch or tree.branch will be used.
193
:param revision_graph: If you have a revision_graph precomputed, pass
194
it in, otherwise it will be created for you.
195
:param tree_branch: The branch associated with tree. If not supplied,
196
tree.branch will be used.
198
if tree_branch is None:
199
tree_branch = tree.branch
200
merger = Merger(tree_branch, this_tree=tree, pb=pb,
201
revision_graph=revision_graph)
202
if other_branch is None:
203
other_branch = tree.branch
204
merger.set_other_revision(other, other_branch)
208
if base_branch is None:
209
base_branch = other_branch
210
merger.set_base_revision(base, base_branch)
213
def revision_tree(self, revision_id, branch=None):
214
if revision_id not in self._cached_trees:
216
branch = self.this_branch
218
tree = self.this_tree.revision_tree(revision_id)
219
except errors.NoSuchRevisionInTree:
220
tree = branch.repository.revision_tree(revision_id)
221
self._cached_trees[revision_id] = tree
222
return self._cached_trees[revision_id]
224
def _get_tree(self, treespec, possible_transports=None):
225
from bzrlib import workingtree
226
location, revno = treespec
228
tree = workingtree.WorkingTree.open_containing(location)[0]
229
return tree.branch, tree
230
branch = Branch.open_containing(location, possible_transports)[0]
232
revision_id = branch.last_revision()
234
revision_id = branch.get_rev_id(revno)
235
revision_id = ensure_null(revision_id)
236
return branch, self.revision_tree(revision_id, branch)
110
def revision_tree(self, revision_id):
111
return self.this_branch.repository.revision_tree(revision_id)
238
113
def ensure_revision_trees(self):
239
114
if self.this_revision_tree is None:
240
self.this_basis_tree = self.revision_tree(self.this_basis)
115
self.this_basis_tree = self.this_branch.repository.revision_tree(
241
117
if self.this_basis == self.this_rev_id:
242
118
self.this_revision_tree = self.this_basis_tree
244
120
if self.other_rev_id is None:
245
121
other_basis_tree = self.revision_tree(self.other_basis)
246
if other_basis_tree.has_changes(self.other_tree):
122
changes = other_basis_tree.changes_from(self.other_tree)
123
if changes.has_changed():
247
124
raise WorkingTreeNotRevision(self.this_tree)
248
125
other_rev_id = self.other_basis
249
126
self.other_tree = other_basis_tree
264
142
def check_basis(self, check_clean, require_commits=True):
265
143
if self.this_basis is None and require_commits is True:
266
raise BzrCommandError("This branch has no commits."
267
" (perhaps you would prefer 'bzr pull')")
144
raise BzrCommandError("This branch has no commits")
269
146
self.compare_basis()
270
147
if self.this_basis != self.this_rev_id:
271
raise errors.UncommittedChanges(self.this_tree)
148
raise BzrCommandError("Working tree has uncommitted changes.")
273
150
def compare_basis(self):
275
basis_tree = self.revision_tree(self.this_tree.last_revision())
276
except errors.NoSuchRevision:
277
basis_tree = self.this_tree.basis_tree()
278
if not self.this_tree.has_changes(basis_tree):
151
changes = self.this_tree.changes_from(self.this_tree.basis_tree())
152
if not changes.has_changed():
279
153
self.this_rev_id = self.this_basis
281
155
def set_interesting_files(self, file_list):
282
self.interesting_files = file_list
157
self._set_interesting_files(file_list)
158
except NotVersionedError, e:
159
raise BzrCommandError("%s is not a source file in any"
162
def _set_interesting_files(self, file_list):
163
"""Set the list of interesting ids from a list of files."""
164
if file_list is None:
165
self.interesting_ids = None
168
interesting_ids = set()
169
for path in file_list:
171
for tree in (self.this_tree, self.base_tree, self.other_tree):
172
file_id = tree.inventory.path2id(path)
173
if file_id is not None:
174
interesting_ids.add(file_id)
177
raise NotVersionedError(path=path)
178
self.interesting_ids = interesting_ids
284
180
def set_pending(self):
285
if not self.base_is_ancestor or not self.base_is_other_ancestor or self.other_rev_id is None:
289
def _add_parent(self):
290
new_parents = self.this_tree.get_parent_ids() + [self.other_rev_id]
291
new_parent_trees = []
292
for revision_id in new_parents:
294
tree = self.revision_tree(revision_id)
295
except errors.NoSuchRevision:
299
new_parent_trees.append((revision_id, tree))
301
self.this_tree.set_parent_trees(new_parent_trees,
302
allow_leftmost_as_ghost=True)
304
for _revision_id, tree in new_parent_trees:
308
def set_other(self, other_revision, possible_transports=None):
181
if not self.base_is_ancestor:
183
if self.other_rev_id is None:
185
ancestry = self.this_branch.repository.get_ancestry(self.this_basis)
186
if self.other_rev_id in ancestry:
188
self.this_tree.add_parent_tree((self.other_rev_id, self.other_tree))
190
def set_other(self, other_revision):
309
191
"""Set the revision and tree to merge from.
311
193
This sets the other_tree, other_rev_id, other_basis attributes.
313
195
:param other_revision: The [path, revision] list to merge from.
315
self.other_branch, self.other_tree = self._get_tree(other_revision,
197
other_branch, self.other_tree = _get_tree(other_revision,
317
199
if other_revision[1] == -1:
318
self.other_rev_id = _mod_revision.ensure_null(
319
self.other_branch.last_revision())
320
if _mod_revision.is_null(self.other_rev_id):
321
raise NoCommits(self.other_branch)
200
self.other_rev_id = other_branch.last_revision()
201
if self.other_rev_id is None:
202
raise NoCommits(other_branch)
322
203
self.other_basis = self.other_rev_id
323
204
elif other_revision[1] is not None:
324
self.other_rev_id = self.other_branch.get_rev_id(other_revision[1])
205
self.other_rev_id = other_branch.get_rev_id(other_revision[1])
325
206
self.other_basis = self.other_rev_id
327
208
self.other_rev_id = None
328
self.other_basis = self.other_branch.last_revision()
209
self.other_basis = other_branch.last_revision()
329
210
if self.other_basis is None:
330
raise NoCommits(self.other_branch)
331
if self.other_rev_id is not None:
332
self._cached_trees[self.other_rev_id] = self.other_tree
333
self._maybe_fetch(self.other_branch,self.this_branch, self.other_basis)
335
def set_other_revision(self, revision_id, other_branch):
336
"""Set 'other' based on a branch and revision id
338
:param revision_id: The revision to use for a tree
339
:param other_branch: The branch containing this tree
341
self.other_rev_id = revision_id
342
self.other_branch = other_branch
343
self._maybe_fetch(other_branch, self.this_branch, self.other_rev_id)
344
self.other_tree = self.revision_tree(revision_id)
345
self.other_basis = revision_id
347
def set_base_revision(self, revision_id, branch):
348
"""Set 'base' based on a branch and revision id
350
:param revision_id: The revision to use for a tree
351
:param branch: The branch containing this tree
353
self.base_rev_id = revision_id
354
self.base_branch = branch
355
self._maybe_fetch(branch, self.this_branch, revision_id)
356
self.base_tree = self.revision_tree(revision_id)
358
def _maybe_fetch(self, source, target, revision_id):
359
if not source.repository.has_same_location(target.repository):
360
target.fetch(source, revision_id)
211
raise NoCommits(other_branch)
212
if other_branch.base != self.this_branch.base:
213
self.this_branch.fetch(other_branch, last_revision=self.other_basis)
362
215
def find_base(self):
363
revisions = [ensure_null(self.this_basis),
364
ensure_null(self.other_basis)]
365
if NULL_REVISION in revisions:
366
self.base_rev_id = NULL_REVISION
367
self.base_tree = self.revision_tree(self.base_rev_id)
368
self._is_criss_cross = False
370
lcas = self.revision_graph.find_lca(revisions[0], revisions[1])
371
self._is_criss_cross = False
373
self.base_rev_id = NULL_REVISION
375
self.base_rev_id = list(lcas)[0]
376
else: # len(lcas) > 1
378
# find_unique_lca can only handle 2 nodes, so we have to
379
# start back at the beginning. It is a shame to traverse
380
# the graph again, but better than re-implementing
382
self.base_rev_id = self.revision_graph.find_unique_lca(
383
revisions[0], revisions[1])
385
self.base_rev_id = self.revision_graph.find_unique_lca(
387
self._is_criss_cross = True
388
if self.base_rev_id == NULL_REVISION:
389
raise UnrelatedBranches()
390
if self._is_criss_cross:
391
warning('Warning: criss-cross merge encountered. See bzr'
392
' help criss-cross.')
393
mutter('Criss-cross lcas: %r' % lcas)
394
interesting_revision_ids = [self.base_rev_id]
395
interesting_revision_ids.extend(lcas)
396
interesting_trees = dict((t.get_revision_id(), t)
397
for t in self.this_branch.repository.revision_trees(
398
interesting_revision_ids))
399
self._cached_trees.update(interesting_trees)
400
self.base_tree = interesting_trees.pop(self.base_rev_id)
401
sorted_lca_keys = self.revision_graph.find_merge_order(
403
self._lca_trees = [interesting_trees[key]
404
for key in sorted_lca_keys]
406
self.base_tree = self.revision_tree(self.base_rev_id)
407
self.base_is_ancestor = True
408
self.base_is_other_ancestor = True
409
mutter('Base revid: %r' % self.base_rev_id)
216
self.set_base([None, None])
411
218
def set_base(self, base_revision):
412
219
"""Set the base revision to use for the merge.
445
266
kwargs['show_base'] = self.show_base
446
267
elif self.show_base:
447
268
raise BzrError("Showing base is not supported for this"
448
" merge type. %s" % self.merge_type)
449
if (not getattr(self.merge_type, 'supports_reverse_cherrypick', True)
450
and not self.base_is_other_ancestor):
451
raise errors.CannotReverseCherrypick()
452
if self.merge_type.supports_cherrypick:
453
kwargs['cherrypick'] = (not self.base_is_ancestor or
454
not self.base_is_other_ancestor)
455
if self._is_criss_cross and getattr(self.merge_type,
456
'supports_lca_trees', False):
457
kwargs['lca_trees'] = self._lca_trees
458
return self.merge_type(pb=self._pb,
459
change_reporter=self.change_reporter,
462
def _do_merge_to(self, merge):
463
if self.other_branch is not None:
464
self.other_branch.update_references(self.this_branch)
466
if self.recurse == 'down':
467
for relpath, file_id in self.this_tree.iter_references():
468
sub_tree = self.this_tree.get_nested_tree(file_id, relpath)
469
other_revision = self.other_tree.get_reference_revision(
471
if other_revision == sub_tree.last_revision():
473
sub_merge = Merger(sub_tree.branch, this_tree=sub_tree)
474
sub_merge.merge_type = self.merge_type
475
other_branch = self.other_branch.reference_parent(file_id, relpath)
476
sub_merge.set_other_revision(other_revision, other_branch)
477
base_revision = self.base_tree.get_reference_revision(file_id)
478
sub_merge.base_tree = \
479
sub_tree.branch.repository.revision_tree(base_revision)
480
sub_merge.base_rev_id = base_revision
484
self.this_tree.lock_tree_write()
486
if self.base_tree is not None:
487
self.base_tree.lock_read()
489
if self.other_tree is not None:
490
self.other_tree.lock_read()
492
merge = self.make_merger()
493
self._do_merge_to(merge)
495
if self.other_tree is not None:
496
self.other_tree.unlock()
498
if self.base_tree is not None:
499
self.base_tree.unlock()
501
self.this_tree.unlock()
269
" merge type. %s" % self.merge_type)
270
merge = self.merge_type(pb=self._pb, **kwargs)
502
271
if len(merge.cooked_conflicts) == 0:
503
if not self.ignore_zero and not is_quiet():
272
if not self.ignore_zero:
504
273
note("All changes applied successfully.")
506
275
note("%d conflicts encountered." % len(merge.cooked_conflicts))
508
277
return len(merge.cooked_conflicts)
511
class _InventoryNoneEntry(object):
512
"""This represents an inventory entry which *isn't there*.
514
It simplifies the merging logic if we always have an InventoryEntry, even
515
if it isn't actually present
522
symlink_target = None
525
_none_entry = _InventoryNoneEntry()
279
def regen_inventory(self, new_entries):
280
old_entries = self.this_tree.read_working_inventory()
284
for path, file_id in new_entries:
287
new_entries_map[file_id] = path
289
def id2path(file_id):
290
path = new_entries_map.get(file_id)
293
entry = old_entries[file_id]
294
if entry.parent_id is None:
296
return pathjoin(id2path(entry.parent_id), entry.name)
298
for file_id in old_entries:
299
entry = old_entries[file_id]
300
path = id2path(file_id)
301
if file_id in self.base_tree.inventory:
302
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
304
executable = getattr(entry, 'executable', False)
305
new_inventory[file_id] = (path, file_id, entry.parent_id,
306
entry.kind, executable)
308
by_path[path] = file_id
313
for path, file_id in new_entries:
315
del new_inventory[file_id]
318
new_path_list.append((path, file_id))
319
if file_id not in old_entries:
321
# Ensure no file is added before its parent
323
for path, file_id in new_path_list:
327
parent = by_path[os.path.dirname(path)]
328
abspath = pathjoin(self.this_tree.basedir, path)
329
kind = osutils.file_kind(abspath)
330
if file_id in self.base_tree.inventory:
331
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
334
new_inventory[file_id] = (path, file_id, parent, kind, executable)
335
by_path[path] = file_id
337
# Get a list in insertion order
338
new_inventory_list = new_inventory.values()
339
mutter ("""Inventory regeneration:
340
old length: %i insertions: %i deletions: %i new_length: %i"""\
341
% (len(old_entries), insertions, deletions,
342
len(new_inventory_list)))
343
assert len(new_inventory_list) == len(old_entries) + insertions\
345
new_inventory_list.sort()
346
return new_inventory_list
528
349
class Merge3Merger(object):
579
365
self.cooked_conflicts = []
580
366
self.reprocess = reprocess
581
367
self.show_base = show_base
582
self._lca_trees = lca_trees
583
# Uncommenting this will change the default algorithm to always use
584
# _entries_lca. This can be useful for running the test suite and
585
# making sure we haven't missed any corner cases.
586
# if lca_trees is None:
587
# self._lca_trees = [self.base_tree]
590
self.change_reporter = change_reporter
591
self.cherrypick = cherrypick
592
370
if self.pp is None:
593
371
self.pp = ProgressPhase("Merge phase", 3, self.pb)
598
self.this_tree.lock_tree_write()
599
self.base_tree.lock_read()
600
self.other_tree.lock_read()
601
self.tt = TreeTransform(self.this_tree, self.pb)
373
if interesting_ids is not None:
374
all_ids = interesting_ids
376
all_ids = set(base_tree)
377
all_ids.update(other_tree)
378
working_tree.lock_tree_write()
379
self.tt = TreeTransform(working_tree, self.pb)
603
381
self.pp.next_phase()
604
self._compute_transform()
606
results = self.tt.apply(no_conflicts=True)
382
child_pb = ui.ui_factory.nested_progress_bar()
384
for num, file_id in enumerate(all_ids):
385
child_pb.update('Preparing file merge', num, len(all_ids))
386
self.merge_names(file_id)
387
file_status = self.merge_contents(file_id)
388
self.merge_executable(file_id, file_status)
393
child_pb = ui.ui_factory.nested_progress_bar()
395
fs_conflicts = resolve_conflicts(self.tt, child_pb)
398
self.cook_conflicts(fs_conflicts)
399
for conflict in self.cooked_conflicts:
402
results = self.tt.apply()
607
403
self.write_modified(results)
609
self.this_tree.add_conflicts(self.cooked_conflicts)
405
working_tree.add_conflicts(self.cooked_conflicts)
610
406
except UnsupportedOperation:
613
409
self.tt.finalize()
614
self.other_tree.unlock()
615
self.base_tree.unlock()
616
self.this_tree.unlock()
619
def make_preview_transform(self):
620
self.base_tree.lock_read()
621
self.other_tree.lock_read()
622
self.tt = TransformPreview(self.this_tree)
625
self._compute_transform()
628
self.other_tree.unlock()
629
self.base_tree.unlock()
633
def _compute_transform(self):
634
if self._lca_trees is None:
635
entries = self._entries3()
636
resolver = self._three_way
638
entries = self._entries_lca()
639
resolver = self._lca_multi_way
640
child_pb = ui.ui_factory.nested_progress_bar()
642
for num, (file_id, changed, parents3, names3,
643
executable3) in enumerate(entries):
644
child_pb.update('Preparing file merge', num, len(entries))
645
self._merge_names(file_id, parents3, names3, resolver=resolver)
647
file_status = self.merge_contents(file_id)
649
file_status = 'unmodified'
650
self._merge_executable(file_id,
651
executable3, file_status, resolver=resolver)
656
child_pb = ui.ui_factory.nested_progress_bar()
658
fs_conflicts = resolve_conflicts(self.tt, child_pb,
659
lambda t, c: conflict_pass(t, c, self.other_tree))
662
if self.change_reporter is not None:
663
from bzrlib import delta
664
delta.report_changes(
665
self.tt.iter_changes(), self.change_reporter)
666
self.cook_conflicts(fs_conflicts)
667
for conflict in self.cooked_conflicts:
671
"""Gather data about files modified between three trees.
673
Return a list of tuples of file_id, changed, parents3, names3,
674
executable3. changed is a boolean indicating whether the file contents
675
or kind were changed. parents3 is a tuple of parent ids for base,
676
other and this. names3 is a tuple of names for base, other and this.
677
executable3 is a tuple of execute-bit values for base, other and this.
680
iterator = self.other_tree.iter_changes(self.base_tree,
681
include_unchanged=True, specific_files=self.interesting_files,
682
extra_trees=[self.this_tree])
683
this_entries = dict((e.file_id, e) for p, e in
684
self.this_tree.iter_entries_by_dir(
685
self.interesting_ids))
686
for (file_id, paths, changed, versioned, parents, names, kind,
687
executable) in iterator:
688
if (self.interesting_ids is not None and
689
file_id not in self.interesting_ids):
691
entry = this_entries.get(file_id)
692
if entry is not None:
693
this_name = entry.name
694
this_parent = entry.parent_id
695
this_executable = entry.executable
699
this_executable = None
700
parents3 = parents + (this_parent,)
701
names3 = names + (this_name,)
702
executable3 = executable + (this_executable,)
703
result.append((file_id, changed, parents3, names3, executable3))
706
def _entries_lca(self):
707
"""Gather data about files modified between multiple trees.
709
This compares OTHER versus all LCA trees, and for interesting entries,
710
it then compares with THIS and BASE.
712
For the multi-valued entries, the format will be (BASE, [lca1, lca2])
713
:return: [(file_id, changed, parents, names, executable)]
714
file_id Simple file_id of the entry
715
changed Boolean, True if the kind or contents changed
717
parents ((base, [parent_id, in, lcas]), parent_id_other,
719
names ((base, [name, in, lcas]), name_in_other, name_in_this)
720
executable ((base, [exec, in, lcas]), exec_in_other, exec_in_this)
722
if self.interesting_files is not None:
723
lookup_trees = [self.this_tree, self.base_tree]
724
lookup_trees.extend(self._lca_trees)
725
# I think we should include the lca trees as well
726
interesting_ids = self.other_tree.paths2ids(self.interesting_files,
729
interesting_ids = self.interesting_ids
731
walker = _mod_tree.MultiWalker(self.other_tree, self._lca_trees)
733
base_inventory = self.base_tree.inventory
734
this_inventory = self.this_tree.inventory
735
for path, file_id, other_ie, lca_values in walker.iter_all():
736
# Is this modified at all from any of the other trees?
738
other_ie = _none_entry
739
if interesting_ids is not None and file_id not in interesting_ids:
742
# If other_revision is found in any of the lcas, that means this
743
# node is uninteresting. This is because when merging, if there are
744
# multiple heads(), we have to create a new node. So if we didn't,
745
# we know that the ancestry is linear, and that OTHER did not
747
# See doc/developers/lca_merge_resolution.txt for details
748
other_revision = other_ie.revision
749
if other_revision is not None:
750
# We can't use this shortcut when other_revision is None,
751
# because it may be None because things are WorkingTrees, and
752
# not because it is *actually* None.
753
is_unmodified = False
754
for lca_path, ie in lca_values:
755
if ie is not None and ie.revision == other_revision:
762
for lca_path, lca_ie in lca_values:
764
lca_entries.append(_none_entry)
766
lca_entries.append(lca_ie)
768
if file_id in base_inventory:
769
base_ie = base_inventory[file_id]
771
base_ie = _none_entry
773
if file_id in this_inventory:
774
this_ie = this_inventory[file_id]
776
this_ie = _none_entry
782
for lca_ie in lca_entries:
783
lca_kinds.append(lca_ie.kind)
784
lca_parent_ids.append(lca_ie.parent_id)
785
lca_names.append(lca_ie.name)
786
lca_executable.append(lca_ie.executable)
788
kind_winner = self._lca_multi_way(
789
(base_ie.kind, lca_kinds),
790
other_ie.kind, this_ie.kind)
791
parent_id_winner = self._lca_multi_way(
792
(base_ie.parent_id, lca_parent_ids),
793
other_ie.parent_id, this_ie.parent_id)
794
name_winner = self._lca_multi_way(
795
(base_ie.name, lca_names),
796
other_ie.name, this_ie.name)
798
content_changed = True
799
if kind_winner == 'this':
800
# No kind change in OTHER, see if there are *any* changes
801
if other_ie.kind == 'directory':
802
if parent_id_winner == 'this' and name_winner == 'this':
803
# No change for this directory in OTHER, skip
805
content_changed = False
806
elif other_ie.kind is None or other_ie.kind == 'file':
807
def get_sha1(ie, tree):
808
if ie.kind != 'file':
810
return tree.get_file_sha1(file_id)
811
base_sha1 = get_sha1(base_ie, self.base_tree)
812
lca_sha1s = [get_sha1(ie, tree) for ie, tree
813
in zip(lca_entries, self._lca_trees)]
814
this_sha1 = get_sha1(this_ie, self.this_tree)
815
other_sha1 = get_sha1(other_ie, self.other_tree)
816
sha1_winner = self._lca_multi_way(
817
(base_sha1, lca_sha1s), other_sha1, this_sha1,
818
allow_overriding_lca=False)
819
exec_winner = self._lca_multi_way(
820
(base_ie.executable, lca_executable),
821
other_ie.executable, this_ie.executable)
822
if (parent_id_winner == 'this' and name_winner == 'this'
823
and sha1_winner == 'this' and exec_winner == 'this'):
824
# No kind, parent, name, exec, or content change for
825
# OTHER, so this node is not considered interesting
827
if sha1_winner == 'this':
828
content_changed = False
829
elif other_ie.kind == 'symlink':
830
def get_target(ie, tree):
831
if ie.kind != 'symlink':
833
return tree.get_symlink_target(file_id)
834
base_target = get_target(base_ie, self.base_tree)
835
lca_targets = [get_target(ie, tree) for ie, tree
836
in zip(lca_entries, self._lca_trees)]
837
this_target = get_target(this_ie, self.this_tree)
838
other_target = get_target(other_ie, self.other_tree)
839
target_winner = self._lca_multi_way(
840
(base_target, lca_targets),
841
other_target, this_target)
842
if (parent_id_winner == 'this' and name_winner == 'this'
843
and target_winner == 'this'):
844
# No kind, parent, name, or symlink target change
847
if target_winner == 'this':
848
content_changed = False
849
elif other_ie.kind == 'tree-reference':
850
# The 'changed' information seems to be handled at a higher
851
# level. At least, _entries3 returns False for content
852
# changed, even when at a new revision_id.
853
content_changed = False
854
if (parent_id_winner == 'this' and name_winner == 'this'):
855
# Nothing interesting
858
raise AssertionError('unhandled kind: %s' % other_ie.kind)
859
# XXX: We need to handle kind == 'symlink'
861
# If we have gotten this far, that means something has changed
862
result.append((file_id, content_changed,
863
((base_ie.parent_id, lca_parent_ids),
864
other_ie.parent_id, this_ie.parent_id),
865
((base_ie.name, lca_names),
866
other_ie.name, this_ie.name),
867
((base_ie.executable, lca_executable),
868
other_ie.executable, this_ie.executable)
410
working_tree.unlock()
873
413
def fix_root(self):
1481
924
def merge_inner(this_branch, other_tree, base_tree, ignore_zero=False,
1483
merge_type=Merge3Merger,
1484
interesting_ids=None,
926
merge_type=Merge3Merger,
927
interesting_ids=None,
1487
930
other_rev_id=None,
1488
931
interesting_files=None,
1491
change_reporter=None):
1492
"""Primary interface for merging.
934
"""Primary interface for merging.
1494
typical use is probably
936
typical use is probably
1495
937
'merge_inner(branch, branch.get_revision_tree(other_revision),
1496
938
branch.get_revision_tree(base_revision))'
1498
940
if this_tree is None:
1499
raise BzrError("bzrlib.merge.merge_inner requires a this_tree "
1500
"parameter as of bzrlib version 0.8.")
1501
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1502
pb=pb, change_reporter=change_reporter)
941
warnings.warn("bzrlib.merge.merge_inner requires a this_tree parameter as of "
942
"bzrlib version 0.8.",
945
this_tree = this_branch.bzrdir.open_workingtree()
946
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1503
948
merger.backup_files = backup_files
1504
949
merger.merge_type = merge_type
1505
950
merger.interesting_ids = interesting_ids
1506
951
merger.ignore_zero = ignore_zero
1507
952
if interesting_files:
1509
raise ValueError('Only supply interesting_ids'
1510
' or interesting_files')
1511
merger.interesting_files = interesting_files
953
assert not interesting_ids, ('Only supply interesting_ids'
954
' or interesting_files')
955
merger._set_interesting_files(interesting_files)
1512
956
merger.show_base = show_base
1513
957
merger.reprocess = reprocess
1514
958
merger.other_rev_id = other_rev_id
1515
959
merger.other_basis = other_rev_id
1516
get_revision_id = getattr(base_tree, 'get_revision_id', None)
1517
if get_revision_id is None:
1518
get_revision_id = base_tree.last_revision
1519
merger.set_base_revision(get_revision_id(), this_branch)
1520
960
return merger.do_merge()
1522
def get_merge_type_registry():
1523
"""Merge type registry is in bzrlib.option to avoid circular imports.
1525
This method provides a sanctioned way to retrieve it.
1527
from bzrlib import option
1528
return option._merge_type_registry
1531
def _plan_annotate_merge(annotated_a, annotated_b, ancestors_a, ancestors_b):
1532
def status_a(revision, text):
1533
if revision in ancestors_b:
1534
return 'killed-b', text
1536
return 'new-a', text
1538
def status_b(revision, text):
1539
if revision in ancestors_a:
1540
return 'killed-a', text
1542
return 'new-b', text
1544
plain_a = [t for (a, t) in annotated_a]
1545
plain_b = [t for (a, t) in annotated_b]
1546
matcher = patiencediff.PatienceSequenceMatcher(None, plain_a, plain_b)
1547
blocks = matcher.get_matching_blocks()
1550
for ai, bi, l in blocks:
1551
# process all mismatched sections
1552
# (last mismatched section is handled because blocks always
1553
# includes a 0-length last block)
1554
for revision, text in annotated_a[a_cur:ai]:
1555
yield status_a(revision, text)
1556
for revision, text in annotated_b[b_cur:bi]:
1557
yield status_b(revision, text)
1558
# and now the matched section
1561
for text_a in plain_a[ai:a_cur]:
1562
yield "unchanged", text_a
1565
class _PlanMergeBase(object):
1567
def __init__(self, a_rev, b_rev, vf, key_prefix):
1570
:param a_rev: Revision-id of one revision to merge
1571
:param b_rev: Revision-id of the other revision to merge
1572
:param vf: A VersionedFiles containing both revisions
1573
:param key_prefix: A prefix for accessing keys in vf, typically
1579
self._last_lines = None
1580
self._last_lines_revision_id = None
1581
self._cached_matching_blocks = {}
1582
self._key_prefix = key_prefix
1583
self._precache_tip_lines()
1585
def _precache_tip_lines(self):
1586
lines = self.get_lines([self.a_rev, self.b_rev])
1587
self.lines_a = lines[self.a_rev]
1588
self.lines_b = lines[self.b_rev]
1590
def get_lines(self, revisions):
1591
"""Get lines for revisions from the backing VersionedFiles.
1593
:raises RevisionNotPresent: on absent texts.
1595
keys = [(self._key_prefix + (rev,)) for rev in revisions]
1597
for record in self.vf.get_record_stream(keys, 'unordered', True):
1598
if record.storage_kind == 'absent':
1599
raise errors.RevisionNotPresent(record.key, self.vf)
1600
result[record.key[-1]] = osutils.chunks_to_lines(
1601
record.get_bytes_as('chunked'))
1604
def plan_merge(self):
1605
"""Generate a 'plan' for merging the two revisions.
1607
This involves comparing their texts and determining the cause of
1608
differences. If text A has a line and text B does not, then either the
1609
line was added to text A, or it was deleted from B. Once the causes
1610
are combined, they are written out in the format described in
1611
VersionedFile.plan_merge
1613
blocks = self._get_matching_blocks(self.a_rev, self.b_rev)
1614
unique_a, unique_b = self._unique_lines(blocks)
1615
new_a, killed_b = self._determine_status(self.a_rev, unique_a)
1616
new_b, killed_a = self._determine_status(self.b_rev, unique_b)
1617
return self._iter_plan(blocks, new_a, killed_b, new_b, killed_a)
1619
def _iter_plan(self, blocks, new_a, killed_b, new_b, killed_a):
1622
for i, j, n in blocks:
1623
for a_index in range(last_i, i):
1624
if a_index in new_a:
1625
if a_index in killed_b:
1626
yield 'conflicted-a', self.lines_a[a_index]
1628
yield 'new-a', self.lines_a[a_index]
1630
yield 'killed-b', self.lines_a[a_index]
1631
for b_index in range(last_j, j):
1632
if b_index in new_b:
1633
if b_index in killed_a:
1634
yield 'conflicted-b', self.lines_b[b_index]
1636
yield 'new-b', self.lines_b[b_index]
1638
yield 'killed-a', self.lines_b[b_index]
1639
# handle common lines
1640
for a_index in range(i, i+n):
1641
yield 'unchanged', self.lines_a[a_index]
1645
def _get_matching_blocks(self, left_revision, right_revision):
1646
"""Return a description of which sections of two revisions match.
1648
See SequenceMatcher.get_matching_blocks
1650
cached = self._cached_matching_blocks.get((left_revision,
1652
if cached is not None:
1654
if self._last_lines_revision_id == left_revision:
1655
left_lines = self._last_lines
1656
right_lines = self.get_lines([right_revision])[right_revision]
1658
lines = self.get_lines([left_revision, right_revision])
1659
left_lines = lines[left_revision]
1660
right_lines = lines[right_revision]
1661
self._last_lines = right_lines
1662
self._last_lines_revision_id = right_revision
1663
matcher = patiencediff.PatienceSequenceMatcher(None, left_lines,
1665
return matcher.get_matching_blocks()
1667
def _unique_lines(self, matching_blocks):
1668
"""Analyse matching_blocks to determine which lines are unique
1670
:return: a tuple of (unique_left, unique_right), where the values are
1671
sets of line numbers of unique lines.
1677
for i, j, n in matching_blocks:
1678
unique_left.extend(range(last_i, i))
1679
unique_right.extend(range(last_j, j))
1682
return unique_left, unique_right
1685
def _subtract_plans(old_plan, new_plan):
1686
"""Remove changes from new_plan that came from old_plan.
1688
It is assumed that the difference between the old_plan and new_plan
1689
is their choice of 'b' text.
1691
All lines from new_plan that differ from old_plan are emitted
1692
verbatim. All lines from new_plan that match old_plan but are
1693
not about the 'b' revision are emitted verbatim.
1695
Lines that match and are about the 'b' revision are the lines we
1696
don't want, so we convert 'killed-b' -> 'unchanged', and 'new-b'
1697
is skipped entirely.
1699
matcher = patiencediff.PatienceSequenceMatcher(None, old_plan,
1702
for i, j, n in matcher.get_matching_blocks():
1703
for jj in range(last_j, j):
1705
for jj in range(j, j+n):
1706
plan_line = new_plan[jj]
1707
if plan_line[0] == 'new-b':
1709
elif plan_line[0] == 'killed-b':
1710
yield 'unchanged', plan_line[1]
1716
class _PlanMerge(_PlanMergeBase):
1717
"""Plan an annotate merge using on-the-fly annotation"""
1719
def __init__(self, a_rev, b_rev, vf, key_prefix):
1720
super(_PlanMerge, self).__init__(a_rev, b_rev, vf, key_prefix)
1721
self.a_key = self._key_prefix + (self.a_rev,)
1722
self.b_key = self._key_prefix + (self.b_rev,)
1723
self.graph = Graph(self.vf)
1724
heads = self.graph.heads((self.a_key, self.b_key))
1726
# one side dominates, so we can just return its values, yay for
1728
# Ideally we would know that before we get this far
1729
self._head_key = heads.pop()
1730
if self._head_key == self.a_key:
1734
mutter('found dominating revision for %s\n%s > %s', self.vf,
1735
self._head_key[-1], other)
1738
self._head_key = None
1741
def _precache_tip_lines(self):
1742
# Turn this into a no-op, because we will do this later
1745
def _find_recursive_lcas(self):
1746
"""Find all the ancestors back to a unique lca"""
1747
cur_ancestors = (self.a_key, self.b_key)
1748
# graph.find_lca(uncommon, keys) now returns plain NULL_REVISION,
1749
# rather than a key tuple. We will just map that directly to no common
1753
next_lcas = self.graph.find_lca(*cur_ancestors)
1754
# Map a plain NULL_REVISION to a simple no-ancestors
1755
if next_lcas == set([NULL_REVISION]):
1757
# Order the lca's based on when they were merged into the tip
1758
# While the actual merge portion of weave merge uses a set() of
1759
# active revisions, the order of insertion *does* effect the
1760
# implicit ordering of the texts.
1761
for rev_key in cur_ancestors:
1762
ordered_parents = tuple(self.graph.find_merge_order(rev_key,
1764
parent_map[rev_key] = ordered_parents
1765
if len(next_lcas) == 0:
1767
elif len(next_lcas) == 1:
1768
parent_map[list(next_lcas)[0]] = ()
1770
elif len(next_lcas) > 2:
1771
# More than 2 lca's, fall back to grabbing all nodes between
1772
# this and the unique lca.
1773
mutter('More than 2 LCAs, falling back to all nodes for:'
1774
' %s, %s\n=> %s', self.a_key, self.b_key, cur_ancestors)
1775
cur_lcas = next_lcas
1776
while len(cur_lcas) > 1:
1777
cur_lcas = self.graph.find_lca(*cur_lcas)
1778
if len(cur_lcas) == 0:
1779
# No common base to find, use the full ancestry
1782
unique_lca = list(cur_lcas)[0]
1783
if unique_lca == NULL_REVISION:
1784
# find_lca will return a plain 'NULL_REVISION' rather
1785
# than a key tuple when there is no common ancestor, we
1786
# prefer to just use None, because it doesn't confuse
1787
# _get_interesting_texts()
1789
parent_map.update(self._find_unique_parents(next_lcas,
1792
cur_ancestors = next_lcas
1795
def _find_unique_parents(self, tip_keys, base_key):
1796
"""Find ancestors of tip that aren't ancestors of base.
1798
:param tip_keys: Nodes that are interesting
1799
:param base_key: Cull all ancestors of this node
1800
:return: The parent map for all revisions between tip_keys and
1801
base_key. base_key will be included. References to nodes outside of
1802
the ancestor set will also be removed.
1804
# TODO: this would be simpler if find_unique_ancestors took a list
1805
# instead of a single tip, internally it supports it, but it
1806
# isn't a "backwards compatible" api change.
1807
if base_key is None:
1808
parent_map = dict(self.graph.iter_ancestry(tip_keys))
1809
# We remove NULL_REVISION because it isn't a proper tuple key, and
1810
# thus confuses things like _get_interesting_texts, and our logic
1811
# to add the texts into the memory weave.
1812
if NULL_REVISION in parent_map:
1813
parent_map.pop(NULL_REVISION)
1816
for tip in tip_keys:
1818
self.graph.find_unique_ancestors(tip, [base_key]))
1819
parent_map = self.graph.get_parent_map(interesting)
1820
parent_map[base_key] = ()
1821
culled_parent_map, child_map, tails = self._remove_external_references(
1823
# Remove all the tails but base_key
1824
if base_key is not None:
1825
tails.remove(base_key)
1826
self._prune_tails(culled_parent_map, child_map, tails)
1827
# Now remove all the uninteresting 'linear' regions
1828
simple_map = _mod_graph.collapse_linear_regions(culled_parent_map)
1832
def _remove_external_references(parent_map):
1833
"""Remove references that go outside of the parent map.
1835
:param parent_map: Something returned from Graph.get_parent_map(keys)
1836
:return: (filtered_parent_map, child_map, tails)
1837
filtered_parent_map is parent_map without external references
1838
child_map is the {parent_key: [child_keys]} mapping
1839
tails is a list of nodes that do not have any parents in the map
1841
# TODO: The basic effect of this function seems more generic than
1842
# _PlanMerge. But the specific details of building a child_map,
1843
# and computing tails seems very specific to _PlanMerge.
1844
# Still, should this be in Graph land?
1845
filtered_parent_map = {}
1848
for key, parent_keys in parent_map.iteritems():
1849
culled_parent_keys = [p for p in parent_keys if p in parent_map]
1850
if not culled_parent_keys:
1852
for parent_key in culled_parent_keys:
1853
child_map.setdefault(parent_key, []).append(key)
1854
# TODO: Do we want to do this, it adds overhead for every node,
1855
# just to say that the node has no children
1856
child_map.setdefault(key, [])
1857
filtered_parent_map[key] = culled_parent_keys
1858
return filtered_parent_map, child_map, tails
1861
def _prune_tails(parent_map, child_map, tails_to_remove):
1862
"""Remove tails from the parent map.
1864
This will remove the supplied revisions until no more children have 0
1867
:param parent_map: A dict of {child: [parents]}, this dictionary will
1868
be modified in place.
1869
:param tails_to_remove: A list of tips that should be removed,
1870
this list will be consumed
1871
:param child_map: The reverse dict of parent_map ({parent: [children]})
1872
this dict will be modified
1873
:return: None, parent_map will be modified in place.
1875
while tails_to_remove:
1876
next = tails_to_remove.pop()
1877
parent_map.pop(next)
1878
children = child_map.pop(next)
1879
for child in children:
1880
child_parents = parent_map[child]
1881
child_parents.remove(next)
1882
if len(child_parents) == 0:
1883
tails_to_remove.append(child)
1885
def _get_interesting_texts(self, parent_map):
1886
"""Return a dict of texts we are interested in.
1888
Note that the input is in key tuples, but the output is in plain
1891
:param parent_map: The output from _find_recursive_lcas
1892
:return: A dict of {'revision_id':lines} as returned by
1893
_PlanMergeBase.get_lines()
1895
all_revision_keys = set(parent_map)
1896
all_revision_keys.add(self.a_key)
1897
all_revision_keys.add(self.b_key)
1899
# Everything else is in 'keys' but get_lines is in 'revision_ids'
1900
all_texts = self.get_lines([k[-1] for k in all_revision_keys])
1903
def _build_weave(self):
1904
from bzrlib import weave
1905
self._weave = weave.Weave(weave_name='in_memory_weave',
1906
allow_reserved=True)
1907
parent_map = self._find_recursive_lcas()
1909
all_texts = self._get_interesting_texts(parent_map)
1911
# Note: Unfortunately, the order given by topo_sort will effect the
1912
# ordering resolution in the output. Specifically, if you add A then B,
1913
# then in the output text A lines will show up before B lines. And, of
1914
# course, topo_sort doesn't guarantee any real ordering.
1915
# So we use merge_sort, and add a fake node on the tip.
1916
# This ensures that left-hand parents will always be inserted into the
1917
# weave before right-hand parents.
1918
tip_key = self._key_prefix + (_mod_revision.CURRENT_REVISION,)
1919
parent_map[tip_key] = (self.a_key, self.b_key)
1921
for seq_num, key, depth, eom in reversed(tsort.merge_sort(parent_map,
1925
# for key in tsort.topo_sort(parent_map):
1926
parent_keys = parent_map[key]
1927
revision_id = key[-1]
1928
parent_ids = [k[-1] for k in parent_keys]
1929
self._weave.add_lines(revision_id, parent_ids,
1930
all_texts[revision_id])
1932
def plan_merge(self):
1933
"""Generate a 'plan' for merging the two revisions.
1935
This involves comparing their texts and determining the cause of
1936
differences. If text A has a line and text B does not, then either the
1937
line was added to text A, or it was deleted from B. Once the causes
1938
are combined, they are written out in the format described in
1939
VersionedFile.plan_merge
1941
if self._head_key is not None: # There was a single head
1942
if self._head_key == self.a_key:
1945
if self._head_key != self.b_key:
1946
raise AssertionError('There was an invalid head: %s != %s'
1947
% (self.b_key, self._head_key))
1949
head_rev = self._head_key[-1]
1950
lines = self.get_lines([head_rev])[head_rev]
1951
return ((plan, line) for line in lines)
1952
return self._weave.plan_merge(self.a_rev, self.b_rev)
1955
class _PlanLCAMerge(_PlanMergeBase):
1957
This merge algorithm differs from _PlanMerge in that:
1958
1. comparisons are done against LCAs only
1959
2. cases where a contested line is new versus one LCA but old versus
1960
another are marked as conflicts, by emitting the line as conflicted-a
1963
This is faster, and hopefully produces more useful output.
1966
def __init__(self, a_rev, b_rev, vf, key_prefix, graph):
1967
_PlanMergeBase.__init__(self, a_rev, b_rev, vf, key_prefix)
1968
lcas = graph.find_lca(key_prefix + (a_rev,), key_prefix + (b_rev,))
1971
if lca == NULL_REVISION:
1974
self.lcas.add(lca[-1])
1975
for lca in self.lcas:
1976
if _mod_revision.is_null(lca):
1979
lca_lines = self.get_lines([lca])[lca]
1980
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_a,
1982
blocks = list(matcher.get_matching_blocks())
1983
self._cached_matching_blocks[(a_rev, lca)] = blocks
1984
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_b,
1986
blocks = list(matcher.get_matching_blocks())
1987
self._cached_matching_blocks[(b_rev, lca)] = blocks
1989
def _determine_status(self, revision_id, unique_line_numbers):
1990
"""Determines the status unique lines versus all lcas.
1992
Basically, determines why the line is unique to this revision.
1994
A line may be determined new, killed, or both.
1996
If a line is determined new, that means it was not present in at least
1997
one LCA, and is not present in the other merge revision.
1999
If a line is determined killed, that means the line was present in
2002
If a line is killed and new, this indicates that the two merge
2003
revisions contain differing conflict resolutions.
2004
:param revision_id: The id of the revision in which the lines are
2006
:param unique_line_numbers: The line numbers of unique lines.
2007
:return a tuple of (new_this, killed_other):
2011
unique_line_numbers = set(unique_line_numbers)
2012
for lca in self.lcas:
2013
blocks = self._get_matching_blocks(revision_id, lca)
2014
unique_vs_lca, _ignored = self._unique_lines(blocks)
2015
new.update(unique_line_numbers.intersection(unique_vs_lca))
2016
killed.update(unique_line_numbers.difference(unique_vs_lca))
963
merge_types = { "merge3": (Merge3Merger, "Native diff3-style merge"),
964
"diff3": (Diff3Merger, "Merge using external diff3"),
965
'weave': (WeaveMerger, "Weave-based merge")
969
def merge_type_help():
970
templ = '%s%%7s: %%s' % (' '*12)
971
lines = [templ % (f[0], f[1][1]) for f in merge_types.iteritems()]
972
return '\n'.join(lines)