13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19
branch as _mod_branch,
20
conflicts as _mod_conflicts,
28
revision as _mod_revision,
37
from bzrlib.symbol_versioning import (
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20
from tempfile import mkdtemp
23
from bzrlib.branch import Branch
24
from bzrlib.conflicts import ConflictList, Conflict
25
from bzrlib.delta import compare_trees
26
from bzrlib.errors import (BzrCommandError,
36
WorkingTreeNotRevision,
39
from bzrlib.merge3 import Merge3
41
from bzrlib.osutils import rename, pathjoin, rmtree
42
from progress import DummyProgress, ProgressPhase
43
from bzrlib.revision import common_ancestor, is_ancestor, NULL_REVISION
44
from bzrlib.textfile import check_text_lines
45
from bzrlib.trace import mutter, warning, note
46
from bzrlib.transform import (TreeTransform, resolve_conflicts, cook_conflicts,
47
FinalPaths, create_by_entry, unique_add)
48
from bzrlib.versionedfile import WeaveMerge
41
51
# TODO: Report back as changes are merged in
53
def _get_tree(treespec, local_branch=None):
54
location, revno = treespec
55
branch = Branch.open_containing(location)[0]
59
revision = branch.last_revision()
61
revision = branch.get_rev_id(revno)
63
revision = NULL_REVISION
64
return branch, _get_revid_tree(branch, revision, local_branch)
67
def _get_revid_tree(branch, revision, local_branch):
69
base_tree = branch.bzrdir.open_workingtree()
71
if local_branch is not None:
72
if local_branch.base != branch.base:
73
local_branch.fetch(branch, revision)
74
base_tree = local_branch.repository.revision_tree(revision)
76
base_tree = branch.repository.revision_tree(revision)
44
80
def transform_tree(from_tree, to_tree, interesting_ids=None):
45
from_tree.lock_tree_write()
47
merge_inner(from_tree.branch, to_tree, from_tree, ignore_zero=True,
48
interesting_ids=interesting_ids, this_tree=from_tree)
81
merge_inner(from_tree.branch, to_tree, from_tree, ignore_zero=True,
82
interesting_ids=interesting_ids, this_tree=from_tree)
53
85
class Merger(object):
54
def __init__(self, this_branch, other_tree=None, base_tree=None,
55
this_tree=None, pb=None, change_reporter=None,
56
recurse='down', revision_graph=None):
86
def __init__(self, this_branch, other_tree=None, base_tree=None,
87
this_tree=None, pb=DummyProgress()):
57
88
object.__init__(self)
89
assert this_tree is not None, "this_tree is required"
58
90
self.this_branch = this_branch
59
self.this_basis = _mod_revision.ensure_null(
60
this_branch.last_revision())
91
self.this_basis = this_branch.last_revision()
61
92
self.this_rev_id = None
62
93
self.this_tree = this_tree
63
94
self.this_revision_tree = None
64
95
self.this_basis_tree = None
65
96
self.other_tree = other_tree
66
self.other_branch = None
67
97
self.base_tree = base_tree
68
98
self.ignore_zero = False
69
99
self.backup_files = False
70
100
self.interesting_ids = None
71
self.interesting_files = None
72
101
self.show_base = False
73
102
self.reprocess = False
75
pb = progress.DummyProgress()
78
self.recurse = recurse
79
self.change_reporter = change_reporter
80
self._cached_trees = {}
81
self._revision_graph = revision_graph
82
self._base_is_ancestor = None
83
self._base_is_other_ancestor = None
84
self._is_criss_cross = None
85
self._lca_trees = None
87
def cache_trees_with_revision_ids(self, trees):
88
"""Cache any tree in trees if it has a revision_id."""
89
for maybe_tree in trees:
90
if maybe_tree is None:
93
rev_id = maybe_tree.get_revision_id()
94
except AttributeError:
96
self._cached_trees[rev_id] = maybe_tree
99
def revision_graph(self):
100
if self._revision_graph is None:
101
self._revision_graph = self.this_branch.repository.get_graph()
102
return self._revision_graph
104
def _set_base_is_ancestor(self, value):
105
self._base_is_ancestor = value
107
def _get_base_is_ancestor(self):
108
if self._base_is_ancestor is None:
109
self._base_is_ancestor = self.revision_graph.is_ancestor(
110
self.base_rev_id, self.this_basis)
111
return self._base_is_ancestor
113
base_is_ancestor = property(_get_base_is_ancestor, _set_base_is_ancestor)
115
def _set_base_is_other_ancestor(self, value):
116
self._base_is_other_ancestor = value
118
def _get_base_is_other_ancestor(self):
119
if self._base_is_other_ancestor is None:
120
if self.other_basis is None:
122
self._base_is_other_ancestor = self.revision_graph.is_ancestor(
123
self.base_rev_id, self.other_basis)
124
return self._base_is_other_ancestor
126
base_is_other_ancestor = property(_get_base_is_other_ancestor,
127
_set_base_is_other_ancestor)
130
def from_uncommitted(tree, other_tree, pb=None, base_tree=None):
131
"""Return a Merger for uncommitted changes in other_tree.
133
:param tree: The tree to merge into
134
:param other_tree: The tree to get uncommitted changes from
135
:param pb: A progress indicator
136
:param base_tree: The basis to use for the merge. If unspecified,
137
other_tree.basis_tree() will be used.
139
if base_tree is None:
140
base_tree = other_tree.basis_tree()
141
merger = Merger(tree.branch, other_tree, base_tree, tree, pb)
142
merger.base_rev_id = merger.base_tree.get_revision_id()
143
merger.other_rev_id = None
144
merger.other_basis = merger.base_rev_id
148
def from_mergeable(klass, tree, mergeable, pb):
149
"""Return a Merger for a bundle or merge directive.
151
:param tree: The tree to merge changes into
152
:param mergeable: A merge directive or bundle
153
:param pb: A progress indicator
155
mergeable.install_revisions(tree.branch.repository)
156
base_revision_id, other_revision_id, verified =\
157
mergeable.get_merge_request(tree.branch.repository)
158
revision_graph = tree.branch.repository.get_graph()
159
if base_revision_id is not None:
160
if (base_revision_id != _mod_revision.NULL_REVISION and
161
revision_graph.is_ancestor(
162
base_revision_id, tree.branch.last_revision())):
163
base_revision_id = None
165
trace.warning('Performing cherrypick')
166
merger = klass.from_revision_ids(pb, tree, other_revision_id,
167
base_revision_id, revision_graph=
169
return merger, verified
172
def from_revision_ids(pb, tree, other, base=None, other_branch=None,
173
base_branch=None, revision_graph=None,
175
"""Return a Merger for revision-ids.
177
:param pb: A progress indicator
178
:param tree: The tree to merge changes into
179
:param other: The revision-id to use as OTHER
180
:param base: The revision-id to use as BASE. If not specified, will
182
:param other_branch: A branch containing the other revision-id. If
183
not supplied, tree.branch is used.
184
:param base_branch: A branch containing the base revision-id. If
185
not supplied, other_branch or tree.branch will be used.
186
:param revision_graph: If you have a revision_graph precomputed, pass
187
it in, otherwise it will be created for you.
188
:param tree_branch: The branch associated with tree. If not supplied,
189
tree.branch will be used.
191
if tree_branch is None:
192
tree_branch = tree.branch
193
merger = Merger(tree_branch, this_tree=tree, pb=pb,
194
revision_graph=revision_graph)
195
if other_branch is None:
196
other_branch = tree.branch
197
merger.set_other_revision(other, other_branch)
201
if base_branch is None:
202
base_branch = other_branch
203
merger.set_base_revision(base, base_branch)
206
def revision_tree(self, revision_id, branch=None):
207
if revision_id not in self._cached_trees:
209
branch = self.this_branch
211
tree = self.this_tree.revision_tree(revision_id)
212
except errors.NoSuchRevisionInTree:
213
tree = branch.repository.revision_tree(revision_id)
214
self._cached_trees[revision_id] = tree
215
return self._cached_trees[revision_id]
217
def _get_tree(self, treespec, possible_transports=None):
218
from bzrlib import workingtree
219
location, revno = treespec
221
tree = workingtree.WorkingTree.open_containing(location)[0]
222
return tree.branch, tree
223
branch = _mod_branch.Branch.open_containing(
224
location, possible_transports)[0]
226
revision_id = branch.last_revision()
228
revision_id = branch.get_rev_id(revno)
229
revision_id = _mod_revision.ensure_null(revision_id)
230
return branch, self.revision_tree(revision_id, branch)
232
@deprecated_method(deprecated_in((2, 1, 0)))
107
def revision_tree(self, revision_id):
108
return self.this_branch.repository.revision_tree(revision_id)
233
110
def ensure_revision_trees(self):
234
111
if self.this_revision_tree is None:
235
self.this_basis_tree = self.revision_tree(self.this_basis)
112
self.this_basis_tree = self.this_branch.repository.revision_tree(
236
114
if self.this_basis == self.this_rev_id:
237
115
self.this_revision_tree = self.this_basis_tree
239
117
if self.other_rev_id is None:
240
118
other_basis_tree = self.revision_tree(self.other_basis)
241
if other_basis_tree.has_changes(self.other_tree):
242
raise errors.WorkingTreeNotRevision(self.this_tree)
119
changes = compare_trees(self.other_tree, other_basis_tree)
120
if changes.has_changed():
121
raise WorkingTreeNotRevision(self.this_tree)
243
122
other_rev_id = self.other_basis
244
123
self.other_tree = other_basis_tree
246
@deprecated_method(deprecated_in((2, 1, 0)))
247
125
def file_revisions(self, file_id):
248
126
self.ensure_revision_trees()
249
127
def get_id(tree, file_id):
250
128
revision_id = tree.inventory[file_id].revision
129
assert revision_id is not None
251
130
return revision_id
252
131
if self.this_rev_id is None:
253
132
if self.this_basis_tree.get_file_sha1(file_id) != \
254
133
self.this_tree.get_file_sha1(file_id):
255
raise errors.WorkingTreeNotRevision(self.this_tree)
134
raise WorkingTreeNotRevision(self.this_tree)
257
136
trees = (self.this_basis_tree, self.other_tree)
258
137
return [get_id(tree, file_id) for tree in trees]
260
@deprecated_method(deprecated_in((2, 1, 0)))
261
139
def check_basis(self, check_clean, require_commits=True):
262
140
if self.this_basis is None and require_commits is True:
263
raise errors.BzrCommandError(
264
"This branch has no commits."
265
" (perhaps you would prefer 'bzr pull')")
141
raise BzrCommandError("This branch has no commits")
267
143
self.compare_basis()
268
144
if self.this_basis != self.this_rev_id:
269
raise errors.UncommittedChanges(self.this_tree)
145
raise BzrCommandError("Working tree has uncommitted changes.")
271
@deprecated_method(deprecated_in((2, 1, 0)))
272
147
def compare_basis(self):
274
basis_tree = self.revision_tree(self.this_tree.last_revision())
275
except errors.NoSuchRevision:
276
basis_tree = self.this_tree.basis_tree()
277
if not self.this_tree.has_changes(basis_tree):
148
changes = compare_trees(self.this_tree,
149
self.this_tree.basis_tree(), False)
150
if not changes.has_changed():
278
151
self.this_rev_id = self.this_basis
280
153
def set_interesting_files(self, file_list):
281
self.interesting_files = file_list
155
self._set_interesting_files(file_list)
156
except NotVersionedError, e:
157
raise BzrCommandError("%s is not a source file in any"
160
def _set_interesting_files(self, file_list):
161
"""Set the list of interesting ids from a list of files."""
162
if file_list is None:
163
self.interesting_ids = None
166
interesting_ids = set()
167
for path in file_list:
169
for tree in (self.this_tree, self.base_tree, self.other_tree):
170
file_id = tree.inventory.path2id(path)
171
if file_id is not None:
172
interesting_ids.add(file_id)
175
raise NotVersionedError(path=path)
176
self.interesting_ids = interesting_ids
283
178
def set_pending(self):
284
if (not self.base_is_ancestor or not self.base_is_other_ancestor
285
or self.other_rev_id is None):
289
def _add_parent(self):
290
new_parents = self.this_tree.get_parent_ids() + [self.other_rev_id]
291
new_parent_trees = []
292
for revision_id in new_parents:
294
tree = self.revision_tree(revision_id)
295
except errors.NoSuchRevision:
299
new_parent_trees.append((revision_id, tree))
301
self.this_tree.set_parent_trees(new_parent_trees,
302
allow_leftmost_as_ghost=True)
304
for _revision_id, tree in new_parent_trees:
308
def set_other(self, other_revision, possible_transports=None):
309
"""Set the revision and tree to merge from.
311
This sets the other_tree, other_rev_id, other_basis attributes.
313
:param other_revision: The [path, revision] list to merge from.
315
self.other_branch, self.other_tree = self._get_tree(other_revision,
179
if not self.base_is_ancestor:
181
if self.other_rev_id is None:
183
ancestry = self.this_branch.repository.get_ancestry(self.this_basis)
184
if self.other_rev_id in ancestry:
186
self.this_tree.add_pending_merge(self.other_rev_id)
188
def set_other(self, other_revision):
189
other_branch, self.other_tree = _get_tree(other_revision,
317
191
if other_revision[1] == -1:
318
self.other_rev_id = _mod_revision.ensure_null(
319
self.other_branch.last_revision())
320
if _mod_revision.is_null(self.other_rev_id):
321
raise errors.NoCommits(self.other_branch)
192
self.other_rev_id = other_branch.last_revision()
193
if self.other_rev_id is None:
194
raise NoCommits(other_branch)
322
195
self.other_basis = self.other_rev_id
323
196
elif other_revision[1] is not None:
324
self.other_rev_id = self.other_branch.get_rev_id(other_revision[1])
197
self.other_rev_id = other_branch.get_rev_id(other_revision[1])
325
198
self.other_basis = self.other_rev_id
327
200
self.other_rev_id = None
328
self.other_basis = self.other_branch.last_revision()
201
self.other_basis = other_branch.last_revision()
329
202
if self.other_basis is None:
330
raise errors.NoCommits(self.other_branch)
331
if self.other_rev_id is not None:
332
self._cached_trees[self.other_rev_id] = self.other_tree
333
self._maybe_fetch(self.other_branch,self.this_branch, self.other_basis)
335
def set_other_revision(self, revision_id, other_branch):
336
"""Set 'other' based on a branch and revision id
338
:param revision_id: The revision to use for a tree
339
:param other_branch: The branch containing this tree
341
self.other_rev_id = revision_id
342
self.other_branch = other_branch
343
self._maybe_fetch(other_branch, self.this_branch, self.other_rev_id)
344
self.other_tree = self.revision_tree(revision_id)
345
self.other_basis = revision_id
347
def set_base_revision(self, revision_id, branch):
348
"""Set 'base' based on a branch and revision id
350
:param revision_id: The revision to use for a tree
351
:param branch: The branch containing this tree
353
self.base_rev_id = revision_id
354
self.base_branch = branch
355
self._maybe_fetch(branch, self.this_branch, revision_id)
356
self.base_tree = self.revision_tree(revision_id)
358
def _maybe_fetch(self, source, target, revision_id):
359
if not source.repository.has_same_location(target.repository):
360
target.fetch(source, revision_id)
203
raise NoCommits(other_branch)
204
if other_branch.base != self.this_branch.base:
205
self.this_branch.fetch(other_branch, last_revision=self.other_basis)
362
207
def find_base(self):
363
revisions = [_mod_revision.ensure_null(self.this_basis),
364
_mod_revision.ensure_null(self.other_basis)]
365
if _mod_revision.NULL_REVISION in revisions:
366
self.base_rev_id = _mod_revision.NULL_REVISION
367
self.base_tree = self.revision_tree(self.base_rev_id)
368
self._is_criss_cross = False
370
lcas = self.revision_graph.find_lca(revisions[0], revisions[1])
371
self._is_criss_cross = False
373
self.base_rev_id = _mod_revision.NULL_REVISION
375
self.base_rev_id = list(lcas)[0]
376
else: # len(lcas) > 1
378
# find_unique_lca can only handle 2 nodes, so we have to
379
# start back at the beginning. It is a shame to traverse
380
# the graph again, but better than re-implementing
382
self.base_rev_id = self.revision_graph.find_unique_lca(
383
revisions[0], revisions[1])
385
self.base_rev_id = self.revision_graph.find_unique_lca(
387
self._is_criss_cross = True
388
if self.base_rev_id == _mod_revision.NULL_REVISION:
389
raise errors.UnrelatedBranches()
390
if self._is_criss_cross:
391
trace.warning('Warning: criss-cross merge encountered. See bzr'
392
' help criss-cross.')
393
trace.mutter('Criss-cross lcas: %r' % lcas)
394
interesting_revision_ids = [self.base_rev_id]
395
interesting_revision_ids.extend(lcas)
396
interesting_trees = dict((t.get_revision_id(), t)
397
for t in self.this_branch.repository.revision_trees(
398
interesting_revision_ids))
399
self._cached_trees.update(interesting_trees)
400
self.base_tree = interesting_trees.pop(self.base_rev_id)
401
sorted_lca_keys = self.revision_graph.find_merge_order(
403
self._lca_trees = [interesting_trees[key]
404
for key in sorted_lca_keys]
406
self.base_tree = self.revision_tree(self.base_rev_id)
407
self.base_is_ancestor = True
408
self.base_is_other_ancestor = True
409
trace.mutter('Base revid: %r' % self.base_rev_id)
208
self.set_base([None, None])
411
210
def set_base(self, base_revision):
412
"""Set the base revision to use for the merge.
414
:param base_revision: A 2-list containing a path and revision number.
416
trace.mutter("doing merge() with no base_revision specified")
211
mutter("doing merge() with no base_revision specified")
417
212
if base_revision == [None, None]:
214
pb = bzrlib.ui.ui_factory.nested_progress_bar()
216
this_repo = self.this_branch.repository
217
self.base_rev_id = common_ancestor(self.this_basis,
222
except NoCommonAncestor:
223
raise UnrelatedBranches()
224
self.base_tree = _get_revid_tree(self.this_branch, self.base_rev_id,
226
self.base_is_ancestor = True
420
base_branch, self.base_tree = self._get_tree(base_revision)
228
base_branch, self.base_tree = _get_tree(base_revision)
421
229
if base_revision[1] == -1:
422
230
self.base_rev_id = base_branch.last_revision()
423
231
elif base_revision[1] is None:
424
self.base_rev_id = _mod_revision.NULL_REVISION
232
self.base_rev_id = None
426
self.base_rev_id = _mod_revision.ensure_null(
427
base_branch.get_rev_id(base_revision[1]))
428
self._maybe_fetch(base_branch, self.this_branch, self.base_rev_id)
234
self.base_rev_id = base_branch.get_rev_id(base_revision[1])
235
if self.this_branch.base != base_branch.base:
236
self.this_branch.fetch(base_branch)
237
self.base_is_ancestor = is_ancestor(self.this_basis,
430
def make_merger(self):
431
kwargs = {'working_tree':self.this_tree, 'this_tree': self.this_tree,
432
'other_tree': self.other_tree,
242
kwargs = {'working_tree':self.this_tree, 'this_tree': self.this_tree,
243
'other_tree': self.other_tree,
433
244
'interesting_ids': self.interesting_ids,
434
'interesting_files': self.interesting_files,
437
246
if self.merge_type.requires_base:
438
247
kwargs['base_tree'] = self.base_tree
439
248
if self.merge_type.supports_reprocess:
440
249
kwargs['reprocess'] = self.reprocess
441
250
elif self.reprocess:
442
raise errors.BzrError(
443
"Conflict reduction is not supported for merge"
444
" type %s." % self.merge_type)
251
raise BzrError("Conflict reduction is not supported for merge"
252
" type %s." % self.merge_type)
445
253
if self.merge_type.supports_show_base:
446
254
kwargs['show_base'] = self.show_base
447
255
elif self.show_base:
448
raise errors.BzrError("Showing base is not supported for this"
256
raise BzrError("Showing base is not supported for this"
449
257
" merge type. %s" % self.merge_type)
450
if (not getattr(self.merge_type, 'supports_reverse_cherrypick', True)
451
and not self.base_is_other_ancestor):
452
raise errors.CannotReverseCherrypick()
453
if self.merge_type.supports_cherrypick:
454
kwargs['cherrypick'] = (not self.base_is_ancestor or
455
not self.base_is_other_ancestor)
456
if self._is_criss_cross and getattr(self.merge_type,
457
'supports_lca_trees', False):
458
kwargs['lca_trees'] = self._lca_trees
459
return self.merge_type(pb=self._pb,
460
change_reporter=self.change_reporter,
463
def _do_merge_to(self, merge):
464
if self.other_branch is not None:
465
self.other_branch.update_references(self.this_branch)
467
if self.recurse == 'down':
468
for relpath, file_id in self.this_tree.iter_references():
469
sub_tree = self.this_tree.get_nested_tree(file_id, relpath)
470
other_revision = self.other_tree.get_reference_revision(
472
if other_revision == sub_tree.last_revision():
474
sub_merge = Merger(sub_tree.branch, this_tree=sub_tree)
475
sub_merge.merge_type = self.merge_type
476
other_branch = self.other_branch.reference_parent(file_id, relpath)
477
sub_merge.set_other_revision(other_revision, other_branch)
478
base_revision = self.base_tree.get_reference_revision(file_id)
479
sub_merge.base_tree = \
480
sub_tree.branch.repository.revision_tree(base_revision)
481
sub_merge.base_rev_id = base_revision
485
self.this_tree.lock_tree_write()
487
if self.base_tree is not None:
488
self.base_tree.lock_read()
490
if self.other_tree is not None:
491
self.other_tree.lock_read()
493
merge = self.make_merger()
494
self._do_merge_to(merge)
496
if self.other_tree is not None:
497
self.other_tree.unlock()
499
if self.base_tree is not None:
500
self.base_tree.unlock()
502
self.this_tree.unlock()
258
merge = self.merge_type(pb=self._pb, **kwargs)
503
259
if len(merge.cooked_conflicts) == 0:
504
if not self.ignore_zero and not trace.is_quiet():
505
trace.note("All changes applied successfully.")
260
if not self.ignore_zero:
261
note("All changes applied successfully.")
507
trace.note("%d conflicts encountered."
508
% len(merge.cooked_conflicts))
263
note("%d conflicts encountered." % len(merge.cooked_conflicts))
510
265
return len(merge.cooked_conflicts)
513
class _InventoryNoneEntry(object):
514
"""This represents an inventory entry which *isn't there*.
516
It simplifies the merging logic if we always have an InventoryEntry, even
517
if it isn't actually present
524
symlink_target = None
527
_none_entry = _InventoryNoneEntry()
267
def regen_inventory(self, new_entries):
268
old_entries = self.this_tree.read_working_inventory()
272
for path, file_id in new_entries:
275
new_entries_map[file_id] = path
277
def id2path(file_id):
278
path = new_entries_map.get(file_id)
281
entry = old_entries[file_id]
282
if entry.parent_id is None:
284
return pathjoin(id2path(entry.parent_id), entry.name)
286
for file_id in old_entries:
287
entry = old_entries[file_id]
288
path = id2path(file_id)
289
if file_id in self.base_tree.inventory:
290
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
292
executable = getattr(entry, 'executable', False)
293
new_inventory[file_id] = (path, file_id, entry.parent_id,
294
entry.kind, executable)
296
by_path[path] = file_id
301
for path, file_id in new_entries:
303
del new_inventory[file_id]
306
new_path_list.append((path, file_id))
307
if file_id not in old_entries:
309
# Ensure no file is added before its parent
311
for path, file_id in new_path_list:
315
parent = by_path[os.path.dirname(path)]
316
abspath = pathjoin(self.this_tree.basedir, path)
317
kind = bzrlib.osutils.file_kind(abspath)
318
if file_id in self.base_tree.inventory:
319
executable = getattr(self.base_tree.inventory[file_id], 'executable', False)
322
new_inventory[file_id] = (path, file_id, parent, kind, executable)
323
by_path[path] = file_id
325
# Get a list in insertion order
326
new_inventory_list = new_inventory.values()
327
mutter ("""Inventory regeneration:
328
old length: %i insertions: %i deletions: %i new_length: %i"""\
329
% (len(old_entries), insertions, deletions,
330
len(new_inventory_list)))
331
assert len(new_inventory_list) == len(old_entries) + insertions\
333
new_inventory_list.sort()
334
return new_inventory_list
530
337
class Merge3Merger(object):
581
353
self.cooked_conflicts = []
582
354
self.reprocess = reprocess
583
355
self.show_base = show_base
584
self._lca_trees = lca_trees
585
# Uncommenting this will change the default algorithm to always use
586
# _entries_lca. This can be useful for running the test suite and
587
# making sure we haven't missed any corner cases.
588
# if lca_trees is None:
589
# self._lca_trees = [self.base_tree]
592
self.change_reporter = change_reporter
593
self.cherrypick = cherrypick
594
358
if self.pp is None:
595
self.pp = progress.ProgressPhase("Merge phase", 3, self.pb)
359
self.pp = ProgressPhase("Merge phase", 3, self.pb)
600
self.this_tree.lock_tree_write()
601
self.base_tree.lock_read()
602
self.other_tree.lock_read()
361
if interesting_ids is not None:
362
all_ids = interesting_ids
364
all_ids = set(base_tree)
365
all_ids.update(other_tree)
366
working_tree.lock_write()
367
self.tt = TreeTransform(working_tree, self.pb)
604
self.tt = transform.TreeTransform(self.this_tree, self.pb)
607
self._compute_transform()
609
results = self.tt.apply(no_conflicts=True)
610
self.write_modified(results)
612
self.this_tree.add_conflicts(self.cooked_conflicts)
613
except errors.UnsupportedOperation:
370
child_pb = ui.ui_factory.nested_progress_bar()
372
for num, file_id in enumerate(all_ids):
373
child_pb.update('Preparing file merge', num, len(all_ids))
374
self.merge_names(file_id)
375
file_status = self.merge_contents(file_id)
376
self.merge_executable(file_id, file_status)
381
child_pb = ui.ui_factory.nested_progress_bar()
383
fs_conflicts = resolve_conflicts(self.tt, child_pb)
386
self.cook_conflicts(fs_conflicts)
387
for conflict in self.cooked_conflicts:
390
results = self.tt.apply()
391
self.write_modified(results)
393
working_tree.set_conflicts(ConflictList(self.cooked_conflicts))
394
except UnsupportedOperation:
616
398
self.tt.finalize()
618
self.other_tree.unlock()
619
self.base_tree.unlock()
620
self.this_tree.unlock()
623
def make_preview_transform(self):
624
self.base_tree.lock_read()
625
self.other_tree.lock_read()
626
self.tt = transform.TransformPreview(self.this_tree)
629
self._compute_transform()
632
self.other_tree.unlock()
633
self.base_tree.unlock()
637
def _compute_transform(self):
638
if self._lca_trees is None:
639
entries = self._entries3()
640
resolver = self._three_way
642
entries = self._entries_lca()
643
resolver = self._lca_multi_way
644
child_pb = ui.ui_factory.nested_progress_bar()
646
for num, (file_id, changed, parents3, names3,
647
executable3) in enumerate(entries):
648
child_pb.update('Preparing file merge', num, len(entries))
649
self._merge_names(file_id, parents3, names3, resolver=resolver)
651
file_status = self.merge_contents(file_id)
653
file_status = 'unmodified'
654
self._merge_executable(file_id,
655
executable3, file_status, resolver=resolver)
660
child_pb = ui.ui_factory.nested_progress_bar()
662
fs_conflicts = transform.resolve_conflicts(self.tt, child_pb,
663
lambda t, c: transform.conflict_pass(t, c, self.other_tree))
666
if self.change_reporter is not None:
667
from bzrlib import delta
668
delta.report_changes(
669
self.tt.iter_changes(), self.change_reporter)
670
self.cook_conflicts(fs_conflicts)
671
for conflict in self.cooked_conflicts:
672
trace.warning(conflict)
675
"""Gather data about files modified between three trees.
677
Return a list of tuples of file_id, changed, parents3, names3,
678
executable3. changed is a boolean indicating whether the file contents
679
or kind were changed. parents3 is a tuple of parent ids for base,
680
other and this. names3 is a tuple of names for base, other and this.
681
executable3 is a tuple of execute-bit values for base, other and this.
684
iterator = self.other_tree.iter_changes(self.base_tree,
685
include_unchanged=True, specific_files=self.interesting_files,
686
extra_trees=[self.this_tree])
687
this_entries = dict((e.file_id, e) for p, e in
688
self.this_tree.iter_entries_by_dir(
689
self.interesting_ids))
690
for (file_id, paths, changed, versioned, parents, names, kind,
691
executable) in iterator:
692
if (self.interesting_ids is not None and
693
file_id not in self.interesting_ids):
695
entry = this_entries.get(file_id)
696
if entry is not None:
697
this_name = entry.name
698
this_parent = entry.parent_id
699
this_executable = entry.executable
703
this_executable = None
704
parents3 = parents + (this_parent,)
705
names3 = names + (this_name,)
706
executable3 = executable + (this_executable,)
707
result.append((file_id, changed, parents3, names3, executable3))
710
def _entries_lca(self):
711
"""Gather data about files modified between multiple trees.
713
This compares OTHER versus all LCA trees, and for interesting entries,
714
it then compares with THIS and BASE.
716
For the multi-valued entries, the format will be (BASE, [lca1, lca2])
717
:return: [(file_id, changed, parents, names, executable)]
718
file_id Simple file_id of the entry
719
changed Boolean, True if the kind or contents changed
721
parents ((base, [parent_id, in, lcas]), parent_id_other,
723
names ((base, [name, in, lcas]), name_in_other, name_in_this)
724
executable ((base, [exec, in, lcas]), exec_in_other, exec_in_this)
726
if self.interesting_files is not None:
727
lookup_trees = [self.this_tree, self.base_tree]
728
lookup_trees.extend(self._lca_trees)
729
# I think we should include the lca trees as well
730
interesting_ids = self.other_tree.paths2ids(self.interesting_files,
733
interesting_ids = self.interesting_ids
735
walker = _mod_tree.MultiWalker(self.other_tree, self._lca_trees)
737
base_inventory = self.base_tree.inventory
738
this_inventory = self.this_tree.inventory
739
for path, file_id, other_ie, lca_values in walker.iter_all():
740
# Is this modified at all from any of the other trees?
742
other_ie = _none_entry
743
if interesting_ids is not None and file_id not in interesting_ids:
746
# If other_revision is found in any of the lcas, that means this
747
# node is uninteresting. This is because when merging, if there are
748
# multiple heads(), we have to create a new node. So if we didn't,
749
# we know that the ancestry is linear, and that OTHER did not
751
# See doc/developers/lca_merge_resolution.txt for details
752
other_revision = other_ie.revision
753
if other_revision is not None:
754
# We can't use this shortcut when other_revision is None,
755
# because it may be None because things are WorkingTrees, and
756
# not because it is *actually* None.
757
is_unmodified = False
758
for lca_path, ie in lca_values:
759
if ie is not None and ie.revision == other_revision:
766
for lca_path, lca_ie in lca_values:
768
lca_entries.append(_none_entry)
770
lca_entries.append(lca_ie)
772
if file_id in base_inventory:
773
base_ie = base_inventory[file_id]
775
base_ie = _none_entry
777
if file_id in this_inventory:
778
this_ie = this_inventory[file_id]
780
this_ie = _none_entry
786
for lca_ie in lca_entries:
787
lca_kinds.append(lca_ie.kind)
788
lca_parent_ids.append(lca_ie.parent_id)
789
lca_names.append(lca_ie.name)
790
lca_executable.append(lca_ie.executable)
792
kind_winner = self._lca_multi_way(
793
(base_ie.kind, lca_kinds),
794
other_ie.kind, this_ie.kind)
795
parent_id_winner = self._lca_multi_way(
796
(base_ie.parent_id, lca_parent_ids),
797
other_ie.parent_id, this_ie.parent_id)
798
name_winner = self._lca_multi_way(
799
(base_ie.name, lca_names),
800
other_ie.name, this_ie.name)
802
content_changed = True
803
if kind_winner == 'this':
804
# No kind change in OTHER, see if there are *any* changes
805
if other_ie.kind == 'directory':
806
if parent_id_winner == 'this' and name_winner == 'this':
807
# No change for this directory in OTHER, skip
809
content_changed = False
810
elif other_ie.kind is None or other_ie.kind == 'file':
811
def get_sha1(ie, tree):
812
if ie.kind != 'file':
814
return tree.get_file_sha1(file_id)
815
base_sha1 = get_sha1(base_ie, self.base_tree)
816
lca_sha1s = [get_sha1(ie, tree) for ie, tree
817
in zip(lca_entries, self._lca_trees)]
818
this_sha1 = get_sha1(this_ie, self.this_tree)
819
other_sha1 = get_sha1(other_ie, self.other_tree)
820
sha1_winner = self._lca_multi_way(
821
(base_sha1, lca_sha1s), other_sha1, this_sha1,
822
allow_overriding_lca=False)
823
exec_winner = self._lca_multi_way(
824
(base_ie.executable, lca_executable),
825
other_ie.executable, this_ie.executable)
826
if (parent_id_winner == 'this' and name_winner == 'this'
827
and sha1_winner == 'this' and exec_winner == 'this'):
828
# No kind, parent, name, exec, or content change for
829
# OTHER, so this node is not considered interesting
831
if sha1_winner == 'this':
832
content_changed = False
833
elif other_ie.kind == 'symlink':
834
def get_target(ie, tree):
835
if ie.kind != 'symlink':
837
return tree.get_symlink_target(file_id)
838
base_target = get_target(base_ie, self.base_tree)
839
lca_targets = [get_target(ie, tree) for ie, tree
840
in zip(lca_entries, self._lca_trees)]
841
this_target = get_target(this_ie, self.this_tree)
842
other_target = get_target(other_ie, self.other_tree)
843
target_winner = self._lca_multi_way(
844
(base_target, lca_targets),
845
other_target, this_target)
846
if (parent_id_winner == 'this' and name_winner == 'this'
847
and target_winner == 'this'):
848
# No kind, parent, name, or symlink target change
851
if target_winner == 'this':
852
content_changed = False
853
elif other_ie.kind == 'tree-reference':
854
# The 'changed' information seems to be handled at a higher
855
# level. At least, _entries3 returns False for content
856
# changed, even when at a new revision_id.
857
content_changed = False
858
if (parent_id_winner == 'this' and name_winner == 'this'):
859
# Nothing interesting
862
raise AssertionError('unhandled kind: %s' % other_ie.kind)
863
# XXX: We need to handle kind == 'symlink'
865
# If we have gotten this far, that means something has changed
866
result.append((file_id, content_changed,
867
((base_ie.parent_id, lca_parent_ids),
868
other_ie.parent_id, this_ie.parent_id),
869
((base_ie.name, lca_names),
870
other_ie.name, this_ie.name),
871
((base_ie.executable, lca_executable),
872
other_ie.executable, this_ie.executable)
879
self.tt.final_kind(self.tt.root)
880
except errors.NoSuchFile:
881
self.tt.cancel_deletion(self.tt.root)
882
if self.tt.final_file_id(self.tt.root) is None:
883
self.tt.version_file(self.tt.tree_file_id(self.tt.root),
885
other_root_file_id = self.other_tree.get_root_id()
886
if other_root_file_id is None:
888
other_root = self.tt.trans_id_file_id(other_root_file_id)
889
if other_root == self.tt.root:
892
self.tt.final_kind(other_root)
893
except errors.NoSuchFile:
895
if self.other_tree.inventory.root.file_id in self.this_tree.inventory:
896
# the other tree's root is a non-root in the current tree
898
self.reparent_children(self.other_tree.inventory.root, self.tt.root)
899
self.tt.cancel_creation(other_root)
900
self.tt.cancel_versioning(other_root)
902
def reparent_children(self, ie, target):
903
for thing, child in ie.children.iteritems():
904
trans_id = self.tt.trans_id_file_id(child.file_id)
905
self.tt.adjust_path(self.tt.final_name(trans_id), target, trans_id)
401
working_tree.unlock()
907
404
def write_modified(self, results):
908
405
modified_hashes = {}
1253
652
determined automatically. If set_version is true, the .OTHER, .THIS
1254
653
or .BASE (in that order) will be created as versioned files.
1256
data = [('OTHER', self.other_tree, other_lines),
655
data = [('OTHER', self.other_tree, other_lines),
1257
656
('THIS', self.this_tree, this_lines)]
1259
658
data.append(('BASE', self.base_tree, base_lines))
1261
# We need to use the actual path in the working tree of the file here,
1262
# ignoring the conflict suffixes
1264
if wt.supports_content_filtering():
1266
filter_tree_path = wt.id2path(file_id)
1267
except errors.NoSuchId:
1268
# file has been deleted
1269
filter_tree_path = None
1271
# Skip the id2path lookup for older formats
1272
filter_tree_path = None
1274
659
versioned = False
1276
661
for suffix, tree, lines in data:
1277
662
if file_id in tree:
1278
663
trans_id = self._conflict_file(name, parent_id, tree, file_id,
1279
suffix, lines, filter_tree_path)
1280
665
file_group.append(trans_id)
1281
666
if set_version and not versioned:
1282
667
self.tt.version_file(file_id, trans_id)
1283
668
versioned = True
1284
669
return file_group
1286
def _conflict_file(self, name, parent_id, tree, file_id, suffix,
1287
lines=None, filter_tree_path=None):
671
def _conflict_file(self, name, parent_id, tree, file_id, suffix,
1288
673
"""Emit a single conflict file."""
1289
674
name = name + '.' + suffix
1290
675
trans_id = self.tt.create_path(name, parent_id)
1291
transform.create_from_tree(self.tt, trans_id, tree, file_id, lines,
676
entry = tree.inventory[file_id]
677
create_by_entry(self.tt, entry, tree, trans_id, lines)
1295
680
def merge_executable(self, file_id, file_status):
1296
681
"""Perform a merge on the execute bit."""
1297
executable = [self.executable(t, file_id) for t in (self.base_tree,
1298
self.other_tree, self.this_tree)]
1299
self._merge_executable(file_id, executable, file_status,
1300
resolver=self._three_way)
1302
def _merge_executable(self, file_id, executable, file_status,
1304
"""Perform a merge on the execute bit."""
1305
base_executable, other_executable, this_executable = executable
1306
682
if file_status == "deleted":
1308
winner = resolver(*executable)
684
trans_id = self.tt.trans_id_file_id(file_id)
686
if self.tt.final_kind(trans_id) != "file":
690
winner = self.scalar_three_way(self.this_tree, self.base_tree,
691
self.other_tree, file_id,
1309
693
if winner == "conflict":
1310
694
# There must be a None in here, if we have a conflict, but we
1311
695
# need executability since file status was not deleted.
1312
if self.executable(self.other_tree, file_id) is None:
696
if self.other_tree.is_executable(file_id) is None:
1315
699
winner = "other"
1316
if winner == 'this' and file_status != "modified":
1318
trans_id = self.tt.trans_id_file_id(file_id)
1320
if self.tt.final_kind(trans_id) != "file":
1322
except errors.NoSuchFile:
1324
700
if winner == "this":
1325
executability = this_executable
701
if file_status == "modified":
702
executability = self.this_tree.is_executable(file_id)
703
if executability is not None:
704
trans_id = self.tt.trans_id_file_id(file_id)
705
self.tt.set_executability(executability, trans_id)
707
assert winner == "other"
1327
708
if file_id in self.other_tree:
1328
executability = other_executable
709
executability = self.other_tree.is_executable(file_id)
1329
710
elif file_id in self.this_tree:
1330
executability = this_executable
711
executability = self.this_tree.is_executable(file_id)
1331
712
elif file_id in self.base_tree:
1332
executability = base_executable
1333
if executability is not None:
1334
trans_id = self.tt.trans_id_file_id(file_id)
1335
self.tt.set_executability(executability, trans_id)
713
executability = self.base_tree.is_executable(file_id)
714
if executability is not None:
715
trans_id = self.tt.trans_id_file_id(file_id)
716
self.tt.set_executability(executability, trans_id)
1337
718
def cook_conflicts(self, fs_conflicts):
1338
719
"""Convert all conflicts into a form that doesn't depend on trans_id"""
720
from conflicts import Conflict
1339
721
name_conflicts = {}
1340
self.cooked_conflicts.extend(transform.cook_conflicts(
1341
fs_conflicts, self.tt))
1342
fp = transform.FinalPaths(self.tt)
722
self.cooked_conflicts.extend(cook_conflicts(fs_conflicts, self.tt))
723
fp = FinalPaths(self.tt)
1343
724
for conflict in self._raw_conflicts:
1344
725
conflict_type = conflict[0]
1345
726
if conflict_type in ('name conflict', 'parent conflict'):
1359
740
if path.endswith(suffix):
1360
741
path = path[:-len(suffix)]
1362
c = _mod_conflicts.Conflict.factory(conflict_type,
1363
path=path, file_id=file_id)
743
c = Conflict.factory(conflict_type, path=path, file_id=file_id)
1364
744
self.cooked_conflicts.append(c)
1365
745
if conflict_type == 'text conflict':
1366
746
trans_id = conflict[1]
1367
747
path = fp.get_path(trans_id)
1368
748
file_id = self.tt.final_file_id(trans_id)
1369
c = _mod_conflicts.Conflict.factory(conflict_type,
1370
path=path, file_id=file_id)
749
c = Conflict.factory(conflict_type, path=path, file_id=file_id)
1371
750
self.cooked_conflicts.append(c)
1373
752
for trans_id, conflicts in name_conflicts.iteritems():
1375
754
this_parent, other_parent = conflicts['parent conflict']
1376
if this_parent == other_parent:
1377
raise AssertionError()
755
assert this_parent != other_parent
1378
756
except KeyError:
1379
757
this_parent = other_parent = \
1380
758
self.tt.final_file_id(self.tt.final_parent(trans_id))
1382
760
this_name, other_name = conflicts['name conflict']
1383
if this_name == other_name:
1384
raise AssertionError()
761
assert this_name != other_name
1385
762
except KeyError:
1386
763
this_name = other_name = self.tt.final_name(trans_id)
1387
764
other_path = fp.get_path(trans_id)
1388
if this_parent is not None and this_name is not None:
765
if this_parent is not None:
1389
766
this_parent_path = \
1390
767
fp.get_path(self.tt.trans_id_file_id(this_parent))
1391
this_path = osutils.pathjoin(this_parent_path, this_name)
768
this_path = pathjoin(this_parent_path, this_name)
1393
770
this_path = "<deleted>"
1394
771
file_id = self.tt.final_file_id(trans_id)
1395
c = _mod_conflicts.Conflict.factory('path conflict', path=this_path,
1396
conflict_path=other_path,
772
c = Conflict.factory('path conflict', path=this_path,
773
conflict_path=other_path, file_id=file_id)
1398
774
self.cooked_conflicts.append(c)
1399
self.cooked_conflicts.sort(key=_mod_conflicts.Conflict.sort_key)
775
self.cooked_conflicts.sort(key=Conflict.sort_key)
1402
778
class WeaveMerger(Merge3Merger):
1403
779
"""Three-way tree merger, text weave merger."""
1404
780
supports_reprocess = True
1405
781
supports_show_base = False
1406
supports_reverse_cherrypick = False
1407
history_based = True
1409
def _generate_merge_plan(self, file_id, base):
1410
return self.this_tree.plan_file_merge(file_id, self.other_tree,
783
def __init__(self, working_tree, this_tree, base_tree, other_tree,
784
interesting_ids=None, pb=DummyProgress(), pp=None,
786
self.this_revision_tree = self._get_revision_tree(this_tree)
787
self.other_revision_tree = self._get_revision_tree(other_tree)
788
super(WeaveMerger, self).__init__(working_tree, this_tree,
789
base_tree, other_tree,
790
interesting_ids=interesting_ids,
791
pb=pb, pp=pp, reprocess=reprocess)
793
def _get_revision_tree(self, tree):
794
"""Return a revision tree related to this tree.
795
If the tree is a WorkingTree, the basis will be returned.
797
if getattr(tree, 'get_weave', False) is False:
798
# If we have a WorkingTree, try using the basis
799
return tree.branch.basis_tree()
803
def _check_file(self, file_id):
804
"""Check that the revision tree's version of the file matches."""
805
for tree, rt in ((self.this_tree, self.this_revision_tree),
806
(self.other_tree, self.other_revision_tree)):
809
if tree.get_file_sha1(file_id) != rt.get_file_sha1(file_id):
810
raise WorkingTreeNotRevision(self.this_tree)
1413
812
def _merged_lines(self, file_id):
1414
813
"""Generate the merged lines.
1415
814
There is no distinction between lines that are meant to contain <<<<<<<
1419
base = self.base_tree
1422
plan = self._generate_merge_plan(file_id, base)
1423
if 'merge' in debug.debug_flags:
1425
trans_id = self.tt.trans_id_file_id(file_id)
1426
name = self.tt.final_name(trans_id) + '.plan'
1427
contents = ('%11s|%s' % l for l in plan)
1428
self.tt.new_file(name, self.tt.final_parent(trans_id), contents)
1429
textmerge = versionedfile.PlanWeaveMerge(plan, '<<<<<<< TREE\n',
1430
'>>>>>>> MERGE-SOURCE\n')
1431
lines, conflicts = textmerge.merge_lines(self.reprocess)
1433
base_lines = textmerge.base_from_plan()
1436
return lines, base_lines
817
weave = self.this_revision_tree.get_weave(file_id)
818
this_revision_id = self.this_revision_tree.inventory[file_id].revision
819
other_revision_id = \
820
self.other_revision_tree.inventory[file_id].revision
821
wm = WeaveMerge(weave, this_revision_id, other_revision_id,
822
'<<<<<<< TREE\n', '>>>>>>> MERGE-SOURCE\n')
823
return wm.merge_lines(self.reprocess)
1438
825
def text_merge(self, file_id, trans_id):
1439
826
"""Perform a (weave) text merge for a given file and file-id.
1440
827
If conflicts are encountered, .THIS and .OTHER files will be emitted,
1441
828
and a conflict will be noted.
1443
lines, base_lines = self._merged_lines(file_id)
830
self._check_file(file_id)
831
lines, conflicts = self._merged_lines(file_id)
1444
832
lines = list(lines)
1445
# Note we're checking whether the OUTPUT is binary in this case,
833
# Note we're checking whether the OUTPUT is binary in this case,
1446
834
# because we don't want to get into weave merge guts.
1447
textfile.check_text_lines(lines)
835
check_text_lines(lines)
1448
836
self.tt.create_file(lines, trans_id)
1449
if base_lines is not None:
1451
838
self._raw_conflicts.append(('text conflict', trans_id))
1452
839
name = self.tt.final_name(trans_id)
1453
840
parent_id = self.tt.final_parent(trans_id)
1454
file_group = self._dump_conflicts(name, parent_id, file_id,
1456
base_lines=base_lines)
841
file_group = self._dump_conflicts(name, parent_id, file_id,
1457
843
file_group.append(trans_id)
1460
class LCAMerger(WeaveMerger):
1462
def _generate_merge_plan(self, file_id, base):
1463
return self.this_tree.plan_file_lca_merge(file_id, self.other_tree,
1466
846
class Diff3Merger(Merge3Merger):
1467
847
"""Three-way merger using external diff3 for text merging"""
1469
848
def dump_file(self, temp_dir, name, tree, file_id):
1470
out_path = osutils.pathjoin(temp_dir, name)
1471
out_file = open(out_path, "wb")
1473
in_file = tree.get_file(file_id)
1474
for line in in_file:
1475
out_file.write(line)
849
out_path = pathjoin(temp_dir, name)
850
out_file = file(out_path, "wb")
851
in_file = tree.get_file(file_id)
1480
856
def text_merge(self, file_id, trans_id):
1483
859
will be dumped, and a will be conflict noted.
1485
861
import bzrlib.patch
1486
temp_dir = osutils.mkdtemp(prefix="bzr-")
862
temp_dir = mkdtemp(prefix="bzr-")
1488
new_file = osutils.pathjoin(temp_dir, "new")
864
new_file = pathjoin(temp_dir, "new")
1489
865
this = self.dump_file(temp_dir, "this", self.this_tree, file_id)
1490
866
base = self.dump_file(temp_dir, "base", self.base_tree, file_id)
1491
867
other = self.dump_file(temp_dir, "other", self.other_tree, file_id)
1492
868
status = bzrlib.patch.diff3(new_file, this, base, other)
1493
869
if status not in (0, 1):
1494
raise errors.BzrError("Unhandled diff3 exit code")
1495
f = open(new_file, 'rb')
1497
self.tt.create_file(f, trans_id)
870
raise BzrError("Unhandled diff3 exit code")
871
self.tt.create_file(file(new_file, "rb"), trans_id)
1501
873
name = self.tt.final_name(trans_id)
1502
874
parent_id = self.tt.final_parent(trans_id)
1503
875
self._dump_conflicts(name, parent_id, file_id)
1504
self._raw_conflicts.append(('text conflict', trans_id))
876
self._raw_conflicts.append(('text conflict', trans_id))
1506
osutils.rmtree(temp_dir)
1509
881
def merge_inner(this_branch, other_tree, base_tree, ignore_zero=False,
1511
merge_type=Merge3Merger,
1512
interesting_ids=None,
883
merge_type=Merge3Merger,
884
interesting_ids=None,
1515
887
other_rev_id=None,
1516
888
interesting_files=None,
1518
pb=progress.DummyProgress(),
1519
change_reporter=None):
1520
"""Primary interface for merging.
891
"""Primary interface for merging.
1522
typical use is probably
893
typical use is probably
1523
894
'merge_inner(branch, branch.get_revision_tree(other_revision),
1524
895
branch.get_revision_tree(base_revision))'
1526
897
if this_tree is None:
1527
raise errors.BzrError("bzrlib.merge.merge_inner requires a this_tree "
1528
"parameter as of bzrlib version 0.8.")
1529
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1530
pb=pb, change_reporter=change_reporter)
898
warnings.warn("bzrlib.merge.merge_inner requires a this_tree parameter as of "
899
"bzrlib version 0.8.",
902
this_tree = this_branch.bzrdir.open_workingtree()
903
merger = Merger(this_branch, other_tree, base_tree, this_tree=this_tree,
1531
905
merger.backup_files = backup_files
1532
906
merger.merge_type = merge_type
1533
907
merger.interesting_ids = interesting_ids
1534
908
merger.ignore_zero = ignore_zero
1535
909
if interesting_files:
1537
raise ValueError('Only supply interesting_ids'
1538
' or interesting_files')
1539
merger.interesting_files = interesting_files
1540
merger.show_base = show_base
910
assert not interesting_ids, ('Only supply interesting_ids'
911
' or interesting_files')
912
merger._set_interesting_files(interesting_files)
913
merger.show_base = show_base
1541
914
merger.reprocess = reprocess
1542
915
merger.other_rev_id = other_rev_id
1543
916
merger.other_basis = other_rev_id
1544
get_revision_id = getattr(base_tree, 'get_revision_id', None)
1545
if get_revision_id is None:
1546
get_revision_id = base_tree.last_revision
1547
merger.cache_trees_with_revision_ids([other_tree, base_tree, this_tree])
1548
merger.set_base_revision(get_revision_id(), this_branch)
1549
917
return merger.do_merge()
1551
def get_merge_type_registry():
1552
"""Merge type registry is in bzrlib.option to avoid circular imports.
1554
This method provides a sanctioned way to retrieve it.
1556
from bzrlib import option
1557
return option._merge_type_registry
1560
def _plan_annotate_merge(annotated_a, annotated_b, ancestors_a, ancestors_b):
1561
def status_a(revision, text):
1562
if revision in ancestors_b:
1563
return 'killed-b', text
1565
return 'new-a', text
1567
def status_b(revision, text):
1568
if revision in ancestors_a:
1569
return 'killed-a', text
1571
return 'new-b', text
1573
plain_a = [t for (a, t) in annotated_a]
1574
plain_b = [t for (a, t) in annotated_b]
1575
matcher = patiencediff.PatienceSequenceMatcher(None, plain_a, plain_b)
1576
blocks = matcher.get_matching_blocks()
1579
for ai, bi, l in blocks:
1580
# process all mismatched sections
1581
# (last mismatched section is handled because blocks always
1582
# includes a 0-length last block)
1583
for revision, text in annotated_a[a_cur:ai]:
1584
yield status_a(revision, text)
1585
for revision, text in annotated_b[b_cur:bi]:
1586
yield status_b(revision, text)
1587
# and now the matched section
1590
for text_a in plain_a[ai:a_cur]:
1591
yield "unchanged", text_a
1594
class _PlanMergeBase(object):
1596
def __init__(self, a_rev, b_rev, vf, key_prefix):
1599
:param a_rev: Revision-id of one revision to merge
1600
:param b_rev: Revision-id of the other revision to merge
1601
:param vf: A VersionedFiles containing both revisions
1602
:param key_prefix: A prefix for accessing keys in vf, typically
1608
self._last_lines = None
1609
self._last_lines_revision_id = None
1610
self._cached_matching_blocks = {}
1611
self._key_prefix = key_prefix
1612
self._precache_tip_lines()
1614
def _precache_tip_lines(self):
1615
lines = self.get_lines([self.a_rev, self.b_rev])
1616
self.lines_a = lines[self.a_rev]
1617
self.lines_b = lines[self.b_rev]
1619
def get_lines(self, revisions):
1620
"""Get lines for revisions from the backing VersionedFiles.
1622
:raises RevisionNotPresent: on absent texts.
1624
keys = [(self._key_prefix + (rev,)) for rev in revisions]
1626
for record in self.vf.get_record_stream(keys, 'unordered', True):
1627
if record.storage_kind == 'absent':
1628
raise errors.RevisionNotPresent(record.key, self.vf)
1629
result[record.key[-1]] = osutils.chunks_to_lines(
1630
record.get_bytes_as('chunked'))
1633
def plan_merge(self):
1634
"""Generate a 'plan' for merging the two revisions.
1636
This involves comparing their texts and determining the cause of
1637
differences. If text A has a line and text B does not, then either the
1638
line was added to text A, or it was deleted from B. Once the causes
1639
are combined, they are written out in the format described in
1640
VersionedFile.plan_merge
1642
blocks = self._get_matching_blocks(self.a_rev, self.b_rev)
1643
unique_a, unique_b = self._unique_lines(blocks)
1644
new_a, killed_b = self._determine_status(self.a_rev, unique_a)
1645
new_b, killed_a = self._determine_status(self.b_rev, unique_b)
1646
return self._iter_plan(blocks, new_a, killed_b, new_b, killed_a)
1648
def _iter_plan(self, blocks, new_a, killed_b, new_b, killed_a):
1651
for i, j, n in blocks:
1652
for a_index in range(last_i, i):
1653
if a_index in new_a:
1654
if a_index in killed_b:
1655
yield 'conflicted-a', self.lines_a[a_index]
1657
yield 'new-a', self.lines_a[a_index]
1659
yield 'killed-b', self.lines_a[a_index]
1660
for b_index in range(last_j, j):
1661
if b_index in new_b:
1662
if b_index in killed_a:
1663
yield 'conflicted-b', self.lines_b[b_index]
1665
yield 'new-b', self.lines_b[b_index]
1667
yield 'killed-a', self.lines_b[b_index]
1668
# handle common lines
1669
for a_index in range(i, i+n):
1670
yield 'unchanged', self.lines_a[a_index]
1674
def _get_matching_blocks(self, left_revision, right_revision):
1675
"""Return a description of which sections of two revisions match.
1677
See SequenceMatcher.get_matching_blocks
1679
cached = self._cached_matching_blocks.get((left_revision,
1681
if cached is not None:
1683
if self._last_lines_revision_id == left_revision:
1684
left_lines = self._last_lines
1685
right_lines = self.get_lines([right_revision])[right_revision]
1687
lines = self.get_lines([left_revision, right_revision])
1688
left_lines = lines[left_revision]
1689
right_lines = lines[right_revision]
1690
self._last_lines = right_lines
1691
self._last_lines_revision_id = right_revision
1692
matcher = patiencediff.PatienceSequenceMatcher(None, left_lines,
1694
return matcher.get_matching_blocks()
1696
def _unique_lines(self, matching_blocks):
1697
"""Analyse matching_blocks to determine which lines are unique
1699
:return: a tuple of (unique_left, unique_right), where the values are
1700
sets of line numbers of unique lines.
1706
for i, j, n in matching_blocks:
1707
unique_left.extend(range(last_i, i))
1708
unique_right.extend(range(last_j, j))
1711
return unique_left, unique_right
1714
def _subtract_plans(old_plan, new_plan):
1715
"""Remove changes from new_plan that came from old_plan.
1717
It is assumed that the difference between the old_plan and new_plan
1718
is their choice of 'b' text.
1720
All lines from new_plan that differ from old_plan are emitted
1721
verbatim. All lines from new_plan that match old_plan but are
1722
not about the 'b' revision are emitted verbatim.
1724
Lines that match and are about the 'b' revision are the lines we
1725
don't want, so we convert 'killed-b' -> 'unchanged', and 'new-b'
1726
is skipped entirely.
1728
matcher = patiencediff.PatienceSequenceMatcher(None, old_plan,
1731
for i, j, n in matcher.get_matching_blocks():
1732
for jj in range(last_j, j):
1734
for jj in range(j, j+n):
1735
plan_line = new_plan[jj]
1736
if plan_line[0] == 'new-b':
1738
elif plan_line[0] == 'killed-b':
1739
yield 'unchanged', plan_line[1]
1745
class _PlanMerge(_PlanMergeBase):
1746
"""Plan an annotate merge using on-the-fly annotation"""
1748
def __init__(self, a_rev, b_rev, vf, key_prefix):
1749
super(_PlanMerge, self).__init__(a_rev, b_rev, vf, key_prefix)
1750
self.a_key = self._key_prefix + (self.a_rev,)
1751
self.b_key = self._key_prefix + (self.b_rev,)
1752
self.graph = _mod_graph.Graph(self.vf)
1753
heads = self.graph.heads((self.a_key, self.b_key))
1755
# one side dominates, so we can just return its values, yay for
1757
# Ideally we would know that before we get this far
1758
self._head_key = heads.pop()
1759
if self._head_key == self.a_key:
1763
trace.mutter('found dominating revision for %s\n%s > %s', self.vf,
1764
self._head_key[-1], other)
1767
self._head_key = None
1770
def _precache_tip_lines(self):
1771
# Turn this into a no-op, because we will do this later
1774
def _find_recursive_lcas(self):
1775
"""Find all the ancestors back to a unique lca"""
1776
cur_ancestors = (self.a_key, self.b_key)
1777
# graph.find_lca(uncommon, keys) now returns plain NULL_REVISION,
1778
# rather than a key tuple. We will just map that directly to no common
1782
next_lcas = self.graph.find_lca(*cur_ancestors)
1783
# Map a plain NULL_REVISION to a simple no-ancestors
1784
if next_lcas == set([_mod_revision.NULL_REVISION]):
1786
# Order the lca's based on when they were merged into the tip
1787
# While the actual merge portion of weave merge uses a set() of
1788
# active revisions, the order of insertion *does* effect the
1789
# implicit ordering of the texts.
1790
for rev_key in cur_ancestors:
1791
ordered_parents = tuple(self.graph.find_merge_order(rev_key,
1793
parent_map[rev_key] = ordered_parents
1794
if len(next_lcas) == 0:
1796
elif len(next_lcas) == 1:
1797
parent_map[list(next_lcas)[0]] = ()
1799
elif len(next_lcas) > 2:
1800
# More than 2 lca's, fall back to grabbing all nodes between
1801
# this and the unique lca.
1802
trace.mutter('More than 2 LCAs, falling back to all nodes for:'
1804
self.a_key, self.b_key, cur_ancestors)
1805
cur_lcas = next_lcas
1806
while len(cur_lcas) > 1:
1807
cur_lcas = self.graph.find_lca(*cur_lcas)
1808
if len(cur_lcas) == 0:
1809
# No common base to find, use the full ancestry
1812
unique_lca = list(cur_lcas)[0]
1813
if unique_lca == _mod_revision.NULL_REVISION:
1814
# find_lca will return a plain 'NULL_REVISION' rather
1815
# than a key tuple when there is no common ancestor, we
1816
# prefer to just use None, because it doesn't confuse
1817
# _get_interesting_texts()
1819
parent_map.update(self._find_unique_parents(next_lcas,
1822
cur_ancestors = next_lcas
1825
def _find_unique_parents(self, tip_keys, base_key):
1826
"""Find ancestors of tip that aren't ancestors of base.
1828
:param tip_keys: Nodes that are interesting
1829
:param base_key: Cull all ancestors of this node
1830
:return: The parent map for all revisions between tip_keys and
1831
base_key. base_key will be included. References to nodes outside of
1832
the ancestor set will also be removed.
1834
# TODO: this would be simpler if find_unique_ancestors took a list
1835
# instead of a single tip, internally it supports it, but it
1836
# isn't a "backwards compatible" api change.
1837
if base_key is None:
1838
parent_map = dict(self.graph.iter_ancestry(tip_keys))
1839
# We remove NULL_REVISION because it isn't a proper tuple key, and
1840
# thus confuses things like _get_interesting_texts, and our logic
1841
# to add the texts into the memory weave.
1842
if _mod_revision.NULL_REVISION in parent_map:
1843
parent_map.pop(_mod_revision.NULL_REVISION)
1846
for tip in tip_keys:
1848
self.graph.find_unique_ancestors(tip, [base_key]))
1849
parent_map = self.graph.get_parent_map(interesting)
1850
parent_map[base_key] = ()
1851
culled_parent_map, child_map, tails = self._remove_external_references(
1853
# Remove all the tails but base_key
1854
if base_key is not None:
1855
tails.remove(base_key)
1856
self._prune_tails(culled_parent_map, child_map, tails)
1857
# Now remove all the uninteresting 'linear' regions
1858
simple_map = _mod_graph.collapse_linear_regions(culled_parent_map)
1862
def _remove_external_references(parent_map):
1863
"""Remove references that go outside of the parent map.
1865
:param parent_map: Something returned from Graph.get_parent_map(keys)
1866
:return: (filtered_parent_map, child_map, tails)
1867
filtered_parent_map is parent_map without external references
1868
child_map is the {parent_key: [child_keys]} mapping
1869
tails is a list of nodes that do not have any parents in the map
1871
# TODO: The basic effect of this function seems more generic than
1872
# _PlanMerge. But the specific details of building a child_map,
1873
# and computing tails seems very specific to _PlanMerge.
1874
# Still, should this be in Graph land?
1875
filtered_parent_map = {}
1878
for key, parent_keys in parent_map.iteritems():
1879
culled_parent_keys = [p for p in parent_keys if p in parent_map]
1880
if not culled_parent_keys:
1882
for parent_key in culled_parent_keys:
1883
child_map.setdefault(parent_key, []).append(key)
1884
# TODO: Do we want to do this, it adds overhead for every node,
1885
# just to say that the node has no children
1886
child_map.setdefault(key, [])
1887
filtered_parent_map[key] = culled_parent_keys
1888
return filtered_parent_map, child_map, tails
1891
def _prune_tails(parent_map, child_map, tails_to_remove):
1892
"""Remove tails from the parent map.
1894
This will remove the supplied revisions until no more children have 0
1897
:param parent_map: A dict of {child: [parents]}, this dictionary will
1898
be modified in place.
1899
:param tails_to_remove: A list of tips that should be removed,
1900
this list will be consumed
1901
:param child_map: The reverse dict of parent_map ({parent: [children]})
1902
this dict will be modified
1903
:return: None, parent_map will be modified in place.
1905
while tails_to_remove:
1906
next = tails_to_remove.pop()
1907
parent_map.pop(next)
1908
children = child_map.pop(next)
1909
for child in children:
1910
child_parents = parent_map[child]
1911
child_parents.remove(next)
1912
if len(child_parents) == 0:
1913
tails_to_remove.append(child)
1915
def _get_interesting_texts(self, parent_map):
1916
"""Return a dict of texts we are interested in.
1918
Note that the input is in key tuples, but the output is in plain
1921
:param parent_map: The output from _find_recursive_lcas
1922
:return: A dict of {'revision_id':lines} as returned by
1923
_PlanMergeBase.get_lines()
1925
all_revision_keys = set(parent_map)
1926
all_revision_keys.add(self.a_key)
1927
all_revision_keys.add(self.b_key)
1929
# Everything else is in 'keys' but get_lines is in 'revision_ids'
1930
all_texts = self.get_lines([k[-1] for k in all_revision_keys])
1933
def _build_weave(self):
1934
from bzrlib import weave
1935
self._weave = weave.Weave(weave_name='in_memory_weave',
1936
allow_reserved=True)
1937
parent_map = self._find_recursive_lcas()
1939
all_texts = self._get_interesting_texts(parent_map)
1941
# Note: Unfortunately, the order given by topo_sort will effect the
1942
# ordering resolution in the output. Specifically, if you add A then B,
1943
# then in the output text A lines will show up before B lines. And, of
1944
# course, topo_sort doesn't guarantee any real ordering.
1945
# So we use merge_sort, and add a fake node on the tip.
1946
# This ensures that left-hand parents will always be inserted into the
1947
# weave before right-hand parents.
1948
tip_key = self._key_prefix + (_mod_revision.CURRENT_REVISION,)
1949
parent_map[tip_key] = (self.a_key, self.b_key)
1951
for seq_num, key, depth, eom in reversed(tsort.merge_sort(parent_map,
1955
# for key in tsort.topo_sort(parent_map):
1956
parent_keys = parent_map[key]
1957
revision_id = key[-1]
1958
parent_ids = [k[-1] for k in parent_keys]
1959
self._weave.add_lines(revision_id, parent_ids,
1960
all_texts[revision_id])
1962
def plan_merge(self):
1963
"""Generate a 'plan' for merging the two revisions.
1965
This involves comparing their texts and determining the cause of
1966
differences. If text A has a line and text B does not, then either the
1967
line was added to text A, or it was deleted from B. Once the causes
1968
are combined, they are written out in the format described in
1969
VersionedFile.plan_merge
1971
if self._head_key is not None: # There was a single head
1972
if self._head_key == self.a_key:
1975
if self._head_key != self.b_key:
1976
raise AssertionError('There was an invalid head: %s != %s'
1977
% (self.b_key, self._head_key))
1979
head_rev = self._head_key[-1]
1980
lines = self.get_lines([head_rev])[head_rev]
1981
return ((plan, line) for line in lines)
1982
return self._weave.plan_merge(self.a_rev, self.b_rev)
1985
class _PlanLCAMerge(_PlanMergeBase):
1987
This merge algorithm differs from _PlanMerge in that:
1988
1. comparisons are done against LCAs only
1989
2. cases where a contested line is new versus one LCA but old versus
1990
another are marked as conflicts, by emitting the line as conflicted-a
1993
This is faster, and hopefully produces more useful output.
1996
def __init__(self, a_rev, b_rev, vf, key_prefix, graph):
1997
_PlanMergeBase.__init__(self, a_rev, b_rev, vf, key_prefix)
1998
lcas = graph.find_lca(key_prefix + (a_rev,), key_prefix + (b_rev,))
2001
if lca == _mod_revision.NULL_REVISION:
2004
self.lcas.add(lca[-1])
2005
for lca in self.lcas:
2006
if _mod_revision.is_null(lca):
2009
lca_lines = self.get_lines([lca])[lca]
2010
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_a,
2012
blocks = list(matcher.get_matching_blocks())
2013
self._cached_matching_blocks[(a_rev, lca)] = blocks
2014
matcher = patiencediff.PatienceSequenceMatcher(None, self.lines_b,
2016
blocks = list(matcher.get_matching_blocks())
2017
self._cached_matching_blocks[(b_rev, lca)] = blocks
2019
def _determine_status(self, revision_id, unique_line_numbers):
2020
"""Determines the status unique lines versus all lcas.
2022
Basically, determines why the line is unique to this revision.
2024
A line may be determined new, killed, or both.
2026
If a line is determined new, that means it was not present in at least
2027
one LCA, and is not present in the other merge revision.
2029
If a line is determined killed, that means the line was present in
2032
If a line is killed and new, this indicates that the two merge
2033
revisions contain differing conflict resolutions.
2034
:param revision_id: The id of the revision in which the lines are
2036
:param unique_line_numbers: The line numbers of unique lines.
2037
:return a tuple of (new_this, killed_other):
2041
unique_line_numbers = set(unique_line_numbers)
2042
for lca in self.lcas:
2043
blocks = self._get_matching_blocks(revision_id, lca)
2044
unique_vs_lca, _ignored = self._unique_lines(blocks)
2045
new.update(unique_line_numbers.intersection(unique_vs_lca))
2046
killed.update(unique_line_numbers.difference(unique_vs_lca))
920
merge_types = { "merge3": (Merge3Merger, "Native diff3-style merge"),
921
"diff3": (Diff3Merger, "Merge using external diff3"),
922
'weave': (WeaveMerger, "Weave-based merge")
926
def merge_type_help():
927
templ = '%s%%7s: %%s' % (' '*12)
928
lines = [templ % (f[0], f[1][1]) for f in merge_types.iteritems()]
929
return '\n'.join(lines)