96
110
from bzrlib.filters import filtered_input_file
97
111
from bzrlib.trace import mutter, note
98
112
from bzrlib.transport.local import LocalTransport
99
from bzrlib.revision import CURRENT_REVISION
113
from bzrlib.progress import DummyProgress, ProgressPhase
114
from bzrlib.revision import NULL_REVISION, CURRENT_REVISION
100
115
from bzrlib.rio import RioReader, rio_file, Stanza
101
from bzrlib.symbol_versioning import (
103
DEPRECATED_PARAMETER,
116
from bzrlib.symbol_versioning import (deprecated_passed,
119
DEPRECATED_PARAMETER,
107
123
MERGE_MODIFIED_HEADER_1 = "BZR merge-modified list format 1"
108
# TODO: Modifying the conflict objects or their type is currently nearly
109
# impossible as there is no clear relationship between the working tree format
110
# and the conflict list file format.
111
124
CONFLICT_HEADER_1 = "BZR conflict list format 1"
113
126
ERROR_PATH_NOT_FOUND = 3 # WindowsError errno code, equivalent to ENOENT
351
342
path = osutils.getcwd()
352
343
control, relpath = bzrdir.BzrDir.open_containing(path)
353
345
return control.open_workingtree(), relpath
356
def open_containing_paths(file_list, default_directory='.',
357
canonicalize=True, apply_view=True):
358
"""Open the WorkingTree that contains a set of paths.
360
Fail if the paths given are not all in a single tree.
362
This is used for the many command-line interfaces that take a list of
363
any number of files and that require they all be in the same tree.
365
# recommended replacement for builtins.internal_tree_files
366
if file_list is None or len(file_list) == 0:
367
tree = WorkingTree.open_containing(default_directory)[0]
368
# XXX: doesn't really belong here, and seems to have the strange
369
# side effect of making it return a bunch of files, not the whole
370
# tree -- mbp 20100716
371
if tree.supports_views() and apply_view:
372
view_files = tree.views.lookup_view()
374
file_list = view_files
375
view_str = views.view_display_str(view_files)
376
note("Ignoring files outside view. View is %s" % view_str)
377
return tree, file_list
378
tree = WorkingTree.open_containing(file_list[0])[0]
379
return tree, tree.safe_relpath_files(file_list, canonicalize,
380
apply_view=apply_view)
382
def safe_relpath_files(self, file_list, canonicalize=True, apply_view=True):
383
"""Convert file_list into a list of relpaths in tree.
385
:param self: A tree to operate on.
386
:param file_list: A list of user provided paths or None.
387
:param apply_view: if True and a view is set, apply it or check that
388
specified files are within it
389
:return: A list of relative paths.
390
:raises errors.PathNotChild: When a provided path is in a different self
393
if file_list is None:
395
if self.supports_views() and apply_view:
396
view_files = self.views.lookup_view()
400
# self.relpath exists as a "thunk" to osutils, but canonical_relpath
401
# doesn't - fix that up here before we enter the loop.
403
fixer = lambda p: osutils.canonical_relpath(self.basedir, p)
406
for filename in file_list:
407
relpath = fixer(osutils.dereference_path(filename))
408
if view_files and not osutils.is_inside_any(view_files, relpath):
409
raise errors.FileOutsideView(filename, view_files)
410
new_list.append(relpath)
414
348
def open_downlevel(path=None):
415
349
"""Open an unsupported working tree.
513
447
def get_file_with_stat(self, file_id, path=None, filtered=True,
514
448
_fstat=os.fstat):
515
"""See Tree.get_file_with_stat."""
449
"""See MutableTree.get_file_with_stat."""
517
451
path = self.id2path(file_id)
518
452
file_obj = self.get_file_byname(path, filtered=False)
519
453
stat_value = _fstat(file_obj.fileno())
520
if filtered and self.supports_content_filtering():
454
if self.supports_content_filtering() and filtered:
521
455
filters = self._content_filter_stack(path)
522
456
file_obj = filtered_input_file(file_obj, filters)
523
457
return (file_obj, stat_value)
525
459
def get_file_text(self, file_id, path=None, filtered=True):
526
my_file = self.get_file(file_id, path=path, filtered=filtered)
528
return my_file.read()
460
return self.get_file(file_id, path=path, filtered=filtered).read()
532
462
def get_file_byname(self, filename, filtered=True):
533
463
path = self.abspath(filename)
534
464
f = file(path, 'rb')
535
if filtered and self.supports_content_filtering():
465
if self.supports_content_filtering() and filtered:
536
466
filters = self._content_filter_stack(filename)
537
467
return filtered_input_file(f, filters)
557
487
incorrectly attributed to CURRENT_REVISION (but after committing, the
558
488
attribution will be correct).
560
maybe_file_parent_keys = []
561
for parent_id in self.get_parent_ids():
563
parent_tree = self.revision_tree(parent_id)
564
except errors.NoSuchRevisionInTree:
565
parent_tree = self.branch.repository.revision_tree(parent_id)
566
parent_tree.lock_read()
568
if file_id not in parent_tree:
570
ie = parent_tree.inventory[file_id]
571
if ie.kind != 'file':
572
# Note: this is slightly unnecessary, because symlinks and
573
# directories have a "text" which is the empty text, and we
574
# know that won't mess up annotations. But it seems cleaner
576
parent_text_key = (file_id, ie.revision)
577
if parent_text_key not in maybe_file_parent_keys:
578
maybe_file_parent_keys.append(parent_text_key)
581
graph = _mod_graph.Graph(self.branch.repository.texts)
582
heads = graph.heads(maybe_file_parent_keys)
583
file_parent_keys = []
584
for key in maybe_file_parent_keys:
586
file_parent_keys.append(key)
588
# Now we have the parents of this content
589
annotator = self.branch.repository.texts.get_annotator()
590
text = self.get_file_text(file_id)
591
this_key =(file_id, default_revision)
592
annotator.add_special_text(this_key, file_parent_keys, text)
593
annotations = [(key[-1], line)
594
for key, line in annotator.annotate_flat(this_key)]
490
basis = self.basis_tree()
493
changes = self.iter_changes(basis, True, [self.id2path(file_id)],
494
require_versioned=True).next()
495
changed_content, kind = changes[2], changes[6]
496
if not changed_content:
497
return basis.annotate_iter(file_id)
501
if kind[0] != 'file':
504
old_lines = list(basis.annotate_iter(file_id))
506
for tree in self.branch.repository.revision_trees(
507
self.get_parent_ids()[1:]):
508
if file_id not in tree:
510
old.append(list(tree.annotate_iter(file_id)))
511
return annotate.reannotate(old, self.get_file(file_id).readlines(),
597
516
def _get_ancestors(self, default_revision):
598
517
ancestors = set([default_revision])
832
749
kind = 'tree-reference'
833
750
return kind, None, None, None
834
751
elif kind == 'symlink':
835
target = osutils.readlink(abspath)
836
return ('symlink', None, None, target)
752
return ('symlink', None, None,
753
os.readlink(abspath.encode(osutils._fs_enc)
754
).decode(osutils._fs_enc))
838
756
return (kind, None, None, None)
840
def _file_content_summary(self, path, stat_result):
841
size = stat_result.st_size
842
executable = self._is_executable_from_path_and_stat(path, stat_result)
843
# try for a stat cache lookup
844
return ('file', size, executable, self._sha_from_stat(
847
758
def _check_parents_for_ghosts(self, revision_ids, allow_leftmost_as_ghost):
848
759
"""Common ghost checking functionality from set_parent_*.
979
890
branch.last_revision().
981
892
from bzrlib.merge import Merger, Merge3Merger
982
merger = Merger(self.branch, this_tree=self)
983
# check that there are no local alterations
984
if not force and self.has_changes():
985
raise errors.UncommittedChanges(self)
986
if to_revision is None:
987
to_revision = _mod_revision.ensure_null(branch.last_revision())
988
merger.other_rev_id = to_revision
989
if _mod_revision.is_null(merger.other_rev_id):
990
raise errors.NoCommits(branch)
991
self.branch.fetch(branch, last_revision=merger.other_rev_id)
992
merger.other_basis = merger.other_rev_id
993
merger.other_tree = self.branch.repository.revision_tree(
995
merger.other_branch = branch
996
if from_revision is None:
999
merger.set_base_revision(from_revision, branch)
1000
if merger.base_rev_id == merger.other_rev_id:
1001
raise errors.PointlessMerge
1002
merger.backup_files = False
1003
if merge_type is None:
1004
merger.merge_type = Merge3Merger
1006
merger.merge_type = merge_type
1007
merger.set_interesting_files(None)
1008
merger.show_base = False
1009
merger.reprocess = False
1010
conflicts = merger.do_merge()
1011
merger.set_pending()
893
pb = bzrlib.ui.ui_factory.nested_progress_bar()
895
merger = Merger(self.branch, this_tree=self, pb=pb)
896
merger.pp = ProgressPhase("Merge phase", 5, pb)
897
merger.pp.next_phase()
898
# check that there are no
900
merger.check_basis(check_clean=True, require_commits=False)
901
if to_revision is None:
902
to_revision = _mod_revision.ensure_null(branch.last_revision())
903
merger.other_rev_id = to_revision
904
if _mod_revision.is_null(merger.other_rev_id):
905
raise errors.NoCommits(branch)
906
self.branch.fetch(branch, last_revision=merger.other_rev_id)
907
merger.other_basis = merger.other_rev_id
908
merger.other_tree = self.branch.repository.revision_tree(
910
merger.other_branch = branch
911
merger.pp.next_phase()
912
if from_revision is None:
915
merger.set_base_revision(from_revision, branch)
916
if merger.base_rev_id == merger.other_rev_id:
917
raise errors.PointlessMerge
918
merger.backup_files = False
919
if merge_type is None:
920
merger.merge_type = Merge3Merger
922
merger.merge_type = merge_type
923
merger.set_interesting_files(None)
924
merger.show_base = False
925
merger.reprocess = False
926
conflicts = merger.do_merge()
1012
930
return conflicts
1014
932
@needs_read_lock
1161
1078
tree_transport = self.bzrdir.root_transport.clone(sub_path)
1162
1079
if tree_transport.base != branch_transport.base:
1163
1080
tree_bzrdir = format.initialize_on_transport(tree_transport)
1164
branch.BranchReferenceFormat().initialize(tree_bzrdir,
1165
target_branch=new_branch)
1081
branch.BranchReferenceFormat().initialize(tree_bzrdir, new_branch)
1167
1083
tree_bzrdir = branch_bzrdir
1168
wt = tree_bzrdir.create_workingtree(_mod_revision.NULL_REVISION)
1084
wt = tree_bzrdir.create_workingtree(NULL_REVISION)
1169
1085
wt.set_parent_ids(self.get_parent_ids())
1170
1086
my_inv = self.inventory
1171
child_inv = inventory.Inventory(root_id=None)
1087
child_inv = Inventory(root_id=None)
1172
1088
new_root = my_inv[file_id]
1173
1089
my_inv.remove_recursive_id(file_id)
1174
1090
new_root.parent_id = None
1199
1115
def _kind(self, relpath):
1200
1116
return osutils.file_kind(self.abspath(relpath))
1202
def list_files(self, include_root=False, from_dir=None, recursive=True):
1203
"""List all files as (path, class, kind, id, entry).
1118
def list_files(self, include_root=False):
1119
"""Recursively list all files as (path, class, kind, id, entry).
1205
1121
Lists, but does not descend into unversioned directories.
1206
1123
This does not include files that have been deleted in this
1207
tree. Skips the control directory.
1209
:param include_root: if True, return an entry for the root
1210
:param from_dir: start from this directory or None for the root
1211
:param recursive: whether to recurse into subdirectories or not
1126
Skips the control directory.
1213
1128
# list_files is an iterator, so @needs_read_lock doesn't work properly
1214
1129
# with it. So callers should be careful to always read_lock the tree.
1229
1144
fk_entries = {'directory':TreeDirectory, 'file':TreeFile, 'symlink':TreeLink}
1231
1146
# directory file_id, relative path, absolute path, reverse sorted children
1232
if from_dir is not None:
1233
from_dir_id = inv.path2id(from_dir)
1234
if from_dir_id is None:
1235
# Directory not versioned
1237
from_dir_abspath = pathjoin(self.basedir, from_dir)
1239
from_dir_id = inv.root.file_id
1240
from_dir_abspath = self.basedir
1241
children = os.listdir(from_dir_abspath)
1147
children = os.listdir(self.basedir)
1242
1148
children.sort()
1243
1149
# jam 20060527 The kernel sized tree seems equivalent whether we
1244
1150
# use a deque and popleft to keep them sorted, or if we use a plain
1245
1151
# list and just reverse() them.
1246
1152
children = collections.deque(children)
1247
stack = [(from_dir_id, u'', from_dir_abspath, children)]
1153
stack = [(inv.root.file_id, u'', self.basedir, children)]
1249
1155
from_dir_id, from_dir_relpath, from_dir_abspath, children = stack[-1]
1313
1214
if fk != 'directory':
1316
# But do this child first if recursing down
1318
new_children = os.listdir(fap)
1320
new_children = collections.deque(new_children)
1321
stack.append((f_ie.file_id, fp, fap, new_children))
1322
# Break out of inner loop,
1323
# so that we start outer loop with child
1217
# But do this child first
1218
new_children = os.listdir(fap)
1220
new_children = collections.deque(new_children)
1221
stack.append((f_ie.file_id, fp, fap, new_children))
1222
# Break out of inner loop,
1223
# so that we start outer loop with child
1326
1226
# if we finished all children, pop it off the stack
1329
1229
@needs_tree_write_lock
1330
def move(self, from_paths, to_dir=None, after=False):
1230
def move(self, from_paths, to_dir=None, after=False, **kwargs):
1331
1231
"""Rename files.
1333
1233
to_dir must exist in the inventory.
1556
1462
from_tail = splitpath(from_rel)[-1]
1557
1463
from_id = inv.path2id(from_rel)
1558
1464
if from_id is None:
1559
# if file is missing in the inventory maybe it's in the basis_tree
1560
basis_tree = self.branch.basis_tree()
1561
from_id = basis_tree.path2id(from_rel)
1563
raise errors.BzrRenameFailedError(from_rel,to_rel,
1564
errors.NotVersionedError(path=str(from_rel)))
1565
# put entry back in the inventory so we can rename it
1566
from_entry = basis_tree.inventory[from_id].copy()
1569
from_entry = inv[from_id]
1465
raise errors.BzrRenameFailedError(from_rel,to_rel,
1466
errors.NotVersionedError(path=str(from_rel)))
1467
from_entry = inv[from_id]
1570
1468
from_parent_id = from_entry.parent_id
1571
1469
to_dir, to_tail = os.path.split(to_rel)
1572
1470
to_dir_id = inv.path2id(to_dir)
1664
1562
@needs_write_lock
1665
1563
def pull(self, source, overwrite=False, stop_revision=None,
1666
change_reporter=None, possible_transports=None, local=False):
1564
change_reporter=None, possible_transports=None):
1565
top_pb = bzrlib.ui.ui_factory.nested_progress_bar()
1667
1566
source.lock_read()
1568
pp = ProgressPhase("Pull phase", 2, top_pb)
1669
1570
old_revision_info = self.branch.last_revision_info()
1670
1571
basis_tree = self.basis_tree()
1671
1572
count = self.branch.pull(source, overwrite, stop_revision,
1672
possible_transports=possible_transports,
1573
possible_transports=possible_transports)
1674
1574
new_revision_info = self.branch.last_revision_info()
1675
1575
if new_revision_info != old_revision_info:
1676
1577
repository = self.branch.repository
1578
pb = bzrlib.ui.ui_factory.nested_progress_bar()
1677
1579
basis_tree.lock_read()
1679
1581
new_basis_tree = self.branch.basis_tree()
1861
1762
raise errors.ObjectNotLocked(self)
1863
1764
def lock_read(self):
1864
"""Lock the tree for reading.
1866
This also locks the branch, and can be unlocked via self.unlock().
1868
:return: A bzrlib.lock.LogicalLockResult.
1765
"""See Branch.lock_read, and WorkingTree.unlock."""
1870
1766
if not self.is_locked():
1871
1767
self._reset_data()
1872
1768
self.branch.lock_read()
1874
self._control_files.lock_read()
1875
return LogicalLockResult(self.unlock)
1770
return self._control_files.lock_read()
1877
1772
self.branch.unlock()
1880
1775
def lock_tree_write(self):
1881
"""See MutableTree.lock_tree_write, and WorkingTree.unlock.
1883
:return: A bzrlib.lock.LogicalLockResult.
1776
"""See MutableTree.lock_tree_write, and WorkingTree.unlock."""
1885
1777
if not self.is_locked():
1886
1778
self._reset_data()
1887
1779
self.branch.lock_read()
1889
self._control_files.lock_write()
1890
return LogicalLockResult(self.unlock)
1781
return self._control_files.lock_write()
1892
1783
self.branch.unlock()
1895
1786
def lock_write(self):
1896
"""See MutableTree.lock_write, and WorkingTree.unlock.
1898
:return: A bzrlib.lock.LogicalLockResult.
1787
"""See MutableTree.lock_write, and WorkingTree.unlock."""
1900
1788
if not self.is_locked():
1901
1789
self._reset_data()
1902
1790
self.branch.lock_write()
1904
self._control_files.lock_write()
1905
return LogicalLockResult(self.unlock)
1792
return self._control_files.lock_write()
1907
1794
self.branch.unlock()
2028
all_files = set() # specified and nested files
2029
1908
unknown_nested_files=set()
2031
to_file = sys.stdout
2033
files_to_backup = []
2035
1910
def recurse_directory_to_add_files(directory):
2036
1911
# Recurse directory and add all files
2037
1912
# so we can check if they have changed.
2038
for parent_info, file_infos in self.walkdirs(directory):
1913
for parent_info, file_infos in\
1914
self.walkdirs(directory):
2039
1915
for relpath, basename, kind, lstat, fileid, kind in file_infos:
2040
1916
# Is it versioned or ignored?
2041
if self.path2id(relpath):
1917
if self.path2id(relpath) or self.is_ignored(relpath):
2042
1918
# Add nested content for deletion.
2043
all_files.add(relpath)
1919
new_files.add(relpath)
2045
# Files which are not versioned
1921
# Files which are not versioned and not ignored
2046
1922
# should be treated as unknown.
2047
files_to_backup.append(relpath)
1923
unknown_nested_files.add((relpath, None, kind))
2049
1925
for filename in files:
2050
1926
# Get file name into canonical form.
2051
1927
abspath = self.abspath(filename)
2052
1928
filename = self.relpath(abspath)
2053
1929
if len(filename) > 0:
2054
all_files.add(filename)
1930
new_files.add(filename)
2055
1931
recurse_directory_to_add_files(filename)
2057
files = list(all_files)
1933
files = list(new_files)
2059
1935
if len(files) == 0:
2060
1936
return # nothing to do
2065
1941
# Bail out if we are going to delete files we shouldn't
2066
1942
if not keep_files and not force:
2067
for (file_id, path, content_change, versioned, parent_id, name,
2068
kind, executable) in self.iter_changes(self.basis_tree(),
2069
include_unchanged=True, require_versioned=False,
2070
want_unversioned=True, specific_files=files):
2071
if versioned[0] == False:
2072
# The record is unknown or newly added
2073
files_to_backup.append(path[1])
2074
elif (content_change and (kind[1] is not None) and
2075
osutils.is_inside_any(files, path[1])):
2076
# Versioned and changed, but not deleted, and still
2077
# in one of the dirs to be deleted.
2078
files_to_backup.append(path[1])
2080
def backup(file_to_backup):
2081
backup_name = self.bzrdir.generate_backup_name(file_to_backup)
2082
osutils.rename(abs_path, self.abspath(backup_name))
2083
return "removed %s (but kept a copy: %s)" % (file_to_backup, backup_name)
2085
# Build inv_delta and delete files where applicable,
1943
has_changed_files = len(unknown_nested_files) > 0
1944
if not has_changed_files:
1945
for (file_id, path, content_change, versioned, parent_id, name,
1946
kind, executable) in self.iter_changes(self.basis_tree(),
1947
include_unchanged=True, require_versioned=False,
1948
want_unversioned=True, specific_files=files):
1949
if versioned == (False, False):
1950
# The record is unknown ...
1951
if not self.is_ignored(path[1]):
1952
# ... but not ignored
1953
has_changed_files = True
1955
elif content_change and (kind[1] is not None):
1956
# Versioned and changed, but not deleted
1957
has_changed_files = True
1960
if has_changed_files:
1961
# Make delta show ALL applicable changes in error message.
1962
tree_delta = self.changes_from(self.basis_tree(),
1963
require_versioned=False, want_unversioned=True,
1964
specific_files=files)
1965
for unknown_file in unknown_nested_files:
1966
if unknown_file not in tree_delta.unversioned:
1967
tree_delta.unversioned.extend((unknown_file,))
1968
raise errors.BzrRemoveChangedFilesError(tree_delta)
1970
# Build inv_delta and delete files where applicaple,
2086
1971
# do this before any modifications to inventory.
2087
1972
for f in files:
2088
1973
fid = self.path2id(f)
2325
2197
# We MUST save it even if an error occurs, because otherwise the users
2326
2198
# local work is unreferenced and will appear to have been lost.
2330
2202
last_rev = self.get_parent_ids()[0]
2331
2203
except IndexError:
2332
2204
last_rev = _mod_revision.NULL_REVISION
2333
if revision is None:
2334
revision = self.branch.last_revision()
2336
old_tip = old_tip or _mod_revision.NULL_REVISION
2338
if not _mod_revision.is_null(old_tip) and old_tip != last_rev:
2339
# the branch we are bound to was updated
2340
# merge those changes in first
2341
base_tree = self.basis_tree()
2342
other_tree = self.branch.repository.revision_tree(old_tip)
2343
nb_conflicts = merge.merge_inner(self.branch, other_tree,
2344
base_tree, this_tree=self,
2345
change_reporter=change_reporter)
2347
self.add_parent_tree((old_tip, other_tree))
2348
trace.note('Rerun update after fixing the conflicts.')
2351
if last_rev != _mod_revision.ensure_null(revision):
2352
# the working tree is up to date with the branch
2353
# we can merge the specified revision from master
2354
to_tree = self.branch.repository.revision_tree(revision)
2355
to_root_id = to_tree.get_root_id()
2205
if last_rev != _mod_revision.ensure_null(self.branch.last_revision()):
2206
# merge tree state up to new branch tip.
2357
2207
basis = self.basis_tree()
2358
2208
basis.lock_read()
2360
if (basis.inventory.root is None
2361
or basis.inventory.root.file_id != to_root_id):
2362
self.set_root_id(to_root_id)
2210
to_tree = self.branch.basis_tree()
2211
if basis.inventory.root is None:
2212
self.set_root_id(to_tree.get_root_id())
2214
result += merge.merge_inner(
2219
change_reporter=change_reporter)
2367
# determine the branch point
2368
graph = self.branch.repository.get_graph()
2369
base_rev_id = graph.find_unique_lca(self.branch.last_revision(),
2371
base_tree = self.branch.repository.revision_tree(base_rev_id)
2373
nb_conflicts = merge.merge_inner(self.branch, to_tree, base_tree,
2375
change_reporter=change_reporter)
2376
self.set_last_revision(revision)
2377
2222
# TODO - dedup parents list with things merged by pull ?
2378
2223
# reuse the tree we've updated to to set the basis:
2379
parent_trees = [(revision, to_tree)]
2224
parent_trees = [(self.branch.last_revision(), to_tree)]
2380
2225
merges = self.get_parent_ids()[1:]
2381
2226
# Ideally we ask the tree for the trees here, that way the working
2382
# tree can decide whether to give us the entire tree or give us a
2227
# tree can decide whether to give us teh entire tree or give us a
2383
2228
# lazy initialised tree. dirstate for instance will have the trees
2384
2229
# in ram already, whereas a last-revision + basis-inventory tree
2385
2230
# will not, but also does not need them when setting parents.
2386
2231
for parent in merges:
2387
2232
parent_trees.append(
2388
2233
(parent, self.branch.repository.revision_tree(parent)))
2389
if not _mod_revision.is_null(old_tip):
2234
if (old_tip is not None and not _mod_revision.is_null(old_tip)):
2390
2235
parent_trees.append(
2391
2236
(old_tip, self.branch.repository.revision_tree(old_tip)))
2392
2237
self.set_parent_trees(parent_trees)
2393
2238
last_rev = parent_trees[0][0]
2240
# the working tree had the same last-revision as the master
2241
# branch did. We may still have pivot local work from the local
2242
# branch into old_tip:
2243
if (old_tip is not None and not _mod_revision.is_null(old_tip)):
2244
self.add_parent_tree_id(old_tip)
2245
if (old_tip is not None and not _mod_revision.is_null(old_tip)
2246
and old_tip != last_rev):
2247
# our last revision was not the prior branch last revision
2248
# and we have converted that last revision to a pending merge.
2249
# base is somewhere between the branch tip now
2250
# and the now pending merge
2252
# Since we just modified the working tree and inventory, flush out
2253
# the current state, before we modify it again.
2254
# TODO: jam 20070214 WorkingTree3 doesn't require this, dirstate
2255
# requires it only because TreeTransform directly munges the
2256
# inventory and calls tree._write_inventory(). Ultimately we
2257
# should be able to remove this extra flush.
2259
graph = self.branch.repository.get_graph()
2260
base_rev_id = graph.find_unique_lca(self.branch.last_revision(),
2262
base_tree = self.branch.repository.revision_tree(base_rev_id)
2263
other_tree = self.branch.repository.revision_tree(old_tip)
2264
result += merge.merge_inner(
2269
change_reporter=change_reporter)
2396
2272
def _write_hashcache_if_dirty(self):
2397
2273
"""Write out the hashcache if it is dirty."""