97
91
want_unversioned=want_unversioned,
100
def iter_changes(self, from_tree, include_unchanged=False,
94
def _iter_changes(self, from_tree, include_unchanged=False,
101
95
specific_files=None, pb=None, extra_trees=None,
102
96
require_versioned=True, want_unversioned=False):
103
97
intertree = InterTree.get(from_tree, self)
104
return intertree.iter_changes(include_unchanged, specific_files, pb,
98
return intertree._iter_changes(include_unchanged, specific_files, pb,
105
99
extra_trees, require_versioned, want_unversioned=want_unversioned)
107
101
def conflicts(self):
108
102
"""Get a list of the conflicts in the tree.
110
104
Each conflict is an instance of bzrlib.conflicts.Conflict.
112
return _mod_conflicts.ConflictList()
114
108
def extras(self):
115
109
"""For trees that can have unversioned files, return all such paths."""
118
112
def get_parent_ids(self):
119
"""Get the parent ids for this tree.
113
"""Get the parent ids for this tree.
121
115
:return: a list of parent ids. [] is returned to indicate
122
116
a tree with no parents.
123
117
:raises: BzrError if the parents are not known.
125
119
raise NotImplementedError(self.get_parent_ids)
127
121
def has_filename(self, filename):
128
122
"""True if the tree has given filename."""
129
raise NotImplementedError(self.has_filename)
123
raise NotImplementedError()
131
125
def has_id(self, file_id):
126
file_id = osutils.safe_file_id(file_id)
132
127
return self.inventory.has_id(file_id)
134
def __contains__(self, file_id):
135
return self.has_id(file_id)
129
__contains__ = has_id
137
131
def has_or_had_id(self, file_id):
132
file_id = osutils.safe_file_id(file_id)
138
133
if file_id == self.inventory.root.file_id:
140
135
return self.inventory.has_id(file_id)
344
235
raise NotImplementedError(self.get_symlink_target)
346
def get_canonical_inventory_paths(self, paths):
347
"""Like get_canonical_inventory_path() but works on multiple items.
349
:param paths: A sequence of paths relative to the root of the tree.
350
:return: A list of paths, with each item the corresponding input path
351
adjusted to account for existing elements that match case
354
return list(self._yield_canonical_inventory_paths(paths))
356
def get_canonical_inventory_path(self, path):
357
"""Returns the first inventory item that case-insensitively matches path.
359
If a path matches exactly, it is returned. If no path matches exactly
360
but more than one path matches case-insensitively, it is implementation
361
defined which is returned.
363
If no path matches case-insensitively, the input path is returned, but
364
with as many path entries that do exist changed to their canonical
367
If you need to resolve many names from the same tree, you should
368
use get_canonical_inventory_paths() to avoid O(N) behaviour.
370
:param path: A paths relative to the root of the tree.
371
:return: The input path adjusted to account for existing elements
372
that match case insensitively.
374
return self._yield_canonical_inventory_paths([path]).next()
376
def _yield_canonical_inventory_paths(self, paths):
378
# First, if the path as specified exists exactly, just use it.
379
if self.path2id(path) is not None:
383
cur_id = self.get_root_id()
385
bit_iter = iter(path.split("/"))
388
for child in self.iter_children(cur_id):
390
child_base = os.path.basename(self.id2path(child))
391
if child_base.lower() == lelt:
393
cur_path = osutils.pathjoin(cur_path, child_base)
396
# before a change is committed we can see this error...
399
# got to the end of this directory and no entries matched.
400
# Return what matched so far, plus the rest as specified.
401
cur_path = osutils.pathjoin(cur_path, elt, *list(bit_iter))
406
def get_root_id(self):
407
"""Return the file_id for the root of this tree."""
408
raise NotImplementedError(self.get_root_id)
410
def annotate_iter(self, file_id,
411
default_revision=_mod_revision.CURRENT_REVISION):
412
"""Return an iterator of revision_id, line tuples.
237
def annotate_iter(self, file_id):
238
"""Return an iterator of revision_id, line tuples
414
240
For working trees (and mutable trees in general), the special
415
241
revision_id 'current:' will be used for lines that are new in this
416
242
tree, e.g. uncommitted changes.
417
243
:param file_id: The file to produce an annotated version from
418
:param default_revision: For lines that don't match a basis, mark them
419
with this revision id. Not all implementations will make use of
422
245
raise NotImplementedError(self.annotate_iter)
424
def _get_plan_merge_data(self, file_id, other, base):
425
from bzrlib import merge, versionedfile
426
vf = versionedfile._PlanMergeVersionedFile(file_id)
427
last_revision_a = self._get_file_revision(file_id, vf, 'this:')
428
last_revision_b = other._get_file_revision(file_id, vf, 'other:')
430
last_revision_base = None
432
last_revision_base = base._get_file_revision(file_id, vf, 'base:')
433
return vf, last_revision_a, last_revision_b, last_revision_base
435
def plan_file_merge(self, file_id, other, base=None):
436
"""Generate a merge plan based on annotations.
438
If the file contains uncommitted changes in this tree, they will be
439
attributed to the 'current:' pseudo-revision. If the file contains
440
uncommitted changes in the other tree, they will be assigned to the
441
'other:' pseudo-revision.
443
data = self._get_plan_merge_data(file_id, other, base)
444
vf, last_revision_a, last_revision_b, last_revision_base = data
445
return vf.plan_merge(last_revision_a, last_revision_b,
448
def plan_file_lca_merge(self, file_id, other, base=None):
449
"""Generate a merge plan based lca-newness.
451
If the file contains uncommitted changes in this tree, they will be
452
attributed to the 'current:' pseudo-revision. If the file contains
453
uncommitted changes in the other tree, they will be assigned to the
454
'other:' pseudo-revision.
456
data = self._get_plan_merge_data(file_id, other, base)
457
vf, last_revision_a, last_revision_b, last_revision_base = data
458
return vf.plan_lca_merge(last_revision_a, last_revision_b,
461
def _iter_parent_trees(self):
462
"""Iterate through parent trees, defaulting to Tree.revision_tree."""
463
for revision_id in self.get_parent_ids():
465
yield self.revision_tree(revision_id)
466
except errors.NoSuchRevisionInTree:
467
yield self.repository.revision_tree(revision_id)
470
def _file_revision(revision_tree, file_id):
471
"""Determine the revision associated with a file in a given tree."""
472
revision_tree.lock_read()
474
return revision_tree.inventory[file_id].revision
476
revision_tree.unlock()
478
def _get_file_revision(self, file_id, vf, tree_revision):
479
"""Ensure that file_id, tree_revision is in vf to plan the merge."""
481
if getattr(self, '_repository', None) is None:
482
last_revision = tree_revision
483
parent_keys = [(file_id, self._file_revision(t, file_id)) for t in
484
self._iter_parent_trees()]
485
vf.add_lines((file_id, last_revision), parent_keys,
486
self.get_file(file_id).readlines())
487
repo = self.branch.repository
490
last_revision = self._file_revision(self, file_id)
491
base_vf = self._repository.texts
492
if base_vf not in vf.fallback_versionedfiles:
493
vf.fallback_versionedfiles.append(base_vf)
496
247
inventory = property(_get_inventory,
497
248
doc="Inventory of this Tree")
499
250
def _check_retrieved(self, ie, f):
500
251
if not __debug__:
502
253
fp = fingerprint_file(f)
505
256
if ie.text_size is not None:
506
257
if ie.text_size != fp['size']:
507
258
raise BzrError("mismatched size for file %r in %r" % (ie.file_id, self._store),
617
369
raise NotImplementedError(self.walkdirs)
619
def supports_content_filtering(self):
372
class EmptyTree(Tree):
375
self._inventory = Inventory(root_id=None)
376
symbol_versioning.warn('EmptyTree is deprecated as of bzr 0.9 please'
377
' use repository.revision_tree instead.',
378
DeprecationWarning, stacklevel=2)
380
def get_parent_ids(self):
383
def get_symlink_target(self, file_id):
386
def has_filename(self, filename):
622
def _content_filter_stack(self, path=None, file_id=None):
623
"""The stack of content filters for a path if filtering is supported.
625
Readers will be applied in first-to-last order.
626
Writers will be applied in last-to-first order.
627
Either the path or the file-id needs to be provided.
629
:param path: path relative to the root of the tree
631
:param file_id: file_id or None if unknown
632
:return: the list of filters - [] if there are none
634
filter_pref_names = filters._get_registered_names()
635
if len(filter_pref_names) == 0:
638
path = self.id2path(file_id)
639
prefs = self.iter_search_rules([path], filter_pref_names).next()
640
stk = filters._get_filter_stack_for(prefs)
641
if 'filters' in debug.debug_flags:
642
note("*** %s content-filter: %s => %r" % (path,prefs,stk))
645
def _content_filter_stack_provider(self):
646
"""A function that returns a stack of ContentFilters.
648
The function takes a path (relative to the top of the tree) and a
649
file-id as parameters.
651
:return: None if content filtering is not supported by this tree.
653
if self.supports_content_filtering():
654
return lambda path, file_id: \
655
self._content_filter_stack(path, file_id)
659
def iter_search_rules(self, path_names, pref_names=None,
660
_default_searcher=rules._per_user_searcher):
661
"""Find the preferences for filenames in a tree.
663
:param path_names: an iterable of paths to find attributes for.
664
Paths are given relative to the root of the tree.
665
:param pref_names: the list of preferences to lookup - None for all
666
:param _default_searcher: private parameter to assist testing - don't use
667
:return: an iterator of tuple sequences, one per path-name.
668
See _RulesSearcher.get_items for details on the tuple sequence.
670
searcher = self._get_rules_searcher(_default_searcher)
671
if searcher is not None:
672
if pref_names is not None:
673
for path in path_names:
674
yield searcher.get_selected_items(path, pref_names)
676
for path in path_names:
677
yield searcher.get_items(path)
680
def _get_rules_searcher(self, default_searcher):
681
"""Get the RulesSearcher for this tree given the default one."""
682
searcher = default_searcher
389
def kind(self, file_id):
390
file_id = osutils.safe_file_id(file_id)
391
assert self._inventory[file_id].kind == "directory"
394
def list_files(self, include_root=False):
397
def __contains__(self, file_id):
398
file_id = osutils.safe_file_id(file_id)
399
return (file_id in self._inventory)
401
def get_file_sha1(self, file_id, path=None, stat_value=None):
686
405
######################################################################
1039
756
self.source._comparison_data(from_entry, path)
1040
757
kind = (from_kind, None)
1041
758
executable = (from_executable, None)
1042
changed_content = from_kind is not None
759
changed_content = True
1043
760
# the parent's path is necessarily known at this point.
1044
761
yield(file_id, (path, to_path), changed_content, versioned, parent,
1045
762
name, kind, executable)
1048
class MultiWalker(object):
1049
"""Walk multiple trees simultaneously, getting combined results."""
1051
# Note: This could be written to not assume you can do out-of-order
1052
# lookups. Instead any nodes that don't match in all trees could be
1053
# marked as 'deferred', and then returned in the final cleanup loop.
1054
# For now, I think it is "nicer" to return things as close to the
1055
# "master_tree" order as we can.
1057
def __init__(self, master_tree, other_trees):
1058
"""Create a new MultiWalker.
1060
All trees being walked must implement "iter_entries_by_dir()", such
1061
that they yield (path, object) tuples, where that object will have a
1062
'.file_id' member, that can be used to check equality.
1064
:param master_tree: All trees will be 'slaved' to the master_tree such
1065
that nodes in master_tree will be used as 'first-pass' sync points.
1066
Any nodes that aren't in master_tree will be merged in a second
1068
:param other_trees: A list of other trees to walk simultaneously.
1070
self._master_tree = master_tree
1071
self._other_trees = other_trees
1073
# Keep track of any nodes that were properly processed just out of
1074
# order, that way we don't return them at the end, we don't have to
1075
# track *all* processed file_ids, just the out-of-order ones
1076
self._out_of_order_processed = set()
1079
def _step_one(iterator):
1080
"""Step an iter_entries_by_dir iterator.
1082
:return: (has_more, path, ie)
1083
If has_more is False, path and ie will be None.
1086
path, ie = iterator.next()
1087
except StopIteration:
1088
return False, None, None
1090
return True, path, ie
1093
def _cmp_path_by_dirblock(path1, path2):
1094
"""Compare two paths based on what directory they are in.
1096
This generates a sort order, such that all children of a directory are
1097
sorted together, and grandchildren are in the same order as the
1098
children appear. But all grandchildren come after all children.
1100
:param path1: first path
1101
:param path2: the second path
1102
:return: negative number if ``path1`` comes first,
1103
0 if paths are equal
1104
and a positive number if ``path2`` sorts first
1106
# Shortcut this special case
1109
# This is stolen from _dirstate_helpers_py.py, only switching it to
1110
# Unicode objects. Consider using encode_utf8() and then using the
1111
# optimized versions, or maybe writing optimized unicode versions.
1112
if not isinstance(path1, unicode):
1113
raise TypeError("'path1' must be a unicode string, not %s: %r"
1114
% (type(path1), path1))
1115
if not isinstance(path2, unicode):
1116
raise TypeError("'path2' must be a unicode string, not %s: %r"
1117
% (type(path2), path2))
1118
return cmp(MultiWalker._path_to_key(path1),
1119
MultiWalker._path_to_key(path2))
1122
def _path_to_key(path):
1123
dirname, basename = osutils.split(path)
1124
return (dirname.split(u'/'), basename)
1126
def _lookup_by_file_id(self, extra_entries, other_tree, file_id):
1127
"""Lookup an inventory entry by file_id.
1129
This is called when an entry is missing in the normal order.
1130
Generally this is because a file was either renamed, or it was
1131
deleted/added. If the entry was found in the inventory and not in
1132
extra_entries, it will be added to self._out_of_order_processed
1134
:param extra_entries: A dictionary of {file_id: (path, ie)}. This
1135
should be filled with entries that were found before they were
1136
used. If file_id is present, it will be removed from the
1138
:param other_tree: The Tree to search, in case we didn't find the entry
1140
:param file_id: The file_id to look for
1141
:return: (path, ie) if found or (None, None) if not present.
1143
if file_id in extra_entries:
1144
return extra_entries.pop(file_id)
1145
# TODO: Is id2path better as the first call, or is
1146
# inventory[file_id] better as a first check?
1148
cur_path = other_tree.id2path(file_id)
1149
except errors.NoSuchId:
1151
if cur_path is None:
1154
self._out_of_order_processed.add(file_id)
1155
cur_ie = other_tree.inventory[file_id]
1156
return (cur_path, cur_ie)
1159
"""Match up the values in the different trees."""
1160
for result in self._walk_master_tree():
1162
self._finish_others()
1163
for result in self._walk_others():
1166
def _walk_master_tree(self):
1167
"""First pass, walk all trees in lock-step.
1169
When we are done, all nodes in the master_tree will have been
1170
processed. _other_walkers, _other_entries, and _others_extra will be
1171
set on 'self' for future processing.
1173
# This iterator has the most "inlining" done, because it tends to touch
1174
# every file in the tree, while the others only hit nodes that don't
1176
master_iterator = self._master_tree.iter_entries_by_dir()
1178
other_walkers = [other.iter_entries_by_dir()
1179
for other in self._other_trees]
1180
other_entries = [self._step_one(walker) for walker in other_walkers]
1181
# Track extra nodes in the other trees
1182
others_extra = [{} for i in xrange(len(self._other_trees))]
1184
master_has_more = True
1185
step_one = self._step_one
1186
lookup_by_file_id = self._lookup_by_file_id
1187
out_of_order_processed = self._out_of_order_processed
1189
while master_has_more:
1190
(master_has_more, path, master_ie) = step_one(master_iterator)
1191
if not master_has_more:
1194
file_id = master_ie.file_id
1196
other_values_append = other_values.append
1197
next_other_entries = []
1198
next_other_entries_append = next_other_entries.append
1199
for idx, (other_has_more, other_path, other_ie) in enumerate(other_entries):
1200
if not other_has_more:
1201
other_values_append(lookup_by_file_id(
1202
others_extra[idx], self._other_trees[idx], file_id))
1203
next_other_entries_append((False, None, None))
1204
elif file_id == other_ie.file_id:
1205
# This is the critical code path, as most of the entries
1206
# should match between most trees.
1207
other_values_append((other_path, other_ie))
1208
next_other_entries_append(step_one(other_walkers[idx]))
1210
# This walker did not match, step it until it either
1211
# matches, or we know we are past the current walker.
1212
other_walker = other_walkers[idx]
1213
other_extra = others_extra[idx]
1214
while (other_has_more and
1215
self._cmp_path_by_dirblock(other_path, path) < 0):
1216
other_file_id = other_ie.file_id
1217
if other_file_id not in out_of_order_processed:
1218
other_extra[other_file_id] = (other_path, other_ie)
1219
other_has_more, other_path, other_ie = \
1220
step_one(other_walker)
1221
if other_has_more and other_ie.file_id == file_id:
1222
# We ended up walking to this point, match and step
1224
other_values_append((other_path, other_ie))
1225
other_has_more, other_path, other_ie = \
1226
step_one(other_walker)
1228
# This record isn't in the normal order, see if it
1230
other_values_append(lookup_by_file_id(
1231
other_extra, self._other_trees[idx], file_id))
1232
next_other_entries_append((other_has_more, other_path,
1234
other_entries = next_other_entries
1236
# We've matched all the walkers, yield this datapoint
1237
yield path, file_id, master_ie, other_values
1238
self._other_walkers = other_walkers
1239
self._other_entries = other_entries
1240
self._others_extra = others_extra
1242
def _finish_others(self):
1243
"""Finish walking the other iterators, so we get all entries."""
1244
for idx, info in enumerate(self._other_entries):
1245
other_extra = self._others_extra[idx]
1246
(other_has_more, other_path, other_ie) = info
1247
while other_has_more:
1248
other_file_id = other_ie.file_id
1249
if other_file_id not in self._out_of_order_processed:
1250
other_extra[other_file_id] = (other_path, other_ie)
1251
other_has_more, other_path, other_ie = \
1252
self._step_one(self._other_walkers[idx])
1253
del self._other_entries
1255
def _walk_others(self):
1256
"""Finish up by walking all the 'deferred' nodes."""
1257
# TODO: One alternative would be to grab all possible unprocessed
1258
# file_ids, and then sort by path, and then yield them. That
1259
# might ensure better ordering, in case a caller strictly
1260
# requires parents before children.
1261
for idx, other_extra in enumerate(self._others_extra):
1262
others = sorted(other_extra.itervalues(),
1263
key=lambda x: self._path_to_key(x[0]))
1264
for other_path, other_ie in others:
1265
file_id = other_ie.file_id
1266
# We don't need to check out_of_order_processed here, because
1267
# the lookup_by_file_id will be removing anything processed
1268
# from the extras cache
1269
other_extra.pop(file_id)
1270
other_values = [(None, None) for i in xrange(idx)]
1271
other_values.append((other_path, other_ie))
1272
for alt_idx, alt_extra in enumerate(self._others_extra[idx+1:]):
1273
alt_idx = alt_idx + idx + 1
1274
alt_extra = self._others_extra[alt_idx]
1275
alt_tree = self._other_trees[alt_idx]
1276
other_values.append(self._lookup_by_file_id(
1277
alt_extra, alt_tree, file_id))
1278
yield other_path, file_id, None, other_values
765
# This was deprecated before 0.12, but did not have an official warning
766
@symbol_versioning.deprecated_function(symbol_versioning.zero_twelve)
767
def RevisionTree(*args, **kwargs):
768
"""RevisionTree has moved to bzrlib.revisiontree.RevisionTree()
770
Accessing it as bzrlib.tree.RevisionTree has been deprecated as of
773
from bzrlib.revisiontree import RevisionTree as _RevisionTree
774
return _RevisionTree(*args, **kwargs)