97
91
want_unversioned=want_unversioned,
100
def iter_changes(self, from_tree, include_unchanged=False,
94
def _iter_changes(self, from_tree, include_unchanged=False,
101
95
specific_files=None, pb=None, extra_trees=None,
102
96
require_versioned=True, want_unversioned=False):
103
97
intertree = InterTree.get(from_tree, self)
104
return intertree.iter_changes(include_unchanged, specific_files, pb,
98
return intertree._iter_changes(include_unchanged, specific_files, pb,
105
99
extra_trees, require_versioned, want_unversioned=want_unversioned)
107
101
def conflicts(self):
108
102
"""Get a list of the conflicts in the tree.
110
104
Each conflict is an instance of bzrlib.conflicts.Conflict.
112
return _mod_conflicts.ConflictList()
114
108
def extras(self):
115
109
"""For trees that can have unversioned files, return all such paths."""
118
112
def get_parent_ids(self):
119
"""Get the parent ids for this tree.
113
"""Get the parent ids for this tree.
121
115
:return: a list of parent ids. [] is returned to indicate
122
116
a tree with no parents.
123
117
:raises: BzrError if the parents are not known.
125
119
raise NotImplementedError(self.get_parent_ids)
127
121
def has_filename(self, filename):
128
122
"""True if the tree has given filename."""
129
raise NotImplementedError(self.has_filename)
123
raise NotImplementedError()
131
125
def has_id(self, file_id):
126
file_id = osutils.safe_file_id(file_id)
132
127
return self.inventory.has_id(file_id)
134
def __contains__(self, file_id):
135
return self.has_id(file_id)
129
__contains__ = has_id
137
131
def has_or_had_id(self, file_id):
132
file_id = osutils.safe_file_id(file_id)
138
133
if file_id == self.inventory.root.file_id:
140
135
return self.inventory.has_id(file_id)
344
235
raise NotImplementedError(self.get_symlink_target)
346
def get_canonical_inventory_paths(self, paths):
347
"""Like get_canonical_inventory_path() but works on multiple items.
349
:param paths: A sequence of paths relative to the root of the tree.
350
:return: A list of paths, with each item the corresponding input path
351
adjusted to account for existing elements that match case
354
return list(self._yield_canonical_inventory_paths(paths))
356
def get_canonical_inventory_path(self, path):
357
"""Returns the first inventory item that case-insensitively matches path.
359
If a path matches exactly, it is returned. If no path matches exactly
360
but more than one path matches case-insensitively, it is implementation
361
defined which is returned.
363
If no path matches case-insensitively, the input path is returned, but
364
with as many path entries that do exist changed to their canonical
367
If you need to resolve many names from the same tree, you should
368
use get_canonical_inventory_paths() to avoid O(N) behaviour.
370
:param path: A paths relative to the root of the tree.
371
:return: The input path adjusted to account for existing elements
372
that match case insensitively.
374
return self._yield_canonical_inventory_paths([path]).next()
376
def _yield_canonical_inventory_paths(self, paths):
378
# First, if the path as specified exists exactly, just use it.
379
if self.path2id(path) is not None:
383
cur_id = self.get_root_id()
385
bit_iter = iter(path.split("/"))
388
for child in self.iter_children(cur_id):
390
child_base = os.path.basename(self.id2path(child))
391
if child_base.lower() == lelt:
393
cur_path = osutils.pathjoin(cur_path, child_base)
396
# before a change is committed we can see this error...
399
# got to the end of this directory and no entries matched.
400
# Return what matched so far, plus the rest as specified.
401
cur_path = osutils.pathjoin(cur_path, elt, *list(bit_iter))
406
def get_root_id(self):
407
"""Return the file_id for the root of this tree."""
408
raise NotImplementedError(self.get_root_id)
410
def annotate_iter(self, file_id,
411
default_revision=_mod_revision.CURRENT_REVISION):
412
"""Return an iterator of revision_id, line tuples.
237
def annotate_iter(self, file_id):
238
"""Return an iterator of revision_id, line tuples
414
240
For working trees (and mutable trees in general), the special
415
241
revision_id 'current:' will be used for lines that are new in this
416
242
tree, e.g. uncommitted changes.
417
243
:param file_id: The file to produce an annotated version from
418
:param default_revision: For lines that don't match a basis, mark them
419
with this revision id. Not all implementations will make use of
422
245
raise NotImplementedError(self.annotate_iter)
424
def _get_plan_merge_data(self, file_id, other, base):
425
from bzrlib import merge, versionedfile
426
vf = versionedfile._PlanMergeVersionedFile(file_id)
427
last_revision_a = self._get_file_revision(file_id, vf, 'this:')
428
last_revision_b = other._get_file_revision(file_id, vf, 'other:')
430
last_revision_base = None
432
last_revision_base = base._get_file_revision(file_id, vf, 'base:')
433
return vf, last_revision_a, last_revision_b, last_revision_base
435
def plan_file_merge(self, file_id, other, base=None):
436
"""Generate a merge plan based on annotations.
438
If the file contains uncommitted changes in this tree, they will be
439
attributed to the 'current:' pseudo-revision. If the file contains
440
uncommitted changes in the other tree, they will be assigned to the
441
'other:' pseudo-revision.
443
data = self._get_plan_merge_data(file_id, other, base)
444
vf, last_revision_a, last_revision_b, last_revision_base = data
445
return vf.plan_merge(last_revision_a, last_revision_b,
448
def plan_file_lca_merge(self, file_id, other, base=None):
449
"""Generate a merge plan based lca-newness.
451
If the file contains uncommitted changes in this tree, they will be
452
attributed to the 'current:' pseudo-revision. If the file contains
453
uncommitted changes in the other tree, they will be assigned to the
454
'other:' pseudo-revision.
456
data = self._get_plan_merge_data(file_id, other, base)
457
vf, last_revision_a, last_revision_b, last_revision_base = data
458
return vf.plan_lca_merge(last_revision_a, last_revision_b,
461
def _iter_parent_trees(self):
462
"""Iterate through parent trees, defaulting to Tree.revision_tree."""
463
for revision_id in self.get_parent_ids():
465
yield self.revision_tree(revision_id)
466
except errors.NoSuchRevisionInTree:
467
yield self.repository.revision_tree(revision_id)
470
def _file_revision(revision_tree, file_id):
471
"""Determine the revision associated with a file in a given tree."""
472
revision_tree.lock_read()
474
return revision_tree.inventory[file_id].revision
476
revision_tree.unlock()
478
def _get_file_revision(self, file_id, vf, tree_revision):
479
"""Ensure that file_id, tree_revision is in vf to plan the merge."""
481
if getattr(self, '_repository', None) is None:
482
last_revision = tree_revision
483
parent_keys = [(file_id, self._file_revision(t, file_id)) for t in
484
self._iter_parent_trees()]
485
vf.add_lines((file_id, last_revision), parent_keys,
486
self.get_file(file_id).readlines())
487
repo = self.branch.repository
490
last_revision = self._file_revision(self, file_id)
491
base_vf = self._repository.texts
492
if base_vf not in vf.fallback_versionedfiles:
493
vf.fallback_versionedfiles.append(base_vf)
496
247
inventory = property(_get_inventory,
497
248
doc="Inventory of this Tree")
499
250
def _check_retrieved(self, ie, f):
500
251
if not __debug__:
502
253
fp = fingerprint_file(f)
505
256
if ie.text_size is not None:
506
257
if ie.text_size != fp['size']:
507
258
raise BzrError("mismatched size for file %r in %r" % (ie.file_id, self._store),
617
369
raise NotImplementedError(self.walkdirs)
619
def supports_content_filtering(self):
372
class EmptyTree(Tree):
375
self._inventory = Inventory(root_id=None)
376
symbol_versioning.warn('EmptyTree is deprecated as of bzr 0.9 please'
377
' use repository.revision_tree instead.',
378
DeprecationWarning, stacklevel=2)
380
def get_parent_ids(self):
383
def get_symlink_target(self, file_id):
386
def has_filename(self, filename):
622
def _content_filter_stack(self, path=None, file_id=None):
623
"""The stack of content filters for a path if filtering is supported.
625
Readers will be applied in first-to-last order.
626
Writers will be applied in last-to-first order.
627
Either the path or the file-id needs to be provided.
629
:param path: path relative to the root of the tree
631
:param file_id: file_id or None if unknown
632
:return: the list of filters - [] if there are none
634
filter_pref_names = filters._get_registered_names()
635
if len(filter_pref_names) == 0:
638
path = self.id2path(file_id)
639
prefs = self.iter_search_rules([path], filter_pref_names).next()
640
stk = filters._get_filter_stack_for(prefs)
641
if 'filters' in debug.debug_flags:
642
note("*** %s content-filter: %s => %r" % (path,prefs,stk))
645
def _content_filter_stack_provider(self):
646
"""A function that returns a stack of ContentFilters.
648
The function takes a path (relative to the top of the tree) and a
649
file-id as parameters.
651
:return: None if content filtering is not supported by this tree.
653
if self.supports_content_filtering():
654
return lambda path, file_id: \
655
self._content_filter_stack(path, file_id)
659
def iter_search_rules(self, path_names, pref_names=None,
660
_default_searcher=None):
661
"""Find the preferences for filenames in a tree.
663
:param path_names: an iterable of paths to find attributes for.
664
Paths are given relative to the root of the tree.
665
:param pref_names: the list of preferences to lookup - None for all
666
:param _default_searcher: private parameter to assist testing - don't use
667
:return: an iterator of tuple sequences, one per path-name.
668
See _RulesSearcher.get_items for details on the tuple sequence.
670
if _default_searcher is None:
671
_default_searcher = rules._per_user_searcher
672
searcher = self._get_rules_searcher(_default_searcher)
673
if searcher is not None:
674
if pref_names is not None:
675
for path in path_names:
676
yield searcher.get_selected_items(path, pref_names)
678
for path in path_names:
679
yield searcher.get_items(path)
682
def _get_rules_searcher(self, default_searcher):
683
"""Get the RulesSearcher for this tree given the default one."""
684
searcher = default_searcher
389
def kind(self, file_id):
390
file_id = osutils.safe_file_id(file_id)
391
assert self._inventory[file_id].kind == "directory"
394
def list_files(self, include_root=False):
397
def __contains__(self, file_id):
398
file_id = osutils.safe_file_id(file_id)
399
return (file_id in self._inventory)
401
def get_file_sha1(self, file_id, path=None, stat_value=None):
688
405
######################################################################
1041
756
self.source._comparison_data(from_entry, path)
1042
757
kind = (from_kind, None)
1043
758
executable = (from_executable, None)
1044
changed_content = from_kind is not None
759
changed_content = True
1045
760
# the parent's path is necessarily known at this point.
1046
761
yield(file_id, (path, to_path), changed_content, versioned, parent,
1047
762
name, kind, executable)
1050
class MultiWalker(object):
1051
"""Walk multiple trees simultaneously, getting combined results."""
1053
# Note: This could be written to not assume you can do out-of-order
1054
# lookups. Instead any nodes that don't match in all trees could be
1055
# marked as 'deferred', and then returned in the final cleanup loop.
1056
# For now, I think it is "nicer" to return things as close to the
1057
# "master_tree" order as we can.
1059
def __init__(self, master_tree, other_trees):
1060
"""Create a new MultiWalker.
1062
All trees being walked must implement "iter_entries_by_dir()", such
1063
that they yield (path, object) tuples, where that object will have a
1064
'.file_id' member, that can be used to check equality.
1066
:param master_tree: All trees will be 'slaved' to the master_tree such
1067
that nodes in master_tree will be used as 'first-pass' sync points.
1068
Any nodes that aren't in master_tree will be merged in a second
1070
:param other_trees: A list of other trees to walk simultaneously.
1072
self._master_tree = master_tree
1073
self._other_trees = other_trees
1075
# Keep track of any nodes that were properly processed just out of
1076
# order, that way we don't return them at the end, we don't have to
1077
# track *all* processed file_ids, just the out-of-order ones
1078
self._out_of_order_processed = set()
1081
def _step_one(iterator):
1082
"""Step an iter_entries_by_dir iterator.
1084
:return: (has_more, path, ie)
1085
If has_more is False, path and ie will be None.
1088
path, ie = iterator.next()
1089
except StopIteration:
1090
return False, None, None
1092
return True, path, ie
1095
def _cmp_path_by_dirblock(path1, path2):
1096
"""Compare two paths based on what directory they are in.
1098
This generates a sort order, such that all children of a directory are
1099
sorted together, and grandchildren are in the same order as the
1100
children appear. But all grandchildren come after all children.
1102
:param path1: first path
1103
:param path2: the second path
1104
:return: negative number if ``path1`` comes first,
1105
0 if paths are equal
1106
and a positive number if ``path2`` sorts first
1108
# Shortcut this special case
1111
# This is stolen from _dirstate_helpers_py.py, only switching it to
1112
# Unicode objects. Consider using encode_utf8() and then using the
1113
# optimized versions, or maybe writing optimized unicode versions.
1114
if not isinstance(path1, unicode):
1115
raise TypeError("'path1' must be a unicode string, not %s: %r"
1116
% (type(path1), path1))
1117
if not isinstance(path2, unicode):
1118
raise TypeError("'path2' must be a unicode string, not %s: %r"
1119
% (type(path2), path2))
1120
return cmp(MultiWalker._path_to_key(path1),
1121
MultiWalker._path_to_key(path2))
1124
def _path_to_key(path):
1125
dirname, basename = osutils.split(path)
1126
return (dirname.split(u'/'), basename)
1128
def _lookup_by_file_id(self, extra_entries, other_tree, file_id):
1129
"""Lookup an inventory entry by file_id.
1131
This is called when an entry is missing in the normal order.
1132
Generally this is because a file was either renamed, or it was
1133
deleted/added. If the entry was found in the inventory and not in
1134
extra_entries, it will be added to self._out_of_order_processed
1136
:param extra_entries: A dictionary of {file_id: (path, ie)}. This
1137
should be filled with entries that were found before they were
1138
used. If file_id is present, it will be removed from the
1140
:param other_tree: The Tree to search, in case we didn't find the entry
1142
:param file_id: The file_id to look for
1143
:return: (path, ie) if found or (None, None) if not present.
1145
if file_id in extra_entries:
1146
return extra_entries.pop(file_id)
1147
# TODO: Is id2path better as the first call, or is
1148
# inventory[file_id] better as a first check?
1150
cur_path = other_tree.id2path(file_id)
1151
except errors.NoSuchId:
1153
if cur_path is None:
1156
self._out_of_order_processed.add(file_id)
1157
cur_ie = other_tree.inventory[file_id]
1158
return (cur_path, cur_ie)
1161
"""Match up the values in the different trees."""
1162
for result in self._walk_master_tree():
1164
self._finish_others()
1165
for result in self._walk_others():
1168
def _walk_master_tree(self):
1169
"""First pass, walk all trees in lock-step.
1171
When we are done, all nodes in the master_tree will have been
1172
processed. _other_walkers, _other_entries, and _others_extra will be
1173
set on 'self' for future processing.
1175
# This iterator has the most "inlining" done, because it tends to touch
1176
# every file in the tree, while the others only hit nodes that don't
1178
master_iterator = self._master_tree.iter_entries_by_dir()
1180
other_walkers = [other.iter_entries_by_dir()
1181
for other in self._other_trees]
1182
other_entries = [self._step_one(walker) for walker in other_walkers]
1183
# Track extra nodes in the other trees
1184
others_extra = [{} for i in xrange(len(self._other_trees))]
1186
master_has_more = True
1187
step_one = self._step_one
1188
lookup_by_file_id = self._lookup_by_file_id
1189
out_of_order_processed = self._out_of_order_processed
1191
while master_has_more:
1192
(master_has_more, path, master_ie) = step_one(master_iterator)
1193
if not master_has_more:
1196
file_id = master_ie.file_id
1198
other_values_append = other_values.append
1199
next_other_entries = []
1200
next_other_entries_append = next_other_entries.append
1201
for idx, (other_has_more, other_path, other_ie) in enumerate(other_entries):
1202
if not other_has_more:
1203
other_values_append(lookup_by_file_id(
1204
others_extra[idx], self._other_trees[idx], file_id))
1205
next_other_entries_append((False, None, None))
1206
elif file_id == other_ie.file_id:
1207
# This is the critical code path, as most of the entries
1208
# should match between most trees.
1209
other_values_append((other_path, other_ie))
1210
next_other_entries_append(step_one(other_walkers[idx]))
1212
# This walker did not match, step it until it either
1213
# matches, or we know we are past the current walker.
1214
other_walker = other_walkers[idx]
1215
other_extra = others_extra[idx]
1216
while (other_has_more and
1217
self._cmp_path_by_dirblock(other_path, path) < 0):
1218
other_file_id = other_ie.file_id
1219
if other_file_id not in out_of_order_processed:
1220
other_extra[other_file_id] = (other_path, other_ie)
1221
other_has_more, other_path, other_ie = \
1222
step_one(other_walker)
1223
if other_has_more and other_ie.file_id == file_id:
1224
# We ended up walking to this point, match and step
1226
other_values_append((other_path, other_ie))
1227
other_has_more, other_path, other_ie = \
1228
step_one(other_walker)
1230
# This record isn't in the normal order, see if it
1232
other_values_append(lookup_by_file_id(
1233
other_extra, self._other_trees[idx], file_id))
1234
next_other_entries_append((other_has_more, other_path,
1236
other_entries = next_other_entries
1238
# We've matched all the walkers, yield this datapoint
1239
yield path, file_id, master_ie, other_values
1240
self._other_walkers = other_walkers
1241
self._other_entries = other_entries
1242
self._others_extra = others_extra
1244
def _finish_others(self):
1245
"""Finish walking the other iterators, so we get all entries."""
1246
for idx, info in enumerate(self._other_entries):
1247
other_extra = self._others_extra[idx]
1248
(other_has_more, other_path, other_ie) = info
1249
while other_has_more:
1250
other_file_id = other_ie.file_id
1251
if other_file_id not in self._out_of_order_processed:
1252
other_extra[other_file_id] = (other_path, other_ie)
1253
other_has_more, other_path, other_ie = \
1254
self._step_one(self._other_walkers[idx])
1255
del self._other_entries
1257
def _walk_others(self):
1258
"""Finish up by walking all the 'deferred' nodes."""
1259
# TODO: One alternative would be to grab all possible unprocessed
1260
# file_ids, and then sort by path, and then yield them. That
1261
# might ensure better ordering, in case a caller strictly
1262
# requires parents before children.
1263
for idx, other_extra in enumerate(self._others_extra):
1264
others = sorted(other_extra.itervalues(),
1265
key=lambda x: self._path_to_key(x[0]))
1266
for other_path, other_ie in others:
1267
file_id = other_ie.file_id
1268
# We don't need to check out_of_order_processed here, because
1269
# the lookup_by_file_id will be removing anything processed
1270
# from the extras cache
1271
other_extra.pop(file_id)
1272
other_values = [(None, None) for i in xrange(idx)]
1273
other_values.append((other_path, other_ie))
1274
for alt_idx, alt_extra in enumerate(self._others_extra[idx+1:]):
1275
alt_idx = alt_idx + idx + 1
1276
alt_extra = self._others_extra[alt_idx]
1277
alt_tree = self._other_trees[alt_idx]
1278
other_values.append(self._lookup_by_file_id(
1279
alt_extra, alt_tree, file_id))
1280
yield other_path, file_id, None, other_values
765
# This was deprecated before 0.12, but did not have an official warning
766
@symbol_versioning.deprecated_function(symbol_versioning.zero_twelve)
767
def RevisionTree(*args, **kwargs):
768
"""RevisionTree has moved to bzrlib.revisiontree.RevisionTree()
770
Accessing it as bzrlib.tree.RevisionTree has been deprecated as of
773
from bzrlib.revisiontree import RevisionTree as _RevisionTree
774
return _RevisionTree(*args, **kwargs)