68
82
sha-1 of the text of the file
71
85
size in bytes of the text of the file
73
87
(reading a version 4 tree created a text_id field.)
75
89
>>> i = Inventory()
78
92
>>> i.add(InventoryDirectory('123', 'src', ROOT_ID))
79
InventoryDirectory('123', 'src', parent_id='TREE_ROOT')
93
InventoryDirectory('123', 'src', parent_id='TREE_ROOT', revision=None)
80
94
>>> i.add(InventoryFile('2323', 'hello.c', parent_id='123'))
81
InventoryFile('2323', 'hello.c', parent_id='123')
82
>>> for j in i.iter_entries():
85
('src', InventoryDirectory('123', 'src', parent_id='TREE_ROOT'))
86
('src/hello.c', InventoryFile('2323', 'hello.c', parent_id='123'))
87
>>> i.add(InventoryFile('2323', 'bye.c', '123'))
88
Traceback (most recent call last):
95
InventoryFile('2323', 'hello.c', parent_id='123', sha1=None, len=None, revision=None)
96
>>> shouldbe = {0: '', 1: 'src', 2: 'src/hello.c'}
97
>>> for ix, j in enumerate(i.iter_entries()):
98
... print (j[0] == shouldbe[ix], j[1])
90
BzrError: inventory already contains entry with id {2323}
100
(True, InventoryDirectory('TREE_ROOT', u'', parent_id=None, revision=None))
101
(True, InventoryDirectory('123', 'src', parent_id='TREE_ROOT', revision=None))
102
(True, InventoryFile('2323', 'hello.c', parent_id='123', sha1=None, len=None, revision=None))
91
103
>>> i.add(InventoryFile('2324', 'bye.c', '123'))
92
InventoryFile('2324', 'bye.c', parent_id='123')
104
InventoryFile('2324', 'bye.c', parent_id='123', sha1=None, len=None, revision=None)
93
105
>>> i.add(InventoryDirectory('2325', 'wibble', '123'))
94
InventoryDirectory('2325', 'wibble', parent_id='123')
106
InventoryDirectory('2325', 'wibble', parent_id='123', revision=None)
95
107
>>> i.path2id('src/wibble')
99
109
>>> i.add(InventoryFile('2326', 'wibble.c', '2325'))
100
InventoryFile('2326', 'wibble.c', parent_id='2325')
110
InventoryFile('2326', 'wibble.c', parent_id='2325', sha1=None, len=None, revision=None)
102
InventoryFile('2326', 'wibble.c', parent_id='2325')
112
InventoryFile('2326', 'wibble.c', parent_id='2325', sha1=None, len=None, revision=None)
103
113
>>> for path, entry in i.iter_entries():
104
... print path.replace('\\\\', '/') # for win32 os.sep
105
... assert i.path2id(path)
111
121
src/wibble/wibble.c
112
>>> i.id2path('2326').replace('\\\\', '/')
122
>>> i.id2path('2326')
113
123
'src/wibble/wibble.c'
116
__slots__ = ['text_sha1', 'text_size', 'file_id', 'name', 'kind',
117
'text_id', 'parent_id', 'children', 'executable',
120
def _add_text_to_weave(self, new_lines, parents, weave_store, transaction):
121
weave_store.add_text(self.file_id, self.revision, new_lines, parents,
126
# Constants returned by describe_change()
128
# TODO: These should probably move to some kind of FileChangeDescription
129
# class; that's like what's inside a TreeDelta but we want to be able to
130
# generate them just for one file at a time.
132
MODIFIED_AND_RENAMED = 'modified and renamed'
134
__slots__ = ['file_id', 'revision', 'parent_id', 'name']
136
# Attributes that all InventoryEntry instances are expected to have, but
137
# that don't vary for all kinds of entry. (e.g. symlink_target is only
138
# relevant to InventoryLink, so there's no reason to make every
139
# InventoryFile instance allocate space to hold a value for it.)
140
# Attributes that only vary for files: executable, text_sha1, text_size,
146
# Attributes that only vary for symlinks: symlink_target
147
symlink_target = None
148
# Attributes that only vary for tree-references: reference_revision
149
reference_revision = None
124
152
def detect_changes(self, old_entry):
125
153
"""Return a (text_modified, meta_modified) from this to old_entry.
127
_read_tree_state must have been called on self and old_entry prior to
155
_read_tree_state must have been called on self and old_entry prior to
128
156
calling detect_changes.
130
158
return False, False
132
def diff(self, text_diff, from_label, tree, to_label, to_entry, to_tree,
133
output_to, reverse=False):
134
"""Perform a diff from this to to_entry.
136
text_diff will be used for textual difference calculation.
137
This is a template method, override _diff in child classes.
139
self._read_tree_state(tree.id2path(self.file_id), tree)
141
# cannot diff from one kind to another - you must do a removal
142
# and an addif they do not match.
143
assert self.kind == to_entry.kind
144
to_entry._read_tree_state(to_tree.id2path(to_entry.file_id),
146
self._diff(text_diff, from_label, tree, to_label, to_entry, to_tree,
149
160
def _diff(self, text_diff, from_label, tree, to_label, to_entry, to_tree,
150
161
output_to, reverse=False):
151
162
"""Perform a diff between two entries of the same kind."""
153
def find_previous_heads(self, previous_inventories, entry_weave):
154
"""Return the revisions and entries that directly preceed this.
156
Returned as a map from revision to inventory entry.
158
This is a map containing the file revisions in all parents
159
for which the file exists, and its revision is not a parent of
160
any other. If the file is new, the set will be empty.
164
def parent_candidates(self, previous_inventories):
165
"""Find possible per-file graph parents.
167
This is currently defined by:
168
- Select the last changed revision in the parent inventory.
169
- Do deal with a short lived bug in bzr 0.8's development two entries
170
that have the same last changed but different 'x' bit settings are
162
def get_ancestors(weave, entry):
163
return set(map(weave.idx_to_name,
164
weave.inclusions([weave.lookup(entry.revision)])))
173
# revision:ie mapping for each ie found in previous_inventories.
175
# identify candidate head revision ids.
167
176
for inv in previous_inventories:
168
if self.file_id in inv:
177
if inv.has_id(self.file_id):
169
178
ie = inv[self.file_id]
170
assert ie.file_id == self.file_id
171
if ie.revision in heads:
172
# fixup logic, there was a bug in revision updates.
173
# with x bit support.
179
if ie.revision in candidates:
180
# same revision value in two different inventories:
181
# correct possible inconsistencies:
182
# * there was a bug in revision updates with 'x' bit
175
if heads[ie.revision].executable != ie.executable:
176
heads[ie.revision].executable = False
185
if candidates[ie.revision].executable != ie.executable:
186
candidates[ie.revision].executable = False
177
187
ie.executable = False
178
188
except AttributeError:
180
assert heads[ie.revision] == ie
182
# may want to add it.
183
# may already be covered:
184
already_present = 0 != len(
185
[head for head in heads
186
if ie.revision in head_ancestors[head]])
188
# an ancestor of a known head.
191
ancestors = get_ancestors(entry_weave, ie)
192
# may knock something else out:
193
check_heads = list(heads.keys())
194
for head in check_heads:
195
if head in ancestors:
196
# this head is not really a head
198
head_ancestors[ie.revision] = ancestors
199
heads[ie.revision] = ie
202
def get_tar_item(self, root, dp, now, tree):
203
"""Get a tarfile item and a file stream for its content."""
204
item = tarfile.TarInfo(os.path.join(root, dp))
205
# TODO: would be cool to actually set it to the timestamp of the
206
# revision it was last changed
208
fileobj = self._put_in_tar(item, tree)
191
# add this revision as a candidate.
192
candidates[ie.revision] = ie
211
195
def has_text(self):
212
196
"""Return true if the object this entry represents has textual data.
234
218
Traceback (most recent call last):
235
219
InvalidEntryName: Invalid entry name: src/hello.c
237
assert isinstance(name, basestring), name
238
221
if '/' in name or '\\' in name:
239
raise InvalidEntryName(name=name)
240
self.executable = False
222
raise errors.InvalidEntryName(name=name)
223
self.file_id = file_id
241
224
self.revision = None
242
self.text_sha1 = None
243
self.text_size = None
244
self.file_id = file_id
246
self.text_id = text_id
247
226
self.parent_id = parent_id
248
self.symlink_target = None
250
228
def kind_character(self):
251
229
"""Return a short kind indicator useful for appending to names."""
252
raise BzrError('unknown kind %r' % self.kind)
254
known_kinds = ('file', 'directory', 'symlink', 'root_directory')
256
def _put_in_tar(self, item, tree):
257
"""populate item for stashing in a tar, and return the content stream.
259
If no content is available, return None.
261
raise BzrError("don't know how to export {%s} of kind %r" %
262
(self.file_id, self.kind))
264
def put_on_disk(self, dest, dp, tree):
265
"""Create a representation of self on disk in the prefix dest.
267
This is a template method - implement _put_on_disk in subclasses.
269
fullpath = appendpath(dest, dp)
270
self._put_on_disk(fullpath, tree)
271
mutter(" export {%s} kind %s to %s" % (self.file_id, self.kind, fullpath))
273
def _put_on_disk(self, fullpath, tree):
274
"""Put this entry onto disk at fullpath, from tree tree."""
275
raise BzrError("don't know how to export {%s} of kind %r" % (self.file_id, self.kind))
230
raise errors.BzrError('unknown kind %r' % self.kind)
232
known_kinds = ('file', 'directory', 'symlink')
277
234
def sorted_children(self):
278
l = self.children.items()
235
return sorted(self.children.items())
283
238
def versionable_kind(kind):
284
return kind in ('file', 'directory', 'symlink')
239
return (kind in ('file', 'directory', 'symlink', 'tree-reference'))
286
def check(self, checker, rev_id, inv, tree):
241
def check(self, checker, rev_id, inv):
287
242
"""Check this inventory entry is intact.
289
244
This is a template method, override _check for kind specific
247
:param checker: Check object providing context for the checks;
248
can be used to find out what parts of the repository have already
250
:param rev_id: Revision id from which this InventoryEntry was loaded.
251
Not necessarily the last-changed revision for this file.
252
:param inv: Inventory from which the entry was loaded.
292
if self.parent_id != None:
254
if self.parent_id is not None:
293
255
if not inv.has_id(self.parent_id):
294
raise BzrCheckError('missing parent {%s} in inventory for revision {%s}'
295
% (self.parent_id, rev_id))
296
self._check(checker, rev_id, tree)
256
raise errors.BzrCheckError(
257
'missing parent {%s} in inventory for revision {%s}' % (
258
self.parent_id, rev_id))
259
checker._add_entry_to_text_key_references(inv, self)
260
self._check(checker, rev_id)
298
def _check(self, checker, rev_id, tree):
262
def _check(self, checker, rev_id):
299
263
"""Check this inventory entry for kind specific errors."""
300
raise BzrCheckError('unknown entry kind %r in revision {%s}' %
264
checker._report_items.append(
265
'unknown entry kind %r in revision {%s}' % (self.kind, rev_id))
305
268
"""Clone this inventory entry."""
306
269
raise NotImplementedError
308
def _get_snapshot_change(self, previous_entries):
309
if len(previous_entries) > 1:
311
elif len(previous_entries) == 0:
272
def describe_change(old_entry, new_entry):
273
"""Describe the change between old_entry and this.
275
This smells of being an InterInventoryEntry situation, but as its
276
the first one, we're making it a static method for now.
278
An entry with a different parent, or different name is considered
279
to be renamed. Reparenting is an internal detail.
280
Note that renaming the parent does not trigger a rename for the
283
# TODO: Perhaps return an object rather than just a string
284
if old_entry is new_entry:
285
# also the case of both being None
287
elif old_entry is None:
314
return 'modified/renamed/reparented'
289
elif new_entry is None:
291
if old_entry.kind != new_entry.kind:
293
text_modified, meta_modified = new_entry.detect_changes(old_entry)
294
if text_modified or meta_modified:
298
# TODO 20060511 (mbp, rbc) factor out 'detect_rename' here.
299
if old_entry.parent_id != new_entry.parent_id:
301
elif old_entry.name != new_entry.name:
305
if renamed and not modified:
306
return InventoryEntry.RENAMED
307
if modified and not renamed:
309
if modified and renamed:
310
return InventoryEntry.MODIFIED_AND_RENAMED
316
313
def __repr__(self):
317
return ("%s(%r, %r, parent_id=%r)"
314
return ("%s(%r, %r, parent_id=%r, revision=%r)"
318
315
% (self.__class__.__name__,
323
def snapshot(self, revision, path, previous_entries,
324
work_tree, weave_store, transaction):
325
"""Make a snapshot of this entry which may or may not have changed.
327
This means that all its fields are populated, that it has its
328
text stored in the text store or weave.
330
mutter('new parents of %s are %r', path, previous_entries)
331
self._read_tree_state(path, work_tree)
332
if len(previous_entries) == 1:
333
# cannot be unchanged unless there is only one parent file rev.
334
parent_ie = previous_entries.values()[0]
335
if self._unchanged(parent_ie):
336
mutter("found unchanged entry")
337
self.revision = parent_ie.revision
339
return self.snapshot_revision(revision, previous_entries,
340
work_tree, weave_store, transaction)
342
def snapshot_revision(self, revision, previous_entries, work_tree,
343
weave_store, transaction):
344
"""Record this revision unconditionally."""
345
mutter('new revision for {%s}', self.file_id)
346
self.revision = revision
347
change = self._get_snapshot_change(previous_entries)
348
self._snapshot_text(previous_entries, work_tree, weave_store,
352
def _snapshot_text(self, file_parents, work_tree, weave_store, transaction):
353
"""Record the 'text' of this entry, whatever form that takes.
355
This default implementation simply adds an empty text.
357
mutter('storing file {%s} in revision {%s}',
358
self.file_id, self.revision)
359
self._add_text_to_weave([], file_parents, weave_store, transaction)
361
321
def __eq__(self, other):
323
# For the case when objects are cached
362
325
if not isinstance(other, InventoryEntry):
363
326
return NotImplemented
672
585
return compatible
675
class Inventory(object):
676
"""Inventory of versioned files in a tree.
678
This describes which file_id is present at each point in the tree,
679
and possibly the SHA-1 or other information about the file.
588
class TreeReference(InventoryEntry):
590
__slots__ = ['reference_revision']
592
kind = 'tree-reference'
594
def __init__(self, file_id, name, parent_id, revision=None,
595
reference_revision=None):
596
InventoryEntry.__init__(self, file_id, name, parent_id)
597
self.revision = revision
598
self.reference_revision = reference_revision
601
return TreeReference(self.file_id, self.name, self.parent_id,
602
self.revision, self.reference_revision)
604
def _read_tree_state(self, path, work_tree):
605
"""Populate fields in the inventory entry from the given tree.
607
self.reference_revision = work_tree.get_reference_revision(
610
def _forget_tree_state(self):
611
self.reference_revision = None
613
def _unchanged(self, previous_ie):
614
"""See InventoryEntry._unchanged."""
615
compatible = super(TreeReference, self)._unchanged(previous_ie)
616
if self.reference_revision != previous_ie.reference_revision:
621
class CommonInventory(object):
622
"""Basic inventory logic, defined in terms of primitives like has_id.
624
An inventory is the metadata about the contents of a tree.
626
This is broadly a map from file_id to entries such as directories, files,
627
symlinks and tree references. Each entry maintains its own metadata like
628
SHA1 and length for files, or children for a directory.
680
630
Entries can be looked up either by path or by file_id.
682
The inventory represents a typical unix file tree, with
683
directories containing files and subdirectories. We never store
684
the full path to a file, because renaming a directory implicitly
685
moves all of its contents. This class internally maintains a
632
InventoryEntry objects must not be modified after they are
633
inserted, other than through the Inventory API.
636
def has_filename(self, filename):
637
return bool(self.path2id(filename))
639
def id2path(self, file_id):
640
"""Return as a string the path to file_id.
643
>>> e = i.add(InventoryDirectory('src-id', 'src', ROOT_ID))
644
>>> e = i.add(InventoryFile('foo-id', 'foo.c', parent_id='src-id'))
645
>>> print i.id2path('foo-id')
648
:raises NoSuchId: If file_id is not present in the inventory.
650
# get all names, skipping root
651
return '/'.join(reversed(
652
[parent.name for parent in
653
self._iter_file_id_parents(file_id)][:-1]))
655
def iter_entries(self, from_dir=None, recursive=True):
656
"""Return (path, entry) pairs, in order by name.
658
:param from_dir: if None, start from the root,
659
otherwise start from this directory (either file-id or entry)
660
:param recursive: recurse into directories or not
663
if self.root is None:
667
elif isinstance(from_dir, basestring):
668
from_dir = self[from_dir]
670
# unrolling the recursive called changed the time from
671
# 440ms/663ms (inline/total) to 116ms/116ms
672
children = from_dir.children.items()
675
for name, ie in children:
678
children = collections.deque(children)
679
stack = [(u'', children)]
681
from_dir_relpath, children = stack[-1]
684
name, ie = children.popleft()
686
# we know that from_dir_relpath never ends in a slash
687
# and 'f' doesn't begin with one, we can do a string op, rather
688
# than the checks of pathjoin(), though this means that all paths
690
path = from_dir_relpath + '/' + name
694
if ie.kind != 'directory':
697
# But do this child first
698
new_children = ie.children.items()
700
new_children = collections.deque(new_children)
701
stack.append((path, new_children))
702
# Break out of inner loop, so that we start outer loop with child
705
# if we finished all children, pop it off the stack
708
def _preload_cache(self):
709
"""Populate any caches, we are about to access all items.
711
The default implementation does nothing, because CommonInventory doesn't
716
def iter_entries_by_dir(self, from_dir=None, specific_file_ids=None,
717
yield_parents=False):
718
"""Iterate over the entries in a directory first order.
720
This returns all entries for a directory before returning
721
the entries for children of a directory. This is not
722
lexicographically sorted order, and is a hybrid between
723
depth-first and breadth-first.
725
:param yield_parents: If True, yield the parents from the root leading
726
down to specific_file_ids that have been requested. This has no
727
impact if specific_file_ids is None.
728
:return: This yields (path, entry) pairs
730
if specific_file_ids and not isinstance(specific_file_ids, set):
731
specific_file_ids = set(specific_file_ids)
732
# TODO? Perhaps this should return the from_dir so that the root is
733
# yielded? or maybe an option?
734
if from_dir is None and specific_file_ids is None:
735
# They are iterating from the root, and have not specified any
736
# specific entries to look at. All current callers fully consume the
737
# iterator, so we can safely assume we are accessing all entries
738
self._preload_cache()
740
if self.root is None:
742
# Optimize a common case
743
if (not yield_parents and specific_file_ids is not None and
744
len(specific_file_ids) == 1):
745
file_id = list(specific_file_ids)[0]
746
if self.has_id(file_id):
747
yield self.id2path(file_id), self[file_id]
750
if (specific_file_ids is None or yield_parents or
751
self.root.file_id in specific_file_ids):
753
elif isinstance(from_dir, basestring):
754
from_dir = self[from_dir]
756
if specific_file_ids is not None:
757
# TODO: jam 20070302 This could really be done as a loop rather
758
# than a bunch of recursive calls.
761
def add_ancestors(file_id):
762
if not byid.has_id(file_id):
764
parent_id = byid[file_id].parent_id
765
if parent_id is None:
767
if parent_id not in parents:
768
parents.add(parent_id)
769
add_ancestors(parent_id)
770
for file_id in specific_file_ids:
771
add_ancestors(file_id)
775
stack = [(u'', from_dir)]
777
cur_relpath, cur_dir = stack.pop()
780
for child_name, child_ie in sorted(cur_dir.children.iteritems()):
782
child_relpath = cur_relpath + child_name
784
if (specific_file_ids is None or
785
child_ie.file_id in specific_file_ids or
786
(yield_parents and child_ie.file_id in parents)):
787
yield child_relpath, child_ie
789
if child_ie.kind == 'directory':
790
if parents is None or child_ie.file_id in parents:
791
child_dirs.append((child_relpath+'/', child_ie))
792
stack.extend(reversed(child_dirs))
794
def _make_delta(self, old):
795
"""Make an inventory delta from two inventories."""
798
adds = new_ids - old_ids
799
deletes = old_ids - new_ids
800
common = old_ids.intersection(new_ids)
802
for file_id in deletes:
803
delta.append((old.id2path(file_id), None, file_id, None))
805
delta.append((None, self.id2path(file_id), file_id, self[file_id]))
806
for file_id in common:
807
if old[file_id] != self[file_id]:
808
delta.append((old.id2path(file_id), self.id2path(file_id),
809
file_id, self[file_id]))
812
def make_entry(self, kind, name, parent_id, file_id=None):
813
"""Simple thunk to bzrlib.inventory.make_entry."""
814
return make_entry(kind, name, parent_id, file_id)
817
"""Return list of (path, ie) for all entries except the root.
819
This may be faster than iter_entries.
822
def descend(dir_ie, dir_path):
823
kids = dir_ie.children.items()
825
for name, ie in kids:
826
child_path = osutils.pathjoin(dir_path, name)
827
accum.append((child_path, ie))
828
if ie.kind == 'directory':
829
descend(ie, child_path)
831
if self.root is not None:
832
descend(self.root, u'')
835
def path2id(self, relpath):
836
"""Walk down through directories to return entry of last component.
838
:param relpath: may be either a list of path components, or a single
839
string, in which case it is automatically split.
841
This returns the entry of the last component in the path,
842
which may be either a file or a directory.
844
Returns None IFF the path is not found.
846
if isinstance(relpath, basestring):
847
names = osutils.splitpath(relpath)
853
except errors.NoSuchId:
854
# root doesn't exist yet so nothing else can
860
children = getattr(parent, 'children', None)
869
return parent.file_id
871
def filter(self, specific_fileids):
872
"""Get an inventory view filtered against a set of file-ids.
874
Children of directories and parents are included.
876
The result may or may not reference the underlying inventory
877
so it should be treated as immutable.
879
interesting_parents = set()
880
for fileid in specific_fileids:
882
interesting_parents.update(self.get_idpath(fileid))
883
except errors.NoSuchId:
884
# This fileid is not in the inventory - that's ok
886
entries = self.iter_entries()
887
if self.root is None:
888
return Inventory(root_id=None)
889
other = Inventory(entries.next()[1].file_id)
890
other.root.revision = self.root.revision
891
other.revision_id = self.revision_id
892
directories_to_expand = set()
893
for path, entry in entries:
894
file_id = entry.file_id
895
if (file_id in specific_fileids
896
or entry.parent_id in directories_to_expand):
897
if entry.kind == 'directory':
898
directories_to_expand.add(file_id)
899
elif file_id not in interesting_parents:
901
other.add(entry.copy())
904
def get_idpath(self, file_id):
905
"""Return a list of file_ids for the path to an entry.
907
The list contains one element for each directory followed by
908
the id of the file itself. So the length of the returned list
909
is equal to the depth of the file in the tree, counting the
910
root directory as depth 1.
913
for parent in self._iter_file_id_parents(file_id):
914
p.insert(0, parent.file_id)
918
class Inventory(CommonInventory):
919
"""Mutable dict based in-memory inventory.
921
We never store the full path to a file, because renaming a directory
922
implicitly moves all of its contents. This class internally maintains a
686
923
lookup tree that allows the children under a directory to be
687
924
returned quickly.
689
InventoryEntry objects must not be modified after they are
690
inserted, other than through the Inventory API.
692
926
>>> inv = Inventory()
693
927
>>> inv.add(InventoryFile('123-123', 'hello.c', ROOT_ID))
694
InventoryFile('123-123', 'hello.c', parent_id='TREE_ROOT')
928
InventoryFile('123-123', 'hello.c', parent_id='TREE_ROOT', sha1=None, len=None, revision=None)
695
929
>>> inv['123-123'].name
698
May be treated as an iterator or set to look up file ids:
700
>>> bool(inv.path2id('hello.c'))
705
May also look up by name:
707
>>> [x[0] for x in inv.iter_entries()]
709
>>> inv = Inventory('TREE_ROOT-12345678-12345678')
710
>>> inv.add(InventoryFile('123-123', 'hello.c', ROOT_ID))
711
InventoryFile('123-123', 'hello.c', parent_id='TREE_ROOT-12345678-12345678')
932
Id's may be looked up from paths:
934
>>> inv.path2id('hello.c')
936
>>> inv.has_id('123-123')
939
There are iterators over the contents:
941
>>> [entry[0] for entry in inv.iter_entries()]
713
def __init__(self, root_id=ROOT_ID):
945
def __init__(self, root_id=ROOT_ID, revision_id=None):
714
946
"""Create or read an inventory.
716
948
If a working directory is specified, the inventory is read
720
952
The inventory is created with a default root directory, with
723
# We are letting Branch.initialize() create a unique inventory
724
# root id. Rather than generating a random one here.
726
# root_id = bzrlib.branch.gen_file_id('TREE_ROOT')
727
self.root = RootEntry(root_id)
955
if root_id is not None:
956
self._set_root(InventoryDirectory(root_id, u'', None))
960
self.revision_id = revision_id
963
# More than one page of ouput is not useful anymore to debug
966
contents = repr(self._byid)
967
if len(contents) > max_len:
968
contents = contents[:(max_len-len(closing))] + closing
969
return "<Inventory object at %x, contents=%r>" % (id(self), contents)
971
def apply_delta(self, delta):
972
"""Apply a delta to this inventory.
974
See the inventory developers documentation for the theory behind
977
If delta application fails the inventory is left in an indeterminate
978
state and must not be used.
980
:param delta: A list of changes to apply. After all the changes are
981
applied the final inventory must be internally consistent, but it
982
is ok to supply changes which, if only half-applied would have an
983
invalid result - such as supplying two changes which rename two
984
files, 'A' and 'B' with each other : [('A', 'B', 'A-id', a_entry),
985
('B', 'A', 'B-id', b_entry)].
987
Each change is a tuple, of the form (old_path, new_path, file_id,
990
When new_path is None, the change indicates the removal of an entry
991
from the inventory and new_entry will be ignored (using None is
992
appropriate). If new_path is not None, then new_entry must be an
993
InventoryEntry instance, which will be incorporated into the
994
inventory (and replace any existing entry with the same file id).
996
When old_path is None, the change indicates the addition of
997
a new entry to the inventory.
999
When neither new_path nor old_path are None, the change is a
1000
modification to an entry, such as a rename, reparent, kind change
1003
The children attribute of new_entry is ignored. This is because
1004
this method preserves children automatically across alterations to
1005
the parent of the children, and cases where the parent id of a
1006
child is changing require the child to be passed in as a separate
1007
change regardless. E.g. in the recursive deletion of a directory -
1008
the directory's children must be included in the delta, or the
1009
final inventory will be invalid.
1011
Note that a file_id must only appear once within a given delta.
1012
An AssertionError is raised otherwise.
1014
# Check that the delta is legal. It would be nice if this could be
1015
# done within the loops below but it's safer to validate the delta
1016
# before starting to mutate the inventory, as there isn't a rollback
1018
list(_check_delta_unique_ids(_check_delta_unique_new_paths(
1019
_check_delta_unique_old_paths(_check_delta_ids_match_entry(
1020
_check_delta_ids_are_valid(
1021
_check_delta_new_path_entry_both_or_None(
1025
# Remove all affected items which were in the original inventory,
1026
# starting with the longest paths, thus ensuring parents are examined
1027
# after their children, which means that everything we examine has no
1028
# modified children remaining by the time we examine it.
1029
for old_path, file_id in sorted(((op, f) for op, np, f, e in delta
1030
if op is not None), reverse=True):
1031
# Preserve unaltered children of file_id for later reinsertion.
1032
file_id_children = getattr(self[file_id], 'children', {})
1033
if len(file_id_children):
1034
children[file_id] = file_id_children
1035
if self.id2path(file_id) != old_path:
1036
raise errors.InconsistentDelta(old_path, file_id,
1037
"Entry was at wrong other path %r." % self.id2path(file_id))
1038
# Remove file_id and the unaltered children. If file_id is not
1039
# being deleted it will be reinserted back later.
1040
self.remove_recursive_id(file_id)
1041
# Insert all affected which should be in the new inventory, reattaching
1042
# their children if they had any. This is done from shortest path to
1043
# longest, ensuring that items which were modified and whose parents in
1044
# the resulting inventory were also modified, are inserted after their
1046
for new_path, f, new_entry in sorted((np, f, e) for op, np, f, e in
1047
delta if np is not None):
1048
if new_entry.kind == 'directory':
1049
# Pop the child which to allow detection of children whose
1050
# parents were deleted and which were not reattached to a new
1052
replacement = InventoryDirectory(new_entry.file_id,
1053
new_entry.name, new_entry.parent_id)
1054
replacement.revision = new_entry.revision
1055
replacement.children = children.pop(replacement.file_id, {})
1056
new_entry = replacement
1059
except errors.DuplicateFileId:
1060
raise errors.InconsistentDelta(new_path, new_entry.file_id,
1061
"New id is already present in target.")
1062
except AttributeError:
1063
raise errors.InconsistentDelta(new_path, new_entry.file_id,
1064
"Parent is not a directory.")
1065
if self.id2path(new_entry.file_id) != new_path:
1066
raise errors.InconsistentDelta(new_path, new_entry.file_id,
1067
"New path is not consistent with parent path.")
1069
# Get the parent id that was deleted
1070
parent_id, children = children.popitem()
1071
raise errors.InconsistentDelta("<deleted>", parent_id,
1072
"The file id was deleted but its children were not deleted.")
1074
def create_by_apply_delta(self, inventory_delta, new_revision_id,
1075
propagate_caches=False):
1076
"""See CHKInventory.create_by_apply_delta()"""
1077
new_inv = self.copy()
1078
new_inv.apply_delta(inventory_delta)
1079
new_inv.revision_id = new_revision_id
1082
def _set_root(self, ie):
728
1084
self._byid = {self.root.file_id: self.root}
732
other = Inventory(self.root.file_id)
1087
# TODO: jam 20051218 Should copy also copy the revision_id?
1088
entries = self.iter_entries()
1089
if self.root is None:
1090
return Inventory(root_id=None)
1091
other = Inventory(entries.next()[1].file_id)
1092
other.root.revision = self.root.revision
733
1093
# copy recursively so we know directories will be added before
734
1094
# their children. There are more efficient ways than this...
735
for path, entry in self.iter_entries():
736
if entry == self.root:
1095
for path, entry in entries:
738
1096
other.add(entry.copy())
742
1099
def __iter__(self):
1100
"""Iterate over all file-ids."""
743
1101
return iter(self._byid)
1103
def iter_just_entries(self):
1104
"""Iterate over all entries.
1106
Unlike iter_entries(), just the entries are returned (not (path, ie))
1107
and the order of entries is undefined.
1109
XXX: We may not want to merge this into bzr.dev.
1111
if self.root is None:
1113
for _, ie in self._byid.iteritems():
746
1116
def __len__(self):
747
1117
"""Returns number of entries."""
748
1118
return len(self._byid)
751
def iter_entries(self, from_dir=None):
752
"""Return (path, entry) pairs, in order by name."""
756
elif isinstance(from_dir, basestring):
757
from_dir = self._byid[from_dir]
759
kids = from_dir.children.items()
761
for name, ie in kids:
763
if ie.kind == 'directory':
764
for cn, cie in self.iter_entries(from_dir=ie.file_id):
765
yield os.path.join(name, cn), cie
769
"""Return list of (path, ie) for all entries except the root.
771
This may be faster than iter_entries.
774
def descend(dir_ie, dir_path):
775
kids = dir_ie.children.items()
777
for name, ie in kids:
778
child_path = os.path.join(dir_path, name)
779
accum.append((child_path, ie))
780
if ie.kind == 'directory':
781
descend(ie, child_path)
783
descend(self.root, '')
787
def directories(self):
788
"""Return (path, entry) pairs for all directories, including the root.
791
def descend(parent_ie, parent_path):
792
accum.append((parent_path, parent_ie))
794
kids = [(ie.name, ie) for ie in parent_ie.children.itervalues() if ie.kind == 'directory']
797
for name, child_ie in kids:
798
child_path = os.path.join(parent_path, name)
799
descend(child_ie, child_path)
800
descend(self.root, '')
805
def __contains__(self, file_id):
806
"""True if this entry contains a file with given id.
808
>>> inv = Inventory()
809
>>> inv.add(InventoryFile('123', 'foo.c', ROOT_ID))
810
InventoryFile('123', 'foo.c', parent_id='TREE_ROOT')
816
return file_id in self._byid
819
1120
def __getitem__(self, file_id):
820
1121
"""Return the entry for given file_id.
822
1123
>>> inv = Inventory()
823
1124
>>> inv.add(InventoryFile('123123', 'hello.c', ROOT_ID))
824
InventoryFile('123123', 'hello.c', parent_id='TREE_ROOT')
1125
InventoryFile('123123', 'hello.c', parent_id='TREE_ROOT', sha1=None, len=None, revision=None)
825
1126
>>> inv['123123'].name
829
1130
return self._byid[file_id]
830
1131
except KeyError:
832
raise BzrError("can't look up file_id None")
834
raise BzrError("file_id {%s} not in inventory" % file_id)
1132
# really we're passing an inventory, not a tree...
1133
raise errors.NoSuchId(self, file_id)
837
1135
def get_file_kind(self, file_id):
838
1136
return self._byid[file_id].kind
1058
1334
del old_parent.children[file_ie.name]
1059
1335
new_parent.children[new_name] = file_ie
1061
1337
file_ie.name = new_name
1062
1338
file_ie.parent_id = new_parent_id
1340
def is_root(self, file_id):
1341
return self.root is not None and file_id == self.root.file_id
1344
class CHKInventory(CommonInventory):
1345
"""An inventory persisted in a CHK store.
1347
By design, a CHKInventory is immutable so many of the methods
1348
supported by Inventory - add, rename, apply_delta, etc - are *not*
1349
supported. To create a new CHKInventory, use create_by_apply_delta()
1350
or from_inventory(), say.
1352
Internally, a CHKInventory has one or two CHKMaps:
1354
* id_to_entry - a map from (file_id,) => InventoryEntry as bytes
1355
* parent_id_basename_to_file_id - a map from (parent_id, basename_utf8)
1358
The second map is optional and not present in early CHkRepository's.
1360
No caching is performed: every method call or item access will perform
1361
requests to the storage layer. As such, keep references to objects you
1365
def __init__(self, search_key_name):
1366
CommonInventory.__init__(self)
1367
self._fileid_to_entry_cache = {}
1368
self._fully_cached = False
1369
self._path_to_fileid_cache = {}
1370
self._search_key_name = search_key_name
1373
def __eq__(self, other):
1374
"""Compare two sets by comparing their contents."""
1375
if not isinstance(other, CHKInventory):
1376
return NotImplemented
1378
this_key = self.id_to_entry.key()
1379
other_key = other.id_to_entry.key()
1380
this_pid_key = self.parent_id_basename_to_file_id.key()
1381
other_pid_key = other.parent_id_basename_to_file_id.key()
1382
if None in (this_key, this_pid_key, other_key, other_pid_key):
1384
return this_key == other_key and this_pid_key == other_pid_key
1386
def _entry_to_bytes(self, entry):
1387
"""Serialise entry as a single bytestring.
1389
:param Entry: An inventory entry.
1390
:return: A bytestring for the entry.
1393
ENTRY ::= FILE | DIR | SYMLINK | TREE
1394
FILE ::= "file: " COMMON SEP SHA SEP SIZE SEP EXECUTABLE
1395
DIR ::= "dir: " COMMON
1396
SYMLINK ::= "symlink: " COMMON SEP TARGET_UTF8
1397
TREE ::= "tree: " COMMON REFERENCE_REVISION
1398
COMMON ::= FILE_ID SEP PARENT_ID SEP NAME_UTF8 SEP REVISION
1401
if entry.parent_id is not None:
1402
parent_str = entry.parent_id
1405
name_str = entry.name.encode("utf8")
1406
if entry.kind == 'file':
1407
if entry.executable:
1411
return "file: %s\n%s\n%s\n%s\n%s\n%d\n%s" % (
1412
entry.file_id, parent_str, name_str, entry.revision,
1413
entry.text_sha1, entry.text_size, exec_str)
1414
elif entry.kind == 'directory':
1415
return "dir: %s\n%s\n%s\n%s" % (
1416
entry.file_id, parent_str, name_str, entry.revision)
1417
elif entry.kind == 'symlink':
1418
return "symlink: %s\n%s\n%s\n%s\n%s" % (
1419
entry.file_id, parent_str, name_str, entry.revision,
1420
entry.symlink_target.encode("utf8"))
1421
elif entry.kind == 'tree-reference':
1422
return "tree: %s\n%s\n%s\n%s\n%s" % (
1423
entry.file_id, parent_str, name_str, entry.revision,
1424
entry.reference_revision)
1426
raise ValueError("unknown kind %r" % entry.kind)
1428
def _expand_fileids_to_parents_and_children(self, file_ids):
1429
"""Give a more wholistic view starting with the given file_ids.
1431
For any file_id which maps to a directory, we will include all children
1432
of that directory. We will also include all directories which are
1433
parents of the given file_ids, but we will not include their children.
1440
fringle # fringle-id
1444
if given [foo-id] we will include
1445
TREE_ROOT as interesting parents
1447
foo-id, baz-id, frob-id, fringle-id
1451
# TODO: Pre-pass over the list of fileids to see if anything is already
1452
# deserialized in self._fileid_to_entry_cache
1454
directories_to_expand = set()
1455
children_of_parent_id = {}
1456
# It is okay if some of the fileids are missing
1457
for entry in self._getitems(file_ids):
1458
if entry.kind == 'directory':
1459
directories_to_expand.add(entry.file_id)
1460
interesting.add(entry.parent_id)
1461
children_of_parent_id.setdefault(entry.parent_id, set()
1462
).add(entry.file_id)
1464
# Now, interesting has all of the direct parents, but not the
1465
# parents of those parents. It also may have some duplicates with
1467
remaining_parents = interesting.difference(file_ids)
1468
# When we hit the TREE_ROOT, we'll get an interesting parent of None,
1469
# but we don't actually want to recurse into that
1470
interesting.add(None) # this will auto-filter it in the loop
1471
remaining_parents.discard(None)
1472
while remaining_parents:
1473
next_parents = set()
1474
for entry in self._getitems(remaining_parents):
1475
next_parents.add(entry.parent_id)
1476
children_of_parent_id.setdefault(entry.parent_id, set()
1477
).add(entry.file_id)
1478
# Remove any search tips we've already processed
1479
remaining_parents = next_parents.difference(interesting)
1480
interesting.update(remaining_parents)
1481
# We should probably also .difference(directories_to_expand)
1482
interesting.update(file_ids)
1483
interesting.discard(None)
1484
while directories_to_expand:
1485
# Expand directories by looking in the
1486
# parent_id_basename_to_file_id map
1487
keys = [StaticTuple(f,).intern() for f in directories_to_expand]
1488
directories_to_expand = set()
1489
items = self.parent_id_basename_to_file_id.iteritems(keys)
1490
next_file_ids = set([item[1] for item in items])
1491
next_file_ids = next_file_ids.difference(interesting)
1492
interesting.update(next_file_ids)
1493
for entry in self._getitems(next_file_ids):
1494
if entry.kind == 'directory':
1495
directories_to_expand.add(entry.file_id)
1496
children_of_parent_id.setdefault(entry.parent_id, set()
1497
).add(entry.file_id)
1498
return interesting, children_of_parent_id
1500
def filter(self, specific_fileids):
1501
"""Get an inventory view filtered against a set of file-ids.
1503
Children of directories and parents are included.
1505
The result may or may not reference the underlying inventory
1506
so it should be treated as immutable.
1509
parent_to_children) = self._expand_fileids_to_parents_and_children(
1511
# There is some overlap here, but we assume that all interesting items
1512
# are in the _fileid_to_entry_cache because we had to read them to
1513
# determine if they were a dir we wanted to recurse, or just a file
1514
# This should give us all the entries we'll want to add, so start
1516
other = Inventory(self.root_id)
1517
other.root.revision = self.root.revision
1518
other.revision_id = self.revision_id
1519
if not interesting or not parent_to_children:
1520
# empty filter, or filtering entrys that don't exist
1521
# (if even 1 existed, then we would have populated
1522
# parent_to_children with at least the tree root.)
1524
cache = self._fileid_to_entry_cache
1525
remaining_children = collections.deque(parent_to_children[self.root_id])
1526
while remaining_children:
1527
file_id = remaining_children.popleft()
1529
if ie.kind == 'directory':
1530
ie = ie.copy() # We create a copy to depopulate the .children attribute
1531
# TODO: depending on the uses of 'other' we should probably alwyas
1532
# '.copy()' to prevent someone from mutating other and
1533
# invaliding our internal cache
1535
if file_id in parent_to_children:
1536
remaining_children.extend(parent_to_children[file_id])
1540
def _bytes_to_utf8name_key(bytes):
1541
"""Get the file_id, revision_id key out of bytes."""
1542
# We don't normally care about name, except for times when we want
1543
# to filter out empty names because of non rich-root...
1544
sections = bytes.split('\n')
1545
kind, file_id = sections[0].split(': ')
1546
return (sections[2], intern(file_id), intern(sections[3]))
1548
def _bytes_to_entry(self, bytes):
1549
"""Deserialise a serialised entry."""
1550
sections = bytes.split('\n')
1551
if sections[0].startswith("file: "):
1552
result = InventoryFile(sections[0][6:],
1553
sections[2].decode('utf8'),
1555
result.text_sha1 = sections[4]
1556
result.text_size = int(sections[5])
1557
result.executable = sections[6] == "Y"
1558
elif sections[0].startswith("dir: "):
1559
result = CHKInventoryDirectory(sections[0][5:],
1560
sections[2].decode('utf8'),
1562
elif sections[0].startswith("symlink: "):
1563
result = InventoryLink(sections[0][9:],
1564
sections[2].decode('utf8'),
1566
result.symlink_target = sections[4].decode('utf8')
1567
elif sections[0].startswith("tree: "):
1568
result = TreeReference(sections[0][6:],
1569
sections[2].decode('utf8'),
1571
result.reference_revision = sections[4]
1573
raise ValueError("Not a serialised entry %r" % bytes)
1574
result.file_id = intern(result.file_id)
1575
result.revision = intern(sections[3])
1576
if result.parent_id == '':
1577
result.parent_id = None
1578
self._fileid_to_entry_cache[result.file_id] = result
1581
def create_by_apply_delta(self, inventory_delta, new_revision_id,
1582
propagate_caches=False):
1583
"""Create a new CHKInventory by applying inventory_delta to this one.
1585
See the inventory developers documentation for the theory behind
1588
:param inventory_delta: The inventory delta to apply. See
1589
Inventory.apply_delta for details.
1590
:param new_revision_id: The revision id of the resulting CHKInventory.
1591
:param propagate_caches: If True, the caches for this inventory are
1592
copied to and updated for the result.
1593
:return: The new CHKInventory.
1595
split = osutils.split
1596
result = CHKInventory(self._search_key_name)
1597
if propagate_caches:
1598
# Just propagate the path-to-fileid cache for now
1599
result._path_to_fileid_cache = dict(self._path_to_fileid_cache.iteritems())
1600
search_key_func = chk_map.search_key_registry.get(self._search_key_name)
1601
self.id_to_entry._ensure_root()
1602
maximum_size = self.id_to_entry._root_node.maximum_size
1603
result.revision_id = new_revision_id
1604
result.id_to_entry = chk_map.CHKMap(
1605
self.id_to_entry._store,
1606
self.id_to_entry.key(),
1607
search_key_func=search_key_func)
1608
result.id_to_entry._ensure_root()
1609
result.id_to_entry._root_node.set_maximum_size(maximum_size)
1610
# Change to apply to the parent_id_basename delta. The dict maps
1611
# (parent_id, basename) -> (old_key, new_value). We use a dict because
1612
# when a path has its id replaced (e.g. the root is changed, or someone
1613
# does bzr mv a b, bzr mv c a, we should output a single change to this
1614
# map rather than two.
1615
parent_id_basename_delta = {}
1616
if self.parent_id_basename_to_file_id is not None:
1617
result.parent_id_basename_to_file_id = chk_map.CHKMap(
1618
self.parent_id_basename_to_file_id._store,
1619
self.parent_id_basename_to_file_id.key(),
1620
search_key_func=search_key_func)
1621
result.parent_id_basename_to_file_id._ensure_root()
1622
self.parent_id_basename_to_file_id._ensure_root()
1623
result_p_id_root = result.parent_id_basename_to_file_id._root_node
1624
p_id_root = self.parent_id_basename_to_file_id._root_node
1625
result_p_id_root.set_maximum_size(p_id_root.maximum_size)
1626
result_p_id_root._key_width = p_id_root._key_width
1628
result.parent_id_basename_to_file_id = None
1629
result.root_id = self.root_id
1630
id_to_entry_delta = []
1631
# inventory_delta is only traversed once, so we just update the
1633
# Check for repeated file ids
1634
inventory_delta = _check_delta_unique_ids(inventory_delta)
1635
# Repeated old paths
1636
inventory_delta = _check_delta_unique_old_paths(inventory_delta)
1637
# Check for repeated new paths
1638
inventory_delta = _check_delta_unique_new_paths(inventory_delta)
1639
# Check for entries that don't match the fileid
1640
inventory_delta = _check_delta_ids_match_entry(inventory_delta)
1641
# Check for nonsense fileids
1642
inventory_delta = _check_delta_ids_are_valid(inventory_delta)
1643
# Check for new_path <-> entry consistency
1644
inventory_delta = _check_delta_new_path_entry_both_or_None(
1646
# All changed entries need to have their parents be directories and be
1647
# at the right path. This set contains (path, id) tuples.
1649
# When we delete an item, all the children of it must be either deleted
1650
# or altered in their own right. As we batch process the change via
1651
# CHKMap.apply_delta, we build a set of things to use to validate the
1655
for old_path, new_path, file_id, entry in inventory_delta:
1658
result.root_id = file_id
1659
if new_path is None:
1664
if propagate_caches:
1666
del result._path_to_fileid_cache[old_path]
1669
deletes.add(file_id)
1671
new_key = StaticTuple(file_id,)
1672
new_value = result._entry_to_bytes(entry)
1673
# Update caches. It's worth doing this whether
1674
# we're propagating the old caches or not.
1675
result._path_to_fileid_cache[new_path] = file_id
1676
parents.add((split(new_path)[0], entry.parent_id))
1677
if old_path is None:
1680
old_key = StaticTuple(file_id,)
1681
if self.id2path(file_id) != old_path:
1682
raise errors.InconsistentDelta(old_path, file_id,
1683
"Entry was at wrong other path %r." %
1684
self.id2path(file_id))
1685
altered.add(file_id)
1686
id_to_entry_delta.append(StaticTuple(old_key, new_key, new_value))
1687
if result.parent_id_basename_to_file_id is not None:
1688
# parent_id, basename changes
1689
if old_path is None:
1692
old_entry = self[file_id]
1693
old_key = self._parent_id_basename_key(old_entry)
1694
if new_path is None:
1698
new_key = self._parent_id_basename_key(entry)
1700
# If the two keys are the same, the value will be unchanged
1701
# as its always the file id for this entry.
1702
if old_key != new_key:
1703
# Transform a change into explicit delete/add preserving
1704
# a possible match on the key from a different file id.
1705
if old_key is not None:
1706
parent_id_basename_delta.setdefault(
1707
old_key, [None, None])[0] = old_key
1708
if new_key is not None:
1709
parent_id_basename_delta.setdefault(
1710
new_key, [None, None])[1] = new_value
1711
# validate that deletes are complete.
1712
for file_id in deletes:
1713
entry = self[file_id]
1714
if entry.kind != 'directory':
1716
# This loop could potentially be better by using the id_basename
1717
# map to just get the child file ids.
1718
for child in entry.children.values():
1719
if child.file_id not in altered:
1720
raise errors.InconsistentDelta(self.id2path(child.file_id),
1721
child.file_id, "Child not deleted or reparented when "
1723
result.id_to_entry.apply_delta(id_to_entry_delta)
1724
if parent_id_basename_delta:
1725
# Transform the parent_id_basename delta data into a linear delta
1726
# with only one record for a given key. Optimally this would allow
1727
# re-keying, but its simpler to just output that as a delete+add
1728
# to spend less time calculating the delta.
1730
for key, (old_key, value) in parent_id_basename_delta.iteritems():
1731
if value is not None:
1732
delta_list.append((old_key, key, value))
1734
delta_list.append((old_key, None, None))
1735
result.parent_id_basename_to_file_id.apply_delta(delta_list)
1736
parents.discard(('', None))
1737
for parent_path, parent in parents:
1739
if result[parent].kind != 'directory':
1740
raise errors.InconsistentDelta(result.id2path(parent), parent,
1741
'Not a directory, but given children')
1742
except errors.NoSuchId:
1743
raise errors.InconsistentDelta("<unknown>", parent,
1744
"Parent is not present in resulting inventory.")
1745
if result.path2id(parent_path) != parent:
1746
raise errors.InconsistentDelta(parent_path, parent,
1747
"Parent has wrong path %r." % result.path2id(parent_path))
1751
def deserialise(klass, chk_store, bytes, expected_revision_id):
1752
"""Deserialise a CHKInventory.
1754
:param chk_store: A CHK capable VersionedFiles instance.
1755
:param bytes: The serialised bytes.
1756
:param expected_revision_id: The revision ID we think this inventory is
1758
:return: A CHKInventory
1760
lines = bytes.split('\n')
1762
raise AssertionError('bytes to deserialize must end with an eol')
1764
if lines[0] != 'chkinventory:':
1765
raise ValueError("not a serialised CHKInventory: %r" % bytes)
1767
allowed_keys = frozenset(['root_id', 'revision_id', 'search_key_name',
1768
'parent_id_basename_to_file_id',
1770
for line in lines[1:]:
1771
key, value = line.split(': ', 1)
1772
if key not in allowed_keys:
1773
raise errors.BzrError('Unknown key in inventory: %r\n%r'
1776
raise errors.BzrError('Duplicate key in inventory: %r\n%r'
1779
revision_id = intern(info['revision_id'])
1780
root_id = intern(info['root_id'])
1781
search_key_name = intern(info.get('search_key_name', 'plain'))
1782
parent_id_basename_to_file_id = intern(info.get(
1783
'parent_id_basename_to_file_id', None))
1784
if not parent_id_basename_to_file_id.startswith('sha1:'):
1785
raise ValueError('parent_id_basename_to_file_id should be a sha1'
1786
' key not %r' % (parent_id_basename_to_file_id,))
1787
id_to_entry = info['id_to_entry']
1788
if not id_to_entry.startswith('sha1:'):
1789
raise ValueError('id_to_entry should be a sha1'
1790
' key not %r' % (id_to_entry,))
1792
result = CHKInventory(search_key_name)
1793
result.revision_id = revision_id
1794
result.root_id = root_id
1795
search_key_func = chk_map.search_key_registry.get(
1796
result._search_key_name)
1797
if parent_id_basename_to_file_id is not None:
1798
result.parent_id_basename_to_file_id = chk_map.CHKMap(
1799
chk_store, StaticTuple(parent_id_basename_to_file_id,),
1800
search_key_func=search_key_func)
1802
result.parent_id_basename_to_file_id = None
1804
result.id_to_entry = chk_map.CHKMap(chk_store,
1805
StaticTuple(id_to_entry,),
1806
search_key_func=search_key_func)
1807
if (result.revision_id,) != expected_revision_id:
1808
raise ValueError("Mismatched revision id and expected: %r, %r" %
1809
(result.revision_id, expected_revision_id))
1813
def from_inventory(klass, chk_store, inventory, maximum_size=0, search_key_name='plain'):
1814
"""Create a CHKInventory from an existing inventory.
1816
The content of inventory is copied into the chk_store, and a
1817
CHKInventory referencing that is returned.
1819
:param chk_store: A CHK capable VersionedFiles instance.
1820
:param inventory: The inventory to copy.
1821
:param maximum_size: The CHKMap node size limit.
1822
:param search_key_name: The identifier for the search key function
1824
result = klass(search_key_name)
1825
result.revision_id = inventory.revision_id
1826
result.root_id = inventory.root.file_id
1828
entry_to_bytes = result._entry_to_bytes
1829
parent_id_basename_key = result._parent_id_basename_key
1830
id_to_entry_dict = {}
1831
parent_id_basename_dict = {}
1832
for path, entry in inventory.iter_entries():
1833
key = StaticTuple(entry.file_id,).intern()
1834
id_to_entry_dict[key] = entry_to_bytes(entry)
1835
p_id_key = parent_id_basename_key(entry)
1836
parent_id_basename_dict[p_id_key] = entry.file_id
1838
result._populate_from_dicts(chk_store, id_to_entry_dict,
1839
parent_id_basename_dict, maximum_size=maximum_size)
1842
def _populate_from_dicts(self, chk_store, id_to_entry_dict,
1843
parent_id_basename_dict, maximum_size):
1844
search_key_func = chk_map.search_key_registry.get(self._search_key_name)
1845
root_key = chk_map.CHKMap.from_dict(chk_store, id_to_entry_dict,
1846
maximum_size=maximum_size, key_width=1,
1847
search_key_func=search_key_func)
1848
self.id_to_entry = chk_map.CHKMap(chk_store, root_key,
1850
root_key = chk_map.CHKMap.from_dict(chk_store,
1851
parent_id_basename_dict,
1852
maximum_size=maximum_size, key_width=2,
1853
search_key_func=search_key_func)
1854
self.parent_id_basename_to_file_id = chk_map.CHKMap(chk_store,
1855
root_key, search_key_func)
1857
def _parent_id_basename_key(self, entry):
1858
"""Create a key for a entry in a parent_id_basename_to_file_id index."""
1859
if entry.parent_id is not None:
1860
parent_id = entry.parent_id
1863
return StaticTuple(parent_id, entry.name.encode('utf8')).intern()
1865
def __getitem__(self, file_id):
1866
"""map a single file_id -> InventoryEntry."""
1868
raise errors.NoSuchId(self, file_id)
1869
result = self._fileid_to_entry_cache.get(file_id, None)
1870
if result is not None:
1873
return self._bytes_to_entry(
1874
self.id_to_entry.iteritems([StaticTuple(file_id,)]).next()[1])
1875
except StopIteration:
1876
# really we're passing an inventory, not a tree...
1877
raise errors.NoSuchId(self, file_id)
1879
def _getitems(self, file_ids):
1880
"""Similar to __getitem__, but lets you query for multiple.
1882
The returned order is undefined. And currently if an item doesn't
1883
exist, it isn't included in the output.
1887
for file_id in file_ids:
1888
entry = self._fileid_to_entry_cache.get(file_id, None)
1890
remaining.append(file_id)
1892
result.append(entry)
1893
file_keys = [StaticTuple(f,).intern() for f in remaining]
1894
for file_key, value in self.id_to_entry.iteritems(file_keys):
1895
entry = self._bytes_to_entry(value)
1896
result.append(entry)
1897
self._fileid_to_entry_cache[entry.file_id] = entry
1900
def has_id(self, file_id):
1901
# Perhaps have an explicit 'contains' method on CHKMap ?
1902
if self._fileid_to_entry_cache.get(file_id, None) is not None:
1905
self.id_to_entry.iteritems([StaticTuple(file_id,)]))) == 1
1907
def is_root(self, file_id):
1908
return file_id == self.root_id
1910
def _iter_file_id_parents(self, file_id):
1911
"""Yield the parents of file_id up to the root."""
1912
while file_id is not None:
1916
raise errors.NoSuchId(tree=self, file_id=file_id)
1918
file_id = ie.parent_id
1921
"""Iterate over all file-ids."""
1922
for key, _ in self.id_to_entry.iteritems():
1925
def iter_just_entries(self):
1926
"""Iterate over all entries.
1928
Unlike iter_entries(), just the entries are returned (not (path, ie))
1929
and the order of entries is undefined.
1931
XXX: We may not want to merge this into bzr.dev.
1933
for key, entry in self.id_to_entry.iteritems():
1935
ie = self._fileid_to_entry_cache.get(file_id, None)
1937
ie = self._bytes_to_entry(entry)
1938
self._fileid_to_entry_cache[file_id] = ie
1941
def _preload_cache(self):
1942
"""Make sure all file-ids are in _fileid_to_entry_cache"""
1943
if self._fully_cached:
1944
return # No need to do it again
1945
# The optimal sort order is to use iteritems() directly
1946
cache = self._fileid_to_entry_cache
1947
for key, entry in self.id_to_entry.iteritems():
1949
if file_id not in cache:
1950
ie = self._bytes_to_entry(entry)
1954
last_parent_id = last_parent_ie = None
1955
pid_items = self.parent_id_basename_to_file_id.iteritems()
1956
for key, child_file_id in pid_items:
1957
if key == ('', ''): # This is the root
1958
if child_file_id != self.root_id:
1959
raise ValueError('Data inconsistency detected.'
1960
' We expected data with key ("","") to match'
1961
' the root id, but %s != %s'
1962
% (child_file_id, self.root_id))
1964
parent_id, basename = key
1965
ie = cache[child_file_id]
1966
if parent_id == last_parent_id:
1967
parent_ie = last_parent_ie
1969
parent_ie = cache[parent_id]
1970
if parent_ie.kind != 'directory':
1971
raise ValueError('Data inconsistency detected.'
1972
' An entry in the parent_id_basename_to_file_id map'
1973
' has parent_id {%s} but the kind of that object'
1974
' is %r not "directory"' % (parent_id, parent_ie.kind))
1975
if parent_ie._children is None:
1976
parent_ie._children = {}
1977
basename = basename.decode('utf-8')
1978
if basename in parent_ie._children:
1979
existing_ie = parent_ie._children[basename]
1980
if existing_ie != ie:
1981
raise ValueError('Data inconsistency detected.'
1982
' Two entries with basename %r were found'
1983
' in the parent entry {%s}'
1984
% (basename, parent_id))
1985
if basename != ie.name:
1986
raise ValueError('Data inconsistency detected.'
1987
' In the parent_id_basename_to_file_id map, file_id'
1988
' {%s} is listed as having basename %r, but in the'
1989
' id_to_entry map it is %r'
1990
% (child_file_id, basename, ie.name))
1991
parent_ie._children[basename] = ie
1992
self._fully_cached = True
1994
def iter_changes(self, basis):
1995
"""Generate a Tree.iter_changes change list between this and basis.
1997
:param basis: Another CHKInventory.
1998
:return: An iterator over the changes between self and basis, as per
1999
tree.iter_changes().
2001
# We want: (file_id, (path_in_source, path_in_target),
2002
# changed_content, versioned, parent, name, kind,
2004
for key, basis_value, self_value in \
2005
self.id_to_entry.iter_changes(basis.id_to_entry):
2007
if basis_value is not None:
2008
basis_entry = basis._bytes_to_entry(basis_value)
2009
path_in_source = basis.id2path(file_id)
2010
basis_parent = basis_entry.parent_id
2011
basis_name = basis_entry.name
2012
basis_executable = basis_entry.executable
2014
path_in_source = None
2017
basis_executable = None
2018
if self_value is not None:
2019
self_entry = self._bytes_to_entry(self_value)
2020
path_in_target = self.id2path(file_id)
2021
self_parent = self_entry.parent_id
2022
self_name = self_entry.name
2023
self_executable = self_entry.executable
2025
path_in_target = None
2028
self_executable = None
2029
if basis_value is None:
2031
kind = (None, self_entry.kind)
2032
versioned = (False, True)
2033
elif self_value is None:
2035
kind = (basis_entry.kind, None)
2036
versioned = (True, False)
2038
kind = (basis_entry.kind, self_entry.kind)
2039
versioned = (True, True)
2040
changed_content = False
2041
if kind[0] != kind[1]:
2042
changed_content = True
2043
elif kind[0] == 'file':
2044
if (self_entry.text_size != basis_entry.text_size or
2045
self_entry.text_sha1 != basis_entry.text_sha1):
2046
changed_content = True
2047
elif kind[0] == 'symlink':
2048
if self_entry.symlink_target != basis_entry.symlink_target:
2049
changed_content = True
2050
elif kind[0] == 'tree-reference':
2051
if (self_entry.reference_revision !=
2052
basis_entry.reference_revision):
2053
changed_content = True
2054
parent = (basis_parent, self_parent)
2055
name = (basis_name, self_name)
2056
executable = (basis_executable, self_executable)
2057
if (not changed_content
2058
and parent[0] == parent[1]
2059
and name[0] == name[1]
2060
and executable[0] == executable[1]):
2061
# Could happen when only the revision changed for a directory
2064
yield (file_id, (path_in_source, path_in_target), changed_content,
2065
versioned, parent, name, kind, executable)
2068
"""Return the number of entries in the inventory."""
2069
return len(self.id_to_entry)
2071
def _make_delta(self, old):
2072
"""Make an inventory delta from two inventories."""
2073
if type(old) != CHKInventory:
2074
return CommonInventory._make_delta(self, old)
2076
for key, old_value, self_value in \
2077
self.id_to_entry.iter_changes(old.id_to_entry):
2079
if old_value is not None:
2080
old_path = old.id2path(file_id)
2083
if self_value is not None:
2084
entry = self._bytes_to_entry(self_value)
2085
self._fileid_to_entry_cache[file_id] = entry
2086
new_path = self.id2path(file_id)
2090
delta.append((old_path, new_path, file_id, entry))
2093
def path2id(self, relpath):
2094
"""See CommonInventory.path2id()."""
2095
# TODO: perhaps support negative hits?
2096
if isinstance(relpath, basestring):
2097
names = osutils.splitpath(relpath)
2102
relpath = osutils.pathjoin(*relpath)
2103
result = self._path_to_fileid_cache.get(relpath, None)
2104
if result is not None:
2106
current_id = self.root_id
2107
if current_id is None:
2109
parent_id_index = self.parent_id_basename_to_file_id
2111
for basename in names:
2112
if cur_path is None:
2115
cur_path = cur_path + '/' + basename
2116
basename_utf8 = basename.encode('utf8')
2117
file_id = self._path_to_fileid_cache.get(cur_path, None)
2119
key_filter = [StaticTuple(current_id, basename_utf8)]
2120
items = parent_id_index.iteritems(key_filter)
2121
for (parent_id, name_utf8), file_id in items:
2122
if parent_id != current_id or name_utf8 != basename_utf8:
2123
raise errors.BzrError("corrupt inventory lookup! "
2124
"%r %r %r %r" % (parent_id, current_id, name_utf8,
2129
self._path_to_fileid_cache[cur_path] = file_id
2130
current_id = file_id
2134
"""Serialise the inventory to lines."""
2135
lines = ["chkinventory:\n"]
2136
if self._search_key_name != 'plain':
2137
# custom ordering grouping things that don't change together
2138
lines.append('search_key_name: %s\n' % (self._search_key_name,))
2139
lines.append("root_id: %s\n" % self.root_id)
2140
lines.append('parent_id_basename_to_file_id: %s\n' %
2141
(self.parent_id_basename_to_file_id.key()[0],))
2142
lines.append("revision_id: %s\n" % self.revision_id)
2143
lines.append("id_to_entry: %s\n" % (self.id_to_entry.key()[0],))
2145
lines.append("revision_id: %s\n" % self.revision_id)
2146
lines.append("root_id: %s\n" % self.root_id)
2147
if self.parent_id_basename_to_file_id is not None:
2148
lines.append('parent_id_basename_to_file_id: %s\n' %
2149
(self.parent_id_basename_to_file_id.key()[0],))
2150
lines.append("id_to_entry: %s\n" % (self.id_to_entry.key()[0],))
2155
"""Get the root entry."""
2156
return self[self.root_id]
2159
class CHKInventoryDirectory(InventoryDirectory):
2160
"""A directory in an inventory."""
2162
__slots__ = ['_children', '_chk_inventory']
2164
def __init__(self, file_id, name, parent_id, chk_inventory):
2165
# Don't call InventoryDirectory.__init__ - it isn't right for this
2167
InventoryEntry.__init__(self, file_id, name, parent_id)
2168
self._children = None
2169
self._chk_inventory = chk_inventory
2173
"""Access the list of children of this directory.
2175
With a parent_id_basename_to_file_id index, loads all the children,
2176
without loads the entire index. Without is bad. A more sophisticated
2177
proxy object might be nice, to allow partial loading of children as
2178
well when specific names are accessed. (So path traversal can be
2179
written in the obvious way but not examine siblings.).
2181
if self._children is not None:
2182
return self._children
2183
# No longer supported
2184
if self._chk_inventory.parent_id_basename_to_file_id is None:
2185
raise AssertionError("Inventories without"
2186
" parent_id_basename_to_file_id are no longer supported")
2188
# XXX: Todo - use proxy objects for the children rather than loading
2189
# all when the attribute is referenced.
2190
parent_id_index = self._chk_inventory.parent_id_basename_to_file_id
2192
for (parent_id, name_utf8), file_id in parent_id_index.iteritems(
2193
key_filter=[StaticTuple(self.file_id,)]):
2194
child_keys.add(StaticTuple(file_id,))
2196
for file_id_key in child_keys:
2197
entry = self._chk_inventory._fileid_to_entry_cache.get(
2198
file_id_key[0], None)
2199
if entry is not None:
2200
result[entry.name] = entry
2201
cached.add(file_id_key)
2202
child_keys.difference_update(cached)
2203
# populate; todo: do by name
2204
id_to_entry = self._chk_inventory.id_to_entry
2205
for file_id_key, bytes in id_to_entry.iteritems(child_keys):
2206
entry = self._chk_inventory._bytes_to_entry(bytes)
2207
result[entry.name] = entry
2208
self._chk_inventory._fileid_to_entry_cache[file_id_key[0]] = entry
2209
self._children = result
2213
'directory': InventoryDirectory,
2214
'file': InventoryFile,
2215
'symlink': InventoryLink,
2216
'tree-reference': TreeReference
2219
def make_entry(kind, name, parent_id, file_id=None):
2220
"""Create an inventory entry.
2222
:param kind: the type of inventory entry to create.
2223
:param name: the basename of the entry.
2224
:param parent_id: the parent_id of the entry.
2225
:param file_id: the file_id to use. if None, one will be created.
2228
file_id = generate_ids.gen_file_id(name)
2229
name = ensure_normalized_name(name)
2231
factory = entry_factory[kind]
2233
raise errors.BadFileKindError(name, kind)
2234
return factory(file_id, name, parent_id)
2237
def ensure_normalized_name(name):
2240
:raises InvalidNormalization: When name is not normalized, and cannot be
2241
accessed on this platform by the normalized path.
2242
:return: The NFC normalised version of name.
2244
#------- This has been copied to bzrlib.dirstate.DirState.add, please
2245
# keep them synchronised.
2246
# we dont import normalized_filename directly because we want to be
2247
# able to change the implementation at runtime for tests.
2248
norm_name, can_access = osutils.normalized_filename(name)
2249
if norm_name != name:
2253
# TODO: jam 20060701 This would probably be more useful
2254
# if the error was raised with the full path
2255
raise errors.InvalidNormalization(name)
2259
_NAME_RE = lazy_regex.lazy_compile(r'^[^/\\]+$')
1069
2261
def is_valid_name(name):
1071
if _NAME_RE == None:
1072
_NAME_RE = re.compile(r'^[^/\\]+$')
1074
2262
return bool(_NAME_RE.match(name))
2265
def _check_delta_unique_ids(delta):
2266
"""Decorate a delta and check that the file ids in it are unique.
2268
:return: A generator over delta.
2272
length = len(ids) + 1
2274
if len(ids) != length:
2275
raise errors.InconsistentDelta(item[0] or item[1], item[2],
2280
def _check_delta_unique_new_paths(delta):
2281
"""Decorate a delta and check that the new paths in it are unique.
2283
:return: A generator over delta.
2287
length = len(paths) + 1
2289
if path is not None:
2291
if len(paths) != length:
2292
raise errors.InconsistentDelta(path, item[2], "repeated path")
2296
def _check_delta_unique_old_paths(delta):
2297
"""Decorate a delta and check that the old paths in it are unique.
2299
:return: A generator over delta.
2303
length = len(paths) + 1
2305
if path is not None:
2307
if len(paths) != length:
2308
raise errors.InconsistentDelta(path, item[2], "repeated path")
2312
def _check_delta_ids_are_valid(delta):
2313
"""Decorate a delta and check that the ids in it are valid.
2315
:return: A generator over delta.
2320
raise errors.InconsistentDelta(item[0] or item[1], item[2],
2321
"entry with file_id None %r" % entry)
2322
if type(item[2]) != str:
2323
raise errors.InconsistentDelta(item[0] or item[1], item[2],
2324
"entry with non bytes file_id %r" % entry)
2328
def _check_delta_ids_match_entry(delta):
2329
"""Decorate a delta and check that the ids in it match the entry.file_id.
2331
:return: A generator over delta.
2335
if entry is not None:
2336
if entry.file_id != item[2]:
2337
raise errors.InconsistentDelta(item[0] or item[1], item[2],
2338
"mismatched id with %r" % entry)
2342
def _check_delta_new_path_entry_both_or_None(delta):
2343
"""Decorate a delta and check that the new_path and entry are paired.
2345
:return: A generator over delta.
2350
if new_path is None and entry is not None:
2351
raise errors.InconsistentDelta(item[0], item[1],
2352
"Entry with no new_path")
2353
if new_path is not None and entry is None:
2354
raise errors.InconsistentDelta(new_path, item[1],
2355
"new_path with no entry")
2359
def mutable_inventory_from_tree(tree):
2360
"""Create a new inventory that has the same contents as a specified tree.
2362
:param tree: Revision tree to create inventory from
2364
entries = tree.iter_entries_by_dir()
2365
inv = Inventory(None, tree.get_revision_id())
2366
for path, inv_entry in entries:
2367
inv.add(inv_entry.copy())