1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
|
Directory fingerprints
======================
.. contents:: :local:
Introduction
------------
The basic idea is that for a directory in a tree (committed or otherwise), we
will have a single scalar value. If these values are the same, the contents of
the subtree under that directory are necessarily the same.
This is intended to help with these use cases, by allowing them to quickly skip
over directories with no relevant changes, and to detect when a directory has
changed:
* diff/status (both local trees and historical trees)
* merge
* log -v
* log on a directory
* commit
Use-case oriented APIs
----------------------
Most of this will be hidden behind the Tree interface. This should cover
``log -v``, ``diff``, ``status``, ``merge`` (and implicit merge during
push, pull, update)::
tree.iter_changes(other_tree)
tree.get_file_lines(file_id) # and get_file, get_file_text
``commit``
~~~~~~~~~~
Commit is similar to ``iter_changes``, but different because it needs to
compare to all the trees. Commit currently needs to compare the working
tree to all the parent trees, which is needed to update the last_modified
field and would be unnecessary if we removed that field (for both files
and directories) and did not store per-file graphs.
This would potentially speed up commit after merge.
Verbose commit also displays the merged files, which does
require looking at all parents of files that aren't identical
to the left-hand parent.
``log``
~~~~~~~
Log is interested in two operations: finding the revisions that touched
anything inside a directory, and getting the differences between
consecutive revisions (possibly filtered to a directory)::
find_touching_revisions(branch, file_id) # should be on Branch?
Log shows the revisions that merged a change. At the moment that is not
included in the per-file graph, and it would also not be visible if the
directories were hashed.
Open questions
--------------
* Is this a good idea at all?
If changing a file changes all its parent directories up to the root it
will cause more churn on commit. (We currently update the all-in-one
inventory, but only have to update one line of it.)
Every time a child changes, we'll get a new node in the per-directory
graph. This is generally useful: it allows bzr log to do the default
mode easily, which is to show all changes under that directory. The
less common operation, ``log --no-recursive`` is still possible by
looking only at when the directory itself was renamed, added or removed.
(That is what the directory graph describes in bzr 0.18 and it is rarely
useful.)
* Should these be hashes or revision ids or something else?
Pros of using hashes: hashes are easy to generate by a foreign branch
plugin (e.g. bzr-svn). They don't need to get recursive last-changed
from the foreign branch, or to walk back through history. They just
need the relevant directory state, which any system we support can
answer.
Hashes converge: if you modify and then modify back, you get the same
hash. This is a pro because you can detect that there were ultimately
no significant changes. And also a con: you cannot use these hashes to form a graph
because they get cycles.
* Are the values unique across the whole tree, or only when comparing
different versions of the same object?
If we use last-changed revisions, then they will be very not unique
across the whole tree. To look up the contents, you must pass a
composite key like ``(file_id, last_changed)``.
If we use hashes they will be same only when the two contain the same
contents. Since we say that file ids must be unique, this
means they will match if and only if they are empty. We might relax
that in future when we introduce path tokens.
* Is it reasonable to assume hashes won't collide?
The odds of SHA-1 hashes colliding "accidentally" are vanishingly small.
It is possible that a `preimage attack`_ against SHA-1 may be discovered
in the future. Since we're not proposing in this document to make
revision-ids be SHA-1, if SHA-1 was obsoleted then we could rewrite the
contents of revisions but would not need to rename revisions. So the
impact of such a migration should just be a format upgrade, and a
recommendation (but not requirement) to re-sign revisions.
.. _`preimage attack`: http://tools.ietf.org/html/rfc4270
* If we use hashes, should it be the hash of the
representation stored for a directory?
In other words, should we pun the representation of the directory with
the form used for validation.
If there's some data stored that's not in the hash it's problematic.
The hash in no longer (effectively) uniquely identifies the
representation.
It is desirable that we have a hash that covers all data, to guard
against bugs, transmission errors, or users trying to hand-hack files.
Since we need one hash of everything in the tree, perhaps we should also
use it for the fingerprint.
Testaments explicitly separate the form used for hashing/signing from
the form used for storage. This allows us to change the storage form
without breaking existing GPG signatures. The downside is that we need
to do work O(tree) to make a testament, and this slows down signing,
verifying and generating bundles. It also means that there is some
stored data which is not protected by the signature: this data is less
important, but corruption of it would still cause problems.
We have encountered some specific problems with disagreement between
inventories as to the last-change of files, which is currently unsigned.
These problems can be introduced by ghosts.
If we hash the representation, there is still a way to support old
signatures, assuming that we never discard irreplaceable information.
The signature should say what format it applies to (similar to
testaments), and we could transform in memory the tree back to that
format.
* Is hashing substantially slower than other possible approaches?
We already hash all the plain files. Except in unusual cases, the
directory metadata will be substantially smaller: perhaps 200:1 as a
rule of thumb.
When building a bzr tree, we spend on the order of 100ms hashing all the
source lines to validate them (about 13MB of source).
* Can you calculate one from a directory in the working tree? Without a basis?
This seems possible with either hashes or revision ids.
Using last_changed means that calculating the fingerprint from a working
tree necessarily requires reading the inventory for the basis
revision, so that we know when unchanged files were last changed. With
hashes we could calculate them using the working tree information alone.
It's true that we will often then compare that information to the basis
tree (e.g. for simple ``bzr diff``), but we may only have to compare at
the top level, and sometimes we're comparing to a
different tree. This also touches on whether we should store
``last_modified`` for files, rather than directories.
For revision ids we need to assign a value to use for uncommitted
changes, but see below about the problems of this.
In some ways it would be elegant to say (hypothetical)::
wt.get_root().get_last_modified() == branch.get_last_revision()
to know that nothing was changed; but this may not be much better than
::
wt.get_root().get_hash() ==
branch.get_basis().get_root().get_hash()
* Can you use this to compare (directories from) two working trees?
If you can generate it from a working tree, you should be able to use it
to compare them.
This does rule out for example using ``last_modified=None`` or
``='current:'`` to mean "changed in the working tree." Even if this is
not supported there seems some risk that we would get the same
fingerprint for trees that are actually different.
We could assign a
hypothetical revision id to the tree for uncommitted files. In that
case there is some risk that the not-yet-committed id would become
visible or committed.
* Can we use an "approximate basis"?
When using radix trees, you may need context beyond the specific
directory being compared.
* Can you get the fingerprint of parents directories with only selected file ids
taken from the working tree?
With hashes, we'd want to carry through the unselected files and
directories from the values they had in the parent revision.
* Are unbalanced trees a significant problem? Trees can be unbalanced by having
many directories (deep or wide), or many files per directory.
For small trees like bzr, 744 of 874 are in the bzrlib subtree. In
general, larger trees are more balanced, because humans, editors and
other tools have trouble managing very unbalanced trees. But there are
exceptions: Aaron has one tree with 20,000 generated but versioned
entries in one directory.
* Should we use a radix tree approach where fingerprints are calculated on a synthetic
tree that is by definition balanced, even when the actual tree is unbalanced?
* What are the specific advantages of using recursive-last-modified rather than
hashes?
It may be a smaller step change.
It's a bidirectional link: given a directory text identifier ``(file_id,
last_changed)`` you can look up the revision that last changed it.
From the preceding, even without the per-file graph you can skip through
the history of this file: go to the last-changed revision, look at all
its parents and repeat.
* Is it a smaller change to use recursive-last-modified on directories?
Probably yes:
1. We can just put it into the current inventory format without changing
anything else.
By contrast to use a hash we'd have to either split up the inventory
as stored, or change the sort order for the inventory, or synthesize
per-directory inventories in memory for hashing.
However, xml is somewhat redundant and slow to parse/generate; and
reading the whole thing before comparing some sections is only a
partial win. It may be a smaller change but we'd be preserving
things we want to change.
1. At present we rarely hash storage representations, only file texts.
This is not a large technical change, but it is a conceptual change.
This has some consequences for how we can upgrade it in future: all
the changed directories need to be rewritten up to the revision level.
1. If we address directories by hash we need hash-addressed
storage.
1. If we address directories by hash then for consistency we'd probably
(not necessarily) want to address file texts by hash.
1. The per-file graph can't be indexed by hash because they can converge, so we
need to either rework or dispose of the per-file graph.
* Any possibilities for avoiding hashes recurring?
1. Hash along with an identification of the parents (as in hg). Then you
can't convert a tree without all its basis trees, and there is still
convergence when the same merge is done by two people, and you can't
create it directly from the working tree.
1. Include last-modified revision id in the hash.
1. Index by ``(revision, hash)`` or vice versa.
1. Store a per-file graph and allow it to have repeated keys. The graph
would tell you about all the parent texts ever seen; you would need
to use revision graph information to resolve ambiguities.
* What are the specific disadvantages of using recursive-last-modified rather than
hashes?
To calculate the last-changed revision, given the last-changed
information of the contained files, you need to look at the revision
graph. They're not enough because you need to know the relations
between the mentioned revisions. In a merge it's possible the correct
directory last-modified will not be the same as that of any of the files
within it. This can also happen when a file is removed (deleted or
renamed) from a directory.
* Should we split up storage of the inventories?
This is not quite the same but connected.
* How does this relate to per-file/per-directory hashes?
If the version of a file or directory is identified by a hash, we can't
use that to point into a per-file graph. We can have a graph indexed by
``(file_id, hash, revision_id)``. The last-modified could be stored as
part of this graph.
The graph would no longer be core data; it could be always present but
might be rebuilt. Treating it as non-core data may make some changes
like shallow branches easier?
* How do you ask a tree for a given text?
Right now we say ::
revision_tree.get_file_lines(file_id)
so the choice of storage is hidden behind the revision tree: it could be
accessed by ``(file_id, last_changed)`` or by hash or otherwise.
At the moment the Repository exports a friend api to RevisionTree,
currently usually talking in VersionedFiles.
We probably wouldn't want Repository to expose a ``get_text_for_sha1()``
interface because that would be very difficult to support on old
repositories or on foreign branches.
Conclusions
-----------
Design changes
--------------
API changes
-----------
..
vim: filetype=rst textwidth=78 expandtab spelllang=en spell
|