13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
17
17
from bzrlib.lazy_import import lazy_import
18
18
lazy_import(globals(), """
19
19
from bzrlib import (
22
from bzrlib.store import revision
23
from bzrlib.store.revision.knit import KnitRevisionStore
26
revision as _mod_revision,
38
39
from bzrlib.decorators import needs_read_lock, needs_write_lock
39
40
from bzrlib.repository import (
53
58
def __repr__(self):
54
59
return 'KnitParentsProvider(%r)' % self._knit
56
def get_parent_map(self, keys):
57
"""See graph.StackedParentsProvider.get_parent_map"""
59
for revision_id in keys:
60
if revision_id is None:
61
raise ValueError('get_parent_map(None) is not valid')
61
def get_parents(self, revision_ids):
63
for revision_id in revision_ids:
62
64
if revision_id == _mod_revision.NULL_REVISION:
63
parent_map[revision_id] = ()
67
self._knit.get_parents_with_ghosts(revision_id))
68
parents = self._knit.get_parents_with_ghosts(revision_id)
68
69
except errors.RevisionNotPresent:
71
72
if len(parents) == 0:
72
parents = (_mod_revision.NULL_REVISION,)
73
parent_map[revision_id] = parents
77
class _KnitsParentsProvider(object):
79
def __init__(self, knit, prefix=()):
80
"""Create a parent provider for string keys mapped to tuple keys."""
85
return 'KnitsParentsProvider(%r)' % self._knit
87
def get_parent_map(self, keys):
88
"""See graph.StackedParentsProvider.get_parent_map"""
89
parent_map = self._knit.get_parent_map(
90
[self._prefix + (key,) for key in keys])
92
for key, parents in parent_map.items():
95
parents = (_mod_revision.NULL_REVISION,)
97
parents = tuple(parent[-1] for parent in parents)
98
result[revid] = parents
99
for revision_id in keys:
100
if revision_id == _mod_revision.NULL_REVISION:
101
result[revision_id] = ()
73
parents = [_mod_revision.NULL_REVISION]
74
parents_list.append(parents)
105
78
class KnitRepository(MetaDirRepository):
112
85
_commit_builder_class = None
113
86
_serializer = None
115
def __init__(self, _format, a_bzrdir, control_files, _commit_builder_class,
117
MetaDirRepository.__init__(self, _format, a_bzrdir, control_files)
88
def __init__(self, _format, a_bzrdir, control_files, _revision_store,
89
control_store, text_store, _commit_builder_class, _serializer):
90
MetaDirRepository.__init__(self, _format, a_bzrdir, control_files,
91
_revision_store, control_store, text_store)
118
92
self._commit_builder_class = _commit_builder_class
119
93
self._serializer = _serializer
120
94
self._reconcile_fixes_text_parents = True
96
def _warn_if_deprecated(self):
97
# This class isn't deprecated
100
def _inventory_add_lines(self, inv_vf, revid, parents, lines, check_content):
101
return inv_vf.add_lines_with_ghosts(revid, parents, lines,
102
check_content=check_content)[0]
123
105
def _all_revision_ids(self):
124
106
"""See Repository.all_revision_ids()."""
125
return [key[0] for key in self.revisions.keys()]
127
def _activate_new_inventory(self):
128
"""Put a replacement inventory.new into use as inventories."""
129
# Copy the content across
131
t.copy('inventory.new.kndx', 'inventory.kndx')
133
t.copy('inventory.new.knit', 'inventory.knit')
134
except errors.NoSuchFile:
135
# empty inventories knit
136
t.delete('inventory.knit')
137
# delete the temp inventory
138
t.delete('inventory.new.kndx')
140
t.delete('inventory.new.knit')
141
except errors.NoSuchFile:
142
# empty inventories knit
144
# Force index reload (sanity check)
145
self.inventories._index._reset_cache()
146
self.inventories.keys()
148
def _backup_inventory(self):
150
t.copy('inventory.kndx', 'inventory.backup.kndx')
151
t.copy('inventory.knit', 'inventory.backup.knit')
153
def _move_file_id(self, from_id, to_id):
154
t = self._transport.clone('knits')
155
from_rel_url = self.texts._index._mapper.map((from_id, None))
156
to_rel_url = self.texts._index._mapper.map((to_id, None))
157
# We expect both files to always exist in this case.
158
for suffix in ('.knit', '.kndx'):
159
t.rename(from_rel_url + suffix, to_rel_url + suffix)
161
def _remove_file_id(self, file_id):
162
t = self._transport.clone('knits')
163
rel_url = self.texts._index._mapper.map((file_id, None))
164
for suffix in ('.kndx', '.knit'):
166
t.delete(rel_url + suffix)
167
except errors.NoSuchFile:
170
def _temp_inventories(self):
171
result = self._format._get_inventories(self._transport, self,
173
# Reconciling when the output has no revisions would result in no
174
# writes - but we want to ensure there is an inventory for
175
# compatibility with older clients that don't lazy-load.
176
result.get_parent_map([('A',)])
107
# Knits get the revision graph from the index of the revision knit, so
108
# it's always possible even if they're on an unlistable transport.
109
return self._revision_store.all_revision_ids(self.get_transaction())
179
111
def fileid_involved_between_revs(self, from_revid, to_revid):
180
112
"""Find file_id(s) which are involved in the changes between revisions.
202
134
return self._fileid_involved_by_set(changed)
137
def get_ancestry(self, revision_id, topo_sorted=True):
138
"""Return a list of revision-ids integrated by a revision.
140
This is topologically sorted, unless 'topo_sorted' is specified as
143
if _mod_revision.is_null(revision_id):
145
vf = self._get_revision_vf()
147
return [None] + vf.get_ancestry(revision_id, topo_sorted)
148
except errors.RevisionNotPresent:
149
raise errors.NoSuchRevision(self, revision_id)
152
def get_data_stream(self, revision_ids):
153
"""See Repository.get_data_stream."""
154
item_keys = self.item_keys_introduced_by(revision_ids)
155
for knit_kind, file_id, versions in item_keys:
157
if knit_kind == 'file':
158
name = ('file', file_id)
159
knit = self.weave_store.get_weave_or_empty(
160
file_id, self.get_transaction())
161
elif knit_kind == 'inventory':
162
knit = self.get_inventory_weave()
163
elif knit_kind == 'revisions':
164
knit = self._revision_store.get_revision_file(
165
self.get_transaction())
166
elif knit_kind == 'signatures':
167
knit = self._revision_store.get_signature_file(
168
self.get_transaction())
170
raise AssertionError('Unknown knit kind %r' % (knit_kind,))
171
yield name, _get_stream_as_bytes(knit, versions)
205
174
def get_revision(self, revision_id):
206
175
"""Return the Revision object for a named revision"""
207
176
revision_id = osutils.safe_revision_id(revision_id)
208
177
return self.get_revision_reconcile(revision_id)
210
def _refresh_data(self):
211
if not self.is_locked():
213
# Create a new transaction to force all knits to see the scope change.
214
# This is safe because we're outside a write group.
215
self.control_files._finish_transaction()
216
if self.is_write_locked():
217
self.control_files._set_write_transaction()
219
self.control_files._set_read_transaction()
180
def get_revision_graph(self, revision_id=None):
181
"""Return a dictionary containing the revision graph.
183
:param revision_id: The revision_id to get a graph from. If None, then
184
the entire revision graph is returned. This is a deprecated mode of
185
operation and will be removed in the future.
186
:return: a dictionary of revision_id->revision_parents_list.
188
if 'evil' in debug.debug_flags:
190
"get_revision_graph scales with size of history.")
191
# special case NULL_REVISION
192
if revision_id == _mod_revision.NULL_REVISION:
194
a_weave = self._get_revision_vf()
195
if revision_id is None:
196
return a_weave.get_graph()
197
if revision_id not in a_weave:
198
raise errors.NoSuchRevision(self, revision_id)
200
# add what can be reached from revision_id
201
return a_weave.get_graph([revision_id])
204
def get_revision_graph_with_ghosts(self, revision_ids=None):
205
"""Return a graph of the revisions with ghosts marked as applicable.
207
:param revision_ids: an iterable of revisions to graph or None for all.
208
:return: a Graph object with the graph reachable from revision_ids.
210
if 'evil' in debug.debug_flags:
212
"get_revision_graph_with_ghosts scales with size of history.")
213
result = deprecated_graph.Graph()
214
vf = self._get_revision_vf()
215
versions = set(vf.versions())
217
pending = set(self.all_revision_ids())
220
pending = set(revision_ids)
221
# special case NULL_REVISION
222
if _mod_revision.NULL_REVISION in pending:
223
pending.remove(_mod_revision.NULL_REVISION)
224
required = set(pending)
227
revision_id = pending.pop()
228
if not revision_id in versions:
229
if revision_id in required:
230
raise errors.NoSuchRevision(self, revision_id)
232
result.add_ghost(revision_id)
233
# mark it as done so we don't try for it again.
234
done.add(revision_id)
236
parent_ids = vf.get_parents_with_ghosts(revision_id)
237
for parent_id in parent_ids:
238
# is this queued or done ?
239
if (parent_id not in pending and
240
parent_id not in done):
242
pending.add(parent_id)
243
result.add_node(revision_id, parent_ids)
244
done.add(revision_id)
247
def _get_revision_vf(self):
248
""":return: a versioned file containing the revisions."""
249
vf = self._revision_store.get_revision_file(self.get_transaction())
252
def _get_history_vf(self):
253
"""Get a versionedfile whose history graph reflects all revisions.
255
For knit repositories, this is the revision knit.
257
return self._get_revision_vf()
221
259
@needs_write_lock
222
260
def reconcile(self, other=None, thorough=False):
225
263
reconciler = KnitReconciler(self, thorough=thorough)
226
264
reconciler.reconcile()
227
265
return reconciler
267
def revision_parents(self, revision_id):
268
return self._get_revision_vf().get_parents(revision_id)
229
270
def _make_parents_provider(self):
230
return _KnitsParentsProvider(self.revisions)
271
return _KnitParentsProvider(self._get_revision_vf())
232
def _find_inconsistent_revision_parents(self, revisions_iterator=None):
273
def _find_inconsistent_revision_parents(self):
233
274
"""Find revisions with different parent lists in the revision object
234
275
and in the index graph.
236
:param revisions_iterator: None, or an iterator of (revid,
237
Revision-or-None). This iterator controls the revisions checked.
238
277
:returns: an iterator yielding tuples of (revison-id, parents-in-index,
239
278
parents-in-revision).
241
if not self.is_locked():
242
raise AssertionError()
244
if revisions_iterator is None:
245
revisions_iterator = self._iter_revisions(None)
246
for revid, revision in revisions_iterator:
249
parent_map = vf.get_parent_map([(revid,)])
250
parents_according_to_index = tuple(parent[-1] for parent in
251
parent_map[(revid,)])
280
assert self.is_locked()
281
vf = self._get_revision_vf()
282
for index_version in vf.versions():
283
parents_according_to_index = tuple(vf.get_parents_with_ghosts(
285
revision = self.get_revision(index_version)
252
286
parents_according_to_revision = tuple(revision.parent_ids)
253
287
if parents_according_to_index != parents_according_to_revision:
254
yield (revid, parents_according_to_index,
288
yield (index_version, parents_according_to_index,
255
289
parents_according_to_revision)
257
291
def _check_for_inconsistent_revision_parents(self):
289
323
_commit_builder_class = None
290
324
# Set this attribute in derived clases to control the _serializer that the
291
325
# repository objects will have passed to their constructor.
293
def _serializer(self):
294
return xml5.serializer_v5
326
_serializer = xml5.serializer_v5
295
327
# Knit based repositories handle ghosts reasonably well.
296
328
supports_ghosts = True
297
# External lookups are not supported in this format.
298
supports_external_lookups = False
300
supports_chks = False
301
_fetch_order = 'topological'
302
_fetch_uses_deltas = True
305
def _get_inventories(self, repo_transport, repo, name='inventory'):
306
mapper = versionedfile.ConstantMapper(name)
307
index = _mod_knit._KndxIndex(repo_transport, mapper,
308
repo.get_transaction, repo.is_write_locked, repo.is_locked)
309
access = _mod_knit._KnitKeyAccess(repo_transport, mapper)
310
return _mod_knit.KnitVersionedFiles(index, access, annotated=False)
312
def _get_revisions(self, repo_transport, repo):
313
mapper = versionedfile.ConstantMapper('revisions')
314
index = _mod_knit._KndxIndex(repo_transport, mapper,
315
repo.get_transaction, repo.is_write_locked, repo.is_locked)
316
access = _mod_knit._KnitKeyAccess(repo_transport, mapper)
317
return _mod_knit.KnitVersionedFiles(index, access, max_delta_chain=0,
320
def _get_signatures(self, repo_transport, repo):
321
mapper = versionedfile.ConstantMapper('signatures')
322
index = _mod_knit._KndxIndex(repo_transport, mapper,
323
repo.get_transaction, repo.is_write_locked, repo.is_locked)
324
access = _mod_knit._KnitKeyAccess(repo_transport, mapper)
325
return _mod_knit.KnitVersionedFiles(index, access, max_delta_chain=0,
328
def _get_texts(self, repo_transport, repo):
329
mapper = versionedfile.HashEscapedPrefixMapper()
330
base_transport = repo_transport.clone('knits')
331
index = _mod_knit._KndxIndex(base_transport, mapper,
332
repo.get_transaction, repo.is_write_locked, repo.is_locked)
333
access = _mod_knit._KnitKeyAccess(base_transport, mapper)
334
return _mod_knit.KnitVersionedFiles(index, access, max_delta_chain=200,
330
def _get_control_store(self, repo_transport, control_files):
331
"""Return the control store for this repository."""
332
return VersionedFileStore(
335
file_mode=control_files._file_mode,
336
versionedfile_class=knit.KnitVersionedFile,
337
versionedfile_kwargs={'factory':knit.KnitPlainFactory()},
340
def _get_revision_store(self, repo_transport, control_files):
341
"""See RepositoryFormat._get_revision_store()."""
342
versioned_file_store = VersionedFileStore(
344
file_mode=control_files._file_mode,
347
versionedfile_class=knit.KnitVersionedFile,
348
versionedfile_kwargs={'delta':False,
349
'factory':knit.KnitPlainFactory(),
353
return KnitRevisionStore(versioned_file_store)
355
def _get_text_store(self, transport, control_files):
356
"""See RepositoryFormat._get_text_store()."""
357
return self._get_versioned_file_store('knits',
360
versionedfile_class=knit.KnitVersionedFile,
361
versionedfile_kwargs={
362
'create_parent_dir':True,
364
'dir_mode':control_files._dir_mode,
337
368
def initialize(self, a_bzrdir, shared=False):
338
369
"""Create a knit format 1 repository.
342
373
:param shared: If true the repository will be initialized as a shared
345
trace.mutter('creating repository in %s.', a_bzrdir.transport.base)
376
mutter('creating repository in %s.', a_bzrdir.transport.base)
348
379
utf8_files = [('format', self.get_format_string())]
350
381
self._upload_blank_content(a_bzrdir, dirs, files, utf8_files, shared)
351
382
repo_transport = a_bzrdir.get_repository_transport(None)
352
383
control_files = lockable_files.LockableFiles(repo_transport,
353
384
'lock', lockdir.LockDir)
385
control_store = self._get_control_store(repo_transport, control_files)
354
386
transaction = transactions.WriteTransaction()
355
result = self.open(a_bzrdir=a_bzrdir, _found=True)
387
# trigger a write of the inventory store.
388
control_store.get_weave_or_empty('inventory', transaction)
389
_revision_store = self._get_revision_store(repo_transport, control_files)
357
390
# the revision id here is irrelevant: it will not be stored, and cannot
358
# already exist, we do this to create files on disk for older clients.
359
result.inventories.get_parent_map([('A',)])
360
result.revisions.get_parent_map([('A',)])
361
result.signatures.get_parent_map([('A',)])
363
self._run_post_repo_init_hooks(result, a_bzrdir, shared)
392
_revision_store.has_revision_id('A', transaction)
393
_revision_store.get_signature_file(transaction)
394
return self.open(a_bzrdir=a_bzrdir, _found=True)
366
396
def open(self, a_bzrdir, _found=False, _override_transport=None):
367
397
"""See RepositoryFormat.open().
369
399
:param _override_transport: INTERNAL USE ONLY. Allows opening the
370
400
repository at a slightly different url
371
401
than normal. I.e. during 'upgrade'.
374
404
format = RepositoryFormat.find_format(a_bzrdir)
405
assert format.__class__ == self.__class__
375
406
if _override_transport is not None:
376
407
repo_transport = _override_transport
378
409
repo_transport = a_bzrdir.get_repository_transport(None)
379
410
control_files = lockable_files.LockableFiles(repo_transport,
380
411
'lock', lockdir.LockDir)
381
repo = self.repository_class(_format=self,
412
text_store = self._get_text_store(repo_transport, control_files)
413
control_store = self._get_control_store(repo_transport, control_files)
414
_revision_store = self._get_revision_store(repo_transport, control_files)
415
return self.repository_class(_format=self,
382
416
a_bzrdir=a_bzrdir,
383
417
control_files=control_files,
418
_revision_store=_revision_store,
419
control_store=control_store,
420
text_store=text_store,
384
421
_commit_builder_class=self._commit_builder_class,
385
422
_serializer=self._serializer)
386
repo.revisions = self._get_revisions(repo_transport, repo)
387
repo.signatures = self._get_signatures(repo_transport, repo)
388
repo.inventories = self._get_inventories(repo_transport, repo)
389
repo.texts = self._get_texts(repo_transport, repo)
390
repo.chk_bytes = None
391
repo._transport = repo_transport
395
425
class RepositoryFormatKnit1(RepositoryFormatKnit):
507
546
def get_format_description(self):
508
547
"""See RepositoryFormat.get_format_description()."""
509
548
return "Knit repository format 4"
551
def _get_stream_as_bytes(knit, required_versions):
552
"""Generate a serialised data stream.
554
The format is a bencoding of a list. The first element of the list is a
555
string of the format signature, then each subsequent element is a list
556
corresponding to a record. Those lists contain:
563
:returns: a bencoded list.
565
knit_stream = knit.get_data_stream(required_versions)
566
format_signature, data_list, callable = knit_stream
568
data.append(format_signature)
569
for version, options, length, parents in data_list:
570
data.append([version, options, parents, callable(length)])
571
return bencode.bencode(data)