103
109
# the default CommitBuilder does not manage trees whose root is versioned.
104
110
_versioned_root = False
106
def __init__(self, repository, parents, config, timestamp=None,
112
def __init__(self, repository, parents, config_stack, timestamp=None,
107
113
timezone=None, committer=None, revprops=None,
108
114
revision_id=None, lossy=False):
109
115
super(VersionedFileCommitBuilder, self).__init__(repository,
110
parents, config, timestamp, timezone, committer, revprops,
116
parents, config_stack, timestamp, timezone, committer, revprops,
111
117
revision_id, lossy)
113
119
basis_id = self.parents[0]
194
200
revision_id=self._new_revision_id,
195
201
properties=self._revprops)
196
202
rev.parent_ids = self.parents
197
self.repository.add_revision(self._new_revision_id, rev,
198
self.new_inventory, self._config)
203
if self._config_stack.get('create_signatures') == _mod_config.SIGN_ALWAYS:
204
testament = Testament(rev, self.revision_tree())
205
plaintext = testament.as_short_text()
206
self.repository.store_revision_signature(
207
gpg.GPGStrategy(self._config_stack), plaintext,
208
self._new_revision_id)
209
self.repository._add_revision(rev)
199
210
self._ensure_fallback_inventories()
200
211
self.repository.commit_write_group()
201
212
return self._new_revision_id
419
430
return None, False, None
420
431
# XXX: Friction: parent_candidates should return a list not a dict
421
432
# so that we don't have to walk the inventories again.
422
parent_candiate_entries = ie.parent_candidates(parent_invs)
423
head_set = self._heads(ie.file_id, parent_candiate_entries.keys())
433
parent_candidate_entries = ie.parent_candidates(parent_invs)
434
head_set = self._heads(ie.file_id, parent_candidate_entries.keys())
425
436
for inv in parent_invs:
426
437
if inv.has_id(ie.file_id):
443
454
# There is a single head, look it up for comparison
444
parent_entry = parent_candiate_entries[heads[0]]
455
parent_entry = parent_candidate_entries[heads[0]]
445
456
# if the non-content specific data has changed, we'll be writing a
447
458
if (parent_entry.parent_id != ie.parent_id or
559
570
:param iter_changes: An iter_changes iterator with the changes to apply
560
571
to basis_revision_id. The iterator must not include any items with
561
572
a current kind of None - missing items must be either filtered out
562
or errored-on beefore record_iter_changes sees the item.
573
or errored-on before record_iter_changes sees the item.
563
574
:param _entry_factory: Private method to bind entry_factory locally for
565
576
:return: A generator of (file_id, relpath, fs_hash) tuples for use with
593
604
_mod_revision.NULL_REVISION))
594
605
# The basis inventory from a repository
596
basis_inv = revtrees[0].inventory
607
basis_tree = revtrees[0]
598
basis_inv = self.repository.revision_tree(
599
_mod_revision.NULL_REVISION).inventory
609
basis_tree = self.repository.revision_tree(
610
_mod_revision.NULL_REVISION)
611
basis_inv = basis_tree.root_inventory
600
612
if len(self.parents) > 0:
601
613
if basis_revision_id != self.parents[0] and not ghost_basis:
603
615
"arbitrary basis parents not yet supported with merges")
604
616
for revtree in revtrees[1:]:
605
for change in revtree.inventory._make_delta(basis_inv):
617
for change in revtree.root_inventory._make_delta(basis_inv):
606
618
if change[1] is None:
607
619
# Not present in this parent.
1010
1022
# return a new inventory, but as there is no revision tree cache in
1011
1023
# repository this is safe for now - RBC 20081013
1012
1024
if basis_inv is None:
1013
basis_inv = basis_tree.inventory
1025
basis_inv = basis_tree.root_inventory
1014
1026
basis_inv.apply_delta(delta)
1015
1027
basis_inv.revision_id = new_revision_id
1016
1028
return (self.add_inventory(new_revision_id, basis_inv, parents),
1027
1039
self.inventories._access.flush()
1030
def add_revision(self, revision_id, rev, inv=None, config=None):
1042
def add_revision(self, revision_id, rev, inv=None):
1031
1043
"""Add rev to the revision store as revision_id.
1033
1045
:param revision_id: the revision id to use.
1034
1046
:param rev: The revision object.
1035
1047
:param inv: The inventory for the revision. if None, it will be looked
1036
1048
up in the inventory storer
1037
:param config: If None no digital signature will be created.
1038
If supplied its signature_needed method will be used
1039
to determine if a signature should be made.
1041
1050
# TODO: jam 20070210 Shouldn't we check rev.revision_id and
1042
1051
# rev.parent_ids?
1043
1052
_mod_revision.check_not_reserved_id(revision_id)
1044
if config is not None and config.signature_needed():
1046
inv = self.get_inventory(revision_id)
1047
tree = InventoryRevisionTree(self, inv, revision_id)
1048
testament = Testament(rev, tree)
1049
plaintext = testament.as_short_text()
1050
self.store_revision_signature(
1051
gpg.GPGStrategy(config), plaintext, revision_id)
1052
1053
# check inventory present
1053
1054
if not self.inventories.get_parent_map([(revision_id,)]):
1054
1055
if inv is None:
1087
1088
keys = {'chk_bytes':set(), 'inventories':set(), 'texts':set()}
1088
1089
kinds = ['chk_bytes', 'texts']
1089
1090
count = len(checker.pending_keys)
1090
bar.update("inventories", 0, 2)
1091
bar.update(gettext("inventories"), 0, 2)
1091
1092
current_keys = checker.pending_keys
1092
1093
checker.pending_keys = {}
1093
1094
# Accumulate current checks.
1182
1183
'sha1 mismatch: %s has sha1 %s expected %s referenced by %s' %
1183
1184
(record.key, sha1, item_data[1], item_data[2]))
1187
def _eliminate_revisions_not_present(self, revision_ids):
1188
"""Check every revision id in revision_ids to see if we have it.
1190
Returns a set of the present revisions.
1193
graph = self.get_graph()
1194
parent_map = graph.get_parent_map(revision_ids)
1195
# The old API returned a list, should this actually be a set?
1196
return parent_map.keys()
1185
1198
def __init__(self, _format, a_bzrdir, control_files):
1186
1199
"""Instantiate a VersionedFileRepository.
1188
1201
:param _format: The format of the repository on disk.
1189
:param a_bzrdir: The BzrDir of the repository.
1202
:param controldir: The ControlDir of the repository.
1190
1203
:param control_files: Control files to use for locking, etc.
1192
1205
# In the future we will have a single api for all stores for
1194
1207
# this construct will accept instances of those things.
1195
1208
super(VersionedFileRepository, self).__init__(_format, a_bzrdir,
1210
self._transport = control_files._transport
1211
self.base = self._transport.base
1198
1213
self._reconcile_does_inventory_gc = True
1199
1214
self._reconcile_fixes_text_parents = False
1204
1219
# rather copying them?
1205
1220
self._safe_to_return_from_cache = False
1222
def fetch(self, source, revision_id=None, find_ghosts=False,
1224
"""Fetch the content required to construct revision_id from source.
1226
If revision_id is None and fetch_spec is None, then all content is
1229
fetch() may not be used when the repository is in a write group -
1230
either finish the current write group before using fetch, or use
1231
fetch before starting the write group.
1233
:param find_ghosts: Find and copy revisions in the source that are
1234
ghosts in the target (and not reachable directly by walking out to
1235
the first-present revision in target from revision_id).
1236
:param revision_id: If specified, all the content needed for this
1237
revision ID will be copied to the target. Fetch will determine for
1238
itself which content needs to be copied.
1239
:param fetch_spec: If specified, a SearchResult or
1240
PendingAncestryResult that describes which revisions to copy. This
1241
allows copying multiple heads at once. Mutually exclusive with
1244
if fetch_spec is not None and revision_id is not None:
1245
raise AssertionError(
1246
"fetch_spec and revision_id are mutually exclusive.")
1247
if self.is_in_write_group():
1248
raise errors.InternalBzrError(
1249
"May not fetch while in a write group.")
1250
# fast path same-url fetch operations
1251
# TODO: lift out to somewhere common with RemoteRepository
1252
# <https://bugs.launchpad.net/bzr/+bug/401646>
1253
if (self.has_same_location(source)
1254
and fetch_spec is None
1255
and self._has_same_fallbacks(source)):
1256
# check that last_revision is in 'from' and then return a
1258
if (revision_id is not None and
1259
not _mod_revision.is_null(revision_id)):
1260
self.get_revision(revision_id)
1262
inter = InterRepository.get(source, self)
1263
if (fetch_spec is not None and
1264
not getattr(inter, "supports_fetch_spec", False)):
1265
raise errors.UnsupportedOperation(
1266
"fetch_spec not supported for %r" % inter)
1267
return inter.fetch(revision_id=revision_id,
1268
find_ghosts=find_ghosts, fetch_spec=fetch_spec)
1207
1270
@needs_read_lock
1208
1271
def gather_stats(self, revid=None, committers=None):
1209
1272
"""See Repository.gather_stats()."""
1218
1281
# result['size'] = t
1221
def get_commit_builder(self, branch, parents, config, timestamp=None,
1284
def get_commit_builder(self, branch, parents, config_stack, timestamp=None,
1222
1285
timezone=None, committer=None, revprops=None,
1223
1286
revision_id=None, lossy=False):
1224
1287
"""Obtain a CommitBuilder for this repository.
1226
1289
:param branch: Branch to commit to.
1227
1290
:param parents: Revision ids of the parents of the new revision.
1228
:param config: Configuration to use.
1291
:param config_stack: Configuration stack to use.
1229
1292
:param timestamp: Optional timestamp recorded for commit.
1230
1293
:param timezone: Optional timezone for timestamp.
1231
1294
:param committer: Optional committer to set for commit.
1238
1301
raise errors.BzrError("Cannot commit directly to a stacked branch"
1239
1302
" in pre-2a formats. See "
1240
1303
"https://bugs.launchpad.net/bzr/+bug/375013 for details.")
1241
result = self._commit_builder_class(self, parents, config,
1304
result = self._commit_builder_class(self, parents, config_stack,
1242
1305
timestamp, timezone, committer, revprops, revision_id,
1244
1307
self.start_write_group()
1500
1563
text_keys[(file_id, revision_id)] = callable_data
1501
1564
for record in self.texts.get_record_stream(text_keys, 'unordered', True):
1502
1565
if record.storage_kind == 'absent':
1503
raise errors.RevisionNotPresent(record.key, self)
1566
raise errors.RevisionNotPresent(record.key[1], record.key[0])
1504
1567
yield text_keys[record.key], record.get_bytes_as('chunked')
1506
1569
def _generate_text_key_index(self, text_key_references=None,
1556
1619
batch_size = 10 # should be ~150MB on a 55K path tree
1557
1620
batch_count = len(revision_order) / batch_size + 1
1558
1621
processed_texts = 0
1559
pb.update("Calculating text parents", processed_texts, text_count)
1622
pb.update(gettext("Calculating text parents"), processed_texts, text_count)
1560
1623
for offset in xrange(batch_count):
1561
1624
to_query = revision_order[offset * batch_size:(offset + 1) *
1565
1628
for revision_id in to_query:
1566
1629
parent_ids = ancestors[revision_id]
1567
1630
for text_key in revision_keys[revision_id]:
1568
pb.update("Calculating text parents", processed_texts)
1631
pb.update(gettext("Calculating text parents"), processed_texts)
1569
1632
processed_texts += 1
1570
1633
candidate_parents = []
1571
1634
for parent_id in parent_ids:
1641
1704
num_file_ids = len(file_ids)
1642
1705
for file_id, altered_versions in file_ids.iteritems():
1643
1706
if pb is not None:
1644
pb.update("Fetch texts", count, num_file_ids)
1707
pb.update(gettext("Fetch texts"), count, num_file_ids)
1646
1709
yield ("file", file_id, altered_versions)
1684
1747
if ((None in revision_ids)
1685
1748
or (_mod_revision.NULL_REVISION in revision_ids)):
1686
1749
raise ValueError('cannot get null revision inventory')
1687
return self._iter_inventories(revision_ids, ordering)
1750
for inv, revid in self._iter_inventories(revision_ids, ordering):
1752
raise errors.NoSuchRevision(self, revid)
1689
1755
def _iter_inventories(self, revision_ids, ordering):
1690
1756
"""single-document based inventory iteration."""
1691
1757
inv_xmls = self._iter_inventory_xmls(revision_ids, ordering)
1692
1758
for text, revision_id in inv_xmls:
1693
yield self._deserialise_inventory(revision_id, text)
1760
yield None, revision_id
1762
yield self._deserialise_inventory(revision_id, text), revision_id
1695
1764
def _iter_inventory_xmls(self, revision_ids, ordering):
1696
1765
if ordering is None:
1715
1784
yield ''.join(chunks), record.key[-1]
1717
raise errors.NoSuchRevision(self, record.key)
1786
yield None, record.key[-1]
1718
1787
if order_as_requested:
1719
1788
# Yield as many results as we can while preserving order.
1720
1789
while next_key in text_chunks:
1749
1818
def _get_inventory_xml(self, revision_id):
1750
1819
"""Get serialized inventory as a string."""
1751
1820
texts = self._iter_inventory_xmls([revision_id], 'unordered')
1753
text, revision_id = texts.next()
1754
except StopIteration:
1755
raise errors.HistoryMissing(self, 'inventory', revision_id)
1821
text, revision_id = texts.next()
1823
raise errors.NoSuchRevision(self, revision_id)
1758
1826
@needs_read_lock
1833
1901
"""Return the graph walker for text revisions."""
1834
1902
return graph.Graph(self.texts)
1904
def revision_ids_to_search_result(self, result_set):
1905
"""Convert a set of revision ids to a graph SearchResult."""
1906
result_parents = set()
1907
for parents in self.get_graph().get_parent_map(
1908
result_set).itervalues():
1909
result_parents.update(parents)
1910
included_keys = result_set.intersection(result_parents)
1911
start_keys = result_set.difference(included_keys)
1912
exclude_keys = result_parents.difference(result_set)
1913
result = vf_search.SearchResult(start_keys, exclude_keys,
1914
len(result_set), result_set)
1836
1917
def _get_versioned_file_checker(self, text_key_references=None,
1837
1918
ancestors=None):
1838
1919
"""Return an object suitable for checking versioned files.
2361
2442
invs_sent_so_far = set([_mod_revision.NULL_REVISION])
2362
2443
inventory_cache = lru_cache.LRUCache(50)
2363
2444
null_inventory = from_repo.revision_tree(
2364
_mod_revision.NULL_REVISION).inventory
2445
_mod_revision.NULL_REVISION).root_inventory
2365
2446
# XXX: ideally the rich-root/tree-refs flags would be per-revision, not
2366
2447
# per-repo (e.g. streaming a non-rich-root revision out of a rich-root
2367
2448
# repo back into a non-rich-root repo ought to be allowed)
2452
2533
self.text_index.iterkeys()])
2453
2534
# text keys is now grouped by file_id
2454
2535
n_versions = len(self.text_index)
2455
progress_bar.update('loading text store', 0, n_versions)
2536
progress_bar.update(gettext('loading text store'), 0, n_versions)
2456
2537
parent_map = self.repository.texts.get_parent_map(self.text_index)
2457
2538
# On unlistable transports this could well be empty/error...
2458
2539
text_keys = self.repository.texts.keys()
2459
2540
unused_keys = frozenset(text_keys) - set(self.text_index)
2460
2541
for num, key in enumerate(self.text_index.iterkeys()):
2461
progress_bar.update('checking text graph', num, n_versions)
2542
progress_bar.update(gettext('checking text graph'), num, n_versions)
2462
2543
correct_parents = self.calculate_file_version_parents(key)
2464
2545
knit_parents = parent_map[key]
2485
2568
content is copied.
2488
ui.ui_factory.warn_experimental_format_fetch(self)
2571
if self.target._format.experimental:
2572
ui.ui_factory.show_user_warning('experimental_format_fetch',
2573
from_format=self.source._format,
2574
to_format=self.target._format)
2489
2575
from bzrlib.fetch import RepoFetcher
2490
2576
# See <https://launchpad.net/bugs/456077> asking for a warning here
2491
2577
if self.source._format.network_name() != self.target._format.network_name():
2552
2638
searcher.stop_searching_any(stop_revs)
2553
2639
if searcher_exhausted:
2555
return searcher.get_result()
2641
(started_keys, excludes, included_keys) = searcher.get_state()
2642
return vf_search.SearchResult(started_keys, excludes,
2643
len(included_keys), included_keys)
2557
2645
@needs_read_lock
2558
2646
def search_missing_revision_ids(self,
2709
2797
# Generate deltas against each tree, to find the shortest.
2798
# FIXME: Support nested trees
2710
2799
texts_possibly_new_in_tree = set()
2711
2800
for basis_id, basis_tree in possible_trees:
2712
delta = tree.inventory._make_delta(basis_tree.inventory)
2801
delta = tree.root_inventory._make_delta(basis_tree.root_inventory)
2713
2802
for old_path, new_path, file_id, new_entry in delta:
2714
2803
if new_path is None:
2715
2804
# This file_id isn't present in the new rev, so we don't
2752
2841
parents_parents = [key[-1] for key in parents_parents_keys]
2753
2842
basis_id = _mod_revision.NULL_REVISION
2754
2843
basis_tree = self.source.revision_tree(basis_id)
2755
delta = parent_tree.inventory._make_delta(basis_tree.inventory)
2844
delta = parent_tree.root_inventory._make_delta(
2845
basis_tree.root_inventory)
2756
2846
self.target.add_inventory_by_delta(
2757
2847
basis_id, delta, current_revision_id, parents_parents)
2758
2848
cache[current_revision_id] = parent_tree
2817
2907
kind = entry.kind
2818
2908
texts_possibly_new_in_tree.add((file_id, entry.revision))
2819
2909
for basis_id, basis_tree in possible_trees:
2820
basis_inv = basis_tree.inventory
2910
basis_inv = basis_tree.root_inventory
2821
2911
for file_key in list(texts_possibly_new_in_tree):
2822
2912
file_id, file_revision = file_key
2905
2995
for offset in range(0, len(revision_ids), batch_size):
2906
2996
self.target.start_write_group()
2908
pb.update('Transferring revisions', offset,
2998
pb.update(gettext('Transferring revisions'), offset,
2909
2999
len(revision_ids))
2910
3000
batch = revision_ids[offset:offset+batch_size]
2911
3001
basis_id = self._fetch_batch(batch, basis_id, cache)
2930
3020
revision_ids = fetch_spec.get_keys()
2932
3022
revision_ids = None
2933
ui.ui_factory.warn_experimental_format_fetch(self)
3023
if self.source._format.experimental:
3024
ui.ui_factory.show_user_warning('experimental_format_fetch',
3025
from_format=self.source._format,
3026
to_format=self.target._format)
2934
3027
if (not self.source.supports_rich_root()
2935
3028
and self.target.supports_rich_root()):
2936
3029
self._converting_to_rich_root = True
3031
3124
_install_revision(repository, revision, revision_tree, signature,
3032
3125
inventory_cache)
3033
3126
if pb is not None:
3034
pb.update('Transferring revisions', n + 1, num_revisions)
3127
pb.update(gettext('Transferring revisions'), n + 1, num_revisions)
3036
3129
repository.abort_write_group()
3052
3145
parent_trees[p_id] = repository.revision_tree(
3053
3146
_mod_revision.NULL_REVISION)
3055
inv = revision_tree.inventory
3148
# FIXME: Support nested trees
3149
inv = revision_tree.root_inventory
3056
3150
entries = inv.iter_entries()
3057
3151
# backwards compatibility hack: skip the root id.
3058
3152
if not repository.supports_rich_root():