4988.10.3
by John Arbash Meinel
Merge bzr.dev 5007, resolve conflict, update NEWS |
1 |
# Copyright (C) 2005-2010 Canonical Ltd
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
2 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
3 |
# This program is free software; you can redistribute it and/or modify
|
4 |
# it under the terms of the GNU General Public License as published by
|
|
5 |
# the Free Software Foundation; either version 2 of the License, or
|
|
6 |
# (at your option) any later version.
|
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
7 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
8 |
# This program is distributed in the hope that it will be useful,
|
9 |
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
10 |
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
11 |
# GNU General Public License for more details.
|
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
12 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
13 |
# You should have received a copy of the GNU General Public License
|
14 |
# along with this program; if not, write to the Free Software
|
|
4183.7.1
by Sabin Iacob
update FSF mailing address |
15 |
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
1218
by Martin Pool
- fix up import |
16 |
|
1231
by Martin Pool
- more progress on fetch on top of weaves |
17 |
|
18 |
"""Copying of history from one branch to another.
|
|
19 |
||
20 |
The basic plan is that every branch knows the history of everything
|
|
21 |
that has merged into it. As the first step of a merge, pull, or
|
|
22 |
branch operation we copy history from the source into the destination
|
|
23 |
branch.
|
|
24 |
"""
|
|
25 |
||
3350.6.4
by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores. |
26 |
import operator |
27 |
||
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
28 |
from bzrlib.lazy_import import lazy_import |
29 |
lazy_import(globals(), """ |
|
30 |
from bzrlib import (
|
|
31 |
tsort,
|
|
32 |
versionedfile,
|
|
33 |
)
|
|
34 |
""") |
|
1534.1.31
by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo. |
35 |
import bzrlib |
4110.2.4
by Martin Pool
Deprecate passing a pb in to RepoFetcher |
36 |
from bzrlib import ( |
37 |
errors, |
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
38 |
ui, |
4110.2.4
by Martin Pool
Deprecate passing a pb in to RepoFetcher |
39 |
)
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
40 |
from bzrlib.revision import NULL_REVISION |
2094.3.5
by John Arbash Meinel
Fix imports to ensure modules are loaded before they are used |
41 |
from bzrlib.trace import mutter |
1534.1.31
by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo. |
42 |
|
1238
by Martin Pool
- remove a lot of dead code from fetch |
43 |
|
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
44 |
class RepoFetcher(object): |
45 |
"""Pull revisions and texts from one repository to another.
|
|
46 |
||
2592.4.5
by Martin Pool
Add Repository.base on all repositories. |
47 |
This should not be used directly, it's essential a object to encapsulate
|
1534.1.33
by Robert Collins
Move copy_content_into into InterRepository and InterWeaveRepo, and disable the default codepath test as we have optimised paths for all current combinations. |
48 |
the logic in InterRepository.fetch().
|
1260
by Martin Pool
- some updates for fetch/update function |
49 |
"""
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
50 |
|
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
51 |
def __init__(self, to_repository, from_repository, last_revision=None, |
4961.2.3
by Martin Pool
Delete deprecated pb parameter to RepoFetcher |
52 |
find_ghosts=True, fetch_spec=None): |
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
53 |
"""Create a repo fetcher.
|
54 |
||
4110.2.2
by Martin Pool
Remove obsolete comments |
55 |
:param last_revision: If set, try to limit to the data this revision
|
56 |
references.
|
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
57 |
:param find_ghosts: If True search the entire history for ghosts.
|
58 |
"""
|
|
4509.3.18
by Martin Pool
RepoFetcher relies on Repository.fetch to shortcircuit no-op fetches |
59 |
# repository.fetch has the responsibility for short-circuiting
|
60 |
# attempts to copy between a repository and itself.
|
|
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
61 |
self.to_repository = to_repository |
62 |
self.from_repository = from_repository |
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
63 |
self.sink = to_repository._get_sink() |
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
64 |
# must not mutate self._last_revision as its potentially a shared instance
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
65 |
self._last_revision = last_revision |
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
66 |
self._fetch_spec = fetch_spec |
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
67 |
self.find_ghosts = find_ghosts |
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
68 |
self.from_repository.lock_read() |
4110.2.22
by Martin Pool
Re-add mutter calls during fetch |
69 |
mutter("Using fetch logic to copy between %s(%s) and %s(%s)", |
70 |
self.from_repository, self.from_repository._format, |
|
71 |
self.to_repository, self.to_repository._format) |
|
3842.3.5
by Andrew Bennetts
Remove some debugging cruft, make more tests pass. |
72 |
try: |
4110.2.3
by Martin Pool
Remove redundant variable from fetch. |
73 |
self.__fetch() |
3842.3.5
by Andrew Bennetts
Remove some debugging cruft, make more tests pass. |
74 |
finally: |
75 |
self.from_repository.unlock() |
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
76 |
|
77 |
def __fetch(self): |
|
78 |
"""Primary worker function.
|
|
79 |
||
3943.8.1
by Marius Kruger
remove all trailing whitespace from bzr source |
80 |
This initialises all the needed variables, and then fetches the
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
81 |
requested revisions, finally clearing the progress bar.
|
82 |
"""
|
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
83 |
# Roughly this is what we're aiming for fetch to become:
|
84 |
#
|
|
85 |
# missing = self.sink.insert_stream(self.source.get_stream(search))
|
|
86 |
# if missing:
|
|
87 |
# missing = self.sink.insert_stream(self.source.get_items(missing))
|
|
88 |
# assert not missing
|
|
1240
by Martin Pool
- clean up fetch code and add progress bar |
89 |
self.count_total = 0 |
1185.33.55
by Martin Pool
[patch] weave fetch optimizations (Goffredo Baroncelli) |
90 |
self.file_ids_names = {} |
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
91 |
pb = ui.ui_factory.nested_progress_bar() |
4110.2.14
by Martin Pool
Small fetch progress tweaks |
92 |
pb.show_pct = pb.show_count = False |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
93 |
try: |
4110.2.14
by Martin Pool
Small fetch progress tweaks |
94 |
pb.update("Finding revisions", 0, 2) |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
95 |
search = self._revids_to_fetch() |
96 |
if search is None: |
|
97 |
return
|
|
4110.2.14
by Martin Pool
Small fetch progress tweaks |
98 |
pb.update("Fetching revisions", 1, 2) |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
99 |
self._fetch_everything_for_search(search) |
100 |
finally: |
|
101 |
pb.finished() |
|
2535.3.6
by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository. |
102 |
|
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
103 |
def _fetch_everything_for_search(self, search): |
2535.3.6
by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository. |
104 |
"""Fetch all data for the given set of revisions."""
|
2535.3.9
by Andrew Bennetts
More comments. |
105 |
# The first phase is "file". We pass the progress bar for it directly
|
2668.2.8
by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by. |
106 |
# into item_keys_introduced_by, which has more information about how
|
2535.3.9
by Andrew Bennetts
More comments. |
107 |
# that phase is progressing than we do. Progress updates for the other
|
108 |
# phases are taken care of in this function.
|
|
109 |
# XXX: there should be a clear owner of the progress reporting. Perhaps
|
|
2668.2.8
by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by. |
110 |
# item_keys_introduced_by should have a richer API than it does at the
|
111 |
# moment, so that it can feed the progress information back to this
|
|
2535.3.9
by Andrew Bennetts
More comments. |
112 |
# function?
|
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
113 |
if (self.from_repository._format.rich_root_data and |
114 |
not self.to_repository._format.rich_root_data): |
|
115 |
raise errors.IncompatibleRepositories( |
|
116 |
self.from_repository, self.to_repository, |
|
117 |
"different rich-root support") |
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
118 |
pb = ui.ui_factory.nested_progress_bar() |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
119 |
try: |
4110.2.12
by Martin Pool
Add more fetch progress |
120 |
pb.update("Get stream source") |
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
121 |
source = self.from_repository._get_source( |
122 |
self.to_repository._format) |
|
123 |
stream = source.get_stream(search) |
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
124 |
from_format = self.from_repository._format |
4110.2.12
by Martin Pool
Add more fetch progress |
125 |
pb.update("Inserting stream") |
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
126 |
resume_tokens, missing_keys = self.sink.insert_stream( |
5195.3.14
by Parth Malwankar
optimized to use "revisions" insert_record_stream for counting #records |
127 |
stream, from_format, []) |
4257.3.2
by Andrew Bennetts
Check during fetch if we are going to be missing data necessary to calculate altered fileids for stacked revisions. |
128 |
if self.to_repository._fallback_repositories: |
4241.18.1
by Andrew Bennetts
Cherry pick stacking push fix from bzr.dev r4289, make a 'Changes from RC1 to Final' section in NEWS. |
129 |
missing_keys.update( |
4257.4.12
by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher. |
130 |
self._parent_inventories(search.get_keys())) |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
131 |
if missing_keys: |
4110.2.12
by Martin Pool
Add more fetch progress |
132 |
pb.update("Missing keys") |
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
133 |
stream = source.get_stream_for_missing_keys(missing_keys) |
4110.2.12
by Martin Pool
Add more fetch progress |
134 |
pb.update("Inserting missing keys") |
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
135 |
resume_tokens, missing_keys = self.sink.insert_stream( |
5195.3.14
by Parth Malwankar
optimized to use "revisions" insert_record_stream for counting #records |
136 |
stream, from_format, resume_tokens) |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
137 |
if missing_keys: |
138 |
raise AssertionError( |
|
139 |
"second push failed to complete a fetch %r." % ( |
|
140 |
missing_keys,)) |
|
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
141 |
if resume_tokens: |
142 |
raise AssertionError( |
|
143 |
"second push failed to commit the fetch %r." % ( |
|
144 |
resume_tokens,)) |
|
4110.2.12
by Martin Pool
Add more fetch progress |
145 |
pb.update("Finishing stream") |
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
146 |
self.sink.finished() |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
147 |
finally: |
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
148 |
pb.finished() |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
149 |
|
1185.65.30
by Robert Collins
Merge integration. |
150 |
def _revids_to_fetch(self): |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
151 |
"""Determines the exact revisions needed from self.from_repository to
|
152 |
install self._last_revision in self.to_repository.
|
|
153 |
||
154 |
If no revisions need to be fetched, then this just returns None.
|
|
155 |
"""
|
|
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
156 |
if self._fetch_spec is not None: |
157 |
return self._fetch_spec |
|
4110.2.22
by Martin Pool
Re-add mutter calls during fetch |
158 |
mutter('fetch up to rev {%s}', self._last_revision) |
1534.4.50
by Robert Collins
Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running. |
159 |
if self._last_revision is NULL_REVISION: |
160 |
# explicit limit of no revisions needed
|
|
3184.1.9
by Robert Collins
* ``Repository.get_data_stream`` is now deprecated in favour of |
161 |
return None |
4316.1.3
by Jonathan Lange
Don't bother wrapping the NoSuchRevision in an InstallFailed. |
162 |
return self.to_repository.search_missing_revision_ids( |
163 |
self.from_repository, self._last_revision, |
|
164 |
find_ghosts=self.find_ghosts) |
|
1185.64.3
by Goffredo Baroncelli
This patch changes the fetch code. Before, the original code expanded every inventory and |
165 |
|
4257.4.12
by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher. |
166 |
def _parent_inventories(self, revision_ids): |
167 |
# Find all the parent revisions referenced by the stream, but
|
|
4257.4.13
by Andrew Bennetts
Tweak comment. |
168 |
# not present in the stream, and make sure we send their
|
4257.4.12
by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher. |
169 |
# inventories.
|
170 |
parent_maps = self.to_repository.get_parent_map(revision_ids) |
|
171 |
parents = set() |
|
172 |
map(parents.update, parent_maps.itervalues()) |
|
173 |
parents.discard(NULL_REVISION) |
|
174 |
parents.difference_update(revision_ids) |
|
175 |
missing_keys = set(('inventories', rev_id) for rev_id in parents) |
|
176 |
return missing_keys |
|
177 |
||
3565.3.3
by Robert Collins
* Fetching data between repositories that have the same model but no |
178 |
|
1910.2.24
by Aaron Bentley
Got intra-repository fetch working between model1 and 2 for all types |
179 |
class Inter1and2Helper(object): |
1910.2.48
by Aaron Bentley
Update from review comments |
180 |
"""Helper for operations that convert data from model 1 and 2
|
3943.8.1
by Marius Kruger
remove all trailing whitespace from bzr source |
181 |
|
1910.2.48
by Aaron Bentley
Update from review comments |
182 |
This is for use by fetchers and converters.
|
183 |
"""
|
|
184 |
||
5050.32.1
by Andrew Bennetts
Fix fetching more than 100 revisions from non-rich-root to rich-root repositories. |
185 |
# This is a class variable so that the test suite can override it.
|
186 |
known_graph_threshold = 100 |
|
187 |
||
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
188 |
def __init__(self, source): |
1910.2.48
by Aaron Bentley
Update from review comments |
189 |
"""Constructor.
|
190 |
||
191 |
:param source: The repository data comes from
|
|
192 |
"""
|
|
193 |
self.source = source |
|
194 |
||
195 |
def iter_rev_trees(self, revs): |
|
196 |
"""Iterate through RevisionTrees efficiently.
|
|
197 |
||
198 |
Additionally, the inventory's revision_id is set if unset.
|
|
199 |
||
200 |
Trees are retrieved in batches of 100, and then yielded in the order
|
|
201 |
they were requested.
|
|
202 |
||
203 |
:param revs: A list of revision ids
|
|
204 |
"""
|
|
3172.4.4
by Robert Collins
Review feedback. |
205 |
# In case that revs is not a list.
|
206 |
revs = list(revs) |
|
1910.2.48
by Aaron Bentley
Update from review comments |
207 |
while revs: |
208 |
for tree in self.source.revision_trees(revs[:100]): |
|
1910.2.44
by Aaron Bentley
Retrieve only 500 revision trees at once |
209 |
if tree.inventory.revision_id is None: |
210 |
tree.inventory.revision_id = tree.get_revision_id() |
|
211 |
yield tree |
|
1910.2.48
by Aaron Bentley
Update from review comments |
212 |
revs = revs[100:] |
1910.2.44
by Aaron Bentley
Retrieve only 500 revision trees at once |
213 |
|
3380.2.4
by Aaron Bentley
Updates from review |
214 |
def _find_root_ids(self, revs, parent_map, graph): |
215 |
revision_root = {} |
|
1910.2.48
by Aaron Bentley
Update from review comments |
216 |
for tree in self.iter_rev_trees(revs): |
1910.2.18
by Aaron Bentley
Implement creation of knits for tree roots |
217 |
revision_id = tree.inventory.root.revision |
2946.3.3
by John Arbash Meinel
Prefer tree.get_root_id() as more explicit than tree.path2id('') |
218 |
root_id = tree.get_root_id() |
3380.1.3
by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed |
219 |
revision_root[revision_id] = root_id |
220 |
# Find out which parents we don't already know root ids for
|
|
221 |
parents = set() |
|
222 |
for revision_parents in parent_map.itervalues(): |
|
223 |
parents.update(revision_parents) |
|
224 |
parents.difference_update(revision_root.keys() + [NULL_REVISION]) |
|
3380.2.7
by Aaron Bentley
Update docs |
225 |
# Limit to revisions present in the versionedfile
|
3380.1.3
by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed |
226 |
parents = graph.get_parent_map(parents).keys() |
227 |
for tree in self.iter_rev_trees(parents): |
|
228 |
root_id = tree.get_root_id() |
|
229 |
revision_root[tree.get_revision_id()] = root_id |
|
4476.3.11
by Andrew Bennetts
All fetch and interrepo tests passing. |
230 |
return revision_root |
3380.2.4
by Aaron Bentley
Updates from review |
231 |
|
232 |
def generate_root_texts(self, revs): |
|
233 |
"""Generate VersionedFiles for all root ids.
|
|
234 |
||
235 |
:param revs: the revisions to include
|
|
236 |
"""
|
|
237 |
graph = self.source.get_graph() |
|
238 |
parent_map = graph.get_parent_map(revs) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
239 |
rev_order = tsort.topo_sort(parent_map) |
4476.3.11
by Andrew Bennetts
All fetch and interrepo tests passing. |
240 |
rev_id_to_root_id = self._find_root_ids(revs, parent_map, graph) |
3350.6.4
by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores. |
241 |
root_id_order = [(rev_id_to_root_id[rev_id], rev_id) for rev_id in |
242 |
rev_order] |
|
243 |
# Guaranteed stable, this groups all the file id operations together
|
|
244 |
# retaining topological order within the revisions of a file id.
|
|
245 |
# File id splits and joins would invalidate this, but they don't exist
|
|
246 |
# yet, and are unlikely to in non-rich-root environments anyway.
|
|
247 |
root_id_order.sort(key=operator.itemgetter(0)) |
|
248 |
# Create a record stream containing the roots to create.
|
|
5050.32.1
by Andrew Bennetts
Fix fetching more than 100 revisions from non-rich-root to rich-root repositories. |
249 |
if len(revs) > self.known_graph_threshold: |
250 |
graph = self.source.get_known_graph_ancestry(revs) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
251 |
new_roots_stream = _new_root_data_stream( |
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
252 |
root_id_order, rev_id_to_root_id, parent_map, self.source, graph) |
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
253 |
return [('texts', new_roots_stream)] |
254 |
||
255 |
||
256 |
def _new_root_data_stream( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
257 |
root_keys_to_create, rev_id_to_root_id_map, parent_map, repo, graph=None): |
4476.3.69
by Andrew Bennetts
Elaborate some docstrings. |
258 |
"""Generate a texts substream of synthesised root entries.
|
259 |
||
260 |
Used in fetches that do rich-root upgrades.
|
|
261 |
|
|
262 |
:param root_keys_to_create: iterable of (root_id, rev_id) pairs describing
|
|
263 |
the root entries to create.
|
|
264 |
:param rev_id_to_root_id_map: dict of known rev_id -> root_id mappings for
|
|
265 |
calculating the parents. If a parent rev_id is not found here then it
|
|
266 |
will be recalculated.
|
|
267 |
:param parent_map: a parent map for all the revisions in
|
|
268 |
root_keys_to_create.
|
|
269 |
:param graph: a graph to use instead of repo.get_graph().
|
|
270 |
"""
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
271 |
for root_key in root_keys_to_create: |
272 |
root_id, rev_id = root_key |
|
273 |
parent_keys = _parent_keys_for_root_version( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
274 |
root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph) |
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
275 |
yield versionedfile.FulltextContentFactory( |
276 |
root_key, parent_keys, None, '') |
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
277 |
|
278 |
||
279 |
def _parent_keys_for_root_version( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
280 |
root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph=None): |
4476.3.69
by Andrew Bennetts
Elaborate some docstrings. |
281 |
"""Get the parent keys for a given root id.
|
282 |
|
|
283 |
A helper function for _new_root_data_stream.
|
|
284 |
"""
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
285 |
# Include direct parents of the revision, but only if they used the same
|
286 |
# root_id and are heads.
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
287 |
rev_parents = parent_map[rev_id] |
288 |
parent_ids = [] |
|
289 |
for parent_id in rev_parents: |
|
290 |
if parent_id == NULL_REVISION: |
|
291 |
continue
|
|
292 |
if parent_id not in rev_id_to_root_id_map: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
293 |
# We probably didn't read this revision, go spend the extra effort
|
294 |
# to actually check
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
295 |
try: |
296 |
tree = repo.revision_tree(parent_id) |
|
297 |
except errors.NoSuchRevision: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
298 |
# Ghost, fill out rev_id_to_root_id in case we encounter this
|
299 |
# again.
|
|
300 |
# But set parent_root_id to None since we don't really know
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
301 |
parent_root_id = None |
302 |
else: |
|
303 |
parent_root_id = tree.get_root_id() |
|
304 |
rev_id_to_root_id_map[parent_id] = None |
|
4476.3.21
by Andrew Bennetts
Clarify some code and comments, and s/1.17/1.18/ in a few places. |
305 |
# XXX: why not:
|
306 |
# rev_id_to_root_id_map[parent_id] = parent_root_id
|
|
307 |
# memory consumption maybe?
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
308 |
else: |
309 |
parent_root_id = rev_id_to_root_id_map[parent_id] |
|
310 |
if root_id == parent_root_id: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
311 |
# With stacking we _might_ want to refer to a non-local revision,
|
312 |
# but this code path only applies when we have the full content
|
|
313 |
# available, so ghosts really are ghosts, not just the edge of
|
|
314 |
# local data.
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
315 |
parent_ids.append(parent_id) |
316 |
else: |
|
317 |
# root_id may be in the parent anyway.
|
|
318 |
try: |
|
319 |
tree = repo.revision_tree(parent_id) |
|
320 |
except errors.NoSuchRevision: |
|
321 |
# ghost, can't refer to it.
|
|
322 |
pass
|
|
323 |
else: |
|
324 |
try: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
325 |
parent_ids.append(tree.inventory[root_id].revision) |
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
326 |
except errors.NoSuchId: |
327 |
# not in the tree
|
|
328 |
pass
|
|
329 |
# Drop non-head parents
|
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
330 |
if graph is None: |
331 |
graph = repo.get_graph() |
|
332 |
heads = graph.heads(parent_ids) |
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
333 |
selected_ids = [] |
334 |
for parent_id in parent_ids: |
|
335 |
if parent_id in heads and parent_id not in selected_ids: |
|
336 |
selected_ids.append(parent_id) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
337 |
parent_keys = [(root_id, parent_id) for parent_id in selected_ids] |
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
338 |
return parent_keys |