4110.2.2
by Martin Pool
Remove obsolete comments |
1 |
# Copyright (C) 2005, 2006, 2008, 2009 Canonical Ltd
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
2 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
3 |
# This program is free software; you can redistribute it and/or modify
|
4 |
# it under the terms of the GNU General Public License as published by
|
|
5 |
# the Free Software Foundation; either version 2 of the License, or
|
|
6 |
# (at your option) any later version.
|
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
7 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
8 |
# This program is distributed in the hope that it will be useful,
|
9 |
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
10 |
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
11 |
# GNU General Public License for more details.
|
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
12 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
13 |
# You should have received a copy of the GNU General Public License
|
14 |
# along with this program; if not, write to the Free Software
|
|
4183.7.1
by Sabin Iacob
update FSF mailing address |
15 |
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
1218
by Martin Pool
- fix up import |
16 |
|
1231
by Martin Pool
- more progress on fetch on top of weaves |
17 |
|
18 |
"""Copying of history from one branch to another.
|
|
19 |
||
20 |
The basic plan is that every branch knows the history of everything
|
|
21 |
that has merged into it. As the first step of a merge, pull, or
|
|
22 |
branch operation we copy history from the source into the destination
|
|
23 |
branch.
|
|
24 |
"""
|
|
25 |
||
3350.6.4
by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores. |
26 |
import operator |
27 |
||
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
28 |
from bzrlib.lazy_import import lazy_import |
29 |
lazy_import(globals(), """ |
|
30 |
from bzrlib import (
|
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
31 |
graph as _mod_graph,
|
32 |
static_tuple,
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
33 |
tsort,
|
34 |
versionedfile,
|
|
35 |
)
|
|
36 |
""") |
|
1534.1.31
by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo. |
37 |
import bzrlib |
4110.2.4
by Martin Pool
Deprecate passing a pb in to RepoFetcher |
38 |
from bzrlib import ( |
39 |
errors, |
|
40 |
symbol_versioning, |
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
41 |
ui, |
4110.2.4
by Martin Pool
Deprecate passing a pb in to RepoFetcher |
42 |
)
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
43 |
from bzrlib.revision import NULL_REVISION |
2094.3.5
by John Arbash Meinel
Fix imports to ensure modules are loaded before they are used |
44 |
from bzrlib.trace import mutter |
1534.1.31
by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo. |
45 |
|
1238
by Martin Pool
- remove a lot of dead code from fetch |
46 |
|
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
47 |
class RepoFetcher(object): |
48 |
"""Pull revisions and texts from one repository to another.
|
|
49 |
||
2592.4.5
by Martin Pool
Add Repository.base on all repositories. |
50 |
This should not be used directly, it's essential a object to encapsulate
|
1534.1.33
by Robert Collins
Move copy_content_into into InterRepository and InterWeaveRepo, and disable the default codepath test as we have optimised paths for all current combinations. |
51 |
the logic in InterRepository.fetch().
|
1260
by Martin Pool
- some updates for fetch/update function |
52 |
"""
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
53 |
|
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
54 |
def __init__(self, to_repository, from_repository, last_revision=None, |
55 |
pb=None, find_ghosts=True, fetch_spec=None): |
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
56 |
"""Create a repo fetcher.
|
57 |
||
4110.2.2
by Martin Pool
Remove obsolete comments |
58 |
:param last_revision: If set, try to limit to the data this revision
|
59 |
references.
|
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
60 |
:param find_ghosts: If True search the entire history for ghosts.
|
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
61 |
:param pb: ProgressBar object to use; deprecated and ignored.
|
62 |
This method will just create one on top of the stack.
|
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
63 |
"""
|
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
64 |
if pb is not None: |
65 |
symbol_versioning.warn( |
|
66 |
symbol_versioning.deprecated_in((1, 14, 0)) |
|
67 |
% "pb parameter to RepoFetcher.__init__") |
|
68 |
# and for simplicity it is in fact ignored
|
|
4509.3.18
by Martin Pool
RepoFetcher relies on Repository.fetch to shortcircuit no-op fetches |
69 |
# repository.fetch has the responsibility for short-circuiting
|
70 |
# attempts to copy between a repository and itself.
|
|
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
71 |
self.to_repository = to_repository |
72 |
self.from_repository = from_repository |
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
73 |
self.sink = to_repository._get_sink() |
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
74 |
# must not mutate self._last_revision as its potentially a shared instance
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
75 |
self._last_revision = last_revision |
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
76 |
self._fetch_spec = fetch_spec |
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
77 |
self.find_ghosts = find_ghosts |
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
78 |
self.from_repository.lock_read() |
4110.2.22
by Martin Pool
Re-add mutter calls during fetch |
79 |
mutter("Using fetch logic to copy between %s(%s) and %s(%s)", |
80 |
self.from_repository, self.from_repository._format, |
|
81 |
self.to_repository, self.to_repository._format) |
|
3842.3.5
by Andrew Bennetts
Remove some debugging cruft, make more tests pass. |
82 |
try: |
4110.2.3
by Martin Pool
Remove redundant variable from fetch. |
83 |
self.__fetch() |
3842.3.5
by Andrew Bennetts
Remove some debugging cruft, make more tests pass. |
84 |
finally: |
85 |
self.from_repository.unlock() |
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
86 |
|
87 |
def __fetch(self): |
|
88 |
"""Primary worker function.
|
|
89 |
||
3943.8.1
by Marius Kruger
remove all trailing whitespace from bzr source |
90 |
This initialises all the needed variables, and then fetches the
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
91 |
requested revisions, finally clearing the progress bar.
|
92 |
"""
|
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
93 |
# Roughly this is what we're aiming for fetch to become:
|
94 |
#
|
|
95 |
# missing = self.sink.insert_stream(self.source.get_stream(search))
|
|
96 |
# if missing:
|
|
97 |
# missing = self.sink.insert_stream(self.source.get_items(missing))
|
|
98 |
# assert not missing
|
|
1240
by Martin Pool
- clean up fetch code and add progress bar |
99 |
self.count_total = 0 |
1185.33.55
by Martin Pool
[patch] weave fetch optimizations (Goffredo Baroncelli) |
100 |
self.file_ids_names = {} |
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
101 |
pb = ui.ui_factory.nested_progress_bar() |
4110.2.14
by Martin Pool
Small fetch progress tweaks |
102 |
pb.show_pct = pb.show_count = False |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
103 |
try: |
4110.2.14
by Martin Pool
Small fetch progress tweaks |
104 |
pb.update("Finding revisions", 0, 2) |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
105 |
search = self._revids_to_fetch() |
106 |
if search is None: |
|
107 |
return
|
|
4110.2.14
by Martin Pool
Small fetch progress tweaks |
108 |
pb.update("Fetching revisions", 1, 2) |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
109 |
self._fetch_everything_for_search(search) |
110 |
finally: |
|
111 |
pb.finished() |
|
2535.3.6
by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository. |
112 |
|
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
113 |
def _fetch_everything_for_search(self, search): |
2535.3.6
by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository. |
114 |
"""Fetch all data for the given set of revisions."""
|
2535.3.9
by Andrew Bennetts
More comments. |
115 |
# The first phase is "file". We pass the progress bar for it directly
|
2668.2.8
by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by. |
116 |
# into item_keys_introduced_by, which has more information about how
|
2535.3.9
by Andrew Bennetts
More comments. |
117 |
# that phase is progressing than we do. Progress updates for the other
|
118 |
# phases are taken care of in this function.
|
|
119 |
# XXX: there should be a clear owner of the progress reporting. Perhaps
|
|
2668.2.8
by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by. |
120 |
# item_keys_introduced_by should have a richer API than it does at the
|
121 |
# moment, so that it can feed the progress information back to this
|
|
2535.3.9
by Andrew Bennetts
More comments. |
122 |
# function?
|
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
123 |
if (self.from_repository._format.rich_root_data and |
124 |
not self.to_repository._format.rich_root_data): |
|
125 |
raise errors.IncompatibleRepositories( |
|
126 |
self.from_repository, self.to_repository, |
|
127 |
"different rich-root support") |
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
128 |
pb = ui.ui_factory.nested_progress_bar() |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
129 |
try: |
4110.2.12
by Martin Pool
Add more fetch progress |
130 |
pb.update("Get stream source") |
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
131 |
source = self.from_repository._get_source( |
132 |
self.to_repository._format) |
|
133 |
stream = source.get_stream(search) |
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
134 |
from_format = self.from_repository._format |
4110.2.12
by Martin Pool
Add more fetch progress |
135 |
pb.update("Inserting stream") |
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
136 |
resume_tokens, missing_keys = self.sink.insert_stream( |
137 |
stream, from_format, []) |
|
4257.3.2
by Andrew Bennetts
Check during fetch if we are going to be missing data necessary to calculate altered fileids for stacked revisions. |
138 |
if self.to_repository._fallback_repositories: |
4241.18.1
by Andrew Bennetts
Cherry pick stacking push fix from bzr.dev r4289, make a 'Changes from RC1 to Final' section in NEWS. |
139 |
missing_keys.update( |
4257.4.12
by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher. |
140 |
self._parent_inventories(search.get_keys())) |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
141 |
if missing_keys: |
4110.2.12
by Martin Pool
Add more fetch progress |
142 |
pb.update("Missing keys") |
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
143 |
stream = source.get_stream_for_missing_keys(missing_keys) |
4110.2.12
by Martin Pool
Add more fetch progress |
144 |
pb.update("Inserting missing keys") |
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
145 |
resume_tokens, missing_keys = self.sink.insert_stream( |
146 |
stream, from_format, resume_tokens) |
|
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
147 |
if missing_keys: |
148 |
raise AssertionError( |
|
149 |
"second push failed to complete a fetch %r." % ( |
|
150 |
missing_keys,)) |
|
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
151 |
if resume_tokens: |
152 |
raise AssertionError( |
|
153 |
"second push failed to commit the fetch %r." % ( |
|
154 |
resume_tokens,)) |
|
4110.2.12
by Martin Pool
Add more fetch progress |
155 |
pb.update("Finishing stream") |
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
156 |
self.sink.finished() |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
157 |
finally: |
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
158 |
pb.finished() |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
159 |
|
1185.65.30
by Robert Collins
Merge integration. |
160 |
def _revids_to_fetch(self): |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
161 |
"""Determines the exact revisions needed from self.from_repository to
|
162 |
install self._last_revision in self.to_repository.
|
|
163 |
||
164 |
If no revisions need to be fetched, then this just returns None.
|
|
165 |
"""
|
|
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
166 |
if self._fetch_spec is not None: |
167 |
return self._fetch_spec |
|
4110.2.22
by Martin Pool
Re-add mutter calls during fetch |
168 |
mutter('fetch up to rev {%s}', self._last_revision) |
1534.4.50
by Robert Collins
Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running. |
169 |
if self._last_revision is NULL_REVISION: |
170 |
# explicit limit of no revisions needed
|
|
3184.1.9
by Robert Collins
* ``Repository.get_data_stream`` is now deprecated in favour of |
171 |
return None |
4316.1.3
by Jonathan Lange
Don't bother wrapping the NoSuchRevision in an InstallFailed. |
172 |
return self.to_repository.search_missing_revision_ids( |
173 |
self.from_repository, self._last_revision, |
|
174 |
find_ghosts=self.find_ghosts) |
|
1185.64.3
by Goffredo Baroncelli
This patch changes the fetch code. Before, the original code expanded every inventory and |
175 |
|
4257.4.12
by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher. |
176 |
def _parent_inventories(self, revision_ids): |
177 |
# Find all the parent revisions referenced by the stream, but
|
|
4257.4.13
by Andrew Bennetts
Tweak comment. |
178 |
# not present in the stream, and make sure we send their
|
4257.4.12
by Andrew Bennetts
Move _parent_inventories helper to RepoFetcher. |
179 |
# inventories.
|
180 |
parent_maps = self.to_repository.get_parent_map(revision_ids) |
|
181 |
parents = set() |
|
182 |
map(parents.update, parent_maps.itervalues()) |
|
183 |
parents.discard(NULL_REVISION) |
|
184 |
parents.difference_update(revision_ids) |
|
185 |
missing_keys = set(('inventories', rev_id) for rev_id in parents) |
|
186 |
return missing_keys |
|
187 |
||
3565.3.3
by Robert Collins
* Fetching data between repositories that have the same model but no |
188 |
|
1910.2.24
by Aaron Bentley
Got intra-repository fetch working between model1 and 2 for all types |
189 |
class Inter1and2Helper(object): |
1910.2.48
by Aaron Bentley
Update from review comments |
190 |
"""Helper for operations that convert data from model 1 and 2
|
3943.8.1
by Marius Kruger
remove all trailing whitespace from bzr source |
191 |
|
1910.2.48
by Aaron Bentley
Update from review comments |
192 |
This is for use by fetchers and converters.
|
193 |
"""
|
|
194 |
||
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
195 |
def __init__(self, source): |
1910.2.48
by Aaron Bentley
Update from review comments |
196 |
"""Constructor.
|
197 |
||
198 |
:param source: The repository data comes from
|
|
199 |
"""
|
|
200 |
self.source = source |
|
201 |
||
202 |
def iter_rev_trees(self, revs): |
|
203 |
"""Iterate through RevisionTrees efficiently.
|
|
204 |
||
205 |
Additionally, the inventory's revision_id is set if unset.
|
|
206 |
||
207 |
Trees are retrieved in batches of 100, and then yielded in the order
|
|
208 |
they were requested.
|
|
209 |
||
210 |
:param revs: A list of revision ids
|
|
211 |
"""
|
|
3172.4.4
by Robert Collins
Review feedback. |
212 |
# In case that revs is not a list.
|
213 |
revs = list(revs) |
|
1910.2.48
by Aaron Bentley
Update from review comments |
214 |
while revs: |
215 |
for tree in self.source.revision_trees(revs[:100]): |
|
1910.2.44
by Aaron Bentley
Retrieve only 500 revision trees at once |
216 |
if tree.inventory.revision_id is None: |
217 |
tree.inventory.revision_id = tree.get_revision_id() |
|
218 |
yield tree |
|
1910.2.48
by Aaron Bentley
Update from review comments |
219 |
revs = revs[100:] |
1910.2.44
by Aaron Bentley
Retrieve only 500 revision trees at once |
220 |
|
3380.2.4
by Aaron Bentley
Updates from review |
221 |
def _find_root_ids(self, revs, parent_map, graph): |
222 |
revision_root = {} |
|
1910.2.48
by Aaron Bentley
Update from review comments |
223 |
for tree in self.iter_rev_trees(revs): |
1910.2.18
by Aaron Bentley
Implement creation of knits for tree roots |
224 |
revision_id = tree.inventory.root.revision |
2946.3.3
by John Arbash Meinel
Prefer tree.get_root_id() as more explicit than tree.path2id('') |
225 |
root_id = tree.get_root_id() |
3380.1.3
by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed |
226 |
revision_root[revision_id] = root_id |
227 |
# Find out which parents we don't already know root ids for
|
|
228 |
parents = set() |
|
229 |
for revision_parents in parent_map.itervalues(): |
|
230 |
parents.update(revision_parents) |
|
231 |
parents.difference_update(revision_root.keys() + [NULL_REVISION]) |
|
3380.2.7
by Aaron Bentley
Update docs |
232 |
# Limit to revisions present in the versionedfile
|
3380.1.3
by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed |
233 |
parents = graph.get_parent_map(parents).keys() |
234 |
for tree in self.iter_rev_trees(parents): |
|
235 |
root_id = tree.get_root_id() |
|
236 |
revision_root[tree.get_revision_id()] = root_id |
|
4476.3.11
by Andrew Bennetts
All fetch and interrepo tests passing. |
237 |
return revision_root |
3380.2.4
by Aaron Bentley
Updates from review |
238 |
|
239 |
def generate_root_texts(self, revs): |
|
240 |
"""Generate VersionedFiles for all root ids.
|
|
241 |
||
242 |
:param revs: the revisions to include
|
|
243 |
"""
|
|
244 |
graph = self.source.get_graph() |
|
245 |
parent_map = graph.get_parent_map(revs) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
246 |
rev_order = tsort.topo_sort(parent_map) |
4476.3.11
by Andrew Bennetts
All fetch and interrepo tests passing. |
247 |
rev_id_to_root_id = self._find_root_ids(revs, parent_map, graph) |
3350.6.4
by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores. |
248 |
root_id_order = [(rev_id_to_root_id[rev_id], rev_id) for rev_id in |
249 |
rev_order] |
|
250 |
# Guaranteed stable, this groups all the file id operations together
|
|
251 |
# retaining topological order within the revisions of a file id.
|
|
252 |
# File id splits and joins would invalidate this, but they don't exist
|
|
253 |
# yet, and are unlikely to in non-rich-root environments anyway.
|
|
254 |
root_id_order.sort(key=operator.itemgetter(0)) |
|
255 |
# Create a record stream containing the roots to create.
|
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
256 |
if len(revs) > 100: |
257 |
graph = _get_rich_root_heads_graph(self.source_repo, revs) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
258 |
new_roots_stream = _new_root_data_stream( |
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
259 |
root_id_order, rev_id_to_root_id, parent_map, self.source, graph) |
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
260 |
return [('texts', new_roots_stream)] |
261 |
||
262 |
||
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
263 |
def _get_rich_root_heads_graph(source_repo, revision_ids): |
264 |
"""Get a Graph object suitable for asking heads() for new rich roots."""
|
|
265 |
st = static_tuple.StaticTuple |
|
266 |
revision_keys = [st(r_id).intern() for r_id in revision_ids] |
|
267 |
known_graph = source_repo.revisions.get_known_graph_ancestry( |
|
268 |
revision_keys) |
|
4849.1.1
by John Arbash Meinel
Typo fix, not caught by pqm because it requires 100 revs. |
269 |
return _mod_graph.GraphThunkIdsToKeys(known_graph) |
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
270 |
|
271 |
||
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
272 |
def _new_root_data_stream( |
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
273 |
root_keys_to_create, rev_id_to_root_id_map, parent_map, repo, graph=None): |
4476.3.69
by Andrew Bennetts
Elaborate some docstrings. |
274 |
"""Generate a texts substream of synthesised root entries.
|
275 |
||
276 |
Used in fetches that do rich-root upgrades.
|
|
277 |
|
|
278 |
:param root_keys_to_create: iterable of (root_id, rev_id) pairs describing
|
|
279 |
the root entries to create.
|
|
280 |
:param rev_id_to_root_id_map: dict of known rev_id -> root_id mappings for
|
|
281 |
calculating the parents. If a parent rev_id is not found here then it
|
|
282 |
will be recalculated.
|
|
283 |
:param parent_map: a parent map for all the revisions in
|
|
284 |
root_keys_to_create.
|
|
285 |
:param graph: a graph to use instead of repo.get_graph().
|
|
286 |
"""
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
287 |
for root_key in root_keys_to_create: |
288 |
root_id, rev_id = root_key |
|
289 |
parent_keys = _parent_keys_for_root_version( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
290 |
root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph) |
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
291 |
yield versionedfile.FulltextContentFactory( |
292 |
root_key, parent_keys, None, '') |
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
293 |
|
294 |
||
295 |
def _parent_keys_for_root_version( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
296 |
root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph=None): |
4476.3.69
by Andrew Bennetts
Elaborate some docstrings. |
297 |
"""Get the parent keys for a given root id.
|
298 |
|
|
299 |
A helper function for _new_root_data_stream.
|
|
300 |
"""
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
301 |
# Include direct parents of the revision, but only if they used the same
|
302 |
# root_id and are heads.
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
303 |
rev_parents = parent_map[rev_id] |
304 |
parent_ids = [] |
|
305 |
for parent_id in rev_parents: |
|
306 |
if parent_id == NULL_REVISION: |
|
307 |
continue
|
|
308 |
if parent_id not in rev_id_to_root_id_map: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
309 |
# We probably didn't read this revision, go spend the extra effort
|
310 |
# to actually check
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
311 |
try: |
312 |
tree = repo.revision_tree(parent_id) |
|
313 |
except errors.NoSuchRevision: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
314 |
# Ghost, fill out rev_id_to_root_id in case we encounter this
|
315 |
# again.
|
|
316 |
# But set parent_root_id to None since we don't really know
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
317 |
parent_root_id = None |
318 |
else: |
|
319 |
parent_root_id = tree.get_root_id() |
|
320 |
rev_id_to_root_id_map[parent_id] = None |
|
4476.3.21
by Andrew Bennetts
Clarify some code and comments, and s/1.17/1.18/ in a few places. |
321 |
# XXX: why not:
|
322 |
# rev_id_to_root_id_map[parent_id] = parent_root_id
|
|
323 |
# memory consumption maybe?
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
324 |
else: |
325 |
parent_root_id = rev_id_to_root_id_map[parent_id] |
|
326 |
if root_id == parent_root_id: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
327 |
# With stacking we _might_ want to refer to a non-local revision,
|
328 |
# but this code path only applies when we have the full content
|
|
329 |
# available, so ghosts really are ghosts, not just the edge of
|
|
330 |
# local data.
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
331 |
parent_ids.append(parent_id) |
332 |
else: |
|
333 |
# root_id may be in the parent anyway.
|
|
334 |
try: |
|
335 |
tree = repo.revision_tree(parent_id) |
|
336 |
except errors.NoSuchRevision: |
|
337 |
# ghost, can't refer to it.
|
|
338 |
pass
|
|
339 |
else: |
|
340 |
try: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
341 |
parent_ids.append(tree.inventory[root_id].revision) |
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
342 |
except errors.NoSuchId: |
343 |
# not in the tree
|
|
344 |
pass
|
|
345 |
# Drop non-head parents
|
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
346 |
if graph is None: |
347 |
graph = repo.get_graph() |
|
348 |
heads = graph.heads(parent_ids) |
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
349 |
selected_ids = [] |
350 |
for parent_id in parent_ids: |
|
351 |
if parent_id in heads and parent_id not in selected_ids: |
|
352 |
selected_ids.append(parent_id) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
353 |
parent_keys = [(root_id, parent_id) for parent_id in selected_ids] |
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
354 |
return parent_keys |