5557.1.7
by John Arbash Meinel
Merge in the bzr.dev 5582 |
1 |
# Copyright (C) 2005-2011 Canonical Ltd
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
2 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
3 |
# This program is free software; you can redistribute it and/or modify
|
4 |
# it under the terms of the GNU General Public License as published by
|
|
5 |
# the Free Software Foundation; either version 2 of the License, or
|
|
6 |
# (at your option) any later version.
|
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
7 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
8 |
# This program is distributed in the hope that it will be useful,
|
9 |
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
10 |
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
11 |
# GNU General Public License for more details.
|
|
1887.1.1
by Adeodato Simó
Do not separate paragraphs in the copyright statement with blank lines, |
12 |
#
|
974.1.27
by aaron.bentley at utoronto
Initial greedy fetch work |
13 |
# You should have received a copy of the GNU General Public License
|
14 |
# along with this program; if not, write to the Free Software
|
|
4183.7.1
by Sabin Iacob
update FSF mailing address |
15 |
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
1218
by Martin Pool
- fix up import |
16 |
|
1231
by Martin Pool
- more progress on fetch on top of weaves |
17 |
|
18 |
"""Copying of history from one branch to another.
|
|
19 |
||
20 |
The basic plan is that every branch knows the history of everything
|
|
21 |
that has merged into it. As the first step of a merge, pull, or
|
|
22 |
branch operation we copy history from the source into the destination
|
|
23 |
branch.
|
|
24 |
"""
|
|
25 |
||
3350.6.4
by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores. |
26 |
import operator |
27 |
||
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
28 |
from bzrlib.lazy_import import lazy_import |
29 |
lazy_import(globals(), """ |
|
30 |
from bzrlib import (
|
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
31 |
graph as _mod_graph,
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
32 |
tsort,
|
33 |
versionedfile,
|
|
34 |
)
|
|
35 |
""") |
|
4110.2.4
by Martin Pool
Deprecate passing a pb in to RepoFetcher |
36 |
from bzrlib import ( |
37 |
errors, |
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
38 |
ui, |
4110.2.4
by Martin Pool
Deprecate passing a pb in to RepoFetcher |
39 |
)
|
6138.4.2
by Jonathan Riddell
add import gettext |
40 |
from bzrlib.i18n import gettext |
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
41 |
from bzrlib.revision import NULL_REVISION |
2094.3.5
by John Arbash Meinel
Fix imports to ensure modules are loaded before they are used |
42 |
from bzrlib.trace import mutter |
1534.1.31
by Robert Collins
Deprecated fetch.fetch and fetch.greedy_fetch for branch.fetch, and move the Repository.fetch internals to InterRepo and InterWeaveRepo. |
43 |
|
1238
by Martin Pool
- remove a lot of dead code from fetch |
44 |
|
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
45 |
class RepoFetcher(object): |
46 |
"""Pull revisions and texts from one repository to another.
|
|
47 |
||
2592.4.5
by Martin Pool
Add Repository.base on all repositories. |
48 |
This should not be used directly, it's essential a object to encapsulate
|
1534.1.33
by Robert Collins
Move copy_content_into into InterRepository and InterWeaveRepo, and disable the default codepath test as we have optimised paths for all current combinations. |
49 |
the logic in InterRepository.fetch().
|
1260
by Martin Pool
- some updates for fetch/update function |
50 |
"""
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
51 |
|
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
52 |
def __init__(self, to_repository, from_repository, last_revision=None, |
4961.2.3
by Martin Pool
Delete deprecated pb parameter to RepoFetcher |
53 |
find_ghosts=True, fetch_spec=None): |
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
54 |
"""Create a repo fetcher.
|
55 |
||
4110.2.2
by Martin Pool
Remove obsolete comments |
56 |
:param last_revision: If set, try to limit to the data this revision
|
57 |
references.
|
|
5536.3.1
by Andrew Bennetts
Rename get_search_result to execute, require a SearchResult (and not a search) be passed to fetch. |
58 |
:param fetch_spec: A SearchResult specifying which revisions to fetch.
|
59 |
If set, this overrides last_revision.
|
|
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
60 |
:param find_ghosts: If True search the entire history for ghosts.
|
61 |
"""
|
|
4509.3.18
by Martin Pool
RepoFetcher relies on Repository.fetch to shortcircuit no-op fetches |
62 |
# repository.fetch has the responsibility for short-circuiting
|
63 |
# attempts to copy between a repository and itself.
|
|
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
64 |
self.to_repository = to_repository |
65 |
self.from_repository = from_repository |
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
66 |
self.sink = to_repository._get_sink() |
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
67 |
# must not mutate self._last_revision as its potentially a shared instance
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
68 |
self._last_revision = last_revision |
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
69 |
self._fetch_spec = fetch_spec |
3172.4.1
by Robert Collins
* Fetching via bzr+ssh will no longer fill ghosts by default (this is |
70 |
self.find_ghosts = find_ghosts |
1534.4.41
by Robert Collins
Branch now uses BzrDir reasonably sanely. |
71 |
self.from_repository.lock_read() |
4110.2.22
by Martin Pool
Re-add mutter calls during fetch |
72 |
mutter("Using fetch logic to copy between %s(%s) and %s(%s)", |
73 |
self.from_repository, self.from_repository._format, |
|
74 |
self.to_repository, self.to_repository._format) |
|
3842.3.5
by Andrew Bennetts
Remove some debugging cruft, make more tests pass. |
75 |
try: |
4110.2.3
by Martin Pool
Remove redundant variable from fetch. |
76 |
self.__fetch() |
3842.3.5
by Andrew Bennetts
Remove some debugging cruft, make more tests pass. |
77 |
finally: |
78 |
self.from_repository.unlock() |
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
79 |
|
80 |
def __fetch(self): |
|
81 |
"""Primary worker function.
|
|
82 |
||
3943.8.1
by Marius Kruger
remove all trailing whitespace from bzr source |
83 |
This initialises all the needed variables, and then fetches the
|
1185.65.27
by Robert Collins
Tweak storage towards mergability. |
84 |
requested revisions, finally clearing the progress bar.
|
85 |
"""
|
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
86 |
# Roughly this is what we're aiming for fetch to become:
|
87 |
#
|
|
88 |
# missing = self.sink.insert_stream(self.source.get_stream(search))
|
|
89 |
# if missing:
|
|
90 |
# missing = self.sink.insert_stream(self.source.get_items(missing))
|
|
91 |
# assert not missing
|
|
1240
by Martin Pool
- clean up fetch code and add progress bar |
92 |
self.count_total = 0 |
1185.33.55
by Martin Pool
[patch] weave fetch optimizations (Goffredo Baroncelli) |
93 |
self.file_ids_names = {} |
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
94 |
pb = ui.ui_factory.nested_progress_bar() |
4110.2.14
by Martin Pool
Small fetch progress tweaks |
95 |
pb.show_pct = pb.show_count = False |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
96 |
try: |
6138.4.1
by Jonathan Riddell
add gettext to progress bar strings |
97 |
pb.update(gettext("Finding revisions"), 0, 2) |
5536.3.1
by Andrew Bennetts
Rename get_search_result to execute, require a SearchResult (and not a search) be passed to fetch. |
98 |
search_result = self._revids_to_fetch() |
99 |
mutter('fetching: %s', search_result) |
|
100 |
if search_result.is_empty(): |
|
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
101 |
return
|
6138.4.1
by Jonathan Riddell
add gettext to progress bar strings |
102 |
pb.update(gettext("Fetching revisions"), 1, 2) |
5536.3.1
by Andrew Bennetts
Rename get_search_result to execute, require a SearchResult (and not a search) be passed to fetch. |
103 |
self._fetch_everything_for_search(search_result) |
4110.2.9
by Martin Pool
Re-add very basic top-level pb for fetch |
104 |
finally: |
105 |
pb.finished() |
|
2535.3.6
by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository. |
106 |
|
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
107 |
def _fetch_everything_for_search(self, search): |
2535.3.6
by Andrew Bennetts
Move some "what repo data to fetch logic" from RepoFetcher to Repository. |
108 |
"""Fetch all data for the given set of revisions."""
|
2535.3.9
by Andrew Bennetts
More comments. |
109 |
# The first phase is "file". We pass the progress bar for it directly
|
2668.2.8
by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by. |
110 |
# into item_keys_introduced_by, which has more information about how
|
2535.3.9
by Andrew Bennetts
More comments. |
111 |
# that phase is progressing than we do. Progress updates for the other
|
112 |
# phases are taken care of in this function.
|
|
113 |
# XXX: there should be a clear owner of the progress reporting. Perhaps
|
|
2668.2.8
by Andrew Bennetts
Rename get_data_to_fetch_for_revision_ids as item_keys_introduced_by. |
114 |
# item_keys_introduced_by should have a richer API than it does at the
|
115 |
# moment, so that it can feed the progress information back to this
|
|
2535.3.9
by Andrew Bennetts
More comments. |
116 |
# function?
|
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
117 |
if (self.from_repository._format.rich_root_data and |
118 |
not self.to_repository._format.rich_root_data): |
|
119 |
raise errors.IncompatibleRepositories( |
|
120 |
self.from_repository, self.to_repository, |
|
121 |
"different rich-root support") |
|
4819.2.4
by John Arbash Meinel
Factor out the common code into a helper so that smart streaming also benefits. |
122 |
pb = ui.ui_factory.nested_progress_bar() |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
123 |
try: |
4110.2.12
by Martin Pool
Add more fetch progress |
124 |
pb.update("Get stream source") |
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
125 |
source = self.from_repository._get_source( |
126 |
self.to_repository._format) |
|
127 |
stream = source.get_stream(search) |
|
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
128 |
from_format = self.from_repository._format |
4110.2.12
by Martin Pool
Add more fetch progress |
129 |
pb.update("Inserting stream") |
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
130 |
resume_tokens, missing_keys = self.sink.insert_stream( |
5195.3.14
by Parth Malwankar
optimized to use "revisions" insert_record_stream for counting #records |
131 |
stream, from_format, []) |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
132 |
if missing_keys: |
4110.2.12
by Martin Pool
Add more fetch progress |
133 |
pb.update("Missing keys") |
4060.1.3
by Robert Collins
Implement the separate source component for fetch - repository.StreamSource. |
134 |
stream = source.get_stream_for_missing_keys(missing_keys) |
4110.2.12
by Martin Pool
Add more fetch progress |
135 |
pb.update("Inserting missing keys") |
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
136 |
resume_tokens, missing_keys = self.sink.insert_stream( |
5195.3.14
by Parth Malwankar
optimized to use "revisions" insert_record_stream for counting #records |
137 |
stream, from_format, resume_tokens) |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
138 |
if missing_keys: |
139 |
raise AssertionError( |
|
140 |
"second push failed to complete a fetch %r." % ( |
|
141 |
missing_keys,)) |
|
4032.3.7
by Robert Collins
Move write locking and write group responsibilities into the Sink objects themselves, allowing complete avoidance of unnecessary calls when the sink is a RemoteSink. |
142 |
if resume_tokens: |
143 |
raise AssertionError( |
|
144 |
"second push failed to commit the fetch %r." % ( |
|
145 |
resume_tokens,)) |
|
4110.2.12
by Martin Pool
Add more fetch progress |
146 |
pb.update("Finishing stream") |
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
147 |
self.sink.finished() |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
148 |
finally: |
4110.2.6
by Martin Pool
Remove more progressbar cruft from fetch |
149 |
pb.finished() |
4029.2.1
by Robert Collins
Support streaming push to stacked branches. |
150 |
|
1185.65.30
by Robert Collins
Merge integration. |
151 |
def _revids_to_fetch(self): |
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
152 |
"""Determines the exact revisions needed from self.from_repository to
|
153 |
install self._last_revision in self.to_repository.
|
|
154 |
||
5539.2.8
by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'. |
155 |
:returns: A SearchResult of some sort. (Possibly a
|
5535.3.21
by Andrew Bennetts
Cosmetic tweaks. |
156 |
PendingAncestryResult, EmptySearchResult, etc.)
|
2535.3.7
by Andrew Bennetts
Remove now unused _fetch_weave_texts, make progress reporting closer to how it was before I refactored __fetch. |
157 |
"""
|
5536.3.1
by Andrew Bennetts
Rename get_search_result to execute, require a SearchResult (and not a search) be passed to fetch. |
158 |
if self._fetch_spec is not None: |
5539.2.8
by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'. |
159 |
# The fetch spec is already a concrete search result.
|
4070.9.2
by Andrew Bennetts
Rough prototype of allowing a SearchResult to be passed to fetch, and using that to improve network conversations. |
160 |
return self._fetch_spec |
5539.2.8
by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'. |
161 |
elif self._last_revision == NULL_REVISION: |
162 |
# fetch_spec is None + last_revision is null => empty fetch.
|
|
1534.4.50
by Robert Collins
Got the bzrdir api straightened out, plenty of refactoring to use it pending, but the api is up and running. |
163 |
# explicit limit of no revisions needed
|
5852.1.6
by Jelmer Vernooij
Add extra test for Repository.search_missing_revision_ids. |
164 |
return _mod_graph.EmptySearchResult() |
5539.2.8
by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'. |
165 |
elif self._last_revision is not None: |
5852.1.6
by Jelmer Vernooij
Add extra test for Repository.search_missing_revision_ids. |
166 |
return _mod_graph.NotInOtherForRevs(self.to_repository, |
5539.2.10
by Andrew Bennetts
s/NotInOtherForRev/NotInOtherForRevs/, and allow passing multiple revision_ids to search_missing_revision_ids. |
167 |
self.from_repository, [self._last_revision], |
5536.3.1
by Andrew Bennetts
Rename get_search_result to execute, require a SearchResult (and not a search) be passed to fetch. |
168 |
find_ghosts=self.find_ghosts).execute() |
5539.2.8
by Andrew Bennetts
Refactor call to search_missing_revision_ids out from RepoFetcher._revids_to_fetch into the 'fetch spec'. |
169 |
else: # self._last_revision is None: |
5852.1.6
by Jelmer Vernooij
Add extra test for Repository.search_missing_revision_ids. |
170 |
return _mod_graph.EverythingNotInOther(self.to_repository, |
5539.2.19
by Andrew Bennetts
Define SearchResult/Search interfaces with explicit abstract base classes, add some docstrings and change a method name. |
171 |
self.from_repository, |
5536.3.1
by Andrew Bennetts
Rename get_search_result to execute, require a SearchResult (and not a search) be passed to fetch. |
172 |
find_ghosts=self.find_ghosts).execute() |
1185.64.3
by Goffredo Baroncelli
This patch changes the fetch code. Before, the original code expanded every inventory and |
173 |
|
3565.3.3
by Robert Collins
* Fetching data between repositories that have the same model but no |
174 |
|
1910.2.24
by Aaron Bentley
Got intra-repository fetch working between model1 and 2 for all types |
175 |
class Inter1and2Helper(object): |
1910.2.48
by Aaron Bentley
Update from review comments |
176 |
"""Helper for operations that convert data from model 1 and 2
|
3943.8.1
by Marius Kruger
remove all trailing whitespace from bzr source |
177 |
|
1910.2.48
by Aaron Bentley
Update from review comments |
178 |
This is for use by fetchers and converters.
|
179 |
"""
|
|
180 |
||
5050.32.1
by Andrew Bennetts
Fix fetching more than 100 revisions from non-rich-root to rich-root repositories. |
181 |
# This is a class variable so that the test suite can override it.
|
182 |
known_graph_threshold = 100 |
|
183 |
||
4022.1.1
by Robert Collins
Refactoring of fetch to have a sender and sink component enabling splitting the logic over a network stream. (Robert Collins, Andrew Bennetts) |
184 |
def __init__(self, source): |
1910.2.48
by Aaron Bentley
Update from review comments |
185 |
"""Constructor.
|
186 |
||
187 |
:param source: The repository data comes from
|
|
188 |
"""
|
|
189 |
self.source = source |
|
190 |
||
191 |
def iter_rev_trees(self, revs): |
|
192 |
"""Iterate through RevisionTrees efficiently.
|
|
193 |
||
194 |
Additionally, the inventory's revision_id is set if unset.
|
|
195 |
||
196 |
Trees are retrieved in batches of 100, and then yielded in the order
|
|
197 |
they were requested.
|
|
198 |
||
199 |
:param revs: A list of revision ids
|
|
200 |
"""
|
|
3172.4.4
by Robert Collins
Review feedback. |
201 |
# In case that revs is not a list.
|
202 |
revs = list(revs) |
|
1910.2.48
by Aaron Bentley
Update from review comments |
203 |
while revs: |
204 |
for tree in self.source.revision_trees(revs[:100]): |
|
1910.2.44
by Aaron Bentley
Retrieve only 500 revision trees at once |
205 |
if tree.inventory.revision_id is None: |
206 |
tree.inventory.revision_id = tree.get_revision_id() |
|
207 |
yield tree |
|
1910.2.48
by Aaron Bentley
Update from review comments |
208 |
revs = revs[100:] |
1910.2.44
by Aaron Bentley
Retrieve only 500 revision trees at once |
209 |
|
3380.2.4
by Aaron Bentley
Updates from review |
210 |
def _find_root_ids(self, revs, parent_map, graph): |
211 |
revision_root = {} |
|
1910.2.48
by Aaron Bentley
Update from review comments |
212 |
for tree in self.iter_rev_trees(revs): |
1910.2.18
by Aaron Bentley
Implement creation of knits for tree roots |
213 |
revision_id = tree.inventory.root.revision |
2946.3.3
by John Arbash Meinel
Prefer tree.get_root_id() as more explicit than tree.path2id('') |
214 |
root_id = tree.get_root_id() |
3380.1.3
by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed |
215 |
revision_root[revision_id] = root_id |
216 |
# Find out which parents we don't already know root ids for
|
|
217 |
parents = set() |
|
218 |
for revision_parents in parent_map.itervalues(): |
|
219 |
parents.update(revision_parents) |
|
220 |
parents.difference_update(revision_root.keys() + [NULL_REVISION]) |
|
3380.2.7
by Aaron Bentley
Update docs |
221 |
# Limit to revisions present in the versionedfile
|
3380.1.3
by Aaron Bentley
Fix model-change fetching with ghosts and when fetch is resumed |
222 |
parents = graph.get_parent_map(parents).keys() |
223 |
for tree in self.iter_rev_trees(parents): |
|
224 |
root_id = tree.get_root_id() |
|
225 |
revision_root[tree.get_revision_id()] = root_id |
|
4476.3.11
by Andrew Bennetts
All fetch and interrepo tests passing. |
226 |
return revision_root |
3380.2.4
by Aaron Bentley
Updates from review |
227 |
|
228 |
def generate_root_texts(self, revs): |
|
229 |
"""Generate VersionedFiles for all root ids.
|
|
230 |
||
231 |
:param revs: the revisions to include
|
|
232 |
"""
|
|
233 |
graph = self.source.get_graph() |
|
234 |
parent_map = graph.get_parent_map(revs) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
235 |
rev_order = tsort.topo_sort(parent_map) |
4476.3.11
by Andrew Bennetts
All fetch and interrepo tests passing. |
236 |
rev_id_to_root_id = self._find_root_ids(revs, parent_map, graph) |
3350.6.4
by Robert Collins
First cut at pluralised VersionedFiles. Some rather massive API incompatabilities, primarily because of the difficulty of coherence among competing stores. |
237 |
root_id_order = [(rev_id_to_root_id[rev_id], rev_id) for rev_id in |
238 |
rev_order] |
|
239 |
# Guaranteed stable, this groups all the file id operations together
|
|
240 |
# retaining topological order within the revisions of a file id.
|
|
241 |
# File id splits and joins would invalidate this, but they don't exist
|
|
242 |
# yet, and are unlikely to in non-rich-root environments anyway.
|
|
243 |
root_id_order.sort(key=operator.itemgetter(0)) |
|
244 |
# Create a record stream containing the roots to create.
|
|
5050.32.1
by Andrew Bennetts
Fix fetching more than 100 revisions from non-rich-root to rich-root repositories. |
245 |
if len(revs) > self.known_graph_threshold: |
246 |
graph = self.source.get_known_graph_ancestry(revs) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
247 |
new_roots_stream = _new_root_data_stream( |
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
248 |
root_id_order, rev_id_to_root_id, parent_map, self.source, graph) |
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
249 |
return [('texts', new_roots_stream)] |
250 |
||
251 |
||
252 |
def _new_root_data_stream( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
253 |
root_keys_to_create, rev_id_to_root_id_map, parent_map, repo, graph=None): |
4476.3.69
by Andrew Bennetts
Elaborate some docstrings. |
254 |
"""Generate a texts substream of synthesised root entries.
|
255 |
||
256 |
Used in fetches that do rich-root upgrades.
|
|
257 |
|
|
258 |
:param root_keys_to_create: iterable of (root_id, rev_id) pairs describing
|
|
259 |
the root entries to create.
|
|
260 |
:param rev_id_to_root_id_map: dict of known rev_id -> root_id mappings for
|
|
261 |
calculating the parents. If a parent rev_id is not found here then it
|
|
262 |
will be recalculated.
|
|
263 |
:param parent_map: a parent map for all the revisions in
|
|
264 |
root_keys_to_create.
|
|
265 |
:param graph: a graph to use instead of repo.get_graph().
|
|
266 |
"""
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
267 |
for root_key in root_keys_to_create: |
268 |
root_id, rev_id = root_key |
|
269 |
parent_keys = _parent_keys_for_root_version( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
270 |
root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph) |
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
271 |
yield versionedfile.FulltextContentFactory( |
272 |
root_key, parent_keys, None, '') |
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
273 |
|
274 |
||
275 |
def _parent_keys_for_root_version( |
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
276 |
root_id, rev_id, rev_id_to_root_id_map, parent_map, repo, graph=None): |
4476.3.69
by Andrew Bennetts
Elaborate some docstrings. |
277 |
"""Get the parent keys for a given root id.
|
278 |
|
|
279 |
A helper function for _new_root_data_stream.
|
|
280 |
"""
|
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
281 |
# Include direct parents of the revision, but only if they used the same
|
282 |
# root_id and are heads.
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
283 |
rev_parents = parent_map[rev_id] |
284 |
parent_ids = [] |
|
285 |
for parent_id in rev_parents: |
|
286 |
if parent_id == NULL_REVISION: |
|
287 |
continue
|
|
288 |
if parent_id not in rev_id_to_root_id_map: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
289 |
# We probably didn't read this revision, go spend the extra effort
|
290 |
# to actually check
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
291 |
try: |
292 |
tree = repo.revision_tree(parent_id) |
|
293 |
except errors.NoSuchRevision: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
294 |
# Ghost, fill out rev_id_to_root_id in case we encounter this
|
295 |
# again.
|
|
296 |
# But set parent_root_id to None since we don't really know
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
297 |
parent_root_id = None |
298 |
else: |
|
299 |
parent_root_id = tree.get_root_id() |
|
300 |
rev_id_to_root_id_map[parent_id] = None |
|
4476.3.21
by Andrew Bennetts
Clarify some code and comments, and s/1.17/1.18/ in a few places. |
301 |
# XXX: why not:
|
302 |
# rev_id_to_root_id_map[parent_id] = parent_root_id
|
|
303 |
# memory consumption maybe?
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
304 |
else: |
305 |
parent_root_id = rev_id_to_root_id_map[parent_id] |
|
306 |
if root_id == parent_root_id: |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
307 |
# With stacking we _might_ want to refer to a non-local revision,
|
308 |
# but this code path only applies when we have the full content
|
|
309 |
# available, so ghosts really are ghosts, not just the edge of
|
|
310 |
# local data.
|
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
311 |
parent_ids.append(parent_id) |
312 |
else: |
|
313 |
# root_id may be in the parent anyway.
|
|
314 |
try: |
|
315 |
tree = repo.revision_tree(parent_id) |
|
316 |
except errors.NoSuchRevision: |
|
317 |
# ghost, can't refer to it.
|
|
318 |
pass
|
|
319 |
else: |
|
320 |
try: |
|
5793.2.3
by Jelmer Vernooij
Add a RevisionTree.get_file_revision() method. |
321 |
parent_ids.append(tree.get_file_revision(root_id)) |
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
322 |
except errors.NoSuchId: |
323 |
# not in the tree
|
|
324 |
pass
|
|
325 |
# Drop non-head parents
|
|
4476.3.41
by Andrew Bennetts
Use FrozenHeadsCache to speed up root generation. |
326 |
if graph is None: |
327 |
graph = repo.get_graph() |
|
328 |
heads = graph.heads(parent_ids) |
|
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
329 |
selected_ids = [] |
330 |
for parent_id in parent_ids: |
|
331 |
if parent_id in heads and parent_id not in selected_ids: |
|
332 |
selected_ids.append(parent_id) |
|
4476.3.9
by Andrew Bennetts
Further reduce duplication. |
333 |
parent_keys = [(root_id, parent_id) for parent_id in selected_ids] |
4476.3.6
by Andrew Bennetts
Refactor out duplicated get parent keys logic from Inter1and2Helper and InterDifferingSerializer. |
334 |
return parent_keys |
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
335 |
|
336 |
||
337 |
class TargetRepoKinds(object): |
|
338 |
"""An enum-like set of constants.
|
|
339 |
|
|
340 |
They are the possible values of FetchSpecFactory.target_repo_kinds.
|
|
341 |
"""
|
|
342 |
||
343 |
PREEXISTING = 'preexisting' |
|
344 |
STACKED = 'stacked' |
|
345 |
EMPTY = 'empty' |
|
346 |
||
347 |
||
348 |
class FetchSpecFactory(object): |
|
349 |
"""A helper for building the best fetch spec for a sprout call.
|
|
350 |
||
351 |
Factors that go into determining the sort of fetch to perform:
|
|
352 |
* did the caller specify any revision IDs?
|
|
5672.1.3
by Andrew Bennetts
Rename a variable, update a docstring. |
353 |
* did the caller specify a source branch (need to fetch its
|
354 |
heads_to_fetch(), usually the tip + tags)
|
|
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
355 |
* is there an existing target repo (don't need to refetch revs it
|
356 |
already has)
|
|
357 |
* target is stacked? (similar to pre-existing target repo: even if
|
|
358 |
the target itself is new don't want to refetch existing revs)
|
|
359 |
||
360 |
:ivar source_branch: the source branch if one specified, else None.
|
|
361 |
:ivar source_branch_stop_revision_id: fetch up to this revision of
|
|
362 |
source_branch, rather than its tip.
|
|
363 |
:ivar source_repo: the source repository if one found, else None.
|
|
364 |
:ivar target_repo: the target repository acquired by sprout.
|
|
365 |
:ivar target_repo_kind: one of the TargetRepoKinds constants.
|
|
366 |
"""
|
|
367 |
||
368 |
def __init__(self): |
|
369 |
self._explicit_rev_ids = set() |
|
370 |
self.source_branch = None |
|
371 |
self.source_branch_stop_revision_id = None |
|
372 |
self.source_repo = None |
|
373 |
self.target_repo = None |
|
374 |
self.target_repo_kind = None |
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
375 |
self.limit = None |
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
376 |
|
377 |
def add_revision_ids(self, revision_ids): |
|
378 |
"""Add revision_ids to the set of revision_ids to be fetched."""
|
|
379 |
self._explicit_rev_ids.update(revision_ids) |
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
380 |
|
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
381 |
def make_fetch_spec(self): |
382 |
"""Build a SearchResult or PendingAncestryResult or etc."""
|
|
383 |
if self.target_repo_kind is None or self.source_repo is None: |
|
384 |
raise AssertionError( |
|
385 |
'Incomplete FetchSpecFactory: %r' % (self.__dict__,)) |
|
386 |
if len(self._explicit_rev_ids) == 0 and self.source_branch is None: |
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
387 |
if self.limit is not None: |
5852.1.11
by Jelmer Vernooij
Review feedback from Andrew: use NotImplementedError instead of UnsupportedOperation, use simpler build_commit rather than build_snapshot. |
388 |
raise NotImplementedError( |
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
389 |
"limit is only supported with a source branch set") |
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
390 |
# Caller hasn't specified any revisions or source branch
|
391 |
if self.target_repo_kind == TargetRepoKinds.EMPTY: |
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
392 |
return _mod_graph.EverythingResult(self.source_repo) |
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
393 |
else: |
394 |
# We want everything not already in the target (or target's
|
|
395 |
# fallbacks).
|
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
396 |
return _mod_graph.EverythingNotInOther( |
5535.4.25
by Andrew Bennetts
Update some more code paths for the change to only accepting SearchResults to fetch(). |
397 |
self.target_repo, self.source_repo).execute() |
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
398 |
heads_to_fetch = set(self._explicit_rev_ids) |
399 |
if self.source_branch is not None: |
|
5741.1.11
by Jelmer Vernooij
Don't make heads_to_fetch() take a stop_revision. |
400 |
must_fetch, if_present_fetch = self.source_branch.heads_to_fetch() |
401 |
if self.source_branch_stop_revision_id is not None: |
|
402 |
# Replace the tip rev from must_fetch with the stop revision
|
|
403 |
# XXX: this might be wrong if the tip rev is also in the
|
|
404 |
# must_fetch set for other reasons (e.g. it's the tip of
|
|
405 |
# multiple loom threads?), but then it's pretty unclear what it
|
|
406 |
# should mean to specify a stop_revision in that case anyway.
|
|
407 |
must_fetch.discard(self.source_branch.last_revision()) |
|
408 |
must_fetch.add(self.source_branch_stop_revision_id) |
|
5672.1.1
by Andrew Bennetts
Refactor some of FetchSpecFactory into new Branch.heads_to_fetch method so that branch implementations like looms can override it. |
409 |
heads_to_fetch.update(must_fetch) |
410 |
else: |
|
5672.1.3
by Andrew Bennetts
Rename a variable, update a docstring. |
411 |
if_present_fetch = set() |
5535.4.23
by Andrew Bennetts
Move FetchSpecFactory and TargetRepoKinds to bzrlib.fetch (from bzrlib.controldir). |
412 |
if self.target_repo_kind == TargetRepoKinds.EMPTY: |
413 |
# PendingAncestryResult does not raise errors if a requested head
|
|
414 |
# is absent. Ideally it would support the
|
|
415 |
# required_ids/if_present_ids distinction, but in practice
|
|
416 |
# heads_to_fetch will almost certainly be present so this doesn't
|
|
417 |
# matter much.
|
|
5672.1.3
by Andrew Bennetts
Rename a variable, update a docstring. |
418 |
all_heads = heads_to_fetch.union(if_present_fetch) |
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
419 |
ret = _mod_graph.PendingAncestryResult(all_heads, self.source_repo) |
420 |
if self.limit is not None: |
|
421 |
graph = self.source_repo.get_graph() |
|
422 |
topo_order = list(graph.iter_topo_order(ret.get_keys())) |
|
5852.1.6
by Jelmer Vernooij
Add extra test for Repository.search_missing_revision_ids. |
423 |
result_set = topo_order[:self.limit] |
424 |
ret = self.source_repo.revision_ids_to_search_result(result_set) |
|
425 |
return ret |
|
5852.1.5
by Andrew Bennetts, Jelmer Vernooij
Support limit= for fetching between Bazaar branches. |
426 |
else: |
427 |
return _mod_graph.NotInOtherForRevs(self.target_repo, self.source_repo, |
|
428 |
required_ids=heads_to_fetch, if_present_ids=if_present_fetch, |
|
429 |
limit=self.limit).execute() |