74
def recreate_search(self, repository, search_bytes):
78
def recreate_search(self, repository, search_bytes, discard_excess=False):
79
"""Recreate a search from its serialised form.
81
:param discard_excess: If True, and the search refers to data we don't
82
have, just silently accept that fact - the verb calling
83
recreate_search trusts that clients will look for missing things
84
they expected and get it from elsewhere.
75
86
lines = search_bytes.split('\n')
76
87
if lines[0] == 'ancestry-of':
78
89
search_result = graph.PendingAncestryResult(heads, repository)
79
90
return search_result, None
80
91
elif lines[0] == 'search':
81
return self.recreate_search_from_recipe(repository, lines[1:])
92
return self.recreate_search_from_recipe(repository, lines[1:],
93
discard_excess=discard_excess)
83
95
return (None, FailedSmartServerResponse(('BadSearch',)))
85
def recreate_search_from_recipe(self, repository, lines):
97
def recreate_search_from_recipe(self, repository, lines,
98
discard_excess=False):
99
"""Recreate a specific revision search (vs a from-tip search).
101
:param discard_excess: If True, and the search refers to data we don't
102
have, just silently accept that fact - the verb calling
103
recreate_search trusts that clients will look for missing things
104
they expected and get it from elsewhere.
86
106
start_keys = set(lines[0].split(' '))
87
107
exclude_keys = set(lines[1].split(' '))
88
108
revision_count = int(lines[2])
267
288
return SuccessfulSmartServerResponse(('ok', ), '\n'.join(lines))
291
class SmartServerRepositoryGetRevIdForRevno(SmartServerRepositoryReadLocked):
293
def do_readlocked_repository_request(self, repository, revno,
295
"""Find the revid for a given revno, given a known revno/revid pair.
300
found_flag, result = repository.get_rev_id_for_revno(revno, known_pair)
301
except errors.RevisionNotPresent, err:
302
if err.revision_id != known_pair[1]:
303
raise AssertionError(
304
'get_rev_id_for_revno raised RevisionNotPresent for '
305
'non-initial revision: ' + err.revision_id)
306
return FailedSmartServerResponse(
307
('nosuchrevision', err.revision_id))
309
return SuccessfulSmartServerResponse(('ok', result))
311
earliest_revno, earliest_revid = result
312
return SuccessfulSmartServerResponse(
313
('history-incomplete', earliest_revno, earliest_revid))
270
316
class SmartServerRequestHasRevision(SmartServerRepositoryRequest):
272
318
def do_repository_request(self, repository, revision_id):
375
421
self._to_format = network_format_registry.get(to_network_name)
422
if self._should_fake_unknown():
423
return FailedSmartServerResponse(
424
('UnknownMethod', 'Repository.get_stream'))
376
425
return None # Signal that we want a body.
427
def _should_fake_unknown(self):
428
"""Return True if we should return UnknownMethod to the client.
430
This is a workaround for bugs in pre-1.19 clients that claim to
431
support receiving streams of CHK repositories. The pre-1.19 client
432
expects inventory records to be serialized in the format defined by
433
to_network_name, but in pre-1.19 (at least) that format definition
434
tries to use the xml5 serializer, which does not correctly handle
435
rich-roots. After 1.19 the client can also accept inventory-deltas
436
(which avoids this issue), and those clients will use the
437
Repository.get_stream_1.19 verb instead of this one.
438
So: if this repository is CHK, and the to_format doesn't match,
439
we should just fake an UnknownSmartMethod error so that the client
440
will fallback to VFS, rather than sending it a stream we know it
443
from_format = self._repository._format
444
to_format = self._to_format
445
if not from_format.supports_chks:
446
# Source not CHK: that's ok
448
if (to_format.supports_chks and
449
from_format.repository_class is to_format.repository_class and
450
from_format._serializer == to_format._serializer):
451
# Source is CHK, but target matches: that's ok
452
# (e.g. 2a->2a, or CHK2->2a)
454
# Source is CHK, and target is not CHK or incompatible CHK. We can't
455
# generate a compatible stream.
378
458
def do_body(self, body_bytes):
379
459
repository = self._repository
380
460
repository.lock_read()
382
search_result, error = self.recreate_search(repository, body_bytes)
462
search_result, error = self.recreate_search(repository, body_bytes,
383
464
if error is not None:
384
465
repository.unlock()
428
520
yield pack_writer.end()
431
def _byte_stream_to_stream(byte_stream):
523
class _ByteStreamDecoder(object):
524
"""Helper for _byte_stream_to_stream.
526
The expected usage of this class is via the function _byte_stream_to_stream
527
which creates a _ByteStreamDecoder, pops off the stream format and then
528
yields the output of record_stream(), the main entry point to
531
Broadly this class has to unwrap two layers of iterators:
535
This is complicated by wishing to return type, iterator_for_type, but
536
getting the data for iterator_for_type when we find out type: we can't
537
simply pass a generator down to the NetworkRecordStream parser, instead
538
we have a little local state to seed each NetworkRecordStream instance,
539
and gather the type that we'll be yielding.
541
:ivar byte_stream: The byte stream being decoded.
542
:ivar stream_decoder: A pack parser used to decode the bytestream
543
:ivar current_type: The current type, used to join adjacent records of the
544
same type into a single stream.
545
:ivar first_bytes: The first bytes to give the next NetworkRecordStream.
548
def __init__(self, byte_stream, record_counter):
549
"""Create a _ByteStreamDecoder."""
550
self.stream_decoder = pack.ContainerPushParser()
551
self.current_type = None
552
self.first_bytes = None
553
self.byte_stream = byte_stream
554
self._record_counter = record_counter
557
def iter_stream_decoder(self):
558
"""Iterate the contents of the pack from stream_decoder."""
559
# dequeue pending items
560
for record in self.stream_decoder.read_pending_records():
562
# Pull bytes of the wire, decode them to records, yield those records.
563
for bytes in self.byte_stream:
564
self.stream_decoder.accept_bytes(bytes)
565
for record in self.stream_decoder.read_pending_records():
568
def iter_substream_bytes(self):
569
if self.first_bytes is not None:
570
yield self.first_bytes
571
# If we run out of pack records, single the outer layer to stop.
572
self.first_bytes = None
573
for record in self.iter_pack_records:
574
record_names, record_bytes = record
575
record_name, = record_names
576
substream_type = record_name[0]
577
if substream_type != self.current_type:
578
# end of a substream, seed the next substream.
579
self.current_type = substream_type
580
self.first_bytes = record_bytes
584
def record_stream(self):
585
"""Yield substream_type, substream from the byte stream."""
586
def wrap_and_count(pb, rc, substream):
587
"""Yield records from stream while showing progress."""
590
if self.current_type != 'revisions' and self.key_count != 0:
591
# As we know the number of revisions now (in self.key_count)
592
# we can setup and use record_counter (rc).
593
if not rc.is_initialized():
594
rc.setup(self.key_count, self.key_count)
595
for record in substream.read():
597
if rc.is_initialized() and counter == rc.STEP:
598
rc.increment(counter)
599
pb.update('Estimate', rc.current, rc.max)
601
if self.current_type == 'revisions':
602
# Total records is proportional to number of revs
603
# to fetch. With remote, we used self.key_count to
604
# track the number of revs. Once we have the revs
605
# counts in self.key_count, the progress bar changes
606
# from 'Estimating..' to 'Estimate' above.
608
if counter == rc.STEP:
609
pb.update('Estimating..', self.key_count)
615
pb = ui.ui_factory.nested_progress_bar()
616
rc = self._record_counter
617
# Make and consume sub generators, one per substream type:
618
while self.first_bytes is not None:
619
substream = NetworkRecordStream(self.iter_substream_bytes())
620
# after substream is fully consumed, self.current_type is set to
621
# the next type, and self.first_bytes is set to the matching bytes.
622
yield self.current_type, wrap_and_count(pb, rc, substream)
624
pb.update('Done', rc.max, rc.max)
627
def seed_state(self):
628
"""Prepare the _ByteStreamDecoder to decode from the pack stream."""
629
# Set a single generator we can use to get data from the pack stream.
630
self.iter_pack_records = self.iter_stream_decoder()
631
# Seed the very first subiterator with content; after this each one
633
list(self.iter_substream_bytes())
636
def _byte_stream_to_stream(byte_stream, record_counter=None):
432
637
"""Convert a byte stream into a format and a stream.
434
639
:param byte_stream: A bytes iterator, as output by _stream_to_byte_stream.
435
640
:return: (RepositoryFormat, stream_generator)
437
stream_decoder = pack.ContainerPushParser()
439
"""Closure to return the substreams."""
440
# May have fully parsed records already.
441
for record in stream_decoder.read_pending_records():
442
record_names, record_bytes = record
443
record_name, = record_names
444
substream_type = record_name[0]
445
substream = NetworkRecordStream([record_bytes])
446
yield substream_type, substream.read()
447
for bytes in byte_stream:
448
stream_decoder.accept_bytes(bytes)
449
for record in stream_decoder.read_pending_records():
450
record_names, record_bytes = record
451
record_name, = record_names
452
substream_type = record_name[0]
453
substream = NetworkRecordStream([record_bytes])
454
yield substream_type, substream.read()
642
decoder = _ByteStreamDecoder(byte_stream, record_counter)
455
643
for bytes in byte_stream:
456
stream_decoder.accept_bytes(bytes)
457
for record in stream_decoder.read_pending_records(max=1):
644
decoder.stream_decoder.accept_bytes(bytes)
645
for record in decoder.stream_decoder.read_pending_records(max=1):
458
646
record_names, src_format_name = record
459
647
src_format = network_format_registry.get(src_format_name)
460
return src_format, record_stream()
648
return src_format, decoder.record_stream()
463
651
class SmartServerRepositoryUnlock(SmartServerRepositoryRequest):