231
212
result = [prefix + suffix for suffix in suffix_list]
238
214
# we loop over each key because that spreads the inserts across prefixes,
239
215
# which is how commit operates.
240
216
for prefix in prefixes:
241
result.append(files.add_lines(prefix + get_key('origin'), (),
217
result.append(files.add_lines(prefix + ('origin',), (),
242
218
['origin' + last_char]))
243
219
for prefix in prefixes:
244
result.append(files.add_lines(prefix + get_key('base'),
220
result.append(files.add_lines(prefix + ('base',),
245
221
get_parents([('origin',)]), ['base' + last_char]))
246
222
for prefix in prefixes:
247
result.append(files.add_lines(prefix + get_key('left'),
223
result.append(files.add_lines(prefix + ('left',),
248
224
get_parents([('base',)]),
249
225
['base\n', 'left' + last_char]))
250
226
if not left_only:
251
227
for prefix in prefixes:
252
result.append(files.add_lines(prefix + get_key('right'),
228
result.append(files.add_lines(prefix + ('right',),
253
229
get_parents([('base',)]),
254
230
['base\n', 'right' + last_char]))
255
231
for prefix in prefixes:
256
result.append(files.add_lines(prefix + get_key('merged'),
232
result.append(files.add_lines(prefix + ('merged',),
257
233
get_parents([('left',), ('right',)]),
258
234
['base\n', 'left\n', 'right\n', 'merged' + last_char]))
1469
1442
transport.mkdir('.')
1470
1443
files = self.factory(transport)
1471
1444
if self.cleanup is not None:
1472
self.addCleanup(self.cleanup, files)
1445
self.addCleanup(lambda:self.cleanup(files))
1475
def get_simple_key(self, suffix):
1476
"""Return a key for the object under test."""
1477
if self.key_length == 1:
1480
return ('FileA',) + (suffix,)
1482
def test_add_lines(self):
1483
f = self.get_versionedfiles()
1484
key0 = self.get_simple_key('r0')
1485
key1 = self.get_simple_key('r1')
1486
key2 = self.get_simple_key('r2')
1487
keyf = self.get_simple_key('foo')
1488
f.add_lines(key0, [], ['a\n', 'b\n'])
1490
f.add_lines(key1, [key0], ['b\n', 'c\n'])
1492
f.add_lines(key1, [], ['b\n', 'c\n'])
1494
self.assertTrue(key0 in keys)
1495
self.assertTrue(key1 in keys)
1497
for record in f.get_record_stream([key0, key1], 'unordered', True):
1498
records.append((record.key, record.get_bytes_as('fulltext')))
1500
self.assertEqual([(key0, 'a\nb\n'), (key1, 'b\nc\n')], records)
1502
def test__add_text(self):
1503
f = self.get_versionedfiles()
1504
key0 = self.get_simple_key('r0')
1505
key1 = self.get_simple_key('r1')
1506
key2 = self.get_simple_key('r2')
1507
keyf = self.get_simple_key('foo')
1508
f._add_text(key0, [], 'a\nb\n')
1510
f._add_text(key1, [key0], 'b\nc\n')
1512
f._add_text(key1, [], 'b\nc\n')
1514
self.assertTrue(key0 in keys)
1515
self.assertTrue(key1 in keys)
1517
for record in f.get_record_stream([key0, key1], 'unordered', True):
1518
records.append((record.key, record.get_bytes_as('fulltext')))
1520
self.assertEqual([(key0, 'a\nb\n'), (key1, 'b\nc\n')], records)
1522
1448
def test_annotate(self):
1523
1449
files = self.get_versionedfiles()
1524
1450
self.get_diamond_files(files)
1558
1484
self.assertRaises(RevisionNotPresent,
1559
1485
files.annotate, prefix + ('missing-key',))
1561
def test_check_no_parameters(self):
1562
files = self.get_versionedfiles()
1564
def test_check_progressbar_parameter(self):
1565
"""A progress bar can be supplied because check can be a generator."""
1566
pb = ui.ui_factory.nested_progress_bar()
1567
self.addCleanup(pb.finished)
1568
files = self.get_versionedfiles()
1569
files.check(progress_bar=pb)
1571
def test_check_with_keys_becomes_generator(self):
1572
files = self.get_versionedfiles()
1573
self.get_diamond_files(files)
1575
entries = files.check(keys=keys)
1577
# Texts output should be fulltexts.
1578
self.capture_stream(files, entries, seen.add,
1579
files.get_parent_map(keys), require_fulltext=True)
1580
# All texts should be output.
1581
self.assertEqual(set(keys), seen)
1583
def test_clear_cache(self):
1584
files = self.get_versionedfiles()
1587
1487
def test_construct(self):
1588
1488
"""Each parameterised test can be constructed on a transport."""
1589
1489
files = self.get_versionedfiles()
1591
def get_diamond_files(self, files, trailing_eol=True, left_only=False,
1491
def get_diamond_files(self, files, trailing_eol=True, left_only=False):
1593
1492
return get_diamond_files(files, self.key_length,
1594
1493
trailing_eol=trailing_eol, nograph=not self.graph,
1595
left_only=left_only, nokeys=nokeys)
1494
left_only=left_only)
1597
def _add_content_nostoresha(self, add_lines):
1496
def test_add_lines_nostoresha(self):
1598
1497
"""When nostore_sha is supplied using old content raises."""
1599
1498
vf = self.get_versionedfiles()
1600
1499
empty_text = ('a', [])
1602
1501
sample_text_no_nl = ('c', ["foo\n", "bar"])
1604
1503
for version, lines in (empty_text, sample_text_nl, sample_text_no_nl):
1606
sha, _, _ = vf.add_lines(self.get_simple_key(version), [],
1609
sha, _, _ = vf._add_text(self.get_simple_key(version), [],
1504
sha, _, _ = vf.add_lines(self.get_simple_key(version), [], lines)
1611
1505
shas.append(sha)
1612
1506
# we now have a copy of all the lines in the vf.
1613
1507
for sha, (version, lines) in zip(
1616
1510
self.assertRaises(errors.ExistingContent,
1617
1511
vf.add_lines, new_key, [], lines,
1618
1512
nostore_sha=sha)
1619
self.assertRaises(errors.ExistingContent,
1620
vf._add_text, new_key, [], ''.join(lines),
1622
1513
# and no new version should have been added.
1623
1514
record = vf.get_record_stream([new_key], 'unordered', True).next()
1624
1515
self.assertEqual('absent', record.storage_kind)
1626
def test_add_lines_nostoresha(self):
1627
self._add_content_nostoresha(add_lines=True)
1629
def test__add_text_nostoresha(self):
1630
self._add_content_nostoresha(add_lines=False)
1632
1517
def test_add_lines_return(self):
1633
1518
files = self.get_versionedfiles()
1634
1519
# save code by using the stock data insertion helper.
1660
1545
('ed8bce375198ea62444dc71952b22cfc2b09226d', 23)],
1663
def test_add_lines_no_key_generates_chk_key(self):
1664
files = self.get_versionedfiles()
1665
# save code by using the stock data insertion helper.
1666
adds = self.get_diamond_files(files, nokeys=True)
1668
# We can only validate the first 2 elements returned from add_lines.
1670
self.assertEqual(3, len(add))
1671
results.append(add[:2])
1672
if self.key_length == 1:
1674
('00e364d235126be43292ab09cb4686cf703ddc17', 7),
1675
('51c64a6f4fc375daf0d24aafbabe4d91b6f4bb44', 5),
1676
('a8478686da38e370e32e42e8a0c220e33ee9132f', 10),
1677
('9ef09dfa9d86780bdec9219a22560c6ece8e0ef1', 11),
1678
('ed8bce375198ea62444dc71952b22cfc2b09226d', 23)],
1680
# Check the added items got CHK keys.
1681
self.assertEqual(set([
1682
('sha1:00e364d235126be43292ab09cb4686cf703ddc17',),
1683
('sha1:51c64a6f4fc375daf0d24aafbabe4d91b6f4bb44',),
1684
('sha1:9ef09dfa9d86780bdec9219a22560c6ece8e0ef1',),
1685
('sha1:a8478686da38e370e32e42e8a0c220e33ee9132f',),
1686
('sha1:ed8bce375198ea62444dc71952b22cfc2b09226d',),
1689
elif self.key_length == 2:
1691
('00e364d235126be43292ab09cb4686cf703ddc17', 7),
1692
('00e364d235126be43292ab09cb4686cf703ddc17', 7),
1693
('51c64a6f4fc375daf0d24aafbabe4d91b6f4bb44', 5),
1694
('51c64a6f4fc375daf0d24aafbabe4d91b6f4bb44', 5),
1695
('a8478686da38e370e32e42e8a0c220e33ee9132f', 10),
1696
('a8478686da38e370e32e42e8a0c220e33ee9132f', 10),
1697
('9ef09dfa9d86780bdec9219a22560c6ece8e0ef1', 11),
1698
('9ef09dfa9d86780bdec9219a22560c6ece8e0ef1', 11),
1699
('ed8bce375198ea62444dc71952b22cfc2b09226d', 23),
1700
('ed8bce375198ea62444dc71952b22cfc2b09226d', 23)],
1702
# Check the added items got CHK keys.
1703
self.assertEqual(set([
1704
('FileA', 'sha1:00e364d235126be43292ab09cb4686cf703ddc17'),
1705
('FileA', 'sha1:51c64a6f4fc375daf0d24aafbabe4d91b6f4bb44'),
1706
('FileA', 'sha1:9ef09dfa9d86780bdec9219a22560c6ece8e0ef1'),
1707
('FileA', 'sha1:a8478686da38e370e32e42e8a0c220e33ee9132f'),
1708
('FileA', 'sha1:ed8bce375198ea62444dc71952b22cfc2b09226d'),
1709
('FileB', 'sha1:00e364d235126be43292ab09cb4686cf703ddc17'),
1710
('FileB', 'sha1:51c64a6f4fc375daf0d24aafbabe4d91b6f4bb44'),
1711
('FileB', 'sha1:9ef09dfa9d86780bdec9219a22560c6ece8e0ef1'),
1712
('FileB', 'sha1:a8478686da38e370e32e42e8a0c220e33ee9132f'),
1713
('FileB', 'sha1:ed8bce375198ea62444dc71952b22cfc2b09226d'),
1717
1548
def test_empty_lines(self):
1718
1549
"""Empty files can be stored."""
1719
1550
f = self.get_versionedfiles()
1741
1572
f.get_record_stream([key_b], 'unordered', True
1742
1573
).next().get_bytes_as('fulltext'))
1744
def test_get_known_graph_ancestry(self):
1745
f = self.get_versionedfiles()
1747
raise TestNotApplicable('ancestry info only relevant with graph.')
1748
key_a = self.get_simple_key('a')
1749
key_b = self.get_simple_key('b')
1750
key_c = self.get_simple_key('c')
1756
f.add_lines(key_a, [], ['\n'])
1757
f.add_lines(key_b, [key_a], ['\n'])
1758
f.add_lines(key_c, [key_a, key_b], ['\n'])
1759
kg = f.get_known_graph_ancestry([key_c])
1760
self.assertIsInstance(kg, _mod_graph.KnownGraph)
1761
self.assertEqual([key_a, key_b, key_c], list(kg.topo_sort()))
1763
def test_known_graph_with_fallbacks(self):
1764
f = self.get_versionedfiles('files')
1766
raise TestNotApplicable('ancestry info only relevant with graph.')
1767
if getattr(f, 'add_fallback_versioned_files', None) is None:
1768
raise TestNotApplicable("%s doesn't support fallbacks"
1769
% (f.__class__.__name__,))
1770
key_a = self.get_simple_key('a')
1771
key_b = self.get_simple_key('b')
1772
key_c = self.get_simple_key('c')
1773
# A only in fallback
1778
g = self.get_versionedfiles('fallback')
1779
g.add_lines(key_a, [], ['\n'])
1780
f.add_fallback_versioned_files(g)
1781
f.add_lines(key_b, [key_a], ['\n'])
1782
f.add_lines(key_c, [key_a, key_b], ['\n'])
1783
kg = f.get_known_graph_ancestry([key_c])
1784
self.assertEqual([key_a, key_b, key_c], list(kg.topo_sort()))
1786
1575
def test_get_record_stream_empty(self):
1787
1576
"""An empty stream can be requested without error."""
1788
1577
f = self.get_versionedfiles()
1796
1585
'knit-ft', 'knit-delta', 'chunked', 'fulltext',
1797
1586
'knit-annotated-ft-gz', 'knit-annotated-delta-gz', 'knit-ft-gz',
1798
1587
'knit-delta-gz',
1799
'knit-delta-closure', 'knit-delta-closure-ref',
1800
'groupcompress-block', 'groupcompress-block-ref'])
1588
'knit-delta-closure', 'knit-delta-closure-ref'])
1802
def capture_stream(self, f, entries, on_seen, parents,
1803
require_fulltext=False):
1590
def capture_stream(self, f, entries, on_seen, parents):
1804
1591
"""Capture a stream for testing."""
1805
1592
for factory in entries:
1806
1593
on_seen(factory.key)
1807
1594
self.assertValidStorageKind(factory.storage_kind)
1808
if factory.sha1 is not None:
1809
self.assertEqual(f.get_sha1s([factory.key])[factory.key],
1595
self.assertEqual(f.get_sha1s([factory.key])[factory.key],
1811
1597
self.assertEqual(parents[factory.key], factory.parents)
1812
1598
self.assertIsInstance(factory.get_bytes_as(factory.storage_kind),
1814
if require_fulltext:
1815
factory.get_bytes_as('fulltext')
1817
1601
def test_get_record_stream_interface(self):
1818
1602
"""each item in a stream has to provide a regular interface."""
2196
def test_get_annotator(self):
2197
files = self.get_versionedfiles()
2198
self.get_diamond_files(files)
2199
origin_key = self.get_simple_key('origin')
2200
base_key = self.get_simple_key('base')
2201
left_key = self.get_simple_key('left')
2202
right_key = self.get_simple_key('right')
2203
merged_key = self.get_simple_key('merged')
2204
# annotator = files.get_annotator()
2205
# introduced full text
2206
origins, lines = files.get_annotator().annotate(origin_key)
2207
self.assertEqual([(origin_key,)], origins)
2208
self.assertEqual(['origin\n'], lines)
2210
origins, lines = files.get_annotator().annotate(base_key)
2211
self.assertEqual([(base_key,)], origins)
2213
origins, lines = files.get_annotator().annotate(merged_key)
2222
# Without a graph everything is new.
2229
self.assertRaises(RevisionNotPresent,
2230
files.get_annotator().annotate, self.get_simple_key('missing-key'))
2232
1985
def test_get_parent_map(self):
2233
1986
files = self.get_versionedfiles()
2234
1987
if self.key_length == 1:
2438
2191
self.assertIdenticalVersionedFile(source, files)
2440
def test_insert_record_stream_long_parent_chain_out_of_order(self):
2441
"""An out of order stream can either error or work."""
2443
raise TestNotApplicable('ancestry info only relevant with graph.')
2444
# Create a reasonably long chain of records based on each other, where
2445
# most will be deltas.
2446
source = self.get_versionedfiles('source')
2449
content = [('same same %d\n' % n) for n in range(500)]
2450
for letter in 'abcdefghijklmnopqrstuvwxyz':
2451
key = ('key-' + letter,)
2452
if self.key_length == 2:
2453
key = ('prefix',) + key
2454
content.append('content for ' + letter + '\n')
2455
source.add_lines(key, parents, content)
2458
# Create a stream of these records, excluding the first record that the
2459
# rest ultimately depend upon, and insert it into a new vf.
2461
for key in reversed(keys):
2462
streams.append(source.get_record_stream([key], 'unordered', False))
2463
deltas = chain(*streams[:-1])
2464
files = self.get_versionedfiles()
2466
files.insert_record_stream(deltas)
2467
except RevisionNotPresent:
2468
# Must not have corrupted the file.
2471
# Must only report either just the first key as a missing parent,
2472
# no key as missing (for nodelta scenarios).
2473
missing = set(files.get_missing_compression_parent_keys())
2474
missing.discard(keys[0])
2475
self.assertEqual(set(), missing)
2477
2193
def get_knit_delta_source(self):
2478
2194
"""Get a source that can produce a stream with knit delta records,
2479
2195
regardless of this test's scenario.