416
434
self.assertEqual(sorted(nodes), nodes)
417
435
self.assertEqual(16, len(nodes))
419
def test_spill_index_stress_1_1_no_combine(self):
420
builder = btree_index.BTreeBuilder(key_elements=1, spill_at=2)
421
builder.set_optimize(for_size=False, combine_backing_indices=False)
422
nodes = [node[0:2] for node in self.make_nodes(16, 1, 0)]
423
builder.add_node(*nodes[0])
424
# Test the parts of the index that take up memory are doing so
426
self.assertEqual(1, len(builder._nodes))
427
self.assertIs(None, builder._nodes_by_key)
428
builder.add_node(*nodes[1])
429
self.assertEqual(0, len(builder._nodes))
430
self.assertIs(None, builder._nodes_by_key)
431
self.assertEqual(1, len(builder._backing_indices))
432
self.assertEqual(2, builder._backing_indices[0].key_count())
434
builder.add_node(*nodes[2])
435
self.assertEqual(1, len(builder._nodes))
436
self.assertIs(None, builder._nodes_by_key)
437
# And spills to a second backing index but doesn't combine
438
builder.add_node(*nodes[3])
439
self.assertEqual(0, len(builder._nodes))
440
self.assertIs(None, builder._nodes_by_key)
441
self.assertEqual(2, len(builder._backing_indices))
442
for backing_index in builder._backing_indices:
443
self.assertEqual(2, backing_index.key_count())
444
# The next spills to the 3rd slot
445
builder.add_node(*nodes[4])
446
builder.add_node(*nodes[5])
447
self.assertEqual(0, len(builder._nodes))
448
self.assertIs(None, builder._nodes_by_key)
449
self.assertEqual(3, len(builder._backing_indices))
450
for backing_index in builder._backing_indices:
451
self.assertEqual(2, backing_index.key_count())
452
# Now spill a few more, and check that we don't combine
453
builder.add_node(*nodes[6])
454
builder.add_node(*nodes[7])
455
builder.add_node(*nodes[8])
456
builder.add_node(*nodes[9])
457
builder.add_node(*nodes[10])
458
builder.add_node(*nodes[11])
459
builder.add_node(*nodes[12])
460
self.assertEqual(6, len(builder._backing_indices))
461
for backing_index in builder._backing_indices:
462
self.assertEqual(2, backing_index.key_count())
463
# Test that memory and disk are both used for query methods; and that
464
# None is skipped over happily.
465
self.assertEqual([(builder,) + node for node in sorted(nodes[:13])],
466
list(builder.iter_all_entries()))
467
# Two nodes - one memory one disk
468
self.assertEqual(set([(builder,) + node for node in nodes[11:13]]),
469
set(builder.iter_entries([nodes[12][0], nodes[11][0]])))
470
self.assertEqual(13, builder.key_count())
471
self.assertEqual(set([(builder,) + node for node in nodes[11:13]]),
472
set(builder.iter_entries_prefix([nodes[12][0], nodes[11][0]])))
473
builder.add_node(*nodes[13])
474
builder.add_node(*nodes[14])
475
builder.add_node(*nodes[15])
476
self.assertEqual(8, len(builder._backing_indices))
477
for backing_index in builder._backing_indices:
478
self.assertEqual(2, backing_index.key_count())
479
# Now finish, and check we got a correctly ordered tree
480
transport = self.get_transport('')
481
size = transport.put_file('index', builder.finish())
482
index = btree_index.BTreeGraphIndex(transport, 'index', size)
483
nodes = list(index.iter_all_entries())
484
self.assertEqual(sorted(nodes), nodes)
485
self.assertEqual(16, len(nodes))
487
def test_set_optimize(self):
488
builder = btree_index.BTreeBuilder(key_elements=2, reference_lists=2)
489
builder.set_optimize(for_size=True)
490
self.assertTrue(builder._optimize_for_size)
491
builder.set_optimize(for_size=False)
492
self.assertFalse(builder._optimize_for_size)
493
# test that we can set combine_backing_indices without effecting
496
builder._optimize_for_size = obj
497
builder.set_optimize(combine_backing_indices=False)
498
self.assertFalse(builder._combine_backing_indices)
499
self.assertIs(obj, builder._optimize_for_size)
500
builder.set_optimize(combine_backing_indices=True)
501
self.assertTrue(builder._combine_backing_indices)
502
self.assertIs(obj, builder._optimize_for_size)
504
437
def test_spill_index_stress_2_2(self):
505
438
# test that references and longer keys don't confuse things.
506
439
builder = btree_index.BTreeBuilder(key_elements=2, reference_lists=2,
979
833
(index, ('name', 'fin2'), 'beta', ((), ))]),
980
834
set(index.iter_entries_prefix([('name', None)])))
982
# XXX: external_references tests are duplicated in test_index. We
983
# probably should have per_graph_index tests...
984
def test_external_references_no_refs(self):
985
index = self.make_index(ref_lists=0, nodes=[])
986
self.assertRaises(ValueError, index.external_references, 0)
988
def test_external_references_no_results(self):
989
index = self.make_index(ref_lists=1, nodes=[
990
(('key',), 'value', ([],))])
991
self.assertEqual(set(), index.external_references(0))
993
def test_external_references_missing_ref(self):
994
missing_key = ('missing',)
995
index = self.make_index(ref_lists=1, nodes=[
996
(('key',), 'value', ([missing_key],))])
997
self.assertEqual(set([missing_key]), index.external_references(0))
999
def test_external_references_multiple_ref_lists(self):
1000
missing_key = ('missing',)
1001
index = self.make_index(ref_lists=2, nodes=[
1002
(('key',), 'value', ([], [missing_key]))])
1003
self.assertEqual(set([]), index.external_references(0))
1004
self.assertEqual(set([missing_key]), index.external_references(1))
1006
def test_external_references_two_records(self):
1007
index = self.make_index(ref_lists=1, nodes=[
1008
(('key-1',), 'value', ([('key-2',)],)),
1009
(('key-2',), 'value', ([],)),
1011
self.assertEqual(set([]), index.external_references(0))
1013
def test__find_ancestors_one_page(self):
1016
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1017
(key1, 'value', ([key2],)),
1018
(key2, 'value', ([],)),
1021
missing_keys = set()
1022
search_keys = index._find_ancestors([key1], 0, parent_map, missing_keys)
1023
self.assertEqual({key1: (key2,), key2: ()}, parent_map)
1024
self.assertEqual(set(), missing_keys)
1025
self.assertEqual(set(), search_keys)
1027
def test__find_ancestors_one_page_w_missing(self):
1031
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1032
(key1, 'value', ([key2],)),
1033
(key2, 'value', ([],)),
1036
missing_keys = set()
1037
search_keys = index._find_ancestors([key2, key3], 0, parent_map,
1039
self.assertEqual({key2: ()}, parent_map)
1040
# we know that key3 is missing because we read the page that it would
1042
self.assertEqual(set([key3]), missing_keys)
1043
self.assertEqual(set(), search_keys)
1045
def test__find_ancestors_one_parent_missing(self):
1049
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1050
(key1, 'value', ([key2],)),
1051
(key2, 'value', ([key3],)),
1054
missing_keys = set()
1055
search_keys = index._find_ancestors([key1], 0, parent_map,
1057
self.assertEqual({key1: (key2,), key2: (key3,)}, parent_map)
1058
self.assertEqual(set(), missing_keys)
1059
# all we know is that key3 wasn't present on the page we were reading
1060
# but if you look, the last key is key2 which comes before key3, so we
1061
# don't know whether key3 would land on this page or not.
1062
self.assertEqual(set([key3]), search_keys)
1063
search_keys = index._find_ancestors(search_keys, 0, parent_map,
1065
# passing it back in, we are sure it is 'missing'
1066
self.assertEqual({key1: (key2,), key2: (key3,)}, parent_map)
1067
self.assertEqual(set([key3]), missing_keys)
1068
self.assertEqual(set([]), search_keys)
1070
def test__find_ancestors_dont_search_known(self):
1074
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1075
(key1, 'value', ([key2],)),
1076
(key2, 'value', ([key3],)),
1077
(key3, 'value', ([],)),
1079
# We already know about key2, so we won't try to search for key3
1080
parent_map = {key2: (key3,)}
1081
missing_keys = set()
1082
search_keys = index._find_ancestors([key1], 0, parent_map,
1084
self.assertEqual({key1: (key2,), key2: (key3,)}, parent_map)
1085
self.assertEqual(set(), missing_keys)
1086
self.assertEqual(set(), search_keys)
1088
def test__find_ancestors_multiple_pages(self):
1089
# We need to use enough keys that we actually cause a split
1090
start_time = 1249671539
1091
email = "joebob@example.com"
1095
for i in xrange(400):
1096
rev_id = '%s-%s-%s' % (email,
1097
osutils.compact_date(start_time + i),
1098
osutils.rand_chars(16))
1100
nodes.append((rev_key, 'value', ref_lists))
1101
# We have a ref 'list' of length 1, with a list of parents, with 1
1102
# parent which is a key
1103
ref_lists = ((rev_key,),)
1104
rev_keys.append(rev_key)
1105
index = self.make_index(ref_lists=1, key_elements=1, nodes=nodes)
1106
self.assertEqual(400, index.key_count())
1107
self.assertEqual(3, len(index._row_offsets))
1108
nodes = dict(index._read_nodes([1, 2]))
1111
min_l2_key = l2.min_key
1112
max_l1_key = l1.max_key
1113
self.assertTrue(max_l1_key < min_l2_key)
1114
parents_min_l2_key = l2.keys[min_l2_key][1][0]
1115
self.assertEqual((l1.max_key,), parents_min_l2_key)
1116
# Now, whatever key we select that would fall on the second page,
1117
# should give us all the parents until the page break
1118
key_idx = rev_keys.index(min_l2_key)
1119
next_key = rev_keys[key_idx+1]
1120
# So now when we get the parent map, we should get the key we are
1121
# looking for, min_l2_key, and then a reference to go look for the
1122
# parent of that key
1124
missing_keys = set()
1125
search_keys = index._find_ancestors([next_key], 0, parent_map,
1127
self.assertEqual([min_l2_key, next_key], sorted(parent_map))
1128
self.assertEqual(set(), missing_keys)
1129
self.assertEqual(set([max_l1_key]), search_keys)
1131
search_keys = index._find_ancestors([max_l1_key], 0, parent_map,
1133
self.assertEqual(sorted(l1.keys), sorted(parent_map))
1134
self.assertEqual(set(), missing_keys)
1135
self.assertEqual(set(), search_keys)
1137
def test__find_ancestors_empty_index(self):
1138
index = self.make_index(ref_lists=1, key_elements=1, nodes=[])
1140
missing_keys = set()
1141
search_keys = index._find_ancestors([('one',), ('two',)], 0, parent_map,
1143
self.assertEqual(set(), search_keys)
1144
self.assertEqual({}, parent_map)
1145
self.assertEqual(set([('one',), ('two',)]), missing_keys)
1147
def test_supports_unlimited_cache(self):
1148
builder = btree_index.BTreeBuilder(reference_lists=0, key_elements=1)
1149
# We need enough nodes to cause a page split (so we have both an
1150
# internal node and a couple leaf nodes. 500 seems to be enough.)
1151
nodes = self.make_nodes(500, 1, 0)
1153
builder.add_node(*node)
1154
stream = builder.finish()
1155
trans = get_transport(self.get_url())
1156
size = trans.put_file('index', stream)
1157
index = btree_index.BTreeGraphIndex(trans, 'index', size)
1158
self.assertEqual(500, index.key_count())
1159
# We have an internal node
1160
self.assertEqual(2, len(index._row_lengths))
1161
# We have at least 2 leaf nodes
1162
self.assertTrue(index._row_lengths[-1] >= 2)
1163
self.assertIsInstance(index._leaf_node_cache, lru_cache.LRUCache)
1164
self.assertEqual(btree_index._NODE_CACHE_SIZE,
1165
index._leaf_node_cache._max_cache)
1166
self.assertIsInstance(index._internal_node_cache, fifo_cache.FIFOCache)
1167
self.assertEqual(100, index._internal_node_cache._max_cache)
1168
# No change if unlimited_cache=False is passed
1169
index = btree_index.BTreeGraphIndex(trans, 'index', size,
1170
unlimited_cache=False)
1171
self.assertIsInstance(index._leaf_node_cache, lru_cache.LRUCache)
1172
self.assertEqual(btree_index._NODE_CACHE_SIZE,
1173
index._leaf_node_cache._max_cache)
1174
self.assertIsInstance(index._internal_node_cache, fifo_cache.FIFOCache)
1175
self.assertEqual(100, index._internal_node_cache._max_cache)
1176
index = btree_index.BTreeGraphIndex(trans, 'index', size,
1177
unlimited_cache=True)
1178
self.assertIsInstance(index._leaf_node_cache, dict)
1179
self.assertIs(type(index._internal_node_cache), dict)
1180
# Exercise the lookup code
1181
entries = set(index.iter_entries([n[0] for n in nodes]))
1182
self.assertEqual(500, len(entries))
1185
837
class TestBTreeNodes(BTreeTestCase):
839
def restore_parser(self):
840
btree_index._btree_serializer = self.saved_parser
1187
842
def setUp(self):
1188
843
BTreeTestCase.setUp(self)
1189
self.overrideAttr(btree_index, '_btree_serializer', self.parse_btree)
844
self.saved_parser = btree_index._btree_serializer
845
self.addCleanup(self.restore_parser)
846
btree_index._btree_serializer = self.parse_btree
1191
848
def test_LeafNode_1_0(self):
1192
849
node_bytes = ("type=leaf\n"
1339
996
(4, ['g', 'h'])],
1340
997
['a', 'b', 'd', 'e', 'g', 'h'],
1341
998
['c', 'd', 'f', 'g'])
1344
class TestExpandOffsets(tests.TestCase):
1346
def make_index(self, size, recommended_pages=None):
1347
"""Make an index with a generic size.
1349
This doesn't actually create anything on disk, it just primes a
1350
BTreeGraphIndex with the recommended information.
1352
index = btree_index.BTreeGraphIndex(get_transport('memory:///'),
1353
'test-index', size=size)
1354
if recommended_pages is not None:
1355
index._recommended_pages = recommended_pages
1358
def set_cached_offsets(self, index, cached_offsets):
1359
"""Monkeypatch to give a canned answer for _get_offsets_for...()."""
1360
def _get_offsets_to_cached_pages():
1361
cached = set(cached_offsets)
1363
index._get_offsets_to_cached_pages = _get_offsets_to_cached_pages
1365
def prepare_index(self, index, node_ref_lists, key_length, key_count,
1366
row_lengths, cached_offsets):
1367
"""Setup the BTreeGraphIndex with some pre-canned information."""
1368
index.node_ref_lists = node_ref_lists
1369
index._key_length = key_length
1370
index._key_count = key_count
1371
index._row_lengths = row_lengths
1372
index._compute_row_offsets()
1373
index._root_node = btree_index._InternalNode('internal\noffset=0\n')
1374
self.set_cached_offsets(index, cached_offsets)
1376
def make_100_node_index(self):
1377
index = self.make_index(4096*100, 6)
1378
# Consider we've already made a single request at the middle
1379
self.prepare_index(index, node_ref_lists=0, key_length=1,
1380
key_count=1000, row_lengths=[1, 99],
1381
cached_offsets=[0, 50])
1384
def make_1000_node_index(self):
1385
index = self.make_index(4096*1000, 6)
1386
# Pretend we've already made a single request in the middle
1387
self.prepare_index(index, node_ref_lists=0, key_length=1,
1388
key_count=90000, row_lengths=[1, 9, 990],
1389
cached_offsets=[0, 5, 500])
1392
def assertNumPages(self, expected_pages, index, size):
1394
self.assertEqual(expected_pages, index._compute_total_pages_in_index())
1396
def assertExpandOffsets(self, expected, index, offsets):
1397
self.assertEqual(expected, index._expand_offsets(offsets),
1398
'We did not get the expected value after expanding'
1401
def test_default_recommended_pages(self):
1402
index = self.make_index(None)
1403
# local transport recommends 4096 byte reads, which is 1 page
1404
self.assertEqual(1, index._recommended_pages)
1406
def test__compute_total_pages_in_index(self):
1407
index = self.make_index(None)
1408
self.assertNumPages(1, index, 1024)
1409
self.assertNumPages(1, index, 4095)
1410
self.assertNumPages(1, index, 4096)
1411
self.assertNumPages(2, index, 4097)
1412
self.assertNumPages(2, index, 8192)
1413
self.assertNumPages(76, index, 4096*75 + 10)
1415
def test__find_layer_start_and_stop(self):
1416
index = self.make_1000_node_index()
1417
self.assertEqual((0, 1), index._find_layer_first_and_end(0))
1418
self.assertEqual((1, 10), index._find_layer_first_and_end(1))
1419
self.assertEqual((1, 10), index._find_layer_first_and_end(9))
1420
self.assertEqual((10, 1000), index._find_layer_first_and_end(10))
1421
self.assertEqual((10, 1000), index._find_layer_first_and_end(99))
1422
self.assertEqual((10, 1000), index._find_layer_first_and_end(999))
1424
def test_unknown_size(self):
1425
# We should not expand if we don't know the file size
1426
index = self.make_index(None, 10)
1427
self.assertExpandOffsets([0], index, [0])
1428
self.assertExpandOffsets([1, 4, 9], index, [1, 4, 9])
1430
def test_more_than_recommended(self):
1431
index = self.make_index(4096*100, 2)
1432
self.assertExpandOffsets([1, 10], index, [1, 10])
1433
self.assertExpandOffsets([1, 10, 20], index, [1, 10, 20])
1435
def test_read_all_from_root(self):
1436
index = self.make_index(4096*10, 20)
1437
self.assertExpandOffsets(range(10), index, [0])
1439
def test_read_all_when_cached(self):
1440
# We've read enough that we can grab all the rest in a single request
1441
index = self.make_index(4096*10, 5)
1442
self.prepare_index(index, node_ref_lists=0, key_length=1,
1443
key_count=1000, row_lengths=[1, 9],
1444
cached_offsets=[0, 1, 2, 5, 6])
1445
# It should fill the remaining nodes, regardless of the one requested
1446
self.assertExpandOffsets([3, 4, 7, 8, 9], index, [3])
1447
self.assertExpandOffsets([3, 4, 7, 8, 9], index, [8])
1448
self.assertExpandOffsets([3, 4, 7, 8, 9], index, [9])
1450
def test_no_root_node(self):
1451
index = self.make_index(4096*10, 5)
1452
self.assertExpandOffsets([0], index, [0])
1454
def test_include_neighbors(self):
1455
index = self.make_100_node_index()
1456
# We expand in both directions, until we have at least 'recommended'
1458
self.assertExpandOffsets([9, 10, 11, 12, 13, 14, 15], index, [12])
1459
self.assertExpandOffsets([88, 89, 90, 91, 92, 93, 94], index, [91])
1460
# If we hit an 'edge' we continue in the other direction
1461
self.assertExpandOffsets([1, 2, 3, 4, 5, 6], index, [2])
1462
self.assertExpandOffsets([94, 95, 96, 97, 98, 99], index, [98])
1464
# Requesting many nodes will expand all locations equally
1465
self.assertExpandOffsets([1, 2, 3, 80, 81, 82], index, [2, 81])
1466
self.assertExpandOffsets([1, 2, 3, 9, 10, 11, 80, 81, 82], index,
1469
def test_stop_at_cached(self):
1470
index = self.make_100_node_index()
1471
self.set_cached_offsets(index, [0, 10, 19])
1472
self.assertExpandOffsets([11, 12, 13, 14, 15, 16], index, [11])
1473
self.assertExpandOffsets([11, 12, 13, 14, 15, 16], index, [12])
1474
self.assertExpandOffsets([12, 13, 14, 15, 16, 17, 18], index, [15])
1475
self.assertExpandOffsets([13, 14, 15, 16, 17, 18], index, [16])
1476
self.assertExpandOffsets([13, 14, 15, 16, 17, 18], index, [17])
1477
self.assertExpandOffsets([13, 14, 15, 16, 17, 18], index, [18])
1479
def test_cannot_fully_expand(self):
1480
index = self.make_100_node_index()
1481
self.set_cached_offsets(index, [0, 10, 12])
1482
# We don't go into an endless loop if we are bound by cached nodes
1483
self.assertExpandOffsets([11], index, [11])
1485
def test_overlap(self):
1486
index = self.make_100_node_index()
1487
self.assertExpandOffsets([10, 11, 12, 13, 14, 15], index, [12, 13])
1488
self.assertExpandOffsets([10, 11, 12, 13, 14, 15], index, [11, 14])
1490
def test_stay_within_layer(self):
1491
index = self.make_1000_node_index()
1492
# When expanding a request, we won't read nodes from the next layer
1493
self.assertExpandOffsets([1, 2, 3, 4], index, [2])
1494
self.assertExpandOffsets([6, 7, 8, 9], index, [6])
1495
self.assertExpandOffsets([6, 7, 8, 9], index, [9])
1496
self.assertExpandOffsets([10, 11, 12, 13, 14, 15], index, [10])
1497
self.assertExpandOffsets([10, 11, 12, 13, 14, 15, 16], index, [13])
1499
self.set_cached_offsets(index, [0, 4, 12])
1500
self.assertExpandOffsets([5, 6, 7, 8, 9], index, [7])
1501
self.assertExpandOffsets([10, 11], index, [11])
1503
def test_small_requests_unexpanded(self):
1504
index = self.make_100_node_index()
1505
self.set_cached_offsets(index, [0])
1506
self.assertExpandOffsets([1], index, [1])
1507
self.assertExpandOffsets([50], index, [50])
1508
# If we request more than one node, then we'll expand
1509
self.assertExpandOffsets([49, 50, 51, 59, 60, 61], index, [50, 60])
1511
# The first pass does not expand
1512
index = self.make_1000_node_index()
1513
self.set_cached_offsets(index, [0])
1514
self.assertExpandOffsets([1], index, [1])
1515
self.set_cached_offsets(index, [0, 1])
1516
self.assertExpandOffsets([100], index, [100])
1517
self.set_cached_offsets(index, [0, 1, 100])
1518
# But after the first depth, we will expand
1519
self.assertExpandOffsets([2, 3, 4, 5, 6, 7], index, [2])
1520
self.assertExpandOffsets([2, 3, 4, 5, 6, 7], index, [4])
1521
self.set_cached_offsets(index, [0, 1, 2, 3, 4, 5, 6, 7, 100])
1522
self.assertExpandOffsets([102, 103, 104, 105, 106, 107, 108], index,