636
611
size = trans.put_file('index', stream)
637
612
return btree_index.BTreeGraphIndex(trans, 'index', size)
614
def make_index_with_offset(self, ref_lists=1, key_elements=1, nodes=[],
616
builder = btree_index.BTreeBuilder(key_elements=key_elements,
617
reference_lists=ref_lists)
618
builder.add_nodes(nodes)
619
transport = self.get_transport('')
620
# NamedTemporaryFile dies on builder.finish().read(). weird.
621
temp_file = builder.finish()
622
content = temp_file.read()
625
transport.put_bytes('index', (' '*offset)+content)
626
return btree_index.BTreeGraphIndex(transport, 'index', size=size,
629
def test_clear_cache(self):
630
nodes = self.make_nodes(160, 2, 2)
631
index = self.make_index(ref_lists=2, key_elements=2, nodes=nodes)
632
self.assertEqual(1, len(list(index.iter_entries([nodes[30][0]]))))
633
self.assertEqual([1, 4], index._row_lengths)
634
self.assertIsNot(None, index._root_node)
635
internal_node_pre_clear = index._internal_node_cache.keys()
636
self.assertTrue(len(index._leaf_node_cache) > 0)
638
# We don't touch _root_node or _internal_node_cache, both should be
639
# small, and can save a round trip or two
640
self.assertIsNot(None, index._root_node)
641
# NOTE: We don't want to affect the _internal_node_cache, as we expect
642
# it will be small, and if we ever do touch this index again, it
643
# will save round-trips. This assertion isn't very strong,
644
# becuase without a 3-level index, we don't have any internal
646
self.assertEqual(internal_node_pre_clear,
647
index._internal_node_cache.keys())
648
self.assertEqual(0, len(index._leaf_node_cache))
639
650
def test_trivial_constructor(self):
640
651
transport = get_transport('trace+' + self.get_url(''))
641
652
index = btree_index.BTreeGraphIndex(transport, 'index', None)
981
1011
self.assertEqual(set([]), index.external_references(0))
1013
def test__find_ancestors_one_page(self):
1016
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1017
(key1, 'value', ([key2],)),
1018
(key2, 'value', ([],)),
1021
missing_keys = set()
1022
search_keys = index._find_ancestors([key1], 0, parent_map, missing_keys)
1023
self.assertEqual({key1: (key2,), key2: ()}, parent_map)
1024
self.assertEqual(set(), missing_keys)
1025
self.assertEqual(set(), search_keys)
1027
def test__find_ancestors_one_page_w_missing(self):
1031
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1032
(key1, 'value', ([key2],)),
1033
(key2, 'value', ([],)),
1036
missing_keys = set()
1037
search_keys = index._find_ancestors([key2, key3], 0, parent_map,
1039
self.assertEqual({key2: ()}, parent_map)
1040
# we know that key3 is missing because we read the page that it would
1042
self.assertEqual(set([key3]), missing_keys)
1043
self.assertEqual(set(), search_keys)
1045
def test__find_ancestors_one_parent_missing(self):
1049
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1050
(key1, 'value', ([key2],)),
1051
(key2, 'value', ([key3],)),
1054
missing_keys = set()
1055
search_keys = index._find_ancestors([key1], 0, parent_map,
1057
self.assertEqual({key1: (key2,), key2: (key3,)}, parent_map)
1058
self.assertEqual(set(), missing_keys)
1059
# all we know is that key3 wasn't present on the page we were reading
1060
# but if you look, the last key is key2 which comes before key3, so we
1061
# don't know whether key3 would land on this page or not.
1062
self.assertEqual(set([key3]), search_keys)
1063
search_keys = index._find_ancestors(search_keys, 0, parent_map,
1065
# passing it back in, we are sure it is 'missing'
1066
self.assertEqual({key1: (key2,), key2: (key3,)}, parent_map)
1067
self.assertEqual(set([key3]), missing_keys)
1068
self.assertEqual(set([]), search_keys)
1070
def test__find_ancestors_dont_search_known(self):
1074
index = self.make_index(ref_lists=1, key_elements=1, nodes=[
1075
(key1, 'value', ([key2],)),
1076
(key2, 'value', ([key3],)),
1077
(key3, 'value', ([],)),
1079
# We already know about key2, so we won't try to search for key3
1080
parent_map = {key2: (key3,)}
1081
missing_keys = set()
1082
search_keys = index._find_ancestors([key1], 0, parent_map,
1084
self.assertEqual({key1: (key2,), key2: (key3,)}, parent_map)
1085
self.assertEqual(set(), missing_keys)
1086
self.assertEqual(set(), search_keys)
1088
def test__find_ancestors_multiple_pages(self):
1089
# We need to use enough keys that we actually cause a split
1090
start_time = 1249671539
1091
email = "joebob@example.com"
1095
for i in xrange(400):
1096
rev_id = '%s-%s-%s' % (email,
1097
osutils.compact_date(start_time + i),
1098
osutils.rand_chars(16))
1100
nodes.append((rev_key, 'value', ref_lists))
1101
# We have a ref 'list' of length 1, with a list of parents, with 1
1102
# parent which is a key
1103
ref_lists = ((rev_key,),)
1104
rev_keys.append(rev_key)
1105
index = self.make_index(ref_lists=1, key_elements=1, nodes=nodes)
1106
self.assertEqual(400, index.key_count())
1107
self.assertEqual(3, len(index._row_offsets))
1108
nodes = dict(index._read_nodes([1, 2]))
1111
min_l2_key = l2.min_key
1112
max_l1_key = l1.max_key
1113
self.assertTrue(max_l1_key < min_l2_key)
1114
parents_min_l2_key = l2.keys[min_l2_key][1][0]
1115
self.assertEqual((l1.max_key,), parents_min_l2_key)
1116
# Now, whatever key we select that would fall on the second page,
1117
# should give us all the parents until the page break
1118
key_idx = rev_keys.index(min_l2_key)
1119
next_key = rev_keys[key_idx+1]
1120
# So now when we get the parent map, we should get the key we are
1121
# looking for, min_l2_key, and then a reference to go look for the
1122
# parent of that key
1124
missing_keys = set()
1125
search_keys = index._find_ancestors([next_key], 0, parent_map,
1127
self.assertEqual([min_l2_key, next_key], sorted(parent_map))
1128
self.assertEqual(set(), missing_keys)
1129
self.assertEqual(set([max_l1_key]), search_keys)
1131
search_keys = index._find_ancestors([max_l1_key], 0, parent_map,
1133
self.assertEqual(sorted(l1.keys), sorted(parent_map))
1134
self.assertEqual(set(), missing_keys)
1135
self.assertEqual(set(), search_keys)
1137
def test__find_ancestors_empty_index(self):
1138
index = self.make_index(ref_lists=1, key_elements=1, nodes=[])
1140
missing_keys = set()
1141
search_keys = index._find_ancestors([('one',), ('two',)], 0, parent_map,
1143
self.assertEqual(set(), search_keys)
1144
self.assertEqual({}, parent_map)
1145
self.assertEqual(set([('one',), ('two',)]), missing_keys)
1147
def test_supports_unlimited_cache(self):
1148
builder = btree_index.BTreeBuilder(reference_lists=0, key_elements=1)
1149
# We need enough nodes to cause a page split (so we have both an
1150
# internal node and a couple leaf nodes. 500 seems to be enough.)
1151
nodes = self.make_nodes(500, 1, 0)
1153
builder.add_node(*node)
1154
stream = builder.finish()
1155
trans = get_transport(self.get_url())
1156
size = trans.put_file('index', stream)
1157
index = btree_index.BTreeGraphIndex(trans, 'index', size)
1158
self.assertEqual(500, index.key_count())
1159
# We have an internal node
1160
self.assertEqual(2, len(index._row_lengths))
1161
# We have at least 2 leaf nodes
1162
self.assertTrue(index._row_lengths[-1] >= 2)
1163
self.assertIsInstance(index._leaf_node_cache, lru_cache.LRUCache)
1164
self.assertEqual(btree_index._NODE_CACHE_SIZE,
1165
index._leaf_node_cache._max_cache)
1166
self.assertIsInstance(index._internal_node_cache, fifo_cache.FIFOCache)
1167
self.assertEqual(100, index._internal_node_cache._max_cache)
1168
# No change if unlimited_cache=False is passed
1169
index = btree_index.BTreeGraphIndex(trans, 'index', size,
1170
unlimited_cache=False)
1171
self.assertIsInstance(index._leaf_node_cache, lru_cache.LRUCache)
1172
self.assertEqual(btree_index._NODE_CACHE_SIZE,
1173
index._leaf_node_cache._max_cache)
1174
self.assertIsInstance(index._internal_node_cache, fifo_cache.FIFOCache)
1175
self.assertEqual(100, index._internal_node_cache._max_cache)
1176
index = btree_index.BTreeGraphIndex(trans, 'index', size,
1177
unlimited_cache=True)
1178
self.assertIsInstance(index._leaf_node_cache, dict)
1179
self.assertIs(type(index._internal_node_cache), dict)
1180
# Exercise the lookup code
1181
entries = set(index.iter_entries([n[0] for n in nodes]))
1182
self.assertEqual(500, len(entries))
984
1185
class TestBTreeNodes(BTreeTestCase):
986
def restore_parser(self):
987
btree_index._btree_serializer = self.saved_parser
989
1187
def setUp(self):
990
1188
BTreeTestCase.setUp(self)
991
self.saved_parser = btree_index._btree_serializer
992
self.addCleanup(self.restore_parser)
993
btree_index._btree_serializer = self.parse_btree
1189
self.overrideAttr(btree_index, '_btree_serializer', self.parse_btree)
995
1191
def test_LeafNode_1_0(self):
996
1192
node_bytes = ("type=leaf\n"