1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
|
# Copyright (C) 2007 Canonical Ltd
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
"""Indexing facilities."""
__all__ = [
'CombinedGraphIndex',
'GraphIndex',
'GraphIndexBuilder',
'InMemoryGraphIndex',
]
from cStringIO import StringIO
import re
from bzrlib import errors
_OPTION_NODE_REFS = "node_ref_lists="
_SIGNATURE = "Bazaar Graph Index 1\n"
_whitespace_re = re.compile('[\t\n\x0b\x0c\r\x00 ]')
_newline_null_re = re.compile('[\n\0]')
class GraphIndexBuilder(object):
"""A builder that can build a GraphIndex.
The resulting graph has the structure:
_SIGNATURE OPTIONS NODES NEWLINE
_SIGNATURE := 'Bazaar Graph Index 1' NEWLINE
OPTIONS := 'node_ref_lists=' DIGITS NEWLINE
NODES := NODE*
NODE := KEY NULL ABSENT? NULL REFERENCES NULL VALUE NEWLINE
KEY := Not-whitespace-utf8
ABSENT := 'a'
REFERENCES := REFERENCE_LIST (TAB REFERENCE_LIST){node_ref_lists - 1}
REFERENCE_LIST := (REFERENCE (CR REFERENCE)*)?
REFERENCE := DIGITS ; digits is the byte offset in the index of the
; referenced key.
VALUE := no-newline-no-null-bytes
"""
def __init__(self, reference_lists=0):
"""Create a GraphIndex builder.
:param reference_lists: The number of node references lists for each
entry.
"""
self.reference_lists = reference_lists
self._nodes = {}
def add_node(self, key, value, references=()):
"""Add a node to the index.
:param key: The key. keys must be whitespace-free utf8.
:param references: An iterable of iterables of keys. Each is a
reference to another key.
:param value: The value to associate with the key. It may be any
bytes as long as it does not contain \0 or \n.
"""
if not key or _whitespace_re.search(key) is not None:
raise errors.BadIndexKey(key)
if _newline_null_re.search(value) is not None:
raise errors.BadIndexValue(value)
if len(references) != self.reference_lists:
raise errors.BadIndexValue(references)
node_refs = []
for reference_list in references:
for reference in reference_list:
if _whitespace_re.search(reference) is not None:
raise errors.BadIndexKey(reference)
if reference not in self._nodes:
self._nodes[reference] = ('a', (), '')
node_refs.append(tuple(reference_list))
if key in self._nodes and self._nodes[key][0] == '':
raise errors.BadIndexDuplicateKey(key, self)
self._nodes[key] = ('', tuple(node_refs), value)
def finish(self):
lines = [_SIGNATURE]
lines.append(_OPTION_NODE_REFS + str(self.reference_lists) + '\n')
prefix_length = len(lines[0]) + len(lines[1])
# references are byte offsets. To avoid having to do nasty
# polynomial work to resolve offsets (references to later in the
# file cannot be determined until all the inbetween references have
# been calculated too) we pad the offsets with 0's to make them be
# of consistent length. Using binary offsets would break the trivial
# file parsing.
# to calculate the width of zero's needed we do three passes:
# one to gather all the non-reference data and the number of references.
# one to pad all the data with reference-length and determine entry
# addresses.
# One to serialise.
# forward sorted by key. In future we may consider topological sorting,
# at the cost of table scans for direct lookup, or a second index for
# direct lookup
nodes = sorted(self._nodes.items())
# if we do not prepass, we don't know how long it will be up front.
expected_bytes = None
# we only need to pre-pass if we have reference lists at all.
if self.reference_lists:
key_offset_info = []
non_ref_bytes = prefix_length
total_references = 0
# TODO use simple multiplication for the constants in this loop.
for key, (absent, references, value) in nodes:
# record the offset known *so far* for this key:
# the non reference bytes to date, and the total references to
# date - saves reaccumulating on the second pass
key_offset_info.append((key, non_ref_bytes, total_references))
# key is literal, value is literal, there are 3 null's, 1 NL
non_ref_bytes += len(key) + len(value) + 3 + 1
# one byte for absent if set.
if absent:
non_ref_bytes += 1
elif self.reference_lists:
# (ref_lists -1) tabs
non_ref_bytes += self.reference_lists - 1
# (ref-1 cr's per ref_list)
for ref_list in references:
# how many references across the whole file?
total_references += len(ref_list)
# accrue reference separators
if ref_list:
non_ref_bytes += len(ref_list) - 1
# how many digits are needed to represent the total byte count?
digits = 1
possible_total_bytes = non_ref_bytes + total_references*digits
while 10 ** digits < possible_total_bytes:
digits += 1
possible_total_bytes = non_ref_bytes + total_references*digits
expected_bytes = possible_total_bytes + 1 # terminating newline
# resolve key addresses.
key_addresses = {}
for key, non_ref_bytes, total_references in key_offset_info:
key_addresses[key] = non_ref_bytes + total_references*digits
# serialise
format_string = '%%0%sd' % digits
for key, (absent, references, value) in nodes:
flattened_references = []
for ref_list in references:
ref_addresses = []
for reference in ref_list:
ref_addresses.append(format_string % key_addresses[reference])
flattened_references.append('\r'.join(ref_addresses))
lines.append("%s\0%s\0%s\0%s\n" % (key, absent,
'\t'.join(flattened_references), value))
lines.append('\n')
result = StringIO(''.join(lines))
if expected_bytes and len(result.getvalue()) != expected_bytes:
raise errors.BzrError('Failed index creation. Internal error:'
' mismatched output length and expected length: %d %d' %
(len(result.getvalue()), expected_bytes))
return StringIO(''.join(lines))
class GraphIndex(object):
"""An index for data with embedded graphs.
The index maps keys to a list of key reference lists, and a value.
Each node has the same number of key reference lists. Each key reference
list can be empty or an arbitrary length. The value is an opaque NULL
terminated string without any newlines. The storage of the index is
hidden in the interface: keys and key references are always bytestrings,
never the internal representation (e.g. dictionary offsets).
It is presumed that the index will not be mutated - it is static data.
Successive iter_all_entries calls will read the entire index each time.
Additionally, iter_entries calls will read the index linearly until the
desired keys are found. XXX: This must be fixed before the index is
suitable for production use. :XXX
"""
def __init__(self, transport, name):
"""Open an index called name on transport.
:param transport: A bzrlib.transport.Transport.
:param name: A path to provide to transport API calls.
"""
self._transport = transport
self._name = name
def iter_all_entries(self):
"""Iterate over all keys within the index.
:return: An iterable of (key, value) or (key, value, reference_lists).
The former tuple is used when there are no reference lists in the
index, making the API compatible with simple key:value index types.
There is no defined order for the result iteration - it will be in
the most efficient order for the index.
"""
stream = self._transport.get(self._name)
self._read_prefix(stream)
line_count = 0
self.keys_by_offset = {}
trailers = 0
pos = stream.tell()
for line in stream.readlines():
if line == '\n':
trailers += 1
continue
key, absent, references, value = line.split('\0')
value = value[:-1] # remove the newline
ref_lists = []
for ref_string in references.split('\t'):
ref_lists.append(tuple([
int(ref) for ref in ref_string.split('\r') if ref
]))
ref_lists = tuple(ref_lists)
self.keys_by_offset[pos] = (key, absent, ref_lists, value)
pos += len(line)
for key, absent, references, value in self.keys_by_offset.itervalues():
if absent:
continue
# resolve references:
if self.node_ref_lists:
node_refs = []
for ref_list in references:
node_refs.append(tuple([self.keys_by_offset[ref][0] for ref in ref_list]))
yield (key, value, tuple(node_refs))
else:
yield (key, value)
if trailers != 1:
# there must be one line - the empty trailer line.
raise errors.BadIndexData(self)
def _read_prefix(self, stream):
signature = stream.read(len(self._signature()))
if not signature == self._signature():
raise errors.BadIndexFormatSignature(self._name, GraphIndex)
options_line = stream.readline()
if not options_line.startswith(_OPTION_NODE_REFS):
raise errors.BadIndexOptions(self)
try:
self.node_ref_lists = int(options_line[len(_OPTION_NODE_REFS):-1])
except ValueError:
raise errors.BadIndexOptions(self)
def iter_entries(self, keys):
"""Iterate over keys within the index.
:param keys: An iterable providing the keys to be retrieved.
:return: An iterable as per iter_all_entries, but restricted to the
keys supplied. No additional keys will be returned, and every
key supplied that is in the index will be returned.
"""
keys = set(keys)
if not keys:
return
for node in self.iter_all_entries():
if not keys:
return
if node[0] in keys:
yield node
keys.remove(node[0])
def _signature(self):
"""The file signature for this index type."""
return _SIGNATURE
def validate(self):
"""Validate that everything in the index can be accessed."""
# iter_all validates completely at the moment, so just do that.
for node in self.iter_all_entries():
pass
class CombinedGraphIndex(object):
"""A GraphIndex made up from smaller GraphIndices.
The backing indices must implement GraphIndex, and are presumed to be
static data.
Queries against the combined index will be made against the first index,
and then the second and so on. The order of index's can thus influence
performance significantly. For example, if one index is on local disk and a
second on a remote server, the local disk index should be before the other
in the index list.
"""
def __init__(self, indices):
"""Create a CombinedGraphIndex backed by indices.
:param indices: An ordered list of indices to query for data.
"""
self._indices = indices
def insert_index(self, pos, index):
"""Insert a new index in the list of indices to query.
:param pos: The position to insert the index.
:param index: The index to insert.
"""
self._indices.insert(pos, index)
def iter_all_entries(self):
"""Iterate over all keys within the index
Duplicate keys across child indices are presumed to have the same
value and are only reported once.
:return: An iterable of (key, reference_lists, value). There is no
defined order for the result iteration - it will be in the most
efficient order for the index.
"""
seen_keys = set()
for index in self._indices:
for node in index.iter_all_entries():
if node[0] not in seen_keys:
yield node
seen_keys.add(node[0])
def iter_entries(self, keys):
"""Iterate over keys within the index.
Duplicate keys across child indices are presumed to have the same
value and are only reported once.
:param keys: An iterable providing the keys to be retrieved.
:return: An iterable of (key, reference_lists, value). There is no
defined order for the result iteration - it will be in the most
efficient order for the index.
"""
keys = set(keys)
for index in self._indices:
if not keys:
return
for node in index.iter_entries(keys):
keys.remove(node[0])
yield node
def validate(self):
"""Validate that everything in the index can be accessed."""
for index in self._indices:
index.validate()
class InMemoryGraphIndex(GraphIndexBuilder):
"""A GraphIndex which operates entirely out of memory and is mutable.
This is designed to allow the accumulation of GraphIndex entries during a
single write operation, where the accumulated entries need to be immediately
available - for example via a CombinedGraphIndex.
"""
def add_nodes(self, nodes):
"""Add nodes to the index.
:param nodes: An iterable of (key, node_refs, value) entries to add.
"""
for (key, value, node_refs) in nodes:
self.add_node(key, value, node_refs)
def iter_all_entries(self):
"""Iterate over all keys within the index
:return: An iterable of (key, reference_lists, value). There is no
defined order for the result iteration - it will be in the most
efficient order for the index (in this case dictionary hash order).
"""
if self.reference_lists:
for key, (absent, references, value) in self._nodes.iteritems():
if not absent:
yield key, value, references
else:
for key, (absent, references, value) in self._nodes.iteritems():
if not absent:
yield key, value
def iter_entries(self, keys):
"""Iterate over keys within the index.
:param keys: An iterable providing the keys to be retrieved.
:return: An iterable of (key, reference_lists, value). There is no
defined order for the result iteration - it will be in the most
efficient order for the index (keys iteration order in this case).
"""
keys = set(keys)
if self.reference_lists:
for key in keys.intersection(self._nodes):
node = self._nodes[key]
if not node[0]:
yield key, node[2], node[1]
else:
for key in keys.intersection(self._nodes):
node = self._nodes[key]
if not node[0]:
yield key, node[2]
def validate(self):
"""In memory index's have no known corruption at the moment."""
|