~bzr-pqm/bzr/bzr.dev

« back to all changes in this revision

Viewing changes to bzrlib/tests/test_knit.py

  • Committer: Robert Collins
  • Date: 2007-08-06 23:49:18 UTC
  • mto: (2592.3.81 repository)
  • mto: This revision was merged to the branch mainline in revision 2933.
  • Revision ID: robertc@robertcollins.net-20070806234918-xc9w5f86tgjphf9u
Prevent the duplicate additions of names to FileNames collections.

Show diffs side-by-side

added added

removed removed

Lines of Context:
12
12
#
13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
 
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
16
16
 
17
17
"""Tests for Knit data structure"""
18
18
 
19
19
from cStringIO import StringIO
20
20
import difflib
21
21
import gzip
 
22
import sha
22
23
import sys
23
24
 
24
25
from bzrlib import (
25
26
    errors,
26
27
    generate_ids,
27
28
    knit,
28
 
    multiparent,
29
 
    osutils,
30
 
    pack,
31
 
    tests,
32
29
    )
33
30
from bzrlib.errors import (
34
31
    RevisionAlreadyPresent,
38
35
    )
39
36
from bzrlib.index import *
40
37
from bzrlib.knit import (
41
 
    AnnotatedKnitContent,
42
38
    KnitContent,
 
39
    KnitGraphIndex,
 
40
    KnitVersionedFile,
 
41
    KnitPlainFactory,
 
42
    KnitAnnotateFactory,
 
43
    _KnitData,
 
44
    _KnitIndex,
 
45
    WeaveToKnit,
43
46
    KnitSequenceMatcher,
44
 
    KnitVersionedFiles,
45
 
    PlainKnitContent,
46
 
    _VFContentMapGenerator,
47
 
    _DirectPackAccess,
48
 
    _KndxIndex,
49
 
    _KnitGraphIndex,
50
 
    _KnitKeyAccess,
51
 
    make_file_factory,
52
 
    )
53
 
from bzrlib.repofmt import pack_repo
54
 
from bzrlib.tests import (
55
 
    Feature,
56
 
    KnownFailure,
57
 
    TestCase,
58
 
    TestCaseWithMemoryTransport,
59
 
    TestCaseWithTransport,
60
 
    TestNotApplicable,
61
 
    )
62
 
from bzrlib.transport import get_transport
 
47
    )
 
48
from bzrlib.osutils import split_lines
 
49
from bzrlib.tests import TestCase, TestCaseWithTransport, Feature
 
50
from bzrlib.transport import TransportLogger, get_transport
63
51
from bzrlib.transport.memory import MemoryTransport
64
 
from bzrlib.tuned_gzip import GzipFile
65
 
from bzrlib.versionedfile import (
66
 
    AbsentContentFactory,
67
 
    ConstantMapper,
68
 
    network_bytes_to_kind_and_offset,
69
 
    RecordingVersionedFilesDecorator,
70
 
    )
71
 
 
72
 
 
73
 
compiled_knit_feature = tests.ModuleAvailableFeature(
74
 
                            'bzrlib._knit_load_data_pyx')
75
 
 
76
 
 
77
 
class KnitContentTestsMixin(object):
 
52
from bzrlib.weave import Weave
 
53
 
 
54
 
 
55
class _CompiledKnitFeature(Feature):
 
56
 
 
57
    def _probe(self):
 
58
        try:
 
59
            import bzrlib._knit_load_data_c
 
60
        except ImportError:
 
61
            return False
 
62
        return True
 
63
 
 
64
    def feature_name(self):
 
65
        return 'bzrlib._knit_load_data_c'
 
66
 
 
67
CompiledKnitFeature = _CompiledKnitFeature()
 
68
 
 
69
 
 
70
class KnitContentTests(TestCase):
78
71
 
79
72
    def test_constructor(self):
80
 
        content = self._make_content([])
 
73
        content = KnitContent([])
81
74
 
82
75
    def test_text(self):
83
 
        content = self._make_content([])
 
76
        content = KnitContent([])
84
77
        self.assertEqual(content.text(), [])
85
78
 
86
 
        content = self._make_content([("origin1", "text1"), ("origin2", "text2")])
 
79
        content = KnitContent([("origin1", "text1"), ("origin2", "text2")])
87
80
        self.assertEqual(content.text(), ["text1", "text2"])
88
81
 
 
82
    def test_annotate(self):
 
83
        content = KnitContent([])
 
84
        self.assertEqual(content.annotate(), [])
 
85
 
 
86
        content = KnitContent([("origin1", "text1"), ("origin2", "text2")])
 
87
        self.assertEqual(content.annotate(),
 
88
            [("origin1", "text1"), ("origin2", "text2")])
 
89
 
 
90
    def test_annotate_iter(self):
 
91
        content = KnitContent([])
 
92
        it = content.annotate_iter()
 
93
        self.assertRaises(StopIteration, it.next)
 
94
 
 
95
        content = KnitContent([("origin1", "text1"), ("origin2", "text2")])
 
96
        it = content.annotate_iter()
 
97
        self.assertEqual(it.next(), ("origin1", "text1"))
 
98
        self.assertEqual(it.next(), ("origin2", "text2"))
 
99
        self.assertRaises(StopIteration, it.next)
 
100
 
89
101
    def test_copy(self):
90
 
        content = self._make_content([("origin1", "text1"), ("origin2", "text2")])
 
102
        content = KnitContent([("origin1", "text1"), ("origin2", "text2")])
91
103
        copy = content.copy()
92
 
        self.assertIsInstance(copy, content.__class__)
93
 
        self.assertEqual(copy.annotate(), content.annotate())
94
 
 
95
 
    def assertDerivedBlocksEqual(self, source, target, noeol=False):
96
 
        """Assert that the derived matching blocks match real output"""
97
 
        source_lines = source.splitlines(True)
98
 
        target_lines = target.splitlines(True)
99
 
        def nl(line):
100
 
            if noeol and not line.endswith('\n'):
101
 
                return line + '\n'
102
 
            else:
103
 
                return line
104
 
        source_content = self._make_content([(None, nl(l)) for l in source_lines])
105
 
        target_content = self._make_content([(None, nl(l)) for l in target_lines])
106
 
        line_delta = source_content.line_delta(target_content)
107
 
        delta_blocks = list(KnitContent.get_line_delta_blocks(line_delta,
108
 
            source_lines, target_lines))
109
 
        matcher = KnitSequenceMatcher(None, source_lines, target_lines)
110
 
        matcher_blocks = list(list(matcher.get_matching_blocks()))
111
 
        self.assertEqual(matcher_blocks, delta_blocks)
112
 
 
113
 
    def test_get_line_delta_blocks(self):
114
 
        self.assertDerivedBlocksEqual('a\nb\nc\n', 'q\nc\n')
115
 
        self.assertDerivedBlocksEqual(TEXT_1, TEXT_1)
116
 
        self.assertDerivedBlocksEqual(TEXT_1, TEXT_1A)
117
 
        self.assertDerivedBlocksEqual(TEXT_1, TEXT_1B)
118
 
        self.assertDerivedBlocksEqual(TEXT_1B, TEXT_1A)
119
 
        self.assertDerivedBlocksEqual(TEXT_1A, TEXT_1B)
120
 
        self.assertDerivedBlocksEqual(TEXT_1A, '')
121
 
        self.assertDerivedBlocksEqual('', TEXT_1A)
122
 
        self.assertDerivedBlocksEqual('', '')
123
 
        self.assertDerivedBlocksEqual('a\nb\nc', 'a\nb\nc\nd')
124
 
 
125
 
    def test_get_line_delta_blocks_noeol(self):
126
 
        """Handle historical knit deltas safely
127
 
 
128
 
        Some existing knit deltas don't consider the last line to differ
129
 
        when the only difference whether it has a final newline.
130
 
 
131
 
        New knit deltas appear to always consider the last line to differ
132
 
        in this case.
133
 
        """
134
 
        self.assertDerivedBlocksEqual('a\nb\nc', 'a\nb\nc\nd\n', noeol=True)
135
 
        self.assertDerivedBlocksEqual('a\nb\nc\nd\n', 'a\nb\nc', noeol=True)
136
 
        self.assertDerivedBlocksEqual('a\nb\nc\n', 'a\nb\nc', noeol=True)
137
 
        self.assertDerivedBlocksEqual('a\nb\nc', 'a\nb\nc\n', noeol=True)
138
 
 
139
 
 
140
 
TEXT_1 = """\
141
 
Banana cup cakes:
142
 
 
143
 
- bananas
144
 
- eggs
145
 
- broken tea cups
146
 
"""
147
 
 
148
 
TEXT_1A = """\
149
 
Banana cup cake recipe
150
 
(serves 6)
151
 
 
152
 
- bananas
153
 
- eggs
154
 
- broken tea cups
155
 
- self-raising flour
156
 
"""
157
 
 
158
 
TEXT_1B = """\
159
 
Banana cup cake recipe
160
 
 
161
 
- bananas (do not use plantains!!!)
162
 
- broken tea cups
163
 
- flour
164
 
"""
165
 
 
166
 
delta_1_1a = """\
167
 
0,1,2
168
 
Banana cup cake recipe
169
 
(serves 6)
170
 
5,5,1
171
 
- self-raising flour
172
 
"""
173
 
 
174
 
TEXT_2 = """\
175
 
Boeuf bourguignon
176
 
 
177
 
- beef
178
 
- red wine
179
 
- small onions
180
 
- carrot
181
 
- mushrooms
182
 
"""
183
 
 
184
 
 
185
 
class TestPlainKnitContent(TestCase, KnitContentTestsMixin):
186
 
 
187
 
    def _make_content(self, lines):
188
 
        annotated_content = AnnotatedKnitContent(lines)
189
 
        return PlainKnitContent(annotated_content.text(), 'bogus')
190
 
 
191
 
    def test_annotate(self):
192
 
        content = self._make_content([])
193
 
        self.assertEqual(content.annotate(), [])
194
 
 
195
 
        content = self._make_content([("origin1", "text1"), ("origin2", "text2")])
196
 
        self.assertEqual(content.annotate(),
197
 
            [("bogus", "text1"), ("bogus", "text2")])
198
 
 
199
 
    def test_line_delta(self):
200
 
        content1 = self._make_content([("", "a"), ("", "b")])
201
 
        content2 = self._make_content([("", "a"), ("", "a"), ("", "c")])
202
 
        self.assertEqual(content1.line_delta(content2),
203
 
            [(1, 2, 2, ["a", "c"])])
204
 
 
205
 
    def test_line_delta_iter(self):
206
 
        content1 = self._make_content([("", "a"), ("", "b")])
207
 
        content2 = self._make_content([("", "a"), ("", "a"), ("", "c")])
208
 
        it = content1.line_delta_iter(content2)
209
 
        self.assertEqual(it.next(), (1, 2, 2, ["a", "c"]))
210
 
        self.assertRaises(StopIteration, it.next)
211
 
 
212
 
 
213
 
class TestAnnotatedKnitContent(TestCase, KnitContentTestsMixin):
214
 
 
215
 
    def _make_content(self, lines):
216
 
        return AnnotatedKnitContent(lines)
217
 
 
218
 
    def test_annotate(self):
219
 
        content = self._make_content([])
220
 
        self.assertEqual(content.annotate(), [])
221
 
 
222
 
        content = self._make_content([("origin1", "text1"), ("origin2", "text2")])
223
 
        self.assertEqual(content.annotate(),
 
104
        self.assertIsInstance(copy, KnitContent)
 
105
        self.assertEqual(copy.annotate(),
224
106
            [("origin1", "text1"), ("origin2", "text2")])
225
107
 
226
108
    def test_line_delta(self):
227
 
        content1 = self._make_content([("", "a"), ("", "b")])
228
 
        content2 = self._make_content([("", "a"), ("", "a"), ("", "c")])
 
109
        content1 = KnitContent([("", "a"), ("", "b")])
 
110
        content2 = KnitContent([("", "a"), ("", "a"), ("", "c")])
229
111
        self.assertEqual(content1.line_delta(content2),
230
112
            [(1, 2, 2, [("", "a"), ("", "c")])])
231
113
 
232
114
    def test_line_delta_iter(self):
233
 
        content1 = self._make_content([("", "a"), ("", "b")])
234
 
        content2 = self._make_content([("", "a"), ("", "a"), ("", "c")])
 
115
        content1 = KnitContent([("", "a"), ("", "b")])
 
116
        content2 = KnitContent([("", "a"), ("", "a"), ("", "c")])
235
117
        it = content1.line_delta_iter(content2)
236
118
        self.assertEqual(it.next(), (1, 2, 2, [("", "a"), ("", "c")]))
237
119
        self.assertRaises(StopIteration, it.next)
263
145
        return queue_call
264
146
 
265
147
 
266
 
class MockReadvFailingTransport(MockTransport):
267
 
    """Fail in the middle of a readv() result.
268
 
 
269
 
    This Transport will successfully yield the first two requested hunks, but
270
 
    raise NoSuchFile for the rest.
271
 
    """
272
 
 
273
 
    def readv(self, relpath, offsets):
274
 
        count = 0
275
 
        for result in MockTransport.readv(self, relpath, offsets):
276
 
            count += 1
277
 
            # we use 2 because the first offset is the pack header, the second
278
 
            # is the first actual content requset
279
 
            if count > 2:
280
 
                raise errors.NoSuchFile(relpath)
281
 
            yield result
282
 
 
283
 
 
284
 
class KnitRecordAccessTestsMixin(object):
285
 
    """Tests for getting and putting knit records."""
286
 
 
287
 
    def test_add_raw_records(self):
288
 
        """Add_raw_records adds records retrievable later."""
289
 
        access = self.get_access()
290
 
        memos = access.add_raw_records([('key', 10)], '1234567890')
291
 
        self.assertEqual(['1234567890'], list(access.get_raw_records(memos)))
292
 
 
293
 
    def test_add_several_raw_records(self):
294
 
        """add_raw_records with many records and read some back."""
295
 
        access = self.get_access()
296
 
        memos = access.add_raw_records([('key', 10), ('key2', 2), ('key3', 5)],
297
 
            '12345678901234567')
298
 
        self.assertEqual(['1234567890', '12', '34567'],
299
 
            list(access.get_raw_records(memos)))
300
 
        self.assertEqual(['1234567890'],
301
 
            list(access.get_raw_records(memos[0:1])))
302
 
        self.assertEqual(['12'],
303
 
            list(access.get_raw_records(memos[1:2])))
304
 
        self.assertEqual(['34567'],
305
 
            list(access.get_raw_records(memos[2:3])))
306
 
        self.assertEqual(['1234567890', '34567'],
307
 
            list(access.get_raw_records(memos[0:1] + memos[2:3])))
308
 
 
309
 
 
310
 
class TestKnitKnitAccess(TestCaseWithMemoryTransport, KnitRecordAccessTestsMixin):
311
 
    """Tests for the .kndx implementation."""
312
 
 
313
 
    def get_access(self):
314
 
        """Get a .knit style access instance."""
315
 
        mapper = ConstantMapper("foo")
316
 
        access = _KnitKeyAccess(self.get_transport(), mapper)
317
 
        return access
318
 
 
319
 
 
320
 
class _TestException(Exception):
321
 
    """Just an exception for local tests to use."""
322
 
 
323
 
 
324
 
class TestPackKnitAccess(TestCaseWithMemoryTransport, KnitRecordAccessTestsMixin):
325
 
    """Tests for the pack based access."""
326
 
 
327
 
    def get_access(self):
328
 
        return self._get_access()[0]
329
 
 
330
 
    def _get_access(self, packname='packfile', index='FOO'):
331
 
        transport = self.get_transport()
332
 
        def write_data(bytes):
333
 
            transport.append_bytes(packname, bytes)
334
 
        writer = pack.ContainerWriter(write_data)
335
 
        writer.begin()
336
 
        access = _DirectPackAccess({})
337
 
        access.set_writer(writer, index, (transport, packname))
338
 
        return access, writer
339
 
 
340
 
    def make_pack_file(self):
341
 
        """Create a pack file with 2 records."""
342
 
        access, writer = self._get_access(packname='packname', index='foo')
343
 
        memos = []
344
 
        memos.extend(access.add_raw_records([('key1', 10)], '1234567890'))
345
 
        memos.extend(access.add_raw_records([('key2', 5)], '12345'))
346
 
        writer.end()
347
 
        return memos
348
 
 
349
 
    def make_vf_for_retrying(self):
350
 
        """Create 3 packs and a reload function.
351
 
 
352
 
        Originally, 2 pack files will have the data, but one will be missing.
353
 
        And then the third will be used in place of the first two if reload()
354
 
        is called.
355
 
 
356
 
        :return: (versioned_file, reload_counter)
357
 
            versioned_file  a KnitVersionedFiles using the packs for access
358
 
        """
359
 
        builder = self.make_branch_builder('.', format="1.9")
360
 
        builder.start_series()
361
 
        builder.build_snapshot('rev-1', None, [
362
 
            ('add', ('', 'root-id', 'directory', None)),
363
 
            ('add', ('file', 'file-id', 'file', 'content\nrev 1\n')),
364
 
            ])
365
 
        builder.build_snapshot('rev-2', ['rev-1'], [
366
 
            ('modify', ('file-id', 'content\nrev 2\n')),
367
 
            ])
368
 
        builder.build_snapshot('rev-3', ['rev-2'], [
369
 
            ('modify', ('file-id', 'content\nrev 3\n')),
370
 
            ])
371
 
        builder.finish_series()
372
 
        b = builder.get_branch()
373
 
        b.lock_write()
374
 
        self.addCleanup(b.unlock)
375
 
        # Pack these three revisions into another pack file, but don't remove
376
 
        # the originals
377
 
        repo = b.repository
378
 
        collection = repo._pack_collection
379
 
        collection.ensure_loaded()
380
 
        orig_packs = collection.packs
381
 
        packer = pack_repo.Packer(collection, orig_packs, '.testpack')
382
 
        new_pack = packer.pack()
383
 
        # forget about the new pack
384
 
        collection.reset()
385
 
        repo.refresh_data()
386
 
        vf = repo.revisions
387
 
        # Set up a reload() function that switches to using the new pack file
388
 
        new_index = new_pack.revision_index
389
 
        access_tuple = new_pack.access_tuple()
390
 
        reload_counter = [0, 0, 0]
391
 
        def reload():
392
 
            reload_counter[0] += 1
393
 
            if reload_counter[1] > 0:
394
 
                # We already reloaded, nothing more to do
395
 
                reload_counter[2] += 1
396
 
                return False
397
 
            reload_counter[1] += 1
398
 
            vf._index._graph_index._indices[:] = [new_index]
399
 
            vf._access._indices.clear()
400
 
            vf._access._indices[new_index] = access_tuple
401
 
            return True
402
 
        # Delete one of the pack files so the data will need to be reloaded. We
403
 
        # will delete the file with 'rev-2' in it
404
 
        trans, name = orig_packs[1].access_tuple()
405
 
        trans.delete(name)
406
 
        # We don't have the index trigger reloading because we want to test
407
 
        # that we reload when the .pack disappears
408
 
        vf._access._reload_func = reload
409
 
        return vf, reload_counter
410
 
 
411
 
    def make_reload_func(self, return_val=True):
412
 
        reload_called = [0]
413
 
        def reload():
414
 
            reload_called[0] += 1
415
 
            return return_val
416
 
        return reload_called, reload
417
 
 
418
 
    def make_retry_exception(self):
419
 
        # We raise a real exception so that sys.exc_info() is properly
420
 
        # populated
421
 
        try:
422
 
            raise _TestException('foobar')
423
 
        except _TestException, e:
424
 
            retry_exc = errors.RetryWithNewPacks(None, reload_occurred=False,
425
 
                                                 exc_info=sys.exc_info())
426
 
        return retry_exc
427
 
 
428
 
    def test_read_from_several_packs(self):
429
 
        access, writer = self._get_access()
430
 
        memos = []
431
 
        memos.extend(access.add_raw_records([('key', 10)], '1234567890'))
432
 
        writer.end()
433
 
        access, writer = self._get_access('pack2', 'FOOBAR')
434
 
        memos.extend(access.add_raw_records([('key', 5)], '12345'))
435
 
        writer.end()
436
 
        access, writer = self._get_access('pack3', 'BAZ')
437
 
        memos.extend(access.add_raw_records([('key', 5)], 'alpha'))
438
 
        writer.end()
439
 
        transport = self.get_transport()
440
 
        access = _DirectPackAccess({"FOO":(transport, 'packfile'),
441
 
            "FOOBAR":(transport, 'pack2'),
442
 
            "BAZ":(transport, 'pack3')})
443
 
        self.assertEqual(['1234567890', '12345', 'alpha'],
444
 
            list(access.get_raw_records(memos)))
445
 
        self.assertEqual(['1234567890'],
446
 
            list(access.get_raw_records(memos[0:1])))
447
 
        self.assertEqual(['12345'],
448
 
            list(access.get_raw_records(memos[1:2])))
449
 
        self.assertEqual(['alpha'],
450
 
            list(access.get_raw_records(memos[2:3])))
451
 
        self.assertEqual(['1234567890', 'alpha'],
452
 
            list(access.get_raw_records(memos[0:1] + memos[2:3])))
453
 
 
454
 
    def test_set_writer(self):
455
 
        """The writer should be settable post construction."""
456
 
        access = _DirectPackAccess({})
457
 
        transport = self.get_transport()
458
 
        packname = 'packfile'
459
 
        index = 'foo'
460
 
        def write_data(bytes):
461
 
            transport.append_bytes(packname, bytes)
462
 
        writer = pack.ContainerWriter(write_data)
463
 
        writer.begin()
464
 
        access.set_writer(writer, index, (transport, packname))
465
 
        memos = access.add_raw_records([('key', 10)], '1234567890')
466
 
        writer.end()
467
 
        self.assertEqual(['1234567890'], list(access.get_raw_records(memos)))
468
 
 
469
 
    def test_missing_index_raises_retry(self):
470
 
        memos = self.make_pack_file()
471
 
        transport = self.get_transport()
472
 
        reload_called, reload_func = self.make_reload_func()
473
 
        # Note that the index key has changed from 'foo' to 'bar'
474
 
        access = _DirectPackAccess({'bar':(transport, 'packname')},
475
 
                                   reload_func=reload_func)
476
 
        e = self.assertListRaises(errors.RetryWithNewPacks,
477
 
                                  access.get_raw_records, memos)
478
 
        # Because a key was passed in which does not match our index list, we
479
 
        # assume that the listing was already reloaded
480
 
        self.assertTrue(e.reload_occurred)
481
 
        self.assertIsInstance(e.exc_info, tuple)
482
 
        self.assertIs(e.exc_info[0], KeyError)
483
 
        self.assertIsInstance(e.exc_info[1], KeyError)
484
 
 
485
 
    def test_missing_index_raises_key_error_with_no_reload(self):
486
 
        memos = self.make_pack_file()
487
 
        transport = self.get_transport()
488
 
        # Note that the index key has changed from 'foo' to 'bar'
489
 
        access = _DirectPackAccess({'bar':(transport, 'packname')})
490
 
        e = self.assertListRaises(KeyError, access.get_raw_records, memos)
491
 
 
492
 
    def test_missing_file_raises_retry(self):
493
 
        memos = self.make_pack_file()
494
 
        transport = self.get_transport()
495
 
        reload_called, reload_func = self.make_reload_func()
496
 
        # Note that the 'filename' has been changed to 'different-packname'
497
 
        access = _DirectPackAccess({'foo':(transport, 'different-packname')},
498
 
                                   reload_func=reload_func)
499
 
        e = self.assertListRaises(errors.RetryWithNewPacks,
500
 
                                  access.get_raw_records, memos)
501
 
        # The file has gone missing, so we assume we need to reload
502
 
        self.assertFalse(e.reload_occurred)
503
 
        self.assertIsInstance(e.exc_info, tuple)
504
 
        self.assertIs(e.exc_info[0], errors.NoSuchFile)
505
 
        self.assertIsInstance(e.exc_info[1], errors.NoSuchFile)
506
 
        self.assertEqual('different-packname', e.exc_info[1].path)
507
 
 
508
 
    def test_missing_file_raises_no_such_file_with_no_reload(self):
509
 
        memos = self.make_pack_file()
510
 
        transport = self.get_transport()
511
 
        # Note that the 'filename' has been changed to 'different-packname'
512
 
        access = _DirectPackAccess({'foo':(transport, 'different-packname')})
513
 
        e = self.assertListRaises(errors.NoSuchFile,
514
 
                                  access.get_raw_records, memos)
515
 
 
516
 
    def test_failing_readv_raises_retry(self):
517
 
        memos = self.make_pack_file()
518
 
        transport = self.get_transport()
519
 
        failing_transport = MockReadvFailingTransport(
520
 
                                [transport.get_bytes('packname')])
521
 
        reload_called, reload_func = self.make_reload_func()
522
 
        access = _DirectPackAccess({'foo':(failing_transport, 'packname')},
523
 
                                   reload_func=reload_func)
524
 
        # Asking for a single record will not trigger the Mock failure
525
 
        self.assertEqual(['1234567890'],
526
 
            list(access.get_raw_records(memos[:1])))
527
 
        self.assertEqual(['12345'],
528
 
            list(access.get_raw_records(memos[1:2])))
529
 
        # A multiple offset readv() will fail mid-way through
530
 
        e = self.assertListRaises(errors.RetryWithNewPacks,
531
 
                                  access.get_raw_records, memos)
532
 
        # The file has gone missing, so we assume we need to reload
533
 
        self.assertFalse(e.reload_occurred)
534
 
        self.assertIsInstance(e.exc_info, tuple)
535
 
        self.assertIs(e.exc_info[0], errors.NoSuchFile)
536
 
        self.assertIsInstance(e.exc_info[1], errors.NoSuchFile)
537
 
        self.assertEqual('packname', e.exc_info[1].path)
538
 
 
539
 
    def test_failing_readv_raises_no_such_file_with_no_reload(self):
540
 
        memos = self.make_pack_file()
541
 
        transport = self.get_transport()
542
 
        failing_transport = MockReadvFailingTransport(
543
 
                                [transport.get_bytes('packname')])
544
 
        reload_called, reload_func = self.make_reload_func()
545
 
        access = _DirectPackAccess({'foo':(failing_transport, 'packname')})
546
 
        # Asking for a single record will not trigger the Mock failure
547
 
        self.assertEqual(['1234567890'],
548
 
            list(access.get_raw_records(memos[:1])))
549
 
        self.assertEqual(['12345'],
550
 
            list(access.get_raw_records(memos[1:2])))
551
 
        # A multiple offset readv() will fail mid-way through
552
 
        e = self.assertListRaises(errors.NoSuchFile,
553
 
                                  access.get_raw_records, memos)
554
 
 
555
 
    def test_reload_or_raise_no_reload(self):
556
 
        access = _DirectPackAccess({}, reload_func=None)
557
 
        retry_exc = self.make_retry_exception()
558
 
        # Without a reload_func, we will just re-raise the original exception
559
 
        self.assertRaises(_TestException, access.reload_or_raise, retry_exc)
560
 
 
561
 
    def test_reload_or_raise_reload_changed(self):
562
 
        reload_called, reload_func = self.make_reload_func(return_val=True)
563
 
        access = _DirectPackAccess({}, reload_func=reload_func)
564
 
        retry_exc = self.make_retry_exception()
565
 
        access.reload_or_raise(retry_exc)
566
 
        self.assertEqual([1], reload_called)
567
 
        retry_exc.reload_occurred=True
568
 
        access.reload_or_raise(retry_exc)
569
 
        self.assertEqual([2], reload_called)
570
 
 
571
 
    def test_reload_or_raise_reload_no_change(self):
572
 
        reload_called, reload_func = self.make_reload_func(return_val=False)
573
 
        access = _DirectPackAccess({}, reload_func=reload_func)
574
 
        retry_exc = self.make_retry_exception()
575
 
        # If reload_occurred is False, then we consider it an error to have
576
 
        # reload_func() return False (no changes).
577
 
        self.assertRaises(_TestException, access.reload_or_raise, retry_exc)
578
 
        self.assertEqual([1], reload_called)
579
 
        retry_exc.reload_occurred=True
580
 
        # If reload_occurred is True, then we assume nothing changed because
581
 
        # it had changed earlier, but didn't change again
582
 
        access.reload_or_raise(retry_exc)
583
 
        self.assertEqual([2], reload_called)
584
 
 
585
 
    def test_annotate_retries(self):
586
 
        vf, reload_counter = self.make_vf_for_retrying()
587
 
        # It is a little bit bogus to annotate the Revision VF, but it works,
588
 
        # as we have ancestry stored there
589
 
        key = ('rev-3',)
590
 
        reload_lines = vf.annotate(key)
591
 
        self.assertEqual([1, 1, 0], reload_counter)
592
 
        plain_lines = vf.annotate(key)
593
 
        self.assertEqual([1, 1, 0], reload_counter) # No extra reloading
594
 
        if reload_lines != plain_lines:
595
 
            self.fail('Annotation was not identical with reloading.')
596
 
        # Now delete the packs-in-use, which should trigger another reload, but
597
 
        # this time we just raise an exception because we can't recover
598
 
        for trans, name in vf._access._indices.itervalues():
599
 
            trans.delete(name)
600
 
        self.assertRaises(errors.NoSuchFile, vf.annotate, key)
601
 
        self.assertEqual([2, 1, 1], reload_counter)
602
 
 
603
 
    def test__get_record_map_retries(self):
604
 
        vf, reload_counter = self.make_vf_for_retrying()
605
 
        keys = [('rev-1',), ('rev-2',), ('rev-3',)]
606
 
        records = vf._get_record_map(keys)
607
 
        self.assertEqual(keys, sorted(records.keys()))
608
 
        self.assertEqual([1, 1, 0], reload_counter)
609
 
        # Now delete the packs-in-use, which should trigger another reload, but
610
 
        # this time we just raise an exception because we can't recover
611
 
        for trans, name in vf._access._indices.itervalues():
612
 
            trans.delete(name)
613
 
        self.assertRaises(errors.NoSuchFile, vf._get_record_map, keys)
614
 
        self.assertEqual([2, 1, 1], reload_counter)
615
 
 
616
 
    def test_get_record_stream_retries(self):
617
 
        vf, reload_counter = self.make_vf_for_retrying()
618
 
        keys = [('rev-1',), ('rev-2',), ('rev-3',)]
619
 
        record_stream = vf.get_record_stream(keys, 'topological', False)
620
 
        record = record_stream.next()
621
 
        self.assertEqual(('rev-1',), record.key)
622
 
        self.assertEqual([0, 0, 0], reload_counter)
623
 
        record = record_stream.next()
624
 
        self.assertEqual(('rev-2',), record.key)
625
 
        self.assertEqual([1, 1, 0], reload_counter)
626
 
        record = record_stream.next()
627
 
        self.assertEqual(('rev-3',), record.key)
628
 
        self.assertEqual([1, 1, 0], reload_counter)
629
 
        # Now delete all pack files, and see that we raise the right error
630
 
        for trans, name in vf._access._indices.itervalues():
631
 
            trans.delete(name)
632
 
        self.assertListRaises(errors.NoSuchFile,
633
 
            vf.get_record_stream, keys, 'topological', False)
634
 
 
635
 
    def test_iter_lines_added_or_present_in_keys_retries(self):
636
 
        vf, reload_counter = self.make_vf_for_retrying()
637
 
        keys = [('rev-1',), ('rev-2',), ('rev-3',)]
638
 
        # Unfortunately, iter_lines_added_or_present_in_keys iterates the
639
 
        # result in random order (determined by the iteration order from a
640
 
        # set()), so we don't have any solid way to trigger whether data is
641
 
        # read before or after. However we tried to delete the middle node to
642
 
        # exercise the code well.
643
 
        # What we care about is that all lines are always yielded, but not
644
 
        # duplicated
645
 
        count = 0
646
 
        reload_lines = sorted(vf.iter_lines_added_or_present_in_keys(keys))
647
 
        self.assertEqual([1, 1, 0], reload_counter)
648
 
        # Now do it again, to make sure the result is equivalent
649
 
        plain_lines = sorted(vf.iter_lines_added_or_present_in_keys(keys))
650
 
        self.assertEqual([1, 1, 0], reload_counter) # No extra reloading
651
 
        self.assertEqual(plain_lines, reload_lines)
652
 
        self.assertEqual(21, len(plain_lines))
653
 
        # Now delete all pack files, and see that we raise the right error
654
 
        for trans, name in vf._access._indices.itervalues():
655
 
            trans.delete(name)
656
 
        self.assertListRaises(errors.NoSuchFile,
657
 
            vf.iter_lines_added_or_present_in_keys, keys)
658
 
        self.assertEqual([2, 1, 1], reload_counter)
659
 
 
660
 
    def test_get_record_stream_yields_disk_sorted_order(self):
661
 
        # if we get 'unordered' pick a semi-optimal order for reading. The
662
 
        # order should be grouped by pack file, and then by position in file
663
 
        repo = self.make_repository('test', format='pack-0.92')
664
 
        repo.lock_write()
665
 
        self.addCleanup(repo.unlock)
666
 
        repo.start_write_group()
667
 
        vf = repo.texts
668
 
        vf.add_lines(('f-id', 'rev-5'), [('f-id', 'rev-4')], ['lines\n'])
669
 
        vf.add_lines(('f-id', 'rev-1'), [], ['lines\n'])
670
 
        vf.add_lines(('f-id', 'rev-2'), [('f-id', 'rev-1')], ['lines\n'])
671
 
        repo.commit_write_group()
672
 
        # We inserted them as rev-5, rev-1, rev-2, we should get them back in
673
 
        # the same order
674
 
        stream = vf.get_record_stream([('f-id', 'rev-1'), ('f-id', 'rev-5'),
675
 
                                       ('f-id', 'rev-2')], 'unordered', False)
676
 
        keys = [r.key for r in stream]
677
 
        self.assertEqual([('f-id', 'rev-5'), ('f-id', 'rev-1'),
678
 
                          ('f-id', 'rev-2')], keys)
679
 
        repo.start_write_group()
680
 
        vf.add_lines(('f-id', 'rev-4'), [('f-id', 'rev-3')], ['lines\n'])
681
 
        vf.add_lines(('f-id', 'rev-3'), [('f-id', 'rev-2')], ['lines\n'])
682
 
        vf.add_lines(('f-id', 'rev-6'), [('f-id', 'rev-5')], ['lines\n'])
683
 
        repo.commit_write_group()
684
 
        # Request in random order, to make sure the output order isn't based on
685
 
        # the request
686
 
        request_keys = set(('f-id', 'rev-%d' % i) for i in range(1, 7))
687
 
        stream = vf.get_record_stream(request_keys, 'unordered', False)
688
 
        keys = [r.key for r in stream]
689
 
        # We want to get the keys back in disk order, but it doesn't matter
690
 
        # which pack we read from first. So this can come back in 2 orders
691
 
        alt1 = [('f-id', 'rev-%d' % i) for i in [4, 3, 6, 5, 1, 2]]
692
 
        alt2 = [('f-id', 'rev-%d' % i) for i in [5, 1, 2, 4, 3, 6]]
693
 
        if keys != alt1 and keys != alt2:
694
 
            self.fail('Returned key order did not match either expected order.'
695
 
                      ' expected %s or %s, not %s'
696
 
                      % (alt1, alt2, keys))
697
 
 
698
 
 
699
148
class LowLevelKnitDataTests(TestCase):
700
149
 
701
150
    def create_gz_content(self, text):
705
154
        gz_file.close()
706
155
        return sio.getvalue()
707
156
 
708
 
    def make_multiple_records(self):
709
 
        """Create the content for multiple records."""
710
 
        sha1sum = osutils.sha('foo\nbar\n').hexdigest()
711
 
        total_txt = []
712
 
        gz_txt = self.create_gz_content('version rev-id-1 2 %s\n'
713
 
                                        'foo\n'
714
 
                                        'bar\n'
715
 
                                        'end rev-id-1\n'
716
 
                                        % (sha1sum,))
717
 
        record_1 = (0, len(gz_txt), sha1sum)
718
 
        total_txt.append(gz_txt)
719
 
        sha1sum = osutils.sha('baz\n').hexdigest()
720
 
        gz_txt = self.create_gz_content('version rev-id-2 1 %s\n'
721
 
                                        'baz\n'
722
 
                                        'end rev-id-2\n'
723
 
                                        % (sha1sum,))
724
 
        record_2 = (record_1[1], len(gz_txt), sha1sum)
725
 
        total_txt.append(gz_txt)
726
 
        return total_txt, record_1, record_2
727
 
 
728
157
    def test_valid_knit_data(self):
729
 
        sha1sum = osutils.sha('foo\nbar\n').hexdigest()
 
158
        sha1sum = sha.new('foo\nbar\n').hexdigest()
730
159
        gz_txt = self.create_gz_content('version rev-id-1 2 %s\n'
731
160
                                        'foo\n'
732
161
                                        'bar\n'
733
162
                                        'end rev-id-1\n'
734
163
                                        % (sha1sum,))
735
164
        transport = MockTransport([gz_txt])
736
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
737
 
        knit = KnitVersionedFiles(None, access)
738
 
        records = [(('rev-id-1',), (('rev-id-1',), 0, len(gz_txt)))]
739
 
 
740
 
        contents = list(knit._read_records_iter(records))
741
 
        self.assertEqual([(('rev-id-1',), ['foo\n', 'bar\n'],
742
 
            '4e48e2c9a3d2ca8a708cb0cc545700544efb5021')], contents)
743
 
 
744
 
        raw_contents = list(knit._read_records_iter_raw(records))
745
 
        self.assertEqual([(('rev-id-1',), gz_txt, sha1sum)], raw_contents)
746
 
 
747
 
    def test_multiple_records_valid(self):
748
 
        total_txt, record_1, record_2 = self.make_multiple_records()
749
 
        transport = MockTransport([''.join(total_txt)])
750
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
751
 
        knit = KnitVersionedFiles(None, access)
752
 
        records = [(('rev-id-1',), (('rev-id-1',), record_1[0], record_1[1])),
753
 
                   (('rev-id-2',), (('rev-id-2',), record_2[0], record_2[1]))]
754
 
 
755
 
        contents = list(knit._read_records_iter(records))
756
 
        self.assertEqual([(('rev-id-1',), ['foo\n', 'bar\n'], record_1[2]),
757
 
                          (('rev-id-2',), ['baz\n'], record_2[2])],
758
 
                         contents)
759
 
 
760
 
        raw_contents = list(knit._read_records_iter_raw(records))
761
 
        self.assertEqual([(('rev-id-1',), total_txt[0], record_1[2]),
762
 
                          (('rev-id-2',), total_txt[1], record_2[2])],
763
 
                         raw_contents)
 
165
        data = _KnitData(transport, 'filename', mode='r')
 
166
        records = [('rev-id-1', 0, len(gz_txt))]
 
167
 
 
168
        contents = data.read_records(records)
 
169
        self.assertEqual({'rev-id-1':(['foo\n', 'bar\n'], sha1sum)}, contents)
 
170
 
 
171
        raw_contents = list(data.read_records_iter_raw(records))
 
172
        self.assertEqual([('rev-id-1', gz_txt)], raw_contents)
764
173
 
765
174
    def test_not_enough_lines(self):
766
 
        sha1sum = osutils.sha('foo\n').hexdigest()
 
175
        sha1sum = sha.new('foo\n').hexdigest()
767
176
        # record says 2 lines data says 1
768
177
        gz_txt = self.create_gz_content('version rev-id-1 2 %s\n'
769
178
                                        'foo\n'
770
179
                                        'end rev-id-1\n'
771
180
                                        % (sha1sum,))
772
181
        transport = MockTransport([gz_txt])
773
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
774
 
        knit = KnitVersionedFiles(None, access)
775
 
        records = [(('rev-id-1',), (('rev-id-1',), 0, len(gz_txt)))]
776
 
        self.assertRaises(errors.KnitCorrupt, list,
777
 
            knit._read_records_iter(records))
 
182
        data = _KnitData(transport, 'filename', mode='r')
 
183
        records = [('rev-id-1', 0, len(gz_txt))]
 
184
        self.assertRaises(errors.KnitCorrupt, data.read_records, records)
778
185
 
779
186
        # read_records_iter_raw won't detect that sort of mismatch/corruption
780
 
        raw_contents = list(knit._read_records_iter_raw(records))
781
 
        self.assertEqual([(('rev-id-1',),  gz_txt, sha1sum)], raw_contents)
 
187
        raw_contents = list(data.read_records_iter_raw(records))
 
188
        self.assertEqual([('rev-id-1', gz_txt)], raw_contents)
782
189
 
783
190
    def test_too_many_lines(self):
784
 
        sha1sum = osutils.sha('foo\nbar\n').hexdigest()
 
191
        sha1sum = sha.new('foo\nbar\n').hexdigest()
785
192
        # record says 1 lines data says 2
786
193
        gz_txt = self.create_gz_content('version rev-id-1 1 %s\n'
787
194
                                        'foo\n'
789
196
                                        'end rev-id-1\n'
790
197
                                        % (sha1sum,))
791
198
        transport = MockTransport([gz_txt])
792
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
793
 
        knit = KnitVersionedFiles(None, access)
794
 
        records = [(('rev-id-1',), (('rev-id-1',), 0, len(gz_txt)))]
795
 
        self.assertRaises(errors.KnitCorrupt, list,
796
 
            knit._read_records_iter(records))
 
199
        data = _KnitData(transport, 'filename', mode='r')
 
200
        records = [('rev-id-1', 0, len(gz_txt))]
 
201
        self.assertRaises(errors.KnitCorrupt, data.read_records, records)
797
202
 
798
203
        # read_records_iter_raw won't detect that sort of mismatch/corruption
799
 
        raw_contents = list(knit._read_records_iter_raw(records))
800
 
        self.assertEqual([(('rev-id-1',), gz_txt, sha1sum)], raw_contents)
 
204
        raw_contents = list(data.read_records_iter_raw(records))
 
205
        self.assertEqual([('rev-id-1', gz_txt)], raw_contents)
801
206
 
802
207
    def test_mismatched_version_id(self):
803
 
        sha1sum = osutils.sha('foo\nbar\n').hexdigest()
 
208
        sha1sum = sha.new('foo\nbar\n').hexdigest()
804
209
        gz_txt = self.create_gz_content('version rev-id-1 2 %s\n'
805
210
                                        'foo\n'
806
211
                                        'bar\n'
807
212
                                        'end rev-id-1\n'
808
213
                                        % (sha1sum,))
809
214
        transport = MockTransport([gz_txt])
810
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
811
 
        knit = KnitVersionedFiles(None, access)
 
215
        data = _KnitData(transport, 'filename', mode='r')
812
216
        # We are asking for rev-id-2, but the data is rev-id-1
813
 
        records = [(('rev-id-2',), (('rev-id-2',), 0, len(gz_txt)))]
814
 
        self.assertRaises(errors.KnitCorrupt, list,
815
 
            knit._read_records_iter(records))
 
217
        records = [('rev-id-2', 0, len(gz_txt))]
 
218
        self.assertRaises(errors.KnitCorrupt, data.read_records, records)
816
219
 
817
 
        # read_records_iter_raw detects mismatches in the header
 
220
        # read_records_iter_raw will notice if we request the wrong version.
818
221
        self.assertRaises(errors.KnitCorrupt, list,
819
 
            knit._read_records_iter_raw(records))
 
222
                          data.read_records_iter_raw(records))
820
223
 
821
224
    def test_uncompressed_data(self):
822
 
        sha1sum = osutils.sha('foo\nbar\n').hexdigest()
 
225
        sha1sum = sha.new('foo\nbar\n').hexdigest()
823
226
        txt = ('version rev-id-1 2 %s\n'
824
227
               'foo\n'
825
228
               'bar\n'
826
229
               'end rev-id-1\n'
827
230
               % (sha1sum,))
828
231
        transport = MockTransport([txt])
829
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
830
 
        knit = KnitVersionedFiles(None, access)
831
 
        records = [(('rev-id-1',), (('rev-id-1',), 0, len(txt)))]
 
232
        data = _KnitData(transport, 'filename', mode='r')
 
233
        records = [('rev-id-1', 0, len(txt))]
832
234
 
833
235
        # We don't have valid gzip data ==> corrupt
834
 
        self.assertRaises(errors.KnitCorrupt, list,
835
 
            knit._read_records_iter(records))
 
236
        self.assertRaises(errors.KnitCorrupt, data.read_records, records)
836
237
 
837
238
        # read_records_iter_raw will notice the bad data
838
239
        self.assertRaises(errors.KnitCorrupt, list,
839
 
            knit._read_records_iter_raw(records))
 
240
                          data.read_records_iter_raw(records))
840
241
 
841
242
    def test_corrupted_data(self):
842
 
        sha1sum = osutils.sha('foo\nbar\n').hexdigest()
 
243
        sha1sum = sha.new('foo\nbar\n').hexdigest()
843
244
        gz_txt = self.create_gz_content('version rev-id-1 2 %s\n'
844
245
                                        'foo\n'
845
246
                                        'bar\n'
848
249
        # Change 2 bytes in the middle to \xff
849
250
        gz_txt = gz_txt[:10] + '\xff\xff' + gz_txt[12:]
850
251
        transport = MockTransport([gz_txt])
851
 
        access = _KnitKeyAccess(transport, ConstantMapper('filename'))
852
 
        knit = KnitVersionedFiles(None, access)
853
 
        records = [(('rev-id-1',), (('rev-id-1',), 0, len(gz_txt)))]
854
 
        self.assertRaises(errors.KnitCorrupt, list,
855
 
            knit._read_records_iter(records))
856
 
        # read_records_iter_raw will barf on bad gz data
857
 
        self.assertRaises(errors.KnitCorrupt, list,
858
 
            knit._read_records_iter_raw(records))
 
252
        data = _KnitData(transport, 'filename', mode='r')
 
253
        records = [('rev-id-1', 0, len(gz_txt))]
 
254
 
 
255
        self.assertRaises(errors.KnitCorrupt, data.read_records, records)
 
256
 
 
257
        # read_records_iter_raw will notice if we request the wrong version.
 
258
        self.assertRaises(errors.KnitCorrupt, list,
 
259
                          data.read_records_iter_raw(records))
859
260
 
860
261
 
861
262
class LowLevelKnitIndexTests(TestCase):
862
263
 
863
 
    def get_knit_index(self, transport, name, mode):
864
 
        mapper = ConstantMapper(name)
 
264
    def get_knit_index(self, *args, **kwargs):
865
265
        orig = knit._load_data
866
266
        def reset():
867
267
            knit._load_data = orig
868
268
        self.addCleanup(reset)
869
269
        from bzrlib._knit_load_data_py import _load_data_py
870
270
        knit._load_data = _load_data_py
871
 
        allow_writes = lambda: 'w' in mode
872
 
        return _KndxIndex(transport, mapper, lambda:None, allow_writes, lambda:True)
 
271
        return _KnitIndex(*args, **kwargs)
 
272
 
 
273
    def test_no_such_file(self):
 
274
        transport = MockTransport()
 
275
 
 
276
        self.assertRaises(NoSuchFile, self.get_knit_index,
 
277
                          transport, "filename", "r")
 
278
        self.assertRaises(NoSuchFile, self.get_knit_index,
 
279
                          transport, "filename", "w", create=False)
873
280
 
874
281
    def test_create_file(self):
875
282
        transport = MockTransport()
876
 
        index = self.get_knit_index(transport, "filename", "w")
877
 
        index.keys()
878
 
        call = transport.calls.pop(0)
879
 
        # call[1][1] is a StringIO - we can't test it by simple equality.
880
 
        self.assertEqual('put_file_non_atomic', call[0])
881
 
        self.assertEqual('filename.kndx', call[1][0])
882
 
        # With no history, _KndxIndex writes a new index:
883
 
        self.assertEqual(_KndxIndex.HEADER,
884
 
            call[1][1].getvalue())
885
 
        self.assertEqual({'create_parent_dir': True}, call[2])
 
283
 
 
284
        index = self.get_knit_index(transport, "filename", "w",
 
285
            file_mode="wb", create=True)
 
286
        self.assertEqual(
 
287
                ("put_bytes_non_atomic",
 
288
                    ("filename", index.HEADER), {"mode": "wb"}),
 
289
                transport.calls.pop(0))
 
290
 
 
291
    def test_delay_create_file(self):
 
292
        transport = MockTransport()
 
293
 
 
294
        index = self.get_knit_index(transport, "filename", "w",
 
295
            create=True, file_mode="wb", create_parent_dir=True,
 
296
            delay_create=True, dir_mode=0777)
 
297
        self.assertEqual([], transport.calls)
 
298
 
 
299
        index.add_versions([])
 
300
        name, (filename, f), kwargs = transport.calls.pop(0)
 
301
        self.assertEqual("put_file_non_atomic", name)
 
302
        self.assertEqual(
 
303
            {"dir_mode": 0777, "create_parent_dir": True, "mode": "wb"},
 
304
            kwargs)
 
305
        self.assertEqual("filename", filename)
 
306
        self.assertEqual(index.HEADER, f.read())
 
307
 
 
308
        index.add_versions([])
 
309
        self.assertEqual(("append_bytes", ("filename", ""), {}),
 
310
            transport.calls.pop(0))
886
311
 
887
312
    def test_read_utf8_version_id(self):
888
313
        unicode_revision_id = u"version-\N{CYRILLIC CAPITAL LETTER A}"
889
314
        utf8_revision_id = unicode_revision_id.encode('utf-8')
890
315
        transport = MockTransport([
891
 
            _KndxIndex.HEADER,
 
316
            _KnitIndex.HEADER,
892
317
            '%s option 0 1 :' % (utf8_revision_id,)
893
318
            ])
894
319
        index = self.get_knit_index(transport, "filename", "r")
895
 
        # _KndxIndex is a private class, and deals in utf8 revision_ids, not
 
320
        # _KnitIndex is a private class, and deals in utf8 revision_ids, not
896
321
        # Unicode revision_ids.
897
 
        self.assertEqual({(utf8_revision_id,):()},
898
 
            index.get_parent_map(index.keys()))
899
 
        self.assertFalse((unicode_revision_id,) in index.keys())
 
322
        self.assertTrue(index.has_version(utf8_revision_id))
 
323
        self.assertFalse(index.has_version(unicode_revision_id))
900
324
 
901
325
    def test_read_utf8_parents(self):
902
326
        unicode_revision_id = u"version-\N{CYRILLIC CAPITAL LETTER A}"
903
327
        utf8_revision_id = unicode_revision_id.encode('utf-8')
904
328
        transport = MockTransport([
905
 
            _KndxIndex.HEADER,
 
329
            _KnitIndex.HEADER,
906
330
            "version option 0 1 .%s :" % (utf8_revision_id,)
907
331
            ])
908
332
        index = self.get_knit_index(transport, "filename", "r")
909
 
        self.assertEqual({("version",):((utf8_revision_id,),)},
910
 
            index.get_parent_map(index.keys()))
 
333
        self.assertEqual([utf8_revision_id],
 
334
            index.get_parents_with_ghosts("version"))
911
335
 
912
336
    def test_read_ignore_corrupted_lines(self):
913
337
        transport = MockTransport([
914
 
            _KndxIndex.HEADER,
 
338
            _KnitIndex.HEADER,
915
339
            "corrupted",
916
340
            "corrupted options 0 1 .b .c ",
917
341
            "version options 0 1 :"
918
342
            ])
919
343
        index = self.get_knit_index(transport, "filename", "r")
920
 
        self.assertEqual(1, len(index.keys()))
921
 
        self.assertEqual(set([("version",)]), index.keys())
 
344
        self.assertEqual(1, index.num_versions())
 
345
        self.assertTrue(index.has_version("version"))
922
346
 
923
347
    def test_read_corrupted_header(self):
924
348
        transport = MockTransport(['not a bzr knit index header\n'])
925
 
        index = self.get_knit_index(transport, "filename", "r")
926
 
        self.assertRaises(KnitHeaderError, index.keys)
 
349
        self.assertRaises(KnitHeaderError,
 
350
            self.get_knit_index, transport, "filename", "r")
927
351
 
928
352
    def test_read_duplicate_entries(self):
929
353
        transport = MockTransport([
930
 
            _KndxIndex.HEADER,
 
354
            _KnitIndex.HEADER,
931
355
            "parent options 0 1 :",
932
356
            "version options1 0 1 0 :",
933
357
            "version options2 1 2 .other :",
934
358
            "version options3 3 4 0 .other :"
935
359
            ])
936
360
        index = self.get_knit_index(transport, "filename", "r")
937
 
        self.assertEqual(2, len(index.keys()))
 
361
        self.assertEqual(2, index.num_versions())
938
362
        # check that the index used is the first one written. (Specific
939
363
        # to KnitIndex style indices.
940
 
        self.assertEqual("1", index._dictionary_compress([("version",)]))
941
 
        self.assertEqual((("version",), 3, 4), index.get_position(("version",)))
942
 
        self.assertEqual(["options3"], index.get_options(("version",)))
943
 
        self.assertEqual({("version",):(("parent",), ("other",))},
944
 
            index.get_parent_map([("version",)]))
 
364
        self.assertEqual("1", index._version_list_to_index(["version"]))
 
365
        self.assertEqual((3, 4), index.get_position("version"))
 
366
        self.assertEqual(["options3"], index.get_options("version"))
 
367
        self.assertEqual(["parent", "other"],
 
368
            index.get_parents_with_ghosts("version"))
945
369
 
946
370
    def test_read_compressed_parents(self):
947
371
        transport = MockTransport([
948
 
            _KndxIndex.HEADER,
 
372
            _KnitIndex.HEADER,
949
373
            "a option 0 1 :",
950
374
            "b option 0 1 0 :",
951
375
            "c option 0 1 1 0 :",
952
376
            ])
953
377
        index = self.get_knit_index(transport, "filename", "r")
954
 
        self.assertEqual({("b",):(("a",),), ("c",):(("b",), ("a",))},
955
 
            index.get_parent_map([("b",), ("c",)]))
 
378
        self.assertEqual(["a"], index.get_parents("b"))
 
379
        self.assertEqual(["b", "a"], index.get_parents("c"))
956
380
 
957
381
    def test_write_utf8_version_id(self):
958
382
        unicode_revision_id = u"version-\N{CYRILLIC CAPITAL LETTER A}"
959
383
        utf8_revision_id = unicode_revision_id.encode('utf-8')
960
384
        transport = MockTransport([
961
 
            _KndxIndex.HEADER
 
385
            _KnitIndex.HEADER
962
386
            ])
963
387
        index = self.get_knit_index(transport, "filename", "r")
964
 
        index.add_records([
965
 
            ((utf8_revision_id,), ["option"], ((utf8_revision_id,), 0, 1), [])])
966
 
        call = transport.calls.pop(0)
967
 
        # call[1][1] is a StringIO - we can't test it by simple equality.
968
 
        self.assertEqual('put_file_non_atomic', call[0])
969
 
        self.assertEqual('filename.kndx', call[1][0])
970
 
        # With no history, _KndxIndex writes a new index:
971
 
        self.assertEqual(_KndxIndex.HEADER +
972
 
            "\n%s option 0 1  :" % (utf8_revision_id,),
973
 
            call[1][1].getvalue())
974
 
        self.assertEqual({'create_parent_dir': True}, call[2])
 
388
        index.add_version(utf8_revision_id, ["option"], 0, 1, [])
 
389
        self.assertEqual(("append_bytes", ("filename",
 
390
            "\n%s option 0 1  :" % (utf8_revision_id,)),
 
391
            {}),
 
392
            transport.calls.pop(0))
975
393
 
976
394
    def test_write_utf8_parents(self):
977
395
        unicode_revision_id = u"version-\N{CYRILLIC CAPITAL LETTER A}"
978
396
        utf8_revision_id = unicode_revision_id.encode('utf-8')
979
397
        transport = MockTransport([
980
 
            _KndxIndex.HEADER
981
 
            ])
982
 
        index = self.get_knit_index(transport, "filename", "r")
983
 
        index.add_records([
984
 
            (("version",), ["option"], (("version",), 0, 1), [(utf8_revision_id,)])])
985
 
        call = transport.calls.pop(0)
986
 
        # call[1][1] is a StringIO - we can't test it by simple equality.
987
 
        self.assertEqual('put_file_non_atomic', call[0])
988
 
        self.assertEqual('filename.kndx', call[1][0])
989
 
        # With no history, _KndxIndex writes a new index:
990
 
        self.assertEqual(_KndxIndex.HEADER +
991
 
            "\nversion option 0 1 .%s :" % (utf8_revision_id,),
992
 
            call[1][1].getvalue())
993
 
        self.assertEqual({'create_parent_dir': True}, call[2])
994
 
 
995
 
    def test_keys(self):
996
 
        transport = MockTransport([
997
 
            _KndxIndex.HEADER
998
 
            ])
999
 
        index = self.get_knit_index(transport, "filename", "r")
1000
 
 
1001
 
        self.assertEqual(set(), index.keys())
1002
 
 
1003
 
        index.add_records([(("a",), ["option"], (("a",), 0, 1), [])])
1004
 
        self.assertEqual(set([("a",)]), index.keys())
1005
 
 
1006
 
        index.add_records([(("a",), ["option"], (("a",), 0, 1), [])])
1007
 
        self.assertEqual(set([("a",)]), index.keys())
1008
 
 
1009
 
        index.add_records([(("b",), ["option"], (("b",), 0, 1), [])])
1010
 
        self.assertEqual(set([("a",), ("b",)]), index.keys())
1011
 
 
1012
 
    def add_a_b(self, index, random_id=None):
1013
 
        kwargs = {}
1014
 
        if random_id is not None:
1015
 
            kwargs["random_id"] = random_id
1016
 
        index.add_records([
1017
 
            (("a",), ["option"], (("a",), 0, 1), [("b",)]),
1018
 
            (("a",), ["opt"], (("a",), 1, 2), [("c",)]),
1019
 
            (("b",), ["option"], (("b",), 2, 3), [("a",)])
1020
 
            ], **kwargs)
1021
 
 
1022
 
    def assertIndexIsAB(self, index):
1023
 
        self.assertEqual({
1024
 
            ('a',): (('c',),),
1025
 
            ('b',): (('a',),),
1026
 
            },
1027
 
            index.get_parent_map(index.keys()))
1028
 
        self.assertEqual((("a",), 1, 2), index.get_position(("a",)))
1029
 
        self.assertEqual((("b",), 2, 3), index.get_position(("b",)))
1030
 
        self.assertEqual(["opt"], index.get_options(("a",)))
 
398
            _KnitIndex.HEADER
 
399
            ])
 
400
        index = self.get_knit_index(transport, "filename", "r")
 
401
        index.add_version("version", ["option"], 0, 1, [utf8_revision_id])
 
402
        self.assertEqual(("append_bytes", ("filename",
 
403
            "\nversion option 0 1 .%s :" % (utf8_revision_id,)),
 
404
            {}),
 
405
            transport.calls.pop(0))
 
406
 
 
407
    def test_get_graph(self):
 
408
        transport = MockTransport()
 
409
        index = self.get_knit_index(transport, "filename", "w", create=True)
 
410
        self.assertEqual([], index.get_graph())
 
411
 
 
412
        index.add_version("a", ["option"], 0, 1, ["b"])
 
413
        self.assertEqual([("a", ["b"])], index.get_graph())
 
414
 
 
415
        index.add_version("c", ["option"], 0, 1, ["d"])
 
416
        self.assertEqual([("a", ["b"]), ("c", ["d"])],
 
417
            sorted(index.get_graph()))
 
418
 
 
419
    def test_get_ancestry(self):
 
420
        transport = MockTransport([
 
421
            _KnitIndex.HEADER,
 
422
            "a option 0 1 :",
 
423
            "b option 0 1 0 .e :",
 
424
            "c option 0 1 1 0 :",
 
425
            "d option 0 1 2 .f :"
 
426
            ])
 
427
        index = self.get_knit_index(transport, "filename", "r")
 
428
 
 
429
        self.assertEqual([], index.get_ancestry([]))
 
430
        self.assertEqual(["a"], index.get_ancestry(["a"]))
 
431
        self.assertEqual(["a", "b"], index.get_ancestry(["b"]))
 
432
        self.assertEqual(["a", "b", "c"], index.get_ancestry(["c"]))
 
433
        self.assertEqual(["a", "b", "c", "d"], index.get_ancestry(["d"]))
 
434
        self.assertEqual(["a", "b"], index.get_ancestry(["a", "b"]))
 
435
        self.assertEqual(["a", "b", "c"], index.get_ancestry(["a", "c"]))
 
436
 
 
437
        self.assertRaises(RevisionNotPresent, index.get_ancestry, ["e"])
 
438
 
 
439
    def test_get_ancestry_with_ghosts(self):
 
440
        transport = MockTransport([
 
441
            _KnitIndex.HEADER,
 
442
            "a option 0 1 :",
 
443
            "b option 0 1 0 .e :",
 
444
            "c option 0 1 0 .f .g :",
 
445
            "d option 0 1 2 .h .j .k :"
 
446
            ])
 
447
        index = self.get_knit_index(transport, "filename", "r")
 
448
 
 
449
        self.assertEqual([], index.get_ancestry_with_ghosts([]))
 
450
        self.assertEqual(["a"], index.get_ancestry_with_ghosts(["a"]))
 
451
        self.assertEqual(["a", "e", "b"],
 
452
            index.get_ancestry_with_ghosts(["b"]))
 
453
        self.assertEqual(["a", "g", "f", "c"],
 
454
            index.get_ancestry_with_ghosts(["c"]))
 
455
        self.assertEqual(["a", "g", "f", "c", "k", "j", "h", "d"],
 
456
            index.get_ancestry_with_ghosts(["d"]))
 
457
        self.assertEqual(["a", "e", "b"],
 
458
            index.get_ancestry_with_ghosts(["a", "b"]))
 
459
        self.assertEqual(["a", "g", "f", "c"],
 
460
            index.get_ancestry_with_ghosts(["a", "c"]))
 
461
        self.assertEqual(
 
462
            ["a", "g", "f", "c", "e", "b", "k", "j", "h", "d"],
 
463
            index.get_ancestry_with_ghosts(["b", "d"]))
 
464
 
 
465
        self.assertRaises(RevisionNotPresent,
 
466
            index.get_ancestry_with_ghosts, ["e"])
 
467
 
 
468
    def test_iter_parents(self):
 
469
        transport = MockTransport()
 
470
        index = self.get_knit_index(transport, "filename", "w", create=True)
 
471
        # no parents
 
472
        index.add_version('r0', ['option'], 0, 1, [])
 
473
        # 1 parent
 
474
        index.add_version('r1', ['option'], 0, 1, ['r0'])
 
475
        # 2 parents
 
476
        index.add_version('r2', ['option'], 0, 1, ['r1', 'r0'])
 
477
        # XXX TODO a ghost
 
478
        # cases: each sample data individually:
 
479
        self.assertEqual(set([('r0', ())]),
 
480
            set(index.iter_parents(['r0'])))
 
481
        self.assertEqual(set([('r1', ('r0', ))]),
 
482
            set(index.iter_parents(['r1'])))
 
483
        self.assertEqual(set([('r2', ('r1', 'r0'))]),
 
484
            set(index.iter_parents(['r2'])))
 
485
        # no nodes returned for a missing node
 
486
        self.assertEqual(set(),
 
487
            set(index.iter_parents(['missing'])))
 
488
        # 1 node returned with missing nodes skipped
 
489
        self.assertEqual(set([('r1', ('r0', ))]),
 
490
            set(index.iter_parents(['ghost1', 'r1', 'ghost'])))
 
491
        # 2 nodes returned
 
492
        self.assertEqual(set([('r0', ()), ('r1', ('r0', ))]),
 
493
            set(index.iter_parents(['r0', 'r1'])))
 
494
        # 2 nodes returned, missing skipped
 
495
        self.assertEqual(set([('r0', ()), ('r1', ('r0', ))]),
 
496
            set(index.iter_parents(['a', 'r0', 'b', 'r1', 'c'])))
 
497
 
 
498
    def test_num_versions(self):
 
499
        transport = MockTransport([
 
500
            _KnitIndex.HEADER
 
501
            ])
 
502
        index = self.get_knit_index(transport, "filename", "r")
 
503
 
 
504
        self.assertEqual(0, index.num_versions())
 
505
        self.assertEqual(0, len(index))
 
506
 
 
507
        index.add_version("a", ["option"], 0, 1, [])
 
508
        self.assertEqual(1, index.num_versions())
 
509
        self.assertEqual(1, len(index))
 
510
 
 
511
        index.add_version("a", ["option2"], 1, 2, [])
 
512
        self.assertEqual(1, index.num_versions())
 
513
        self.assertEqual(1, len(index))
 
514
 
 
515
        index.add_version("b", ["option"], 0, 1, [])
 
516
        self.assertEqual(2, index.num_versions())
 
517
        self.assertEqual(2, len(index))
 
518
 
 
519
    def test_get_versions(self):
 
520
        transport = MockTransport([
 
521
            _KnitIndex.HEADER
 
522
            ])
 
523
        index = self.get_knit_index(transport, "filename", "r")
 
524
 
 
525
        self.assertEqual([], index.get_versions())
 
526
 
 
527
        index.add_version("a", ["option"], 0, 1, [])
 
528
        self.assertEqual(["a"], index.get_versions())
 
529
 
 
530
        index.add_version("a", ["option"], 0, 1, [])
 
531
        self.assertEqual(["a"], index.get_versions())
 
532
 
 
533
        index.add_version("b", ["option"], 0, 1, [])
 
534
        self.assertEqual(["a", "b"], index.get_versions())
 
535
 
 
536
    def test_add_version(self):
 
537
        transport = MockTransport([
 
538
            _KnitIndex.HEADER
 
539
            ])
 
540
        index = self.get_knit_index(transport, "filename", "r")
 
541
 
 
542
        index.add_version("a", ["option"], 0, 1, ["b"])
 
543
        self.assertEqual(("append_bytes",
 
544
            ("filename", "\na option 0 1 .b :"),
 
545
            {}), transport.calls.pop(0))
 
546
        self.assertTrue(index.has_version("a"))
 
547
        self.assertEqual(1, index.num_versions())
 
548
        self.assertEqual((0, 1), index.get_position("a"))
 
549
        self.assertEqual(["option"], index.get_options("a"))
 
550
        self.assertEqual(["b"], index.get_parents_with_ghosts("a"))
 
551
 
 
552
        index.add_version("a", ["opt"], 1, 2, ["c"])
 
553
        self.assertEqual(("append_bytes",
 
554
            ("filename", "\na opt 1 2 .c :"),
 
555
            {}), transport.calls.pop(0))
 
556
        self.assertTrue(index.has_version("a"))
 
557
        self.assertEqual(1, index.num_versions())
 
558
        self.assertEqual((1, 2), index.get_position("a"))
 
559
        self.assertEqual(["opt"], index.get_options("a"))
 
560
        self.assertEqual(["c"], index.get_parents_with_ghosts("a"))
 
561
 
 
562
        index.add_version("b", ["option"], 2, 3, ["a"])
 
563
        self.assertEqual(("append_bytes",
 
564
            ("filename", "\nb option 2 3 0 :"),
 
565
            {}), transport.calls.pop(0))
 
566
        self.assertTrue(index.has_version("b"))
 
567
        self.assertEqual(2, index.num_versions())
 
568
        self.assertEqual((2, 3), index.get_position("b"))
 
569
        self.assertEqual(["option"], index.get_options("b"))
 
570
        self.assertEqual(["a"], index.get_parents_with_ghosts("b"))
1031
571
 
1032
572
    def test_add_versions(self):
1033
573
        transport = MockTransport([
1034
 
            _KndxIndex.HEADER
 
574
            _KnitIndex.HEADER
1035
575
            ])
1036
576
        index = self.get_knit_index(transport, "filename", "r")
1037
577
 
1038
 
        self.add_a_b(index)
1039
 
        call = transport.calls.pop(0)
1040
 
        # call[1][1] is a StringIO - we can't test it by simple equality.
1041
 
        self.assertEqual('put_file_non_atomic', call[0])
1042
 
        self.assertEqual('filename.kndx', call[1][0])
1043
 
        # With no history, _KndxIndex writes a new index:
1044
 
        self.assertEqual(
1045
 
            _KndxIndex.HEADER +
 
578
        index.add_versions([
 
579
            ("a", ["option"], 0, 1, ["b"]),
 
580
            ("a", ["opt"], 1, 2, ["c"]),
 
581
            ("b", ["option"], 2, 3, ["a"])
 
582
            ])
 
583
        self.assertEqual(("append_bytes", ("filename",
1046
584
            "\na option 0 1 .b :"
1047
585
            "\na opt 1 2 .c :"
1048
 
            "\nb option 2 3 0 :",
1049
 
            call[1][1].getvalue())
1050
 
        self.assertEqual({'create_parent_dir': True}, call[2])
1051
 
        self.assertIndexIsAB(index)
1052
 
 
1053
 
    def test_add_versions_random_id_is_accepted(self):
1054
 
        transport = MockTransport([
1055
 
            _KndxIndex.HEADER
1056
 
            ])
1057
 
        index = self.get_knit_index(transport, "filename", "r")
1058
 
        self.add_a_b(index, random_id=True)
 
586
            "\nb option 2 3 0 :"
 
587
            ), {}), transport.calls.pop(0))
 
588
        self.assertTrue(index.has_version("a"))
 
589
        self.assertTrue(index.has_version("b"))
 
590
        self.assertEqual(2, index.num_versions())
 
591
        self.assertEqual((1, 2), index.get_position("a"))
 
592
        self.assertEqual((2, 3), index.get_position("b"))
 
593
        self.assertEqual(["opt"], index.get_options("a"))
 
594
        self.assertEqual(["option"], index.get_options("b"))
 
595
        self.assertEqual(["c"], index.get_parents_with_ghosts("a"))
 
596
        self.assertEqual(["a"], index.get_parents_with_ghosts("b"))
1059
597
 
1060
598
    def test_delay_create_and_add_versions(self):
1061
599
        transport = MockTransport()
1062
600
 
1063
 
        index = self.get_knit_index(transport, "filename", "w")
1064
 
        # dir_mode=0777)
 
601
        index = self.get_knit_index(transport, "filename", "w",
 
602
            create=True, file_mode="wb", create_parent_dir=True,
 
603
            delay_create=True, dir_mode=0777)
1065
604
        self.assertEqual([], transport.calls)
1066
 
        self.add_a_b(index)
1067
 
        #self.assertEqual(
1068
 
        #[    {"dir_mode": 0777, "create_parent_dir": True, "mode": "wb"},
1069
 
        #    kwargs)
1070
 
        # Two calls: one during which we load the existing index (and when its
1071
 
        # missing create it), then a second where we write the contents out.
1072
 
        self.assertEqual(2, len(transport.calls))
1073
 
        call = transport.calls.pop(0)
1074
 
        self.assertEqual('put_file_non_atomic', call[0])
1075
 
        self.assertEqual('filename.kndx', call[1][0])
1076
 
        # With no history, _KndxIndex writes a new index:
1077
 
        self.assertEqual(_KndxIndex.HEADER, call[1][1].getvalue())
1078
 
        self.assertEqual({'create_parent_dir': True}, call[2])
1079
 
        call = transport.calls.pop(0)
1080
 
        # call[1][1] is a StringIO - we can't test it by simple equality.
1081
 
        self.assertEqual('put_file_non_atomic', call[0])
1082
 
        self.assertEqual('filename.kndx', call[1][0])
1083
 
        # With no history, _KndxIndex writes a new index:
1084
 
        self.assertEqual(
1085
 
            _KndxIndex.HEADER +
 
605
 
 
606
        index.add_versions([
 
607
            ("a", ["option"], 0, 1, ["b"]),
 
608
            ("a", ["opt"], 1, 2, ["c"]),
 
609
            ("b", ["option"], 2, 3, ["a"])
 
610
            ])
 
611
        name, (filename, f), kwargs = transport.calls.pop(0)
 
612
        self.assertEqual("put_file_non_atomic", name)
 
613
        self.assertEqual(
 
614
            {"dir_mode": 0777, "create_parent_dir": True, "mode": "wb"},
 
615
            kwargs)
 
616
        self.assertEqual("filename", filename)
 
617
        self.assertEqual(
 
618
            index.HEADER +
1086
619
            "\na option 0 1 .b :"
1087
620
            "\na opt 1 2 .c :"
1088
621
            "\nb option 2 3 0 :",
1089
 
            call[1][1].getvalue())
1090
 
        self.assertEqual({'create_parent_dir': True}, call[2])
1091
 
 
1092
 
    def assertTotalBuildSize(self, size, keys, positions):
1093
 
        self.assertEqual(size,
1094
 
                         knit._get_total_build_size(None, keys, positions))
1095
 
 
1096
 
    def test__get_total_build_size(self):
1097
 
        positions = {
1098
 
            ('a',): (('fulltext', False), (('a',), 0, 100), None),
1099
 
            ('b',): (('line-delta', False), (('b',), 100, 21), ('a',)),
1100
 
            ('c',): (('line-delta', False), (('c',), 121, 35), ('b',)),
1101
 
            ('d',): (('line-delta', False), (('d',), 156, 12), ('b',)),
1102
 
            }
1103
 
        self.assertTotalBuildSize(100, [('a',)], positions)
1104
 
        self.assertTotalBuildSize(121, [('b',)], positions)
1105
 
        # c needs both a & b
1106
 
        self.assertTotalBuildSize(156, [('c',)], positions)
1107
 
        # we shouldn't count 'b' twice
1108
 
        self.assertTotalBuildSize(156, [('b',), ('c',)], positions)
1109
 
        self.assertTotalBuildSize(133, [('d',)], positions)
1110
 
        self.assertTotalBuildSize(168, [('c',), ('d',)], positions)
 
622
            f.read())
 
623
 
 
624
    def test_has_version(self):
 
625
        transport = MockTransport([
 
626
            _KnitIndex.HEADER,
 
627
            "a option 0 1 :"
 
628
            ])
 
629
        index = self.get_knit_index(transport, "filename", "r")
 
630
 
 
631
        self.assertTrue(index.has_version("a"))
 
632
        self.assertFalse(index.has_version("b"))
1111
633
 
1112
634
    def test_get_position(self):
1113
635
        transport = MockTransport([
1114
 
            _KndxIndex.HEADER,
 
636
            _KnitIndex.HEADER,
1115
637
            "a option 0 1 :",
1116
638
            "b option 1 2 :"
1117
639
            ])
1118
640
        index = self.get_knit_index(transport, "filename", "r")
1119
641
 
1120
 
        self.assertEqual((("a",), 0, 1), index.get_position(("a",)))
1121
 
        self.assertEqual((("b",), 1, 2), index.get_position(("b",)))
 
642
        self.assertEqual((0, 1), index.get_position("a"))
 
643
        self.assertEqual((1, 2), index.get_position("b"))
1122
644
 
1123
645
    def test_get_method(self):
1124
646
        transport = MockTransport([
1125
 
            _KndxIndex.HEADER,
 
647
            _KnitIndex.HEADER,
1126
648
            "a fulltext,unknown 0 1 :",
1127
649
            "b unknown,line-delta 1 2 :",
1128
650
            "c bad 3 4 :"
1135
657
 
1136
658
    def test_get_options(self):
1137
659
        transport = MockTransport([
1138
 
            _KndxIndex.HEADER,
 
660
            _KnitIndex.HEADER,
1139
661
            "a opt1 0 1 :",
1140
662
            "b opt2,opt3 1 2 :"
1141
663
            ])
1144
666
        self.assertEqual(["opt1"], index.get_options("a"))
1145
667
        self.assertEqual(["opt2", "opt3"], index.get_options("b"))
1146
668
 
1147
 
    def test_get_parent_map(self):
1148
 
        transport = MockTransport([
1149
 
            _KndxIndex.HEADER,
1150
 
            "a option 0 1 :",
1151
 
            "b option 1 2 0 .c :",
1152
 
            "c option 1 2 1 0 .e :"
1153
 
            ])
1154
 
        index = self.get_knit_index(transport, "filename", "r")
1155
 
 
1156
 
        self.assertEqual({
1157
 
            ("a",):(),
1158
 
            ("b",):(("a",), ("c",)),
1159
 
            ("c",):(("b",), ("a",), ("e",)),
1160
 
            }, index.get_parent_map(index.keys()))
 
669
    def test_get_parents(self):
 
670
        transport = MockTransport([
 
671
            _KnitIndex.HEADER,
 
672
            "a option 0 1 :",
 
673
            "b option 1 2 0 .c :",
 
674
            "c option 1 2 1 0 .e :"
 
675
            ])
 
676
        index = self.get_knit_index(transport, "filename", "r")
 
677
 
 
678
        self.assertEqual([], index.get_parents("a"))
 
679
        self.assertEqual(["a", "c"], index.get_parents("b"))
 
680
        self.assertEqual(["b", "a"], index.get_parents("c"))
 
681
 
 
682
    def test_get_parents_with_ghosts(self):
 
683
        transport = MockTransport([
 
684
            _KnitIndex.HEADER,
 
685
            "a option 0 1 :",
 
686
            "b option 1 2 0 .c :",
 
687
            "c option 1 2 1 0 .e :"
 
688
            ])
 
689
        index = self.get_knit_index(transport, "filename", "r")
 
690
 
 
691
        self.assertEqual([], index.get_parents_with_ghosts("a"))
 
692
        self.assertEqual(["a", "c"], index.get_parents_with_ghosts("b"))
 
693
        self.assertEqual(["b", "a", "e"],
 
694
            index.get_parents_with_ghosts("c"))
 
695
 
 
696
    def test_check_versions_present(self):
 
697
        transport = MockTransport([
 
698
            _KnitIndex.HEADER,
 
699
            "a option 0 1 :",
 
700
            "b option 0 1 :"
 
701
            ])
 
702
        index = self.get_knit_index(transport, "filename", "r")
 
703
 
 
704
        check = index.check_versions_present
 
705
 
 
706
        check([])
 
707
        check(["a"])
 
708
        check(["b"])
 
709
        check(["a", "b"])
 
710
        self.assertRaises(RevisionNotPresent, check, ["c"])
 
711
        self.assertRaises(RevisionNotPresent, check, ["a", "b", "c"])
1161
712
 
1162
713
    def test_impossible_parent(self):
1163
714
        """Test we get KnitCorrupt if the parent couldn't possibly exist."""
1164
715
        transport = MockTransport([
1165
 
            _KndxIndex.HEADER,
 
716
            _KnitIndex.HEADER,
1166
717
            "a option 0 1 :",
1167
718
            "b option 0 1 4 :"  # We don't have a 4th record
1168
719
            ])
1169
 
        index = self.get_knit_index(transport, 'filename', 'r')
1170
720
        try:
1171
 
            self.assertRaises(errors.KnitCorrupt, index.keys)
 
721
            self.assertRaises(errors.KnitCorrupt,
 
722
                              self.get_knit_index, transport, 'filename', 'r')
1172
723
        except TypeError, e:
1173
724
            if (str(e) == ('exceptions must be strings, classes, or instances,'
1174
725
                           ' not exceptions.IndexError')
1181
732
 
1182
733
    def test_corrupted_parent(self):
1183
734
        transport = MockTransport([
1184
 
            _KndxIndex.HEADER,
 
735
            _KnitIndex.HEADER,
1185
736
            "a option 0 1 :",
1186
737
            "b option 0 1 :",
1187
738
            "c option 0 1 1v :", # Can't have a parent of '1v'
1188
739
            ])
1189
 
        index = self.get_knit_index(transport, 'filename', 'r')
1190
740
        try:
1191
 
            self.assertRaises(errors.KnitCorrupt, index.keys)
 
741
            self.assertRaises(errors.KnitCorrupt,
 
742
                              self.get_knit_index, transport, 'filename', 'r')
1192
743
        except TypeError, e:
1193
744
            if (str(e) == ('exceptions must be strings, classes, or instances,'
1194
745
                           ' not exceptions.ValueError')
1201
752
 
1202
753
    def test_corrupted_parent_in_list(self):
1203
754
        transport = MockTransport([
1204
 
            _KndxIndex.HEADER,
 
755
            _KnitIndex.HEADER,
1205
756
            "a option 0 1 :",
1206
757
            "b option 0 1 :",
1207
758
            "c option 0 1 1 v :", # Can't have a parent of 'v'
1208
759
            ])
1209
 
        index = self.get_knit_index(transport, 'filename', 'r')
1210
760
        try:
1211
 
            self.assertRaises(errors.KnitCorrupt, index.keys)
 
761
            self.assertRaises(errors.KnitCorrupt,
 
762
                              self.get_knit_index, transport, 'filename', 'r')
1212
763
        except TypeError, e:
1213
764
            if (str(e) == ('exceptions must be strings, classes, or instances,'
1214
765
                           ' not exceptions.ValueError')
1221
772
 
1222
773
    def test_invalid_position(self):
1223
774
        transport = MockTransport([
1224
 
            _KndxIndex.HEADER,
 
775
            _KnitIndex.HEADER,
1225
776
            "a option 1v 1 :",
1226
777
            ])
1227
 
        index = self.get_knit_index(transport, 'filename', 'r')
1228
778
        try:
1229
 
            self.assertRaises(errors.KnitCorrupt, index.keys)
 
779
            self.assertRaises(errors.KnitCorrupt,
 
780
                              self.get_knit_index, transport, 'filename', 'r')
1230
781
        except TypeError, e:
1231
782
            if (str(e) == ('exceptions must be strings, classes, or instances,'
1232
783
                           ' not exceptions.ValueError')
1239
790
 
1240
791
    def test_invalid_size(self):
1241
792
        transport = MockTransport([
1242
 
            _KndxIndex.HEADER,
 
793
            _KnitIndex.HEADER,
1243
794
            "a option 1 1v :",
1244
795
            ])
1245
 
        index = self.get_knit_index(transport, 'filename', 'r')
1246
796
        try:
1247
 
            self.assertRaises(errors.KnitCorrupt, index.keys)
 
797
            self.assertRaises(errors.KnitCorrupt,
 
798
                              self.get_knit_index, transport, 'filename', 'r')
1248
799
        except TypeError, e:
1249
800
            if (str(e) == ('exceptions must be strings, classes, or instances,'
1250
801
                           ' not exceptions.ValueError')
1255
806
            else:
1256
807
                raise
1257
808
 
1258
 
    def test_scan_unvalidated_index_not_implemented(self):
1259
 
        transport = MockTransport()
1260
 
        index = self.get_knit_index(transport, 'filename', 'r')
1261
 
        self.assertRaises(
1262
 
            NotImplementedError, index.scan_unvalidated_index,
1263
 
            'dummy graph_index')
1264
 
        self.assertRaises(
1265
 
            NotImplementedError, index.get_missing_compression_parents)
1266
 
 
1267
809
    def test_short_line(self):
1268
810
        transport = MockTransport([
1269
 
            _KndxIndex.HEADER,
 
811
            _KnitIndex.HEADER,
1270
812
            "a option 0 10  :",
1271
813
            "b option 10 10 0", # This line isn't terminated, ignored
1272
814
            ])
1273
815
        index = self.get_knit_index(transport, "filename", "r")
1274
 
        self.assertEqual(set([('a',)]), index.keys())
 
816
        self.assertEqual(['a'], index.get_versions())
1275
817
 
1276
818
    def test_skip_incomplete_record(self):
1277
819
        # A line with bogus data should just be skipped
1278
820
        transport = MockTransport([
1279
 
            _KndxIndex.HEADER,
 
821
            _KnitIndex.HEADER,
1280
822
            "a option 0 10  :",
1281
823
            "b option 10 10 0", # This line isn't terminated, ignored
1282
824
            "c option 20 10 0 :", # Properly terminated, and starts with '\n'
1283
825
            ])
1284
826
        index = self.get_knit_index(transport, "filename", "r")
1285
 
        self.assertEqual(set([('a',), ('c',)]), index.keys())
 
827
        self.assertEqual(['a', 'c'], index.get_versions())
1286
828
 
1287
829
    def test_trailing_characters(self):
1288
830
        # A line with bogus data should just be skipped
1289
831
        transport = MockTransport([
1290
 
            _KndxIndex.HEADER,
 
832
            _KnitIndex.HEADER,
1291
833
            "a option 0 10  :",
1292
834
            "b option 10 10 0 :a", # This line has extra trailing characters
1293
835
            "c option 20 10 0 :", # Properly terminated, and starts with '\n'
1294
836
            ])
1295
837
        index = self.get_knit_index(transport, "filename", "r")
1296
 
        self.assertEqual(set([('a',), ('c',)]), index.keys())
 
838
        self.assertEqual(['a', 'c'], index.get_versions())
1297
839
 
1298
840
 
1299
841
class LowLevelKnitIndexTests_c(LowLevelKnitIndexTests):
1300
842
 
1301
 
    _test_needs_features = [compiled_knit_feature]
 
843
    _test_needs_features = [CompiledKnitFeature]
1302
844
 
1303
 
    def get_knit_index(self, transport, name, mode):
1304
 
        mapper = ConstantMapper(name)
 
845
    def get_knit_index(self, *args, **kwargs):
1305
846
        orig = knit._load_data
1306
847
        def reset():
1307
848
            knit._load_data = orig
1308
849
        self.addCleanup(reset)
1309
 
        from bzrlib._knit_load_data_pyx import _load_data_c
 
850
        from bzrlib._knit_load_data_c import _load_data_c
1310
851
        knit._load_data = _load_data_c
1311
 
        allow_writes = lambda: mode == 'w'
1312
 
        return _KndxIndex(transport, mapper, lambda:None, allow_writes, lambda:True)
1313
 
 
1314
 
 
1315
 
class Test_KnitAnnotator(TestCaseWithMemoryTransport):
1316
 
 
1317
 
    def make_annotator(self):
1318
 
        factory = knit.make_pack_factory(True, True, 1)
1319
 
        vf = factory(self.get_transport())
1320
 
        return knit._KnitAnnotator(vf)
1321
 
 
1322
 
    def test__expand_fulltext(self):
1323
 
        ann = self.make_annotator()
1324
 
        rev_key = ('rev-id',)
1325
 
        ann._num_compression_children[rev_key] = 1
1326
 
        res = ann._expand_record(rev_key, (('parent-id',),), None,
1327
 
                           ['line1\n', 'line2\n'], ('fulltext', True))
1328
 
        # The content object and text lines should be cached appropriately
1329
 
        self.assertEqual(['line1\n', 'line2'], res)
1330
 
        content_obj = ann._content_objects[rev_key]
1331
 
        self.assertEqual(['line1\n', 'line2\n'], content_obj._lines)
1332
 
        self.assertEqual(res, content_obj.text())
1333
 
        self.assertEqual(res, ann._text_cache[rev_key])
1334
 
 
1335
 
    def test__expand_delta_comp_parent_not_available(self):
1336
 
        # Parent isn't available yet, so we return nothing, but queue up this
1337
 
        # node for later processing
1338
 
        ann = self.make_annotator()
1339
 
        rev_key = ('rev-id',)
1340
 
        parent_key = ('parent-id',)
1341
 
        record = ['0,1,1\n', 'new-line\n']
1342
 
        details = ('line-delta', False)
1343
 
        res = ann._expand_record(rev_key, (parent_key,), parent_key,
1344
 
                                 record, details)
1345
 
        self.assertEqual(None, res)
1346
 
        self.assertTrue(parent_key in ann._pending_deltas)
1347
 
        pending = ann._pending_deltas[parent_key]
1348
 
        self.assertEqual(1, len(pending))
1349
 
        self.assertEqual((rev_key, (parent_key,), record, details), pending[0])
1350
 
 
1351
 
    def test__expand_record_tracks_num_children(self):
1352
 
        ann = self.make_annotator()
1353
 
        rev_key = ('rev-id',)
1354
 
        rev2_key = ('rev2-id',)
1355
 
        parent_key = ('parent-id',)
1356
 
        record = ['0,1,1\n', 'new-line\n']
1357
 
        details = ('line-delta', False)
1358
 
        ann._num_compression_children[parent_key] = 2
1359
 
        ann._expand_record(parent_key, (), None, ['line1\n', 'line2\n'],
1360
 
                           ('fulltext', False))
1361
 
        res = ann._expand_record(rev_key, (parent_key,), parent_key,
1362
 
                                 record, details)
1363
 
        self.assertEqual({parent_key: 1}, ann._num_compression_children)
1364
 
        # Expanding the second child should remove the content object, and the
1365
 
        # num_compression_children entry
1366
 
        res = ann._expand_record(rev2_key, (parent_key,), parent_key,
1367
 
                                 record, details)
1368
 
        self.assertFalse(parent_key in ann._content_objects)
1369
 
        self.assertEqual({}, ann._num_compression_children)
1370
 
        # We should not cache the content_objects for rev2 and rev, because
1371
 
        # they do not have compression children of their own.
1372
 
        self.assertEqual({}, ann._content_objects)
1373
 
 
1374
 
    def test__expand_delta_records_blocks(self):
1375
 
        ann = self.make_annotator()
1376
 
        rev_key = ('rev-id',)
1377
 
        parent_key = ('parent-id',)
1378
 
        record = ['0,1,1\n', 'new-line\n']
1379
 
        details = ('line-delta', True)
1380
 
        ann._num_compression_children[parent_key] = 2
1381
 
        ann._expand_record(parent_key, (), None,
1382
 
                           ['line1\n', 'line2\n', 'line3\n'],
1383
 
                           ('fulltext', False))
1384
 
        ann._expand_record(rev_key, (parent_key,), parent_key, record, details)
1385
 
        self.assertEqual({(rev_key, parent_key): [(1, 1, 1), (3, 3, 0)]},
1386
 
                         ann._matching_blocks)
1387
 
        rev2_key = ('rev2-id',)
1388
 
        record = ['0,1,1\n', 'new-line\n']
1389
 
        details = ('line-delta', False)
1390
 
        ann._expand_record(rev2_key, (parent_key,), parent_key, record, details)
1391
 
        self.assertEqual([(1, 1, 2), (3, 3, 0)],
1392
 
                         ann._matching_blocks[(rev2_key, parent_key)])
1393
 
 
1394
 
    def test__get_parent_ann_uses_matching_blocks(self):
1395
 
        ann = self.make_annotator()
1396
 
        rev_key = ('rev-id',)
1397
 
        parent_key = ('parent-id',)
1398
 
        parent_ann = [(parent_key,)]*3
1399
 
        block_key = (rev_key, parent_key)
1400
 
        ann._annotations_cache[parent_key] = parent_ann
1401
 
        ann._matching_blocks[block_key] = [(0, 1, 1), (3, 3, 0)]
1402
 
        # We should not try to access any parent_lines content, because we know
1403
 
        # we already have the matching blocks
1404
 
        par_ann, blocks = ann._get_parent_annotations_and_matches(rev_key,
1405
 
                                        ['1\n', '2\n', '3\n'], parent_key)
1406
 
        self.assertEqual(parent_ann, par_ann)
1407
 
        self.assertEqual([(0, 1, 1), (3, 3, 0)], blocks)
1408
 
        self.assertEqual({}, ann._matching_blocks)
1409
 
 
1410
 
    def test__process_pending(self):
1411
 
        ann = self.make_annotator()
1412
 
        rev_key = ('rev-id',)
1413
 
        p1_key = ('p1-id',)
1414
 
        p2_key = ('p2-id',)
1415
 
        record = ['0,1,1\n', 'new-line\n']
1416
 
        details = ('line-delta', False)
1417
 
        p1_record = ['line1\n', 'line2\n']
1418
 
        ann._num_compression_children[p1_key] = 1
1419
 
        res = ann._expand_record(rev_key, (p1_key,p2_key), p1_key,
1420
 
                                 record, details)
1421
 
        self.assertEqual(None, res)
1422
 
        # self.assertTrue(p1_key in ann._pending_deltas)
1423
 
        self.assertEqual({}, ann._pending_annotation)
1424
 
        # Now insert p1, and we should be able to expand the delta
1425
 
        res = ann._expand_record(p1_key, (), None, p1_record,
1426
 
                                 ('fulltext', False))
1427
 
        self.assertEqual(p1_record, res)
1428
 
        ann._annotations_cache[p1_key] = [(p1_key,)]*2
1429
 
        res = ann._process_pending(p1_key)
1430
 
        self.assertEqual([], res)
1431
 
        self.assertFalse(p1_key in ann._pending_deltas)
1432
 
        self.assertTrue(p2_key in ann._pending_annotation)
1433
 
        self.assertEqual({p2_key: [(rev_key, (p1_key, p2_key))]},
1434
 
                         ann._pending_annotation)
1435
 
        # Now fill in parent 2, and pending annotation should be satisfied
1436
 
        res = ann._expand_record(p2_key, (), None, [], ('fulltext', False))
1437
 
        ann._annotations_cache[p2_key] = []
1438
 
        res = ann._process_pending(p2_key)
1439
 
        self.assertEqual([rev_key], res)
1440
 
        self.assertEqual({}, ann._pending_annotation)
1441
 
        self.assertEqual({}, ann._pending_deltas)
1442
 
 
1443
 
    def test_record_delta_removes_basis(self):
1444
 
        ann = self.make_annotator()
1445
 
        ann._expand_record(('parent-id',), (), None,
1446
 
                           ['line1\n', 'line2\n'], ('fulltext', False))
1447
 
        ann._num_compression_children['parent-id'] = 2
1448
 
 
1449
 
    def test_annotate_special_text(self):
1450
 
        ann = self.make_annotator()
1451
 
        vf = ann._vf
1452
 
        rev1_key = ('rev-1',)
1453
 
        rev2_key = ('rev-2',)
1454
 
        rev3_key = ('rev-3',)
1455
 
        spec_key = ('special:',)
1456
 
        vf.add_lines(rev1_key, [], ['initial content\n'])
1457
 
        vf.add_lines(rev2_key, [rev1_key], ['initial content\n',
1458
 
                                            'common content\n',
1459
 
                                            'content in 2\n'])
1460
 
        vf.add_lines(rev3_key, [rev1_key], ['initial content\n',
1461
 
                                            'common content\n',
1462
 
                                            'content in 3\n'])
1463
 
        spec_text = ('initial content\n'
1464
 
                     'common content\n'
1465
 
                     'content in 2\n'
1466
 
                     'content in 3\n')
1467
 
        ann.add_special_text(spec_key, [rev2_key, rev3_key], spec_text)
1468
 
        anns, lines = ann.annotate(spec_key)
1469
 
        self.assertEqual([(rev1_key,),
1470
 
                          (rev2_key, rev3_key),
1471
 
                          (rev2_key,),
1472
 
                          (rev3_key,),
1473
 
                         ], anns)
1474
 
        self.assertEqualDiff(spec_text, ''.join(lines))
 
852
        return _KnitIndex(*args, **kwargs)
 
853
 
1475
854
 
1476
855
 
1477
856
class KnitTests(TestCaseWithTransport):
1478
857
    """Class containing knit test helper routines."""
1479
858
 
1480
 
    def make_test_knit(self, annotate=False, name='test'):
1481
 
        mapper = ConstantMapper(name)
1482
 
        return make_file_factory(annotate, mapper)(self.get_transport())
1483
 
 
1484
 
 
1485
 
class TestBadShaError(KnitTests):
1486
 
    """Tests for handling of sha errors."""
1487
 
 
1488
 
    def test_sha_exception_has_text(self):
1489
 
        # having the failed text included in the error allows for recovery.
1490
 
        source = self.make_test_knit()
1491
 
        target = self.make_test_knit(name="target")
1492
 
        if not source._max_delta_chain:
1493
 
            raise TestNotApplicable(
1494
 
                "cannot get delta-caused sha failures without deltas.")
1495
 
        # create a basis
1496
 
        basis = ('basis',)
1497
 
        broken = ('broken',)
1498
 
        source.add_lines(basis, (), ['foo\n'])
1499
 
        source.add_lines(broken, (basis,), ['foo\n', 'bar\n'])
1500
 
        # Seed target with a bad basis text
1501
 
        target.add_lines(basis, (), ['gam\n'])
1502
 
        target.insert_record_stream(
1503
 
            source.get_record_stream([broken], 'unordered', False))
1504
 
        err = self.assertRaises(errors.KnitCorrupt,
1505
 
            target.get_record_stream([broken], 'unordered', True
1506
 
            ).next().get_bytes_as, 'chunked')
1507
 
        self.assertEqual(['gam\n', 'bar\n'], err.content)
1508
 
        # Test for formatting with live data
1509
 
        self.assertStartsWith(str(err), "Knit ")
 
859
    def make_test_knit(self, annotate=False, delay_create=False, index=None):
 
860
        if not annotate:
 
861
            factory = KnitPlainFactory()
 
862
        else:
 
863
            factory = None
 
864
        return KnitVersionedFile('test', get_transport('.'), access_mode='w',
 
865
                                 factory=factory, create=True,
 
866
                                 delay_create=delay_create, index=index)
 
867
 
 
868
 
 
869
class BasicKnitTests(KnitTests):
 
870
 
 
871
    def add_stock_one_and_one_a(self, k):
 
872
        k.add_lines('text-1', [], split_lines(TEXT_1))
 
873
        k.add_lines('text-1a', ['text-1'], split_lines(TEXT_1A))
 
874
 
 
875
    def test_knit_constructor(self):
 
876
        """Construct empty k"""
 
877
        self.make_test_knit()
 
878
 
 
879
    def test_make_explicit_index(self):
 
880
        """We can supply an index to use."""
 
881
        knit = KnitVersionedFile('test', get_transport('.'),
 
882
            index='strangelove')
 
883
        self.assertEqual(knit._index, 'strangelove')
 
884
 
 
885
    def test_knit_add(self):
 
886
        """Store one text in knit and retrieve"""
 
887
        k = self.make_test_knit()
 
888
        k.add_lines('text-1', [], split_lines(TEXT_1))
 
889
        self.assertTrue(k.has_version('text-1'))
 
890
        self.assertEqualDiff(''.join(k.get_lines('text-1')), TEXT_1)
 
891
 
 
892
    def test_knit_reload(self):
 
893
        # test that the content in a reloaded knit is correct
 
894
        k = self.make_test_knit()
 
895
        k.add_lines('text-1', [], split_lines(TEXT_1))
 
896
        del k
 
897
        k2 = KnitVersionedFile('test', get_transport('.'), access_mode='r', factory=KnitPlainFactory(), create=True)
 
898
        self.assertTrue(k2.has_version('text-1'))
 
899
        self.assertEqualDiff(''.join(k2.get_lines('text-1')), TEXT_1)
 
900
 
 
901
    def test_knit_several(self):
 
902
        """Store several texts in a knit"""
 
903
        k = self.make_test_knit()
 
904
        k.add_lines('text-1', [], split_lines(TEXT_1))
 
905
        k.add_lines('text-2', [], split_lines(TEXT_2))
 
906
        self.assertEqualDiff(''.join(k.get_lines('text-1')), TEXT_1)
 
907
        self.assertEqualDiff(''.join(k.get_lines('text-2')), TEXT_2)
 
908
        
 
909
    def test_repeated_add(self):
 
910
        """Knit traps attempt to replace existing version"""
 
911
        k = self.make_test_knit()
 
912
        k.add_lines('text-1', [], split_lines(TEXT_1))
 
913
        self.assertRaises(RevisionAlreadyPresent, 
 
914
                k.add_lines,
 
915
                'text-1', [], split_lines(TEXT_1))
 
916
 
 
917
    def test_empty(self):
 
918
        k = self.make_test_knit(True)
 
919
        k.add_lines('text-1', [], [])
 
920
        self.assertEquals(k.get_lines('text-1'), [])
 
921
 
 
922
    def test_incomplete(self):
 
923
        """Test if texts without a ending line-end can be inserted and
 
924
        extracted."""
 
925
        k = KnitVersionedFile('test', get_transport('.'), delta=False, create=True)
 
926
        k.add_lines('text-1', [], ['a\n',    'b'  ])
 
927
        k.add_lines('text-2', ['text-1'], ['a\rb\n', 'b\n'])
 
928
        # reopening ensures maximum room for confusion
 
929
        k = KnitVersionedFile('test', get_transport('.'), delta=False, create=True)
 
930
        self.assertEquals(k.get_lines('text-1'), ['a\n',    'b'  ])
 
931
        self.assertEquals(k.get_lines('text-2'), ['a\rb\n', 'b\n'])
 
932
 
 
933
    def test_delta(self):
 
934
        """Expression of knit delta as lines"""
 
935
        k = self.make_test_knit()
 
936
        KnitContent
 
937
        td = list(line_delta(TEXT_1.splitlines(True),
 
938
                             TEXT_1A.splitlines(True)))
 
939
        self.assertEqualDiff(''.join(td), delta_1_1a)
 
940
        out = apply_line_delta(TEXT_1.splitlines(True), td)
 
941
        self.assertEqualDiff(''.join(out), TEXT_1A)
 
942
 
 
943
    def assertDerivedBlocksEqual(self, source, target, noeol=False):
 
944
        """Assert that the derived matching blocks match real output"""
 
945
        source_lines = source.splitlines(True)
 
946
        target_lines = target.splitlines(True)
 
947
        def nl(line):
 
948
            if noeol and not line.endswith('\n'):
 
949
                return line + '\n'
 
950
            else:
 
951
                return line
 
952
        source_content = KnitContent([(None, nl(l)) for l in source_lines])
 
953
        target_content = KnitContent([(None, nl(l)) for l in target_lines])
 
954
        line_delta = source_content.line_delta(target_content)
 
955
        delta_blocks = list(KnitContent.get_line_delta_blocks(line_delta,
 
956
            source_lines, target_lines))
 
957
        matcher = KnitSequenceMatcher(None, source_lines, target_lines)
 
958
        matcher_blocks = list(list(matcher.get_matching_blocks()))
 
959
        self.assertEqual(matcher_blocks, delta_blocks)
 
960
 
 
961
    def test_get_line_delta_blocks(self):
 
962
        self.assertDerivedBlocksEqual('a\nb\nc\n', 'q\nc\n')
 
963
        self.assertDerivedBlocksEqual(TEXT_1, TEXT_1)
 
964
        self.assertDerivedBlocksEqual(TEXT_1, TEXT_1A)
 
965
        self.assertDerivedBlocksEqual(TEXT_1, TEXT_1B)
 
966
        self.assertDerivedBlocksEqual(TEXT_1B, TEXT_1A)
 
967
        self.assertDerivedBlocksEqual(TEXT_1A, TEXT_1B)
 
968
        self.assertDerivedBlocksEqual(TEXT_1A, '')
 
969
        self.assertDerivedBlocksEqual('', TEXT_1A)
 
970
        self.assertDerivedBlocksEqual('', '')
 
971
        self.assertDerivedBlocksEqual('a\nb\nc', 'a\nb\nc\nd')
 
972
 
 
973
    def test_get_line_delta_blocks_noeol(self):
 
974
        """Handle historical knit deltas safely
 
975
 
 
976
        Some existing knit deltas don't consider the last line to differ
 
977
        when the only difference whether it has a final newline.
 
978
 
 
979
        New knit deltas appear to always consider the last line to differ
 
980
        in this case.
 
981
        """
 
982
        self.assertDerivedBlocksEqual('a\nb\nc', 'a\nb\nc\nd\n', noeol=True)
 
983
        self.assertDerivedBlocksEqual('a\nb\nc\nd\n', 'a\nb\nc', noeol=True)
 
984
        self.assertDerivedBlocksEqual('a\nb\nc\n', 'a\nb\nc', noeol=True)
 
985
        self.assertDerivedBlocksEqual('a\nb\nc', 'a\nb\nc\n', noeol=True)
 
986
 
 
987
    def test_add_with_parents(self):
 
988
        """Store in knit with parents"""
 
989
        k = self.make_test_knit()
 
990
        self.add_stock_one_and_one_a(k)
 
991
        self.assertEquals(k.get_parents('text-1'), [])
 
992
        self.assertEquals(k.get_parents('text-1a'), ['text-1'])
 
993
 
 
994
    def test_ancestry(self):
 
995
        """Store in knit with parents"""
 
996
        k = self.make_test_knit()
 
997
        self.add_stock_one_and_one_a(k)
 
998
        self.assertEquals(set(k.get_ancestry(['text-1a'])), set(['text-1a', 'text-1']))
 
999
 
 
1000
    def test_add_delta(self):
 
1001
        """Store in knit with parents"""
 
1002
        k = KnitVersionedFile('test', get_transport('.'), factory=KnitPlainFactory(),
 
1003
            delta=True, create=True)
 
1004
        self.add_stock_one_and_one_a(k)
 
1005
        k.clear_cache()
 
1006
        self.assertEqualDiff(''.join(k.get_lines('text-1a')), TEXT_1A)
 
1007
 
 
1008
    def test_add_delta_knit_graph_index(self):
 
1009
        """Does adding work with a KnitGraphIndex."""
 
1010
        index = InMemoryGraphIndex(2)
 
1011
        knit_index = KnitGraphIndex(index, add_callback=index.add_nodes,
 
1012
            deltas=True)
 
1013
        k = KnitVersionedFile('test', get_transport('.'),
 
1014
            delta=True, create=True, index=knit_index)
 
1015
        self.add_stock_one_and_one_a(k)
 
1016
        k.clear_cache()
 
1017
        self.assertEqualDiff(''.join(k.get_lines('text-1a')), TEXT_1A)
 
1018
        # check the index had the right data added.
 
1019
        self.assertEqual(set([
 
1020
            (('text-1', ), ' 0 127', ((), ())),
 
1021
            (('text-1a', ), ' 127 140', ((('text-1', ),), (('text-1', ),))),
 
1022
            ]), set(index.iter_all_entries()))
 
1023
        # we should not have a .kndx file
 
1024
        self.assertFalse(get_transport('.').has('test.kndx'))
 
1025
 
 
1026
    def test_annotate(self):
 
1027
        """Annotations"""
 
1028
        k = KnitVersionedFile('knit', get_transport('.'), factory=KnitAnnotateFactory(),
 
1029
            delta=True, create=True)
 
1030
        self.insert_and_test_small_annotate(k)
 
1031
 
 
1032
    def insert_and_test_small_annotate(self, k):
 
1033
        """test annotation with k works correctly."""
 
1034
        k.add_lines('text-1', [], ['a\n', 'b\n'])
 
1035
        k.add_lines('text-2', ['text-1'], ['a\n', 'c\n'])
 
1036
 
 
1037
        origins = k.annotate('text-2')
 
1038
        self.assertEquals(origins[0], ('text-1', 'a\n'))
 
1039
        self.assertEquals(origins[1], ('text-2', 'c\n'))
 
1040
 
 
1041
    def test_annotate_fulltext(self):
 
1042
        """Annotations"""
 
1043
        k = KnitVersionedFile('knit', get_transport('.'), factory=KnitAnnotateFactory(),
 
1044
            delta=False, create=True)
 
1045
        self.insert_and_test_small_annotate(k)
 
1046
 
 
1047
    def test_annotate_merge_1(self):
 
1048
        k = self.make_test_knit(True)
 
1049
        k.add_lines('text-a1', [], ['a\n', 'b\n'])
 
1050
        k.add_lines('text-a2', [], ['d\n', 'c\n'])
 
1051
        k.add_lines('text-am', ['text-a1', 'text-a2'], ['d\n', 'b\n'])
 
1052
        origins = k.annotate('text-am')
 
1053
        self.assertEquals(origins[0], ('text-a2', 'd\n'))
 
1054
        self.assertEquals(origins[1], ('text-a1', 'b\n'))
 
1055
 
 
1056
    def test_annotate_merge_2(self):
 
1057
        k = self.make_test_knit(True)
 
1058
        k.add_lines('text-a1', [], ['a\n', 'b\n', 'c\n'])
 
1059
        k.add_lines('text-a2', [], ['x\n', 'y\n', 'z\n'])
 
1060
        k.add_lines('text-am', ['text-a1', 'text-a2'], ['a\n', 'y\n', 'c\n'])
 
1061
        origins = k.annotate('text-am')
 
1062
        self.assertEquals(origins[0], ('text-a1', 'a\n'))
 
1063
        self.assertEquals(origins[1], ('text-a2', 'y\n'))
 
1064
        self.assertEquals(origins[2], ('text-a1', 'c\n'))
 
1065
 
 
1066
    def test_annotate_merge_9(self):
 
1067
        k = self.make_test_knit(True)
 
1068
        k.add_lines('text-a1', [], ['a\n', 'b\n', 'c\n'])
 
1069
        k.add_lines('text-a2', [], ['x\n', 'y\n', 'z\n'])
 
1070
        k.add_lines('text-am', ['text-a1', 'text-a2'], ['k\n', 'y\n', 'c\n'])
 
1071
        origins = k.annotate('text-am')
 
1072
        self.assertEquals(origins[0], ('text-am', 'k\n'))
 
1073
        self.assertEquals(origins[1], ('text-a2', 'y\n'))
 
1074
        self.assertEquals(origins[2], ('text-a1', 'c\n'))
 
1075
 
 
1076
    def test_annotate_merge_3(self):
 
1077
        k = self.make_test_knit(True)
 
1078
        k.add_lines('text-a1', [], ['a\n', 'b\n', 'c\n'])
 
1079
        k.add_lines('text-a2', [] ,['x\n', 'y\n', 'z\n'])
 
1080
        k.add_lines('text-am', ['text-a1', 'text-a2'], ['k\n', 'y\n', 'z\n'])
 
1081
        origins = k.annotate('text-am')
 
1082
        self.assertEquals(origins[0], ('text-am', 'k\n'))
 
1083
        self.assertEquals(origins[1], ('text-a2', 'y\n'))
 
1084
        self.assertEquals(origins[2], ('text-a2', 'z\n'))
 
1085
 
 
1086
    def test_annotate_merge_4(self):
 
1087
        k = self.make_test_knit(True)
 
1088
        k.add_lines('text-a1', [], ['a\n', 'b\n', 'c\n'])
 
1089
        k.add_lines('text-a2', [], ['x\n', 'y\n', 'z\n'])
 
1090
        k.add_lines('text-a3', ['text-a1'], ['a\n', 'b\n', 'p\n'])
 
1091
        k.add_lines('text-am', ['text-a2', 'text-a3'], ['a\n', 'b\n', 'z\n'])
 
1092
        origins = k.annotate('text-am')
 
1093
        self.assertEquals(origins[0], ('text-a1', 'a\n'))
 
1094
        self.assertEquals(origins[1], ('text-a1', 'b\n'))
 
1095
        self.assertEquals(origins[2], ('text-a2', 'z\n'))
 
1096
 
 
1097
    def test_annotate_merge_5(self):
 
1098
        k = self.make_test_knit(True)
 
1099
        k.add_lines('text-a1', [], ['a\n', 'b\n', 'c\n'])
 
1100
        k.add_lines('text-a2', [], ['d\n', 'e\n', 'f\n'])
 
1101
        k.add_lines('text-a3', [], ['x\n', 'y\n', 'z\n'])
 
1102
        k.add_lines('text-am',
 
1103
                    ['text-a1', 'text-a2', 'text-a3'],
 
1104
                    ['a\n', 'e\n', 'z\n'])
 
1105
        origins = k.annotate('text-am')
 
1106
        self.assertEquals(origins[0], ('text-a1', 'a\n'))
 
1107
        self.assertEquals(origins[1], ('text-a2', 'e\n'))
 
1108
        self.assertEquals(origins[2], ('text-a3', 'z\n'))
 
1109
 
 
1110
    def test_annotate_file_cherry_pick(self):
 
1111
        k = self.make_test_knit(True)
 
1112
        k.add_lines('text-1', [], ['a\n', 'b\n', 'c\n'])
 
1113
        k.add_lines('text-2', ['text-1'], ['d\n', 'e\n', 'f\n'])
 
1114
        k.add_lines('text-3', ['text-2', 'text-1'], ['a\n', 'b\n', 'c\n'])
 
1115
        origins = k.annotate('text-3')
 
1116
        self.assertEquals(origins[0], ('text-1', 'a\n'))
 
1117
        self.assertEquals(origins[1], ('text-1', 'b\n'))
 
1118
        self.assertEquals(origins[2], ('text-1', 'c\n'))
 
1119
 
 
1120
    def test_knit_join(self):
 
1121
        """Store in knit with parents"""
 
1122
        k1 = KnitVersionedFile('test1', get_transport('.'), factory=KnitPlainFactory(), create=True)
 
1123
        k1.add_lines('text-a', [], split_lines(TEXT_1))
 
1124
        k1.add_lines('text-b', ['text-a'], split_lines(TEXT_1))
 
1125
 
 
1126
        k1.add_lines('text-c', [], split_lines(TEXT_1))
 
1127
        k1.add_lines('text-d', ['text-c'], split_lines(TEXT_1))
 
1128
 
 
1129
        k1.add_lines('text-m', ['text-b', 'text-d'], split_lines(TEXT_1))
 
1130
 
 
1131
        k2 = KnitVersionedFile('test2', get_transport('.'), factory=KnitPlainFactory(), create=True)
 
1132
        count = k2.join(k1, version_ids=['text-m'])
 
1133
        self.assertEquals(count, 5)
 
1134
        self.assertTrue(k2.has_version('text-a'))
 
1135
        self.assertTrue(k2.has_version('text-c'))
 
1136
 
 
1137
    def test_reannotate(self):
 
1138
        k1 = KnitVersionedFile('knit1', get_transport('.'),
 
1139
                               factory=KnitAnnotateFactory(), create=True)
 
1140
        # 0
 
1141
        k1.add_lines('text-a', [], ['a\n', 'b\n'])
 
1142
        # 1
 
1143
        k1.add_lines('text-b', ['text-a'], ['a\n', 'c\n'])
 
1144
 
 
1145
        k2 = KnitVersionedFile('test2', get_transport('.'),
 
1146
                               factory=KnitAnnotateFactory(), create=True)
 
1147
        k2.join(k1, version_ids=['text-b'])
 
1148
 
 
1149
        # 2
 
1150
        k1.add_lines('text-X', ['text-b'], ['a\n', 'b\n'])
 
1151
        # 2
 
1152
        k2.add_lines('text-c', ['text-b'], ['z\n', 'c\n'])
 
1153
        # 3
 
1154
        k2.add_lines('text-Y', ['text-b'], ['b\n', 'c\n'])
 
1155
 
 
1156
        # test-c will have index 3
 
1157
        k1.join(k2, version_ids=['text-c'])
 
1158
 
 
1159
        lines = k1.get_lines('text-c')
 
1160
        self.assertEquals(lines, ['z\n', 'c\n'])
 
1161
 
 
1162
        origins = k1.annotate('text-c')
 
1163
        self.assertEquals(origins[0], ('text-c', 'z\n'))
 
1164
        self.assertEquals(origins[1], ('text-b', 'c\n'))
 
1165
 
 
1166
    def test_get_line_delta_texts(self):
 
1167
        """Make sure we can call get_texts on text with reused line deltas"""
 
1168
        k1 = KnitVersionedFile('test1', get_transport('.'), 
 
1169
                               factory=KnitPlainFactory(), create=True)
 
1170
        for t in range(3):
 
1171
            if t == 0:
 
1172
                parents = []
 
1173
            else:
 
1174
                parents = ['%d' % (t-1)]
 
1175
            k1.add_lines('%d' % t, parents, ['hello\n'] * t)
 
1176
        k1.get_texts(('%d' % t) for t in range(3))
 
1177
        
 
1178
    def test_iter_lines_reads_in_order(self):
 
1179
        t = MemoryTransport()
 
1180
        instrumented_t = TransportLogger(t)
 
1181
        k1 = KnitVersionedFile('id', instrumented_t, create=True, delta=True)
 
1182
        self.assertEqual([('id.kndx',)], instrumented_t._calls)
 
1183
        # add texts with no required ordering
 
1184
        k1.add_lines('base', [], ['text\n'])
 
1185
        k1.add_lines('base2', [], ['text2\n'])
 
1186
        k1.clear_cache()
 
1187
        instrumented_t._calls = []
 
1188
        # request a last-first iteration
 
1189
        results = list(k1.iter_lines_added_or_present_in_versions(['base2', 'base']))
 
1190
        self.assertEqual([('id.knit', [(0, 87), (87, 89)])], instrumented_t._calls)
 
1191
        self.assertEqual(['text\n', 'text2\n'], results)
 
1192
 
 
1193
    def test_create_empty_annotated(self):
 
1194
        k1 = self.make_test_knit(True)
 
1195
        # 0
 
1196
        k1.add_lines('text-a', [], ['a\n', 'b\n'])
 
1197
        k2 = k1.create_empty('t', MemoryTransport())
 
1198
        self.assertTrue(isinstance(k2.factory, KnitAnnotateFactory))
 
1199
        self.assertEqual(k1.delta, k2.delta)
 
1200
        # the generic test checks for empty content and file class
 
1201
 
 
1202
    def test_knit_format(self):
 
1203
        # this tests that a new knit index file has the expected content
 
1204
        # and that is writes the data we expect as records are added.
 
1205
        knit = self.make_test_knit(True)
 
1206
        # Now knit files are not created until we first add data to them
 
1207
        self.assertFileEqual("# bzr knit index 8\n", 'test.kndx')
 
1208
        knit.add_lines_with_ghosts('revid', ['a_ghost'], ['a\n'])
 
1209
        self.assertFileEqual(
 
1210
            "# bzr knit index 8\n"
 
1211
            "\n"
 
1212
            "revid fulltext 0 84 .a_ghost :",
 
1213
            'test.kndx')
 
1214
        knit.add_lines_with_ghosts('revid2', ['revid'], ['a\n'])
 
1215
        self.assertFileEqual(
 
1216
            "# bzr knit index 8\n"
 
1217
            "\nrevid fulltext 0 84 .a_ghost :"
 
1218
            "\nrevid2 line-delta 84 82 0 :",
 
1219
            'test.kndx')
 
1220
        # we should be able to load this file again
 
1221
        knit = KnitVersionedFile('test', get_transport('.'), access_mode='r')
 
1222
        self.assertEqual(['revid', 'revid2'], knit.versions())
 
1223
        # write a short write to the file and ensure that its ignored
 
1224
        indexfile = file('test.kndx', 'ab')
 
1225
        indexfile.write('\nrevid3 line-delta 166 82 1 2 3 4 5 .phwoar:demo ')
 
1226
        indexfile.close()
 
1227
        # we should be able to load this file again
 
1228
        knit = KnitVersionedFile('test', get_transport('.'), access_mode='w')
 
1229
        self.assertEqual(['revid', 'revid2'], knit.versions())
 
1230
        # and add a revision with the same id the failed write had
 
1231
        knit.add_lines('revid3', ['revid2'], ['a\n'])
 
1232
        # and when reading it revid3 should now appear.
 
1233
        knit = KnitVersionedFile('test', get_transport('.'), access_mode='r')
 
1234
        self.assertEqual(['revid', 'revid2', 'revid3'], knit.versions())
 
1235
        self.assertEqual(['revid2'], knit.get_parents('revid3'))
 
1236
 
 
1237
    def test_delay_create(self):
 
1238
        """Test that passing delay_create=True creates files late"""
 
1239
        knit = self.make_test_knit(annotate=True, delay_create=True)
 
1240
        self.failIfExists('test.knit')
 
1241
        self.failIfExists('test.kndx')
 
1242
        knit.add_lines_with_ghosts('revid', ['a_ghost'], ['a\n'])
 
1243
        self.failUnlessExists('test.knit')
 
1244
        self.assertFileEqual(
 
1245
            "# bzr knit index 8\n"
 
1246
            "\n"
 
1247
            "revid fulltext 0 84 .a_ghost :",
 
1248
            'test.kndx')
 
1249
 
 
1250
    def test_create_parent_dir(self):
 
1251
        """create_parent_dir can create knits in nonexistant dirs"""
 
1252
        # Has no effect if we don't set 'delay_create'
 
1253
        trans = get_transport('.')
 
1254
        self.assertRaises(NoSuchFile, KnitVersionedFile, 'dir/test',
 
1255
                          trans, access_mode='w', factory=None,
 
1256
                          create=True, create_parent_dir=True)
 
1257
        # Nothing should have changed yet
 
1258
        knit = KnitVersionedFile('dir/test', trans, access_mode='w',
 
1259
                                 factory=None, create=True,
 
1260
                                 create_parent_dir=True,
 
1261
                                 delay_create=True)
 
1262
        self.failIfExists('dir/test.knit')
 
1263
        self.failIfExists('dir/test.kndx')
 
1264
        self.failIfExists('dir')
 
1265
        knit.add_lines('revid', [], ['a\n'])
 
1266
        self.failUnlessExists('dir')
 
1267
        self.failUnlessExists('dir/test.knit')
 
1268
        self.assertFileEqual(
 
1269
            "# bzr knit index 8\n"
 
1270
            "\n"
 
1271
            "revid fulltext 0 84  :",
 
1272
            'dir/test.kndx')
 
1273
 
 
1274
    def test_create_mode_700(self):
 
1275
        trans = get_transport('.')
 
1276
        if not trans._can_roundtrip_unix_modebits():
 
1277
            # Can't roundtrip, so no need to run this test
 
1278
            return
 
1279
        knit = KnitVersionedFile('dir/test', trans, access_mode='w',
 
1280
                                 factory=None, create=True,
 
1281
                                 create_parent_dir=True,
 
1282
                                 delay_create=True,
 
1283
                                 file_mode=0600,
 
1284
                                 dir_mode=0700)
 
1285
        knit.add_lines('revid', [], ['a\n'])
 
1286
        self.assertTransportMode(trans, 'dir', 0700)
 
1287
        self.assertTransportMode(trans, 'dir/test.knit', 0600)
 
1288
        self.assertTransportMode(trans, 'dir/test.kndx', 0600)
 
1289
 
 
1290
    def test_create_mode_770(self):
 
1291
        trans = get_transport('.')
 
1292
        if not trans._can_roundtrip_unix_modebits():
 
1293
            # Can't roundtrip, so no need to run this test
 
1294
            return
 
1295
        knit = KnitVersionedFile('dir/test', trans, access_mode='w',
 
1296
                                 factory=None, create=True,
 
1297
                                 create_parent_dir=True,
 
1298
                                 delay_create=True,
 
1299
                                 file_mode=0660,
 
1300
                                 dir_mode=0770)
 
1301
        knit.add_lines('revid', [], ['a\n'])
 
1302
        self.assertTransportMode(trans, 'dir', 0770)
 
1303
        self.assertTransportMode(trans, 'dir/test.knit', 0660)
 
1304
        self.assertTransportMode(trans, 'dir/test.kndx', 0660)
 
1305
 
 
1306
    def test_create_mode_777(self):
 
1307
        trans = get_transport('.')
 
1308
        if not trans._can_roundtrip_unix_modebits():
 
1309
            # Can't roundtrip, so no need to run this test
 
1310
            return
 
1311
        knit = KnitVersionedFile('dir/test', trans, access_mode='w',
 
1312
                                 factory=None, create=True,
 
1313
                                 create_parent_dir=True,
 
1314
                                 delay_create=True,
 
1315
                                 file_mode=0666,
 
1316
                                 dir_mode=0777)
 
1317
        knit.add_lines('revid', [], ['a\n'])
 
1318
        self.assertTransportMode(trans, 'dir', 0777)
 
1319
        self.assertTransportMode(trans, 'dir/test.knit', 0666)
 
1320
        self.assertTransportMode(trans, 'dir/test.kndx', 0666)
 
1321
 
 
1322
    def test_plan_merge(self):
 
1323
        my_knit = self.make_test_knit(annotate=True)
 
1324
        my_knit.add_lines('text1', [], split_lines(TEXT_1))
 
1325
        my_knit.add_lines('text1a', ['text1'], split_lines(TEXT_1A))
 
1326
        my_knit.add_lines('text1b', ['text1'], split_lines(TEXT_1B))
 
1327
        plan = list(my_knit.plan_merge('text1a', 'text1b'))
 
1328
        for plan_line, expected_line in zip(plan, AB_MERGE):
 
1329
            self.assertEqual(plan_line, expected_line)
 
1330
 
 
1331
 
 
1332
TEXT_1 = """\
 
1333
Banana cup cakes:
 
1334
 
 
1335
- bananas
 
1336
- eggs
 
1337
- broken tea cups
 
1338
"""
 
1339
 
 
1340
TEXT_1A = """\
 
1341
Banana cup cake recipe
 
1342
(serves 6)
 
1343
 
 
1344
- bananas
 
1345
- eggs
 
1346
- broken tea cups
 
1347
- self-raising flour
 
1348
"""
 
1349
 
 
1350
TEXT_1B = """\
 
1351
Banana cup cake recipe
 
1352
 
 
1353
- bananas (do not use plantains!!!)
 
1354
- broken tea cups
 
1355
- flour
 
1356
"""
 
1357
 
 
1358
delta_1_1a = """\
 
1359
0,1,2
 
1360
Banana cup cake recipe
 
1361
(serves 6)
 
1362
5,5,1
 
1363
- self-raising flour
 
1364
"""
 
1365
 
 
1366
TEXT_2 = """\
 
1367
Boeuf bourguignon
 
1368
 
 
1369
- beef
 
1370
- red wine
 
1371
- small onions
 
1372
- carrot
 
1373
- mushrooms
 
1374
"""
 
1375
 
 
1376
AB_MERGE_TEXT="""unchanged|Banana cup cake recipe
 
1377
new-a|(serves 6)
 
1378
unchanged|
 
1379
killed-b|- bananas
 
1380
killed-b|- eggs
 
1381
new-b|- bananas (do not use plantains!!!)
 
1382
unchanged|- broken tea cups
 
1383
new-a|- self-raising flour
 
1384
new-b|- flour
 
1385
"""
 
1386
AB_MERGE=[tuple(l.split('|')) for l in AB_MERGE_TEXT.splitlines(True)]
 
1387
 
 
1388
 
 
1389
def line_delta(from_lines, to_lines):
 
1390
    """Generate line-based delta from one text to another"""
 
1391
    s = difflib.SequenceMatcher(None, from_lines, to_lines)
 
1392
    for op in s.get_opcodes():
 
1393
        if op[0] == 'equal':
 
1394
            continue
 
1395
        yield '%d,%d,%d\n' % (op[1], op[2], op[4]-op[3])
 
1396
        for i in range(op[3], op[4]):
 
1397
            yield to_lines[i]
 
1398
 
 
1399
 
 
1400
def apply_line_delta(basis_lines, delta_lines):
 
1401
    """Apply a line-based perfect diff
 
1402
    
 
1403
    basis_lines -- text to apply the patch to
 
1404
    delta_lines -- diff instructions and content
 
1405
    """
 
1406
    out = basis_lines[:]
 
1407
    i = 0
 
1408
    offset = 0
 
1409
    while i < len(delta_lines):
 
1410
        l = delta_lines[i]
 
1411
        a, b, c = map(long, l.split(','))
 
1412
        i = i + 1
 
1413
        out[offset+a:offset+b] = delta_lines[i:i+c]
 
1414
        i = i + c
 
1415
        offset = offset + (b - a) + c
 
1416
    return out
 
1417
 
 
1418
 
 
1419
class TestWeaveToKnit(KnitTests):
 
1420
 
 
1421
    def test_weave_to_knit_matches(self):
 
1422
        # check that the WeaveToKnit is_compatible function
 
1423
        # registers True for a Weave to a Knit.
 
1424
        w = Weave()
 
1425
        k = self.make_test_knit()
 
1426
        self.failUnless(WeaveToKnit.is_compatible(w, k))
 
1427
        self.failIf(WeaveToKnit.is_compatible(k, w))
 
1428
        self.failIf(WeaveToKnit.is_compatible(w, w))
 
1429
        self.failIf(WeaveToKnit.is_compatible(k, k))
 
1430
 
 
1431
 
 
1432
class TestKnitCaching(KnitTests):
 
1433
    
 
1434
    def create_knit(self, cache_add=False):
 
1435
        k = self.make_test_knit(True)
 
1436
        if cache_add:
 
1437
            k.enable_cache()
 
1438
 
 
1439
        k.add_lines('text-1', [], split_lines(TEXT_1))
 
1440
        k.add_lines('text-2', [], split_lines(TEXT_2))
 
1441
        return k
 
1442
 
 
1443
    def test_no_caching(self):
 
1444
        k = self.create_knit()
 
1445
        # Nothing should be cached without setting 'enable_cache'
 
1446
        self.assertEqual({}, k._data._cache)
 
1447
 
 
1448
    def test_cache_add_and_clear(self):
 
1449
        k = self.create_knit(True)
 
1450
 
 
1451
        self.assertEqual(['text-1', 'text-2'], sorted(k._data._cache.keys()))
 
1452
 
 
1453
        k.clear_cache()
 
1454
        self.assertEqual({}, k._data._cache)
 
1455
 
 
1456
    def test_cache_data_read_raw(self):
 
1457
        k = self.create_knit()
 
1458
 
 
1459
        # Now cache and read
 
1460
        k.enable_cache()
 
1461
 
 
1462
        def read_one_raw(version):
 
1463
            pos_map = k._get_components_positions([version])
 
1464
            method, pos, size, next = pos_map[version]
 
1465
            lst = list(k._data.read_records_iter_raw([(version, pos, size)]))
 
1466
            self.assertEqual(1, len(lst))
 
1467
            return lst[0]
 
1468
 
 
1469
        val = read_one_raw('text-1')
 
1470
        self.assertEqual({'text-1':val[1]}, k._data._cache)
 
1471
 
 
1472
        k.clear_cache()
 
1473
        # After clear, new reads are not cached
 
1474
        self.assertEqual({}, k._data._cache)
 
1475
 
 
1476
        val2 = read_one_raw('text-1')
 
1477
        self.assertEqual(val, val2)
 
1478
        self.assertEqual({}, k._data._cache)
 
1479
 
 
1480
    def test_cache_data_read(self):
 
1481
        k = self.create_knit()
 
1482
 
 
1483
        def read_one(version):
 
1484
            pos_map = k._get_components_positions([version])
 
1485
            method, pos, size, next = pos_map[version]
 
1486
            lst = list(k._data.read_records_iter([(version, pos, size)]))
 
1487
            self.assertEqual(1, len(lst))
 
1488
            return lst[0]
 
1489
 
 
1490
        # Now cache and read
 
1491
        k.enable_cache()
 
1492
 
 
1493
        val = read_one('text-2')
 
1494
        self.assertEqual(['text-2'], k._data._cache.keys())
 
1495
        self.assertEqual('text-2', val[0])
 
1496
        content, digest = k._data._parse_record('text-2',
 
1497
                                                k._data._cache['text-2'])
 
1498
        self.assertEqual(content, val[1])
 
1499
        self.assertEqual(digest, val[2])
 
1500
 
 
1501
        k.clear_cache()
 
1502
        self.assertEqual({}, k._data._cache)
 
1503
 
 
1504
        val2 = read_one('text-2')
 
1505
        self.assertEqual(val, val2)
 
1506
        self.assertEqual({}, k._data._cache)
 
1507
 
 
1508
    def test_cache_read(self):
 
1509
        k = self.create_knit()
 
1510
        k.enable_cache()
 
1511
 
 
1512
        text = k.get_text('text-1')
 
1513
        self.assertEqual(TEXT_1, text)
 
1514
        self.assertEqual(['text-1'], k._data._cache.keys())
 
1515
 
 
1516
        k.clear_cache()
 
1517
        self.assertEqual({}, k._data._cache)
 
1518
 
 
1519
        text = k.get_text('text-1')
 
1520
        self.assertEqual(TEXT_1, text)
 
1521
        self.assertEqual({}, k._data._cache)
1510
1522
 
1511
1523
 
1512
1524
class TestKnitIndex(KnitTests):
1515
1527
        """Adding versions to the index should update the lookup dict"""
1516
1528
        knit = self.make_test_knit()
1517
1529
        idx = knit._index
1518
 
        idx.add_records([(('a-1',), ['fulltext'], (('a-1',), 0, 0), [])])
 
1530
        idx.add_version('a-1', ['fulltext'], 0, 0, [])
1519
1531
        self.check_file_contents('test.kndx',
1520
1532
            '# bzr knit index 8\n'
1521
1533
            '\n'
1522
1534
            'a-1 fulltext 0 0  :'
1523
1535
            )
1524
 
        idx.add_records([
1525
 
            (('a-2',), ['fulltext'], (('a-2',), 0, 0), [('a-1',)]),
1526
 
            (('a-3',), ['fulltext'], (('a-3',), 0, 0), [('a-2',)]),
1527
 
            ])
 
1536
        idx.add_versions([('a-2', ['fulltext'], 0, 0, ['a-1']),
 
1537
                          ('a-3', ['fulltext'], 0, 0, ['a-2']),
 
1538
                         ])
1528
1539
        self.check_file_contents('test.kndx',
1529
1540
            '# bzr knit index 8\n'
1530
1541
            '\n'
1532
1543
            'a-2 fulltext 0 0 0 :\n'
1533
1544
            'a-3 fulltext 0 0 1 :'
1534
1545
            )
1535
 
        self.assertEqual(set([('a-3',), ('a-1',), ('a-2',)]), idx.keys())
1536
 
        self.assertEqual({
1537
 
            ('a-1',): ((('a-1',), 0, 0), None, (), ('fulltext', False)),
1538
 
            ('a-2',): ((('a-2',), 0, 0), None, (('a-1',),), ('fulltext', False)),
1539
 
            ('a-3',): ((('a-3',), 0, 0), None, (('a-2',),), ('fulltext', False)),
1540
 
            }, idx.get_build_details(idx.keys()))
1541
 
        self.assertEqual({('a-1',):(),
1542
 
            ('a-2',):(('a-1',),),
1543
 
            ('a-3',):(('a-2',),),},
1544
 
            idx.get_parent_map(idx.keys()))
 
1546
        self.assertEqual(['a-1', 'a-2', 'a-3'], idx._history)
 
1547
        self.assertEqual({'a-1':('a-1', ['fulltext'], 0, 0, [], 0),
 
1548
                          'a-2':('a-2', ['fulltext'], 0, 0, ['a-1'], 1),
 
1549
                          'a-3':('a-3', ['fulltext'], 0, 0, ['a-2'], 2),
 
1550
                         }, idx._cache)
1545
1551
 
1546
1552
    def test_add_versions_fails_clean(self):
1547
1553
        """If add_versions fails in the middle, it restores a pristine state.
1557
1563
 
1558
1564
        knit = self.make_test_knit()
1559
1565
        idx = knit._index
1560
 
        idx.add_records([(('a-1',), ['fulltext'], (('a-1',), 0, 0), [])])
 
1566
        idx.add_version('a-1', ['fulltext'], 0, 0, [])
1561
1567
 
1562
1568
        class StopEarly(Exception):
1563
1569
            pass
1564
1570
 
1565
1571
        def generate_failure():
1566
1572
            """Add some entries and then raise an exception"""
1567
 
            yield (('a-2',), ['fulltext'], (None, 0, 0), ('a-1',))
1568
 
            yield (('a-3',), ['fulltext'], (None, 0, 0), ('a-2',))
 
1573
            yield ('a-2', ['fulltext'], 0, 0, ['a-1'])
 
1574
            yield ('a-3', ['fulltext'], 0, 0, ['a-2'])
1569
1575
            raise StopEarly()
1570
1576
 
1571
1577
        # Assert the pre-condition
1572
 
        def assertA1Only():
1573
 
            self.assertEqual(set([('a-1',)]), set(idx.keys()))
1574
 
            self.assertEqual(
1575
 
                {('a-1',): ((('a-1',), 0, 0), None, (), ('fulltext', False))},
1576
 
                idx.get_build_details([('a-1',)]))
1577
 
            self.assertEqual({('a-1',):()}, idx.get_parent_map(idx.keys()))
1578
 
 
1579
 
        assertA1Only()
1580
 
        self.assertRaises(StopEarly, idx.add_records, generate_failure())
 
1578
        self.assertEqual(['a-1'], idx._history)
 
1579
        self.assertEqual({'a-1':('a-1', ['fulltext'], 0, 0, [], 0)}, idx._cache)
 
1580
 
 
1581
        self.assertRaises(StopEarly, idx.add_versions, generate_failure())
 
1582
 
1581
1583
        # And it shouldn't be modified
1582
 
        assertA1Only()
 
1584
        self.assertEqual(['a-1'], idx._history)
 
1585
        self.assertEqual({'a-1':('a-1', ['fulltext'], 0, 0, [], 0)}, idx._cache)
1583
1586
 
1584
1587
    def test_knit_index_ignores_empty_files(self):
1585
1588
        # There was a race condition in older bzr, where a ^C at the right time
1594
1597
    def test_knit_index_checks_header(self):
1595
1598
        t = get_transport('.')
1596
1599
        t.put_bytes('test.kndx', '# not really a knit header\n\n')
1597
 
        k = self.make_test_knit()
1598
 
        self.assertRaises(KnitHeaderError, k.keys)
 
1600
 
 
1601
        self.assertRaises(KnitHeaderError, self.make_test_knit)
1599
1602
 
1600
1603
 
1601
1604
class TestGraphIndexKnit(KnitTests):
1607
1610
            builder.add_node(node, references, value)
1608
1611
        stream = builder.finish()
1609
1612
        trans = self.get_transport()
1610
 
        size = trans.put_file(name, stream)
1611
 
        return GraphIndex(trans, name, size)
 
1613
        trans.put_file(name, stream)
 
1614
        return GraphIndex(trans, name)
1612
1615
 
1613
1616
    def two_graph_index(self, deltas=False, catch_adds=False):
1614
1617
        """Build a two-graph index.
1640
1643
            add_callback = self.catch_add
1641
1644
        else:
1642
1645
            add_callback = None
1643
 
        return _KnitGraphIndex(combined_index, lambda:True, deltas=deltas,
 
1646
        return KnitGraphIndex(combined_index, deltas=deltas,
1644
1647
            add_callback=add_callback)
1645
1648
 
1646
 
    def test_keys(self):
1647
 
        index = self.two_graph_index()
1648
 
        self.assertEqual(set([('tail',), ('tip',), ('parent',), ('separate',)]),
1649
 
            set(index.keys()))
 
1649
    def test_get_graph(self):
 
1650
        index = self.two_graph_index()
 
1651
        self.assertEqual(set([
 
1652
            ('tip', ('parent', )),
 
1653
            ('tail', ()),
 
1654
            ('parent', ('tail', 'ghost')),
 
1655
            ('separate', ()),
 
1656
            ]), set(index.get_graph()))
 
1657
 
 
1658
    def test_get_ancestry(self):
 
1659
        # get_ancestry is defined as eliding ghosts, not erroring.
 
1660
        index = self.two_graph_index()
 
1661
        self.assertEqual([], index.get_ancestry([]))
 
1662
        self.assertEqual(['separate'], index.get_ancestry(['separate']))
 
1663
        self.assertEqual(['tail'], index.get_ancestry(['tail']))
 
1664
        self.assertEqual(['tail', 'parent'], index.get_ancestry(['parent']))
 
1665
        self.assertEqual(['tail', 'parent', 'tip'], index.get_ancestry(['tip']))
 
1666
        self.assertTrue(index.get_ancestry(['tip', 'separate']) in
 
1667
            (['tail', 'parent', 'tip', 'separate'],
 
1668
             ['separate', 'tail', 'parent', 'tip'],
 
1669
            ))
 
1670
        # and without topo_sort
 
1671
        self.assertEqual(set(['separate']),
 
1672
            set(index.get_ancestry(['separate'], topo_sorted=False)))
 
1673
        self.assertEqual(set(['tail']),
 
1674
            set(index.get_ancestry(['tail'], topo_sorted=False)))
 
1675
        self.assertEqual(set(['tail', 'parent']),
 
1676
            set(index.get_ancestry(['parent'], topo_sorted=False)))
 
1677
        self.assertEqual(set(['tail', 'parent', 'tip']),
 
1678
            set(index.get_ancestry(['tip'], topo_sorted=False)))
 
1679
        self.assertEqual(set(['separate', 'tail', 'parent', 'tip']),
 
1680
            set(index.get_ancestry(['tip', 'separate'])))
 
1681
        # asking for a ghost makes it go boom.
 
1682
        self.assertRaises(errors.RevisionNotPresent, index.get_ancestry, ['ghost'])
 
1683
 
 
1684
    def test_get_ancestry_with_ghosts(self):
 
1685
        index = self.two_graph_index()
 
1686
        self.assertEqual([], index.get_ancestry_with_ghosts([]))
 
1687
        self.assertEqual(['separate'], index.get_ancestry_with_ghosts(['separate']))
 
1688
        self.assertEqual(['tail'], index.get_ancestry_with_ghosts(['tail']))
 
1689
        self.assertTrue(index.get_ancestry_with_ghosts(['parent']) in
 
1690
            (['tail', 'ghost', 'parent'],
 
1691
             ['ghost', 'tail', 'parent'],
 
1692
            ))
 
1693
        self.assertTrue(index.get_ancestry_with_ghosts(['tip']) in
 
1694
            (['tail', 'ghost', 'parent', 'tip'],
 
1695
             ['ghost', 'tail', 'parent', 'tip'],
 
1696
            ))
 
1697
        self.assertTrue(index.get_ancestry_with_ghosts(['tip', 'separate']) in
 
1698
            (['tail', 'ghost', 'parent', 'tip', 'separate'],
 
1699
             ['ghost', 'tail', 'parent', 'tip', 'separate'],
 
1700
             ['separate', 'tail', 'ghost', 'parent', 'tip'],
 
1701
             ['separate', 'ghost', 'tail', 'parent', 'tip'],
 
1702
            ))
 
1703
        # asking for a ghost makes it go boom.
 
1704
        self.assertRaises(errors.RevisionNotPresent, index.get_ancestry_with_ghosts, ['ghost'])
 
1705
 
 
1706
    def test_num_versions(self):
 
1707
        index = self.two_graph_index()
 
1708
        self.assertEqual(4, index.num_versions())
 
1709
 
 
1710
    def test_get_versions(self):
 
1711
        index = self.two_graph_index()
 
1712
        self.assertEqual(set(['tail', 'tip', 'parent', 'separate']),
 
1713
            set(index.get_versions()))
 
1714
 
 
1715
    def test_has_version(self):
 
1716
        index = self.two_graph_index()
 
1717
        self.assertTrue(index.has_version('tail'))
 
1718
        self.assertFalse(index.has_version('ghost'))
1650
1719
 
1651
1720
    def test_get_position(self):
1652
1721
        index = self.two_graph_index()
1653
 
        self.assertEqual((index._graph_index._indices[0], 0, 100), index.get_position(('tip',)))
1654
 
        self.assertEqual((index._graph_index._indices[1], 100, 78), index.get_position(('parent',)))
 
1722
        self.assertEqual((0, 100), index.get_position('tip'))
 
1723
        self.assertEqual((100, 78), index.get_position('parent'))
1655
1724
 
1656
1725
    def test_get_method_deltas(self):
1657
1726
        index = self.two_graph_index(deltas=True)
1658
 
        self.assertEqual('fulltext', index.get_method(('tip',)))
1659
 
        self.assertEqual('line-delta', index.get_method(('parent',)))
 
1727
        self.assertEqual('fulltext', index.get_method('tip'))
 
1728
        self.assertEqual('line-delta', index.get_method('parent'))
1660
1729
 
1661
1730
    def test_get_method_no_deltas(self):
1662
1731
        # check that the parent-history lookup is ignored with deltas=False.
1663
1732
        index = self.two_graph_index(deltas=False)
1664
 
        self.assertEqual('fulltext', index.get_method(('tip',)))
1665
 
        self.assertEqual('fulltext', index.get_method(('parent',)))
 
1733
        self.assertEqual('fulltext', index.get_method('tip'))
 
1734
        self.assertEqual('fulltext', index.get_method('parent'))
1666
1735
 
1667
1736
    def test_get_options_deltas(self):
1668
1737
        index = self.two_graph_index(deltas=True)
1669
 
        self.assertEqual(['fulltext', 'no-eol'], index.get_options(('tip',)))
1670
 
        self.assertEqual(['line-delta'], index.get_options(('parent',)))
 
1738
        self.assertEqual(['fulltext', 'no-eol'], index.get_options('tip'))
 
1739
        self.assertEqual(['line-delta'], index.get_options('parent'))
1671
1740
 
1672
1741
    def test_get_options_no_deltas(self):
1673
1742
        # check that the parent-history lookup is ignored with deltas=False.
1674
1743
        index = self.two_graph_index(deltas=False)
1675
 
        self.assertEqual(['fulltext', 'no-eol'], index.get_options(('tip',)))
1676
 
        self.assertEqual(['fulltext'], index.get_options(('parent',)))
1677
 
 
1678
 
    def test_get_parent_map(self):
1679
 
        index = self.two_graph_index()
1680
 
        self.assertEqual({('parent',):(('tail',), ('ghost',))},
1681
 
            index.get_parent_map([('parent',), ('ghost',)]))
 
1744
        self.assertEqual(['fulltext', 'no-eol'], index.get_options('tip'))
 
1745
        self.assertEqual(['fulltext'], index.get_options('parent'))
 
1746
 
 
1747
    def test_get_parents(self):
 
1748
        # get_parents ignores ghosts
 
1749
        index = self.two_graph_index()
 
1750
        self.assertEqual(('tail', ), index.get_parents('parent'))
 
1751
        # and errors on ghosts.
 
1752
        self.assertRaises(errors.RevisionNotPresent,
 
1753
            index.get_parents, 'ghost')
 
1754
 
 
1755
    def test_get_parents_with_ghosts(self):
 
1756
        index = self.two_graph_index()
 
1757
        self.assertEqual(('tail', 'ghost'), index.get_parents_with_ghosts('parent'))
 
1758
        # and errors on ghosts.
 
1759
        self.assertRaises(errors.RevisionNotPresent,
 
1760
            index.get_parents_with_ghosts, 'ghost')
 
1761
 
 
1762
    def test_check_versions_present(self):
 
1763
        # ghosts should not be considered present
 
1764
        index = self.two_graph_index()
 
1765
        self.assertRaises(RevisionNotPresent, index.check_versions_present,
 
1766
            ['ghost'])
 
1767
        self.assertRaises(RevisionNotPresent, index.check_versions_present,
 
1768
            ['tail', 'ghost'])
 
1769
        index.check_versions_present(['tail', 'separate'])
1682
1770
 
1683
1771
    def catch_add(self, entries):
1684
1772
        self.caught_entries.append(entries)
1685
1773
 
1686
1774
    def test_add_no_callback_errors(self):
1687
1775
        index = self.two_graph_index()
1688
 
        self.assertRaises(errors.ReadOnlyError, index.add_records,
1689
 
            [(('new',), 'fulltext,no-eol', (None, 50, 60), ['separate'])])
 
1776
        self.assertRaises(errors.ReadOnlyError, index.add_version,
 
1777
            'new', 'fulltext,no-eol', 50, 60, ['separate'])
1690
1778
 
1691
1779
    def test_add_version_smoke(self):
1692
1780
        index = self.two_graph_index(catch_adds=True)
1693
 
        index.add_records([(('new',), 'fulltext,no-eol', (None, 50, 60),
1694
 
            [('separate',)])])
 
1781
        index.add_version('new', 'fulltext,no-eol', 50, 60, ['separate'])
1695
1782
        self.assertEqual([[(('new', ), 'N50 60', ((('separate',),),))]],
1696
1783
            self.caught_entries)
1697
1784
 
1698
1785
    def test_add_version_delta_not_delta_index(self):
1699
1786
        index = self.two_graph_index(catch_adds=True)
1700
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1701
 
            [(('new',), 'no-eol,line-delta', (None, 0, 100), [('parent',)])])
 
1787
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1788
            'new', 'no-eol,line-delta', 0, 100, ['parent'])
1702
1789
        self.assertEqual([], self.caught_entries)
1703
1790
 
1704
1791
    def test_add_version_same_dup(self):
1705
1792
        index = self.two_graph_index(catch_adds=True)
1706
1793
        # options can be spelt two different ways
1707
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 100), [('parent',)])])
1708
 
        index.add_records([(('tip',), 'no-eol,fulltext', (None, 0, 100), [('parent',)])])
1709
 
        # position/length are ignored (because each pack could have fulltext or
1710
 
        # delta, and be at a different position.
1711
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 50, 100),
1712
 
            [('parent',)])])
1713
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 1000),
1714
 
            [('parent',)])])
1715
 
        # but neither should have added data:
1716
 
        self.assertEqual([[], [], [], []], self.caught_entries)
1717
 
 
 
1794
        index.add_version('tip', 'fulltext,no-eol', 0, 100, ['parent'])
 
1795
        index.add_version('tip', 'no-eol,fulltext', 0, 100, ['parent'])
 
1796
        # but neither should have added data.
 
1797
        self.assertEqual([[], []], self.caught_entries)
 
1798
        
1718
1799
    def test_add_version_different_dup(self):
1719
1800
        index = self.two_graph_index(deltas=True, catch_adds=True)
1720
1801
        # change options
1721
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1722
 
            [(('tip',), 'line-delta', (None, 0, 100), [('parent',)])])
1723
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1724
 
            [(('tip',), 'fulltext', (None, 0, 100), [('parent',)])])
 
1802
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1803
            'tip', 'no-eol,line-delta', 0, 100, ['parent'])
 
1804
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1805
            'tip', 'line-delta,no-eol', 0, 100, ['parent'])
 
1806
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1807
            'tip', 'fulltext', 0, 100, ['parent'])
 
1808
        # position/length
 
1809
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1810
            'tip', 'fulltext,no-eol', 50, 100, ['parent'])
 
1811
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1812
            'tip', 'fulltext,no-eol', 0, 1000, ['parent'])
1725
1813
        # parents
1726
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1727
 
            [(('tip',), 'fulltext,no-eol', (None, 0, 100), [])])
 
1814
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
1815
            'tip', 'fulltext,no-eol', 0, 100, [])
1728
1816
        self.assertEqual([], self.caught_entries)
1729
 
 
 
1817
        
1730
1818
    def test_add_versions_nodeltas(self):
1731
1819
        index = self.two_graph_index(catch_adds=True)
1732
 
        index.add_records([
1733
 
                (('new',), 'fulltext,no-eol', (None, 50, 60), [('separate',)]),
1734
 
                (('new2',), 'fulltext', (None, 0, 6), [('new',)]),
 
1820
        index.add_versions([
 
1821
                ('new', 'fulltext,no-eol', 50, 60, ['separate']),
 
1822
                ('new2', 'fulltext', 0, 6, ['new']),
1735
1823
                ])
1736
1824
        self.assertEqual([(('new', ), 'N50 60', ((('separate',),),)),
1737
1825
            (('new2', ), ' 0 6', ((('new',),),))],
1740
1828
 
1741
1829
    def test_add_versions_deltas(self):
1742
1830
        index = self.two_graph_index(deltas=True, catch_adds=True)
1743
 
        index.add_records([
1744
 
                (('new',), 'fulltext,no-eol', (None, 50, 60), [('separate',)]),
1745
 
                (('new2',), 'line-delta', (None, 0, 6), [('new',)]),
 
1831
        index.add_versions([
 
1832
                ('new', 'fulltext,no-eol', 50, 60, ['separate']),
 
1833
                ('new2', 'line-delta', 0, 6, ['new']),
1746
1834
                ])
1747
1835
        self.assertEqual([(('new', ), 'N50 60', ((('separate',),), ())),
1748
1836
            (('new2', ), ' 0 6', ((('new',),), (('new',),), ))],
1751
1839
 
1752
1840
    def test_add_versions_delta_not_delta_index(self):
1753
1841
        index = self.two_graph_index(catch_adds=True)
1754
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1755
 
            [(('new',), 'no-eol,line-delta', (None, 0, 100), [('parent',)])])
 
1842
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1843
            [('new', 'no-eol,line-delta', 0, 100, ['parent'])])
1756
1844
        self.assertEqual([], self.caught_entries)
1757
1845
 
1758
 
    def test_add_versions_random_id_accepted(self):
1759
 
        index = self.two_graph_index(catch_adds=True)
1760
 
        index.add_records([], random_id=True)
1761
 
 
1762
1846
    def test_add_versions_same_dup(self):
1763
1847
        index = self.two_graph_index(catch_adds=True)
1764
1848
        # options can be spelt two different ways
1765
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 100),
1766
 
            [('parent',)])])
1767
 
        index.add_records([(('tip',), 'no-eol,fulltext', (None, 0, 100),
1768
 
            [('parent',)])])
1769
 
        # position/length are ignored (because each pack could have fulltext or
1770
 
        # delta, and be at a different position.
1771
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 50, 100),
1772
 
            [('parent',)])])
1773
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 1000),
1774
 
            [('parent',)])])
 
1849
        index.add_versions([('tip', 'fulltext,no-eol', 0, 100, ['parent'])])
 
1850
        index.add_versions([('tip', 'no-eol,fulltext', 0, 100, ['parent'])])
1775
1851
        # but neither should have added data.
1776
 
        self.assertEqual([[], [], [], []], self.caught_entries)
1777
 
 
 
1852
        self.assertEqual([[], []], self.caught_entries)
 
1853
        
1778
1854
    def test_add_versions_different_dup(self):
1779
1855
        index = self.two_graph_index(deltas=True, catch_adds=True)
1780
1856
        # change options
1781
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1782
 
            [(('tip',), 'line-delta', (None, 0, 100), [('parent',)])])
1783
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1784
 
            [(('tip',), 'fulltext', (None, 0, 100), [('parent',)])])
 
1857
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1858
            [('tip', 'no-eol,line-delta', 0, 100, ['parent'])])
 
1859
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1860
            [('tip', 'line-delta,no-eol', 0, 100, ['parent'])])
 
1861
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1862
            [('tip', 'fulltext', 0, 100, ['parent'])])
 
1863
        # position/length
 
1864
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1865
            [('tip', 'fulltext,no-eol', 50, 100, ['parent'])])
 
1866
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1867
            [('tip', 'fulltext,no-eol', 0, 1000, ['parent'])])
1785
1868
        # parents
1786
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1787
 
            [(('tip',), 'fulltext,no-eol', (None, 0, 100), [])])
 
1869
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1870
            [('tip', 'fulltext,no-eol', 0, 100, [])])
1788
1871
        # change options in the second record
1789
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1790
 
            [(('tip',), 'fulltext,no-eol', (None, 0, 100), [('parent',)]),
1791
 
             (('tip',), 'line-delta', (None, 0, 100), [('parent',)])])
 
1872
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
1873
            [('tip', 'fulltext,no-eol', 0, 100, ['parent']),
 
1874
             ('tip', 'no-eol,line-delta', 0, 100, ['parent'])])
1792
1875
        self.assertEqual([], self.caught_entries)
1793
1876
 
1794
 
    def make_g_index_missing_compression_parent(self):
1795
 
        graph_index = self.make_g_index('missing_comp', 2,
1796
 
            [(('tip', ), ' 100 78',
1797
 
              ([('missing-parent', ), ('ghost', )], [('missing-parent', )]))])
1798
 
        return graph_index
1799
 
 
1800
 
    def make_g_index_missing_parent(self):
1801
 
        graph_index = self.make_g_index('missing_parent', 2,
1802
 
            [(('parent', ), ' 100 78', ([], [])),
1803
 
             (('tip', ), ' 100 78',
1804
 
              ([('parent', ), ('missing-parent', )], [('parent', )])),
1805
 
              ])
1806
 
        return graph_index
1807
 
 
1808
 
    def make_g_index_no_external_refs(self):
1809
 
        graph_index = self.make_g_index('no_external_refs', 2,
1810
 
            [(('rev', ), ' 100 78',
1811
 
              ([('parent', ), ('ghost', )], []))])
1812
 
        return graph_index
1813
 
 
1814
 
    def test_add_good_unvalidated_index(self):
1815
 
        unvalidated = self.make_g_index_no_external_refs()
1816
 
        combined = CombinedGraphIndex([unvalidated])
1817
 
        index = _KnitGraphIndex(combined, lambda: True, deltas=True)
1818
 
        index.scan_unvalidated_index(unvalidated)
1819
 
        self.assertEqual(frozenset(), index.get_missing_compression_parents())
1820
 
 
1821
 
    def test_add_missing_compression_parent_unvalidated_index(self):
1822
 
        unvalidated = self.make_g_index_missing_compression_parent()
1823
 
        combined = CombinedGraphIndex([unvalidated])
1824
 
        index = _KnitGraphIndex(combined, lambda: True, deltas=True)
1825
 
        index.scan_unvalidated_index(unvalidated)
1826
 
        # This also checks that its only the compression parent that is
1827
 
        # examined, otherwise 'ghost' would also be reported as a missing
1828
 
        # parent.
1829
 
        self.assertEqual(
1830
 
            frozenset([('missing-parent',)]),
1831
 
            index.get_missing_compression_parents())
1832
 
 
1833
 
    def test_add_missing_noncompression_parent_unvalidated_index(self):
1834
 
        unvalidated = self.make_g_index_missing_parent()
1835
 
        combined = CombinedGraphIndex([unvalidated])
1836
 
        index = _KnitGraphIndex(combined, lambda: True, deltas=True,
1837
 
            track_external_parent_refs=True)
1838
 
        index.scan_unvalidated_index(unvalidated)
1839
 
        self.assertEqual(
1840
 
            frozenset([('missing-parent',)]), index.get_missing_parents())
1841
 
 
1842
 
    def test_track_external_parent_refs(self):
1843
 
        g_index = self.make_g_index('empty', 2, [])
1844
 
        combined = CombinedGraphIndex([g_index])
1845
 
        index = _KnitGraphIndex(combined, lambda: True, deltas=True,
1846
 
            add_callback=self.catch_add, track_external_parent_refs=True)
1847
 
        self.caught_entries = []
1848
 
        index.add_records([
1849
 
            (('new-key',), 'fulltext,no-eol', (None, 50, 60),
1850
 
             [('parent-1',), ('parent-2',)])])
1851
 
        self.assertEqual(
1852
 
            frozenset([('parent-1',), ('parent-2',)]),
1853
 
            index.get_missing_parents())
1854
 
 
1855
 
    def test_add_unvalidated_index_with_present_external_references(self):
1856
 
        index = self.two_graph_index(deltas=True)
1857
 
        # Ugly hack to get at one of the underlying GraphIndex objects that
1858
 
        # two_graph_index built.
1859
 
        unvalidated = index._graph_index._indices[1]
1860
 
        # 'parent' is an external ref of _indices[1] (unvalidated), but is
1861
 
        # present in _indices[0].
1862
 
        index.scan_unvalidated_index(unvalidated)
1863
 
        self.assertEqual(frozenset(), index.get_missing_compression_parents())
1864
 
 
1865
 
    def make_new_missing_parent_g_index(self, name):
1866
 
        missing_parent = name + '-missing-parent'
1867
 
        graph_index = self.make_g_index(name, 2,
1868
 
            [((name + 'tip', ), ' 100 78',
1869
 
              ([(missing_parent, ), ('ghost', )], [(missing_parent, )]))])
1870
 
        return graph_index
1871
 
 
1872
 
    def test_add_mulitiple_unvalidated_indices_with_missing_parents(self):
1873
 
        g_index_1 = self.make_new_missing_parent_g_index('one')
1874
 
        g_index_2 = self.make_new_missing_parent_g_index('two')
1875
 
        combined = CombinedGraphIndex([g_index_1, g_index_2])
1876
 
        index = _KnitGraphIndex(combined, lambda: True, deltas=True)
1877
 
        index.scan_unvalidated_index(g_index_1)
1878
 
        index.scan_unvalidated_index(g_index_2)
1879
 
        self.assertEqual(
1880
 
            frozenset([('one-missing-parent',), ('two-missing-parent',)]),
1881
 
            index.get_missing_compression_parents())
1882
 
 
1883
 
    def test_add_mulitiple_unvalidated_indices_with_mutual_dependencies(self):
1884
 
        graph_index_a = self.make_g_index('one', 2,
1885
 
            [(('parent-one', ), ' 100 78', ([('non-compression-parent',)], [])),
1886
 
             (('child-of-two', ), ' 100 78',
1887
 
              ([('parent-two',)], [('parent-two',)]))])
1888
 
        graph_index_b = self.make_g_index('two', 2,
1889
 
            [(('parent-two', ), ' 100 78', ([('non-compression-parent',)], [])),
1890
 
             (('child-of-one', ), ' 100 78',
1891
 
              ([('parent-one',)], [('parent-one',)]))])
1892
 
        combined = CombinedGraphIndex([graph_index_a, graph_index_b])
1893
 
        index = _KnitGraphIndex(combined, lambda: True, deltas=True)
1894
 
        index.scan_unvalidated_index(graph_index_a)
1895
 
        index.scan_unvalidated_index(graph_index_b)
1896
 
        self.assertEqual(
1897
 
            frozenset([]), index.get_missing_compression_parents())
 
1877
    def test_iter_parents(self):
 
1878
        index1 = self.make_g_index('1', 1, [
 
1879
        # no parents
 
1880
            (('r0', ), 'N0 100', ([], )),
 
1881
        # 1 parent
 
1882
            (('r1', ), '', ([('r0', )], ))])
 
1883
        index2 = self.make_g_index('2', 1, [
 
1884
        # 2 parents
 
1885
            (('r2', ), 'N0 100', ([('r1', ), ('r0', )], )),
 
1886
            ])
 
1887
        combined_index = CombinedGraphIndex([index1, index2])
 
1888
        index = KnitGraphIndex(combined_index)
 
1889
        # XXX TODO a ghost
 
1890
        # cases: each sample data individually:
 
1891
        self.assertEqual(set([('r0', ())]),
 
1892
            set(index.iter_parents(['r0'])))
 
1893
        self.assertEqual(set([('r1', ('r0', ))]),
 
1894
            set(index.iter_parents(['r1'])))
 
1895
        self.assertEqual(set([('r2', ('r1', 'r0'))]),
 
1896
            set(index.iter_parents(['r2'])))
 
1897
        # no nodes returned for a missing node
 
1898
        self.assertEqual(set(),
 
1899
            set(index.iter_parents(['missing'])))
 
1900
        # 1 node returned with missing nodes skipped
 
1901
        self.assertEqual(set([('r1', ('r0', ))]),
 
1902
            set(index.iter_parents(['ghost1', 'r1', 'ghost'])))
 
1903
        # 2 nodes returned
 
1904
        self.assertEqual(set([('r0', ()), ('r1', ('r0', ))]),
 
1905
            set(index.iter_parents(['r0', 'r1'])))
 
1906
        # 2 nodes returned, missing skipped
 
1907
        self.assertEqual(set([('r0', ()), ('r1', ('r0', ))]),
 
1908
            set(index.iter_parents(['a', 'r0', 'b', 'r1', 'c'])))
1898
1909
 
1899
1910
 
1900
1911
class TestNoParentsGraphIndexKnit(KnitTests):
1901
 
    """Tests for knits using _KnitGraphIndex with no parents."""
 
1912
    """Tests for knits using KnitGraphIndex with no parents."""
1902
1913
 
1903
1914
    def make_g_index(self, name, ref_lists=0, nodes=[]):
1904
1915
        builder = GraphIndexBuilder(ref_lists)
1906
1917
            builder.add_node(node, references)
1907
1918
        stream = builder.finish()
1908
1919
        trans = self.get_transport()
1909
 
        size = trans.put_file(name, stream)
1910
 
        return GraphIndex(trans, name, size)
1911
 
 
1912
 
    def test_add_good_unvalidated_index(self):
1913
 
        unvalidated = self.make_g_index('unvalidated')
1914
 
        combined = CombinedGraphIndex([unvalidated])
1915
 
        index = _KnitGraphIndex(combined, lambda: True, parents=False)
1916
 
        index.scan_unvalidated_index(unvalidated)
1917
 
        self.assertEqual(frozenset(),
1918
 
            index.get_missing_compression_parents())
 
1920
        trans.put_file(name, stream)
 
1921
        return GraphIndex(trans, name)
1919
1922
 
1920
1923
    def test_parents_deltas_incompatible(self):
1921
1924
        index = CombinedGraphIndex([])
1922
 
        self.assertRaises(errors.KnitError, _KnitGraphIndex, lambda:True,
1923
 
            index, deltas=True, parents=False)
 
1925
        self.assertRaises(errors.KnitError, KnitGraphIndex, index,
 
1926
            deltas=True, parents=False)
1924
1927
 
1925
1928
    def two_graph_index(self, catch_adds=False):
1926
1929
        """Build a two-graph index.
1942
1945
            add_callback = self.catch_add
1943
1946
        else:
1944
1947
            add_callback = None
1945
 
        return _KnitGraphIndex(combined_index, lambda:True, parents=False,
 
1948
        return KnitGraphIndex(combined_index, parents=False,
1946
1949
            add_callback=add_callback)
1947
1950
 
1948
 
    def test_keys(self):
1949
 
        index = self.two_graph_index()
1950
 
        self.assertEqual(set([('tail',), ('tip',), ('parent',), ('separate',)]),
1951
 
            set(index.keys()))
 
1951
    def test_get_graph(self):
 
1952
        index = self.two_graph_index()
 
1953
        self.assertEqual(set([
 
1954
            ('tip', ()),
 
1955
            ('tail', ()),
 
1956
            ('parent', ()),
 
1957
            ('separate', ()),
 
1958
            ]), set(index.get_graph()))
 
1959
 
 
1960
    def test_get_ancestry(self):
 
1961
        # with no parents, ancestry is always just the key.
 
1962
        index = self.two_graph_index()
 
1963
        self.assertEqual([], index.get_ancestry([]))
 
1964
        self.assertEqual(['separate'], index.get_ancestry(['separate']))
 
1965
        self.assertEqual(['tail'], index.get_ancestry(['tail']))
 
1966
        self.assertEqual(['parent'], index.get_ancestry(['parent']))
 
1967
        self.assertEqual(['tip'], index.get_ancestry(['tip']))
 
1968
        self.assertTrue(index.get_ancestry(['tip', 'separate']) in
 
1969
            (['tip', 'separate'],
 
1970
             ['separate', 'tip'],
 
1971
            ))
 
1972
        # asking for a ghost makes it go boom.
 
1973
        self.assertRaises(errors.RevisionNotPresent, index.get_ancestry, ['ghost'])
 
1974
 
 
1975
    def test_get_ancestry_with_ghosts(self):
 
1976
        index = self.two_graph_index()
 
1977
        self.assertEqual([], index.get_ancestry_with_ghosts([]))
 
1978
        self.assertEqual(['separate'], index.get_ancestry_with_ghosts(['separate']))
 
1979
        self.assertEqual(['tail'], index.get_ancestry_with_ghosts(['tail']))
 
1980
        self.assertEqual(['parent'], index.get_ancestry_with_ghosts(['parent']))
 
1981
        self.assertEqual(['tip'], index.get_ancestry_with_ghosts(['tip']))
 
1982
        self.assertTrue(index.get_ancestry_with_ghosts(['tip', 'separate']) in
 
1983
            (['tip', 'separate'],
 
1984
             ['separate', 'tip'],
 
1985
            ))
 
1986
        # asking for a ghost makes it go boom.
 
1987
        self.assertRaises(errors.RevisionNotPresent, index.get_ancestry_with_ghosts, ['ghost'])
 
1988
 
 
1989
    def test_num_versions(self):
 
1990
        index = self.two_graph_index()
 
1991
        self.assertEqual(4, index.num_versions())
 
1992
 
 
1993
    def test_get_versions(self):
 
1994
        index = self.two_graph_index()
 
1995
        self.assertEqual(set(['tail', 'tip', 'parent', 'separate']),
 
1996
            set(index.get_versions()))
 
1997
 
 
1998
    def test_has_version(self):
 
1999
        index = self.two_graph_index()
 
2000
        self.assertTrue(index.has_version('tail'))
 
2001
        self.assertFalse(index.has_version('ghost'))
1952
2002
 
1953
2003
    def test_get_position(self):
1954
2004
        index = self.two_graph_index()
1955
 
        self.assertEqual((index._graph_index._indices[0], 0, 100),
1956
 
            index.get_position(('tip',)))
1957
 
        self.assertEqual((index._graph_index._indices[1], 100, 78),
1958
 
            index.get_position(('parent',)))
 
2005
        self.assertEqual((0, 100), index.get_position('tip'))
 
2006
        self.assertEqual((100, 78), index.get_position('parent'))
1959
2007
 
1960
2008
    def test_get_method(self):
1961
2009
        index = self.two_graph_index()
1962
 
        self.assertEqual('fulltext', index.get_method(('tip',)))
1963
 
        self.assertEqual(['fulltext'], index.get_options(('parent',)))
 
2010
        self.assertEqual('fulltext', index.get_method('tip'))
 
2011
        self.assertEqual(['fulltext'], index.get_options('parent'))
1964
2012
 
1965
2013
    def test_get_options(self):
1966
2014
        index = self.two_graph_index()
1967
 
        self.assertEqual(['fulltext', 'no-eol'], index.get_options(('tip',)))
1968
 
        self.assertEqual(['fulltext'], index.get_options(('parent',)))
1969
 
 
1970
 
    def test_get_parent_map(self):
1971
 
        index = self.two_graph_index()
1972
 
        self.assertEqual({('parent',):None},
1973
 
            index.get_parent_map([('parent',), ('ghost',)]))
 
2015
        self.assertEqual(['fulltext', 'no-eol'], index.get_options('tip'))
 
2016
        self.assertEqual(['fulltext'], index.get_options('parent'))
 
2017
 
 
2018
    def test_get_parents(self):
 
2019
        index = self.two_graph_index()
 
2020
        self.assertEqual((), index.get_parents('parent'))
 
2021
        # and errors on ghosts.
 
2022
        self.assertRaises(errors.RevisionNotPresent,
 
2023
            index.get_parents, 'ghost')
 
2024
 
 
2025
    def test_get_parents_with_ghosts(self):
 
2026
        index = self.two_graph_index()
 
2027
        self.assertEqual((), index.get_parents_with_ghosts('parent'))
 
2028
        # and errors on ghosts.
 
2029
        self.assertRaises(errors.RevisionNotPresent,
 
2030
            index.get_parents_with_ghosts, 'ghost')
 
2031
 
 
2032
    def test_check_versions_present(self):
 
2033
        index = self.two_graph_index()
 
2034
        self.assertRaises(RevisionNotPresent, index.check_versions_present,
 
2035
            ['missing'])
 
2036
        self.assertRaises(RevisionNotPresent, index.check_versions_present,
 
2037
            ['tail', 'missing'])
 
2038
        index.check_versions_present(['tail', 'separate'])
1974
2039
 
1975
2040
    def catch_add(self, entries):
1976
2041
        self.caught_entries.append(entries)
1977
2042
 
1978
2043
    def test_add_no_callback_errors(self):
1979
2044
        index = self.two_graph_index()
1980
 
        self.assertRaises(errors.ReadOnlyError, index.add_records,
1981
 
            [(('new',), 'fulltext,no-eol', (None, 50, 60), [('separate',)])])
 
2045
        self.assertRaises(errors.ReadOnlyError, index.add_version,
 
2046
            'new', 'fulltext,no-eol', 50, 60, ['separate'])
1982
2047
 
1983
2048
    def test_add_version_smoke(self):
1984
2049
        index = self.two_graph_index(catch_adds=True)
1985
 
        index.add_records([(('new',), 'fulltext,no-eol', (None, 50, 60), [])])
 
2050
        index.add_version('new', 'fulltext,no-eol', 50, 60, [])
1986
2051
        self.assertEqual([[(('new', ), 'N50 60')]],
1987
2052
            self.caught_entries)
1988
2053
 
1989
2054
    def test_add_version_delta_not_delta_index(self):
1990
2055
        index = self.two_graph_index(catch_adds=True)
1991
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
1992
 
            [(('new',), 'no-eol,line-delta', (None, 0, 100), [])])
 
2056
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2057
            'new', 'no-eol,line-delta', 0, 100, [])
1993
2058
        self.assertEqual([], self.caught_entries)
1994
2059
 
1995
2060
    def test_add_version_same_dup(self):
1996
2061
        index = self.two_graph_index(catch_adds=True)
1997
2062
        # options can be spelt two different ways
1998
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 100), [])])
1999
 
        index.add_records([(('tip',), 'no-eol,fulltext', (None, 0, 100), [])])
2000
 
        # position/length are ignored (because each pack could have fulltext or
2001
 
        # delta, and be at a different position.
2002
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 50, 100), [])])
2003
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 1000), [])])
 
2063
        index.add_version('tip', 'fulltext,no-eol', 0, 100, [])
 
2064
        index.add_version('tip', 'no-eol,fulltext', 0, 100, [])
2004
2065
        # but neither should have added data.
2005
 
        self.assertEqual([[], [], [], []], self.caught_entries)
2006
 
 
 
2066
        self.assertEqual([[], []], self.caught_entries)
 
2067
        
2007
2068
    def test_add_version_different_dup(self):
2008
2069
        index = self.two_graph_index(catch_adds=True)
2009
2070
        # change options
2010
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2011
 
            [(('tip',), 'no-eol,line-delta', (None, 0, 100), [])])
2012
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2013
 
            [(('tip',), 'line-delta,no-eol', (None, 0, 100), [])])
2014
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2015
 
            [(('tip',), 'fulltext', (None, 0, 100), [])])
 
2071
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2072
            'tip', 'no-eol,line-delta', 0, 100, [])
 
2073
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2074
            'tip', 'line-delta,no-eol', 0, 100, [])
 
2075
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2076
            'tip', 'fulltext', 0, 100, [])
 
2077
        # position/length
 
2078
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2079
            'tip', 'fulltext,no-eol', 50, 100, [])
 
2080
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2081
            'tip', 'fulltext,no-eol', 0, 1000, [])
2016
2082
        # parents
2017
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2018
 
            [(('tip',), 'fulltext,no-eol', (None, 0, 100), [('parent',)])])
 
2083
        self.assertRaises(errors.KnitCorrupt, index.add_version,
 
2084
            'tip', 'fulltext,no-eol', 0, 100, ['parent'])
2019
2085
        self.assertEqual([], self.caught_entries)
2020
 
 
 
2086
        
2021
2087
    def test_add_versions(self):
2022
2088
        index = self.two_graph_index(catch_adds=True)
2023
 
        index.add_records([
2024
 
                (('new',), 'fulltext,no-eol', (None, 50, 60), []),
2025
 
                (('new2',), 'fulltext', (None, 0, 6), []),
 
2089
        index.add_versions([
 
2090
                ('new', 'fulltext,no-eol', 50, 60, []),
 
2091
                ('new2', 'fulltext', 0, 6, []),
2026
2092
                ])
2027
2093
        self.assertEqual([(('new', ), 'N50 60'), (('new2', ), ' 0 6')],
2028
2094
            sorted(self.caught_entries[0]))
2030
2096
 
2031
2097
    def test_add_versions_delta_not_delta_index(self):
2032
2098
        index = self.two_graph_index(catch_adds=True)
2033
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2034
 
            [(('new',), 'no-eol,line-delta', (None, 0, 100), [('parent',)])])
 
2099
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2100
            [('new', 'no-eol,line-delta', 0, 100, ['parent'])])
2035
2101
        self.assertEqual([], self.caught_entries)
2036
2102
 
2037
2103
    def test_add_versions_parents_not_parents_index(self):
2038
2104
        index = self.two_graph_index(catch_adds=True)
2039
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2040
 
            [(('new',), 'no-eol,fulltext', (None, 0, 100), [('parent',)])])
 
2105
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2106
            [('new', 'no-eol,fulltext', 0, 100, ['parent'])])
2041
2107
        self.assertEqual([], self.caught_entries)
2042
2108
 
2043
 
    def test_add_versions_random_id_accepted(self):
2044
 
        index = self.two_graph_index(catch_adds=True)
2045
 
        index.add_records([], random_id=True)
2046
 
 
2047
2109
    def test_add_versions_same_dup(self):
2048
2110
        index = self.two_graph_index(catch_adds=True)
2049
2111
        # options can be spelt two different ways
2050
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 100), [])])
2051
 
        index.add_records([(('tip',), 'no-eol,fulltext', (None, 0, 100), [])])
2052
 
        # position/length are ignored (because each pack could have fulltext or
2053
 
        # delta, and be at a different position.
2054
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 50, 100), [])])
2055
 
        index.add_records([(('tip',), 'fulltext,no-eol', (None, 0, 1000), [])])
 
2112
        index.add_versions([('tip', 'fulltext,no-eol', 0, 100, [])])
 
2113
        index.add_versions([('tip', 'no-eol,fulltext', 0, 100, [])])
2056
2114
        # but neither should have added data.
2057
 
        self.assertEqual([[], [], [], []], self.caught_entries)
2058
 
 
 
2115
        self.assertEqual([[], []], self.caught_entries)
 
2116
        
2059
2117
    def test_add_versions_different_dup(self):
2060
2118
        index = self.two_graph_index(catch_adds=True)
2061
2119
        # change options
2062
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2063
 
            [(('tip',), 'no-eol,line-delta', (None, 0, 100), [])])
2064
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2065
 
            [(('tip',), 'line-delta,no-eol', (None, 0, 100), [])])
2066
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2067
 
            [(('tip',), 'fulltext', (None, 0, 100), [])])
 
2120
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2121
            [('tip', 'no-eol,line-delta', 0, 100, [])])
 
2122
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2123
            [('tip', 'line-delta,no-eol', 0, 100, [])])
 
2124
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2125
            [('tip', 'fulltext', 0, 100, [])])
 
2126
        # position/length
 
2127
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2128
            [('tip', 'fulltext,no-eol', 50, 100, [])])
 
2129
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2130
            [('tip', 'fulltext,no-eol', 0, 1000, [])])
2068
2131
        # parents
2069
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2070
 
            [(('tip',), 'fulltext,no-eol', (None, 0, 100), [('parent',)])])
 
2132
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2133
            [('tip', 'fulltext,no-eol', 0, 100, ['parent'])])
2071
2134
        # change options in the second record
2072
 
        self.assertRaises(errors.KnitCorrupt, index.add_records,
2073
 
            [(('tip',), 'fulltext,no-eol', (None, 0, 100), []),
2074
 
             (('tip',), 'no-eol,line-delta', (None, 0, 100), [])])
 
2135
        self.assertRaises(errors.KnitCorrupt, index.add_versions,
 
2136
            [('tip', 'fulltext,no-eol', 0, 100, []),
 
2137
             ('tip', 'no-eol,line-delta', 0, 100, [])])
2075
2138
        self.assertEqual([], self.caught_entries)
2076
2139
 
2077
 
 
2078
 
class TestKnitVersionedFiles(KnitTests):
2079
 
 
2080
 
    def assertGroupKeysForIo(self, exp_groups, keys, non_local_keys,
2081
 
                             positions, _min_buffer_size=None):
2082
 
        kvf = self.make_test_knit()
2083
 
        if _min_buffer_size is None:
2084
 
            _min_buffer_size = knit._STREAM_MIN_BUFFER_SIZE
2085
 
        self.assertEqual(exp_groups, kvf._group_keys_for_io(keys,
2086
 
                                        non_local_keys, positions,
2087
 
                                        _min_buffer_size=_min_buffer_size))
2088
 
 
2089
 
    def assertSplitByPrefix(self, expected_map, expected_prefix_order,
2090
 
                            keys):
2091
 
        split, prefix_order = KnitVersionedFiles._split_by_prefix(keys)
2092
 
        self.assertEqual(expected_map, split)
2093
 
        self.assertEqual(expected_prefix_order, prefix_order)
2094
 
 
2095
 
    def test__group_keys_for_io(self):
2096
 
        ft_detail = ('fulltext', False)
2097
 
        ld_detail = ('line-delta', False)
2098
 
        f_a = ('f', 'a')
2099
 
        f_b = ('f', 'b')
2100
 
        f_c = ('f', 'c')
2101
 
        g_a = ('g', 'a')
2102
 
        g_b = ('g', 'b')
2103
 
        g_c = ('g', 'c')
2104
 
        positions = {
2105
 
            f_a: (ft_detail, (f_a, 0, 100), None),
2106
 
            f_b: (ld_detail, (f_b, 100, 21), f_a),
2107
 
            f_c: (ld_detail, (f_c, 180, 15), f_b),
2108
 
            g_a: (ft_detail, (g_a, 121, 35), None),
2109
 
            g_b: (ld_detail, (g_b, 156, 12), g_a),
2110
 
            g_c: (ld_detail, (g_c, 195, 13), g_a),
2111
 
            }
2112
 
        self.assertGroupKeysForIo([([f_a], set())],
2113
 
                                  [f_a], [], positions)
2114
 
        self.assertGroupKeysForIo([([f_a], set([f_a]))],
2115
 
                                  [f_a], [f_a], positions)
2116
 
        self.assertGroupKeysForIo([([f_a, f_b], set([]))],
2117
 
                                  [f_a, f_b], [], positions)
2118
 
        self.assertGroupKeysForIo([([f_a, f_b], set([f_b]))],
2119
 
                                  [f_a, f_b], [f_b], positions)
2120
 
        self.assertGroupKeysForIo([([f_a, f_b, g_a, g_b], set())],
2121
 
                                  [f_a, g_a, f_b, g_b], [], positions)
2122
 
        self.assertGroupKeysForIo([([f_a, f_b, g_a, g_b], set())],
2123
 
                                  [f_a, g_a, f_b, g_b], [], positions,
2124
 
                                  _min_buffer_size=150)
2125
 
        self.assertGroupKeysForIo([([f_a, f_b], set()), ([g_a, g_b], set())],
2126
 
                                  [f_a, g_a, f_b, g_b], [], positions,
2127
 
                                  _min_buffer_size=100)
2128
 
        self.assertGroupKeysForIo([([f_c], set()), ([g_b], set())],
2129
 
                                  [f_c, g_b], [], positions,
2130
 
                                  _min_buffer_size=125)
2131
 
        self.assertGroupKeysForIo([([g_b, f_c], set())],
2132
 
                                  [g_b, f_c], [], positions,
2133
 
                                  _min_buffer_size=125)
2134
 
 
2135
 
    def test__split_by_prefix(self):
2136
 
        self.assertSplitByPrefix({'f': [('f', 'a'), ('f', 'b')],
2137
 
                                  'g': [('g', 'b'), ('g', 'a')],
2138
 
                                 }, ['f', 'g'],
2139
 
                                 [('f', 'a'), ('g', 'b'),
2140
 
                                  ('g', 'a'), ('f', 'b')])
2141
 
 
2142
 
        self.assertSplitByPrefix({'f': [('f', 'a'), ('f', 'b')],
2143
 
                                  'g': [('g', 'b'), ('g', 'a')],
2144
 
                                 }, ['f', 'g'],
2145
 
                                 [('f', 'a'), ('f', 'b'),
2146
 
                                  ('g', 'b'), ('g', 'a')])
2147
 
 
2148
 
        self.assertSplitByPrefix({'f': [('f', 'a'), ('f', 'b')],
2149
 
                                  'g': [('g', 'b'), ('g', 'a')],
2150
 
                                 }, ['f', 'g'],
2151
 
                                 [('f', 'a'), ('f', 'b'),
2152
 
                                  ('g', 'b'), ('g', 'a')])
2153
 
 
2154
 
        self.assertSplitByPrefix({'f': [('f', 'a'), ('f', 'b')],
2155
 
                                  'g': [('g', 'b'), ('g', 'a')],
2156
 
                                  '': [('a',), ('b',)]
2157
 
                                 }, ['f', 'g', ''],
2158
 
                                 [('f', 'a'), ('g', 'b'),
2159
 
                                  ('a',), ('b',),
2160
 
                                  ('g', 'a'), ('f', 'b')])
2161
 
 
2162
 
 
2163
 
class TestStacking(KnitTests):
2164
 
 
2165
 
    def get_basis_and_test_knit(self):
2166
 
        basis = self.make_test_knit(name='basis')
2167
 
        basis = RecordingVersionedFilesDecorator(basis)
2168
 
        test = self.make_test_knit(name='test')
2169
 
        test.add_fallback_versioned_files(basis)
2170
 
        return basis, test
2171
 
 
2172
 
    def test_add_fallback_versioned_files(self):
2173
 
        basis = self.make_test_knit(name='basis')
2174
 
        test = self.make_test_knit(name='test')
2175
 
        # It must not error; other tests test that the fallback is referred to
2176
 
        # when accessing data.
2177
 
        test.add_fallback_versioned_files(basis)
2178
 
 
2179
 
    def test_add_lines(self):
2180
 
        # lines added to the test are not added to the basis
2181
 
        basis, test = self.get_basis_and_test_knit()
2182
 
        key = ('foo',)
2183
 
        key_basis = ('bar',)
2184
 
        key_cross_border = ('quux',)
2185
 
        key_delta = ('zaphod',)
2186
 
        test.add_lines(key, (), ['foo\n'])
2187
 
        self.assertEqual({}, basis.get_parent_map([key]))
2188
 
        # lines added to the test that reference across the stack do a
2189
 
        # fulltext.
2190
 
        basis.add_lines(key_basis, (), ['foo\n'])
2191
 
        basis.calls = []
2192
 
        test.add_lines(key_cross_border, (key_basis,), ['foo\n'])
2193
 
        self.assertEqual('fulltext', test._index.get_method(key_cross_border))
2194
 
        # we don't even need to look at the basis to see that this should be
2195
 
        # stored as a fulltext
2196
 
        self.assertEqual([], basis.calls)
2197
 
        # Subsequent adds do delta.
2198
 
        basis.calls = []
2199
 
        test.add_lines(key_delta, (key_cross_border,), ['foo\n'])
2200
 
        self.assertEqual('line-delta', test._index.get_method(key_delta))
2201
 
        self.assertEqual([], basis.calls)
2202
 
 
2203
 
    def test_annotate(self):
2204
 
        # annotations from the test knit are answered without asking the basis
2205
 
        basis, test = self.get_basis_and_test_knit()
2206
 
        key = ('foo',)
2207
 
        key_basis = ('bar',)
2208
 
        key_missing = ('missing',)
2209
 
        test.add_lines(key, (), ['foo\n'])
2210
 
        details = test.annotate(key)
2211
 
        self.assertEqual([(key, 'foo\n')], details)
2212
 
        self.assertEqual([], basis.calls)
2213
 
        # But texts that are not in the test knit are looked for in the basis
2214
 
        # directly.
2215
 
        basis.add_lines(key_basis, (), ['foo\n', 'bar\n'])
2216
 
        basis.calls = []
2217
 
        details = test.annotate(key_basis)
2218
 
        self.assertEqual([(key_basis, 'foo\n'), (key_basis, 'bar\n')], details)
2219
 
        # Not optimised to date:
2220
 
        # self.assertEqual([("annotate", key_basis)], basis.calls)
2221
 
        self.assertEqual([('get_parent_map', set([key_basis])),
2222
 
            ('get_parent_map', set([key_basis])),
2223
 
            ('get_record_stream', [key_basis], 'topological', True)],
2224
 
            basis.calls)
2225
 
 
2226
 
    def test_check(self):
2227
 
        # At the moment checking a stacked knit does implicitly check the
2228
 
        # fallback files.
2229
 
        basis, test = self.get_basis_and_test_knit()
2230
 
        test.check()
2231
 
 
2232
 
    def test_get_parent_map(self):
2233
 
        # parents in the test knit are answered without asking the basis
2234
 
        basis, test = self.get_basis_and_test_knit()
2235
 
        key = ('foo',)
2236
 
        key_basis = ('bar',)
2237
 
        key_missing = ('missing',)
2238
 
        test.add_lines(key, (), [])
2239
 
        parent_map = test.get_parent_map([key])
2240
 
        self.assertEqual({key: ()}, parent_map)
2241
 
        self.assertEqual([], basis.calls)
2242
 
        # But parents that are not in the test knit are looked for in the basis
2243
 
        basis.add_lines(key_basis, (), [])
2244
 
        basis.calls = []
2245
 
        parent_map = test.get_parent_map([key, key_basis, key_missing])
2246
 
        self.assertEqual({key: (),
2247
 
            key_basis: ()}, parent_map)
2248
 
        self.assertEqual([("get_parent_map", set([key_basis, key_missing]))],
2249
 
            basis.calls)
2250
 
 
2251
 
    def test_get_record_stream_unordered_fulltexts(self):
2252
 
        # records from the test knit are answered without asking the basis:
2253
 
        basis, test = self.get_basis_and_test_knit()
2254
 
        key = ('foo',)
2255
 
        key_basis = ('bar',)
2256
 
        key_missing = ('missing',)
2257
 
        test.add_lines(key, (), ['foo\n'])
2258
 
        records = list(test.get_record_stream([key], 'unordered', True))
2259
 
        self.assertEqual(1, len(records))
2260
 
        self.assertEqual([], basis.calls)
2261
 
        # Missing (from test knit) objects are retrieved from the basis:
2262
 
        basis.add_lines(key_basis, (), ['foo\n', 'bar\n'])
2263
 
        basis.calls = []
2264
 
        records = list(test.get_record_stream([key_basis, key_missing],
2265
 
            'unordered', True))
2266
 
        self.assertEqual(2, len(records))
2267
 
        calls = list(basis.calls)
2268
 
        for record in records:
2269
 
            self.assertSubset([record.key], (key_basis, key_missing))
2270
 
            if record.key == key_missing:
2271
 
                self.assertIsInstance(record, AbsentContentFactory)
2272
 
            else:
2273
 
                reference = list(basis.get_record_stream([key_basis],
2274
 
                    'unordered', True))[0]
2275
 
                self.assertEqual(reference.key, record.key)
2276
 
                self.assertEqual(reference.sha1, record.sha1)
2277
 
                self.assertEqual(reference.storage_kind, record.storage_kind)
2278
 
                self.assertEqual(reference.get_bytes_as(reference.storage_kind),
2279
 
                    record.get_bytes_as(record.storage_kind))
2280
 
                self.assertEqual(reference.get_bytes_as('fulltext'),
2281
 
                    record.get_bytes_as('fulltext'))
2282
 
        # It's not strictly minimal, but it seems reasonable for now for it to
2283
 
        # ask which fallbacks have which parents.
2284
 
        self.assertEqual([
2285
 
            ("get_parent_map", set([key_basis, key_missing])),
2286
 
            ("get_record_stream", [key_basis], 'unordered', True)],
2287
 
            calls)
2288
 
 
2289
 
    def test_get_record_stream_ordered_fulltexts(self):
2290
 
        # ordering is preserved down into the fallback store.
2291
 
        basis, test = self.get_basis_and_test_knit()
2292
 
        key = ('foo',)
2293
 
        key_basis = ('bar',)
2294
 
        key_basis_2 = ('quux',)
2295
 
        key_missing = ('missing',)
2296
 
        test.add_lines(key, (key_basis,), ['foo\n'])
2297
 
        # Missing (from test knit) objects are retrieved from the basis:
2298
 
        basis.add_lines(key_basis, (key_basis_2,), ['foo\n', 'bar\n'])
2299
 
        basis.add_lines(key_basis_2, (), ['quux\n'])
2300
 
        basis.calls = []
2301
 
        # ask for in non-topological order
2302
 
        records = list(test.get_record_stream(
2303
 
            [key, key_basis, key_missing, key_basis_2], 'topological', True))
2304
 
        self.assertEqual(4, len(records))
2305
 
        results = []
2306
 
        for record in records:
2307
 
            self.assertSubset([record.key],
2308
 
                (key_basis, key_missing, key_basis_2, key))
2309
 
            if record.key == key_missing:
2310
 
                self.assertIsInstance(record, AbsentContentFactory)
2311
 
            else:
2312
 
                results.append((record.key, record.sha1, record.storage_kind,
2313
 
                    record.get_bytes_as('fulltext')))
2314
 
        calls = list(basis.calls)
2315
 
        order = [record[0] for record in results]
2316
 
        self.assertEqual([key_basis_2, key_basis, key], order)
2317
 
        for result in results:
2318
 
            if result[0] == key:
2319
 
                source = test
2320
 
            else:
2321
 
                source = basis
2322
 
            record = source.get_record_stream([result[0]], 'unordered',
2323
 
                True).next()
2324
 
            self.assertEqual(record.key, result[0])
2325
 
            self.assertEqual(record.sha1, result[1])
2326
 
            # We used to check that the storage kind matched, but actually it
2327
 
            # depends on whether it was sourced from the basis, or in a single
2328
 
            # group, because asking for full texts returns proxy objects to a
2329
 
            # _ContentMapGenerator object; so checking the kind is unneeded.
2330
 
            self.assertEqual(record.get_bytes_as('fulltext'), result[3])
2331
 
        # It's not strictly minimal, but it seems reasonable for now for it to
2332
 
        # ask which fallbacks have which parents.
2333
 
        self.assertEqual([
2334
 
            ("get_parent_map", set([key_basis, key_basis_2, key_missing])),
2335
 
            # topological is requested from the fallback, because that is what
2336
 
            # was requested at the top level.
2337
 
            ("get_record_stream", [key_basis_2, key_basis], 'topological', True)],
2338
 
            calls)
2339
 
 
2340
 
    def test_get_record_stream_unordered_deltas(self):
2341
 
        # records from the test knit are answered without asking the basis:
2342
 
        basis, test = self.get_basis_and_test_knit()
2343
 
        key = ('foo',)
2344
 
        key_basis = ('bar',)
2345
 
        key_missing = ('missing',)
2346
 
        test.add_lines(key, (), ['foo\n'])
2347
 
        records = list(test.get_record_stream([key], 'unordered', False))
2348
 
        self.assertEqual(1, len(records))
2349
 
        self.assertEqual([], basis.calls)
2350
 
        # Missing (from test knit) objects are retrieved from the basis:
2351
 
        basis.add_lines(key_basis, (), ['foo\n', 'bar\n'])
2352
 
        basis.calls = []
2353
 
        records = list(test.get_record_stream([key_basis, key_missing],
2354
 
            'unordered', False))
2355
 
        self.assertEqual(2, len(records))
2356
 
        calls = list(basis.calls)
2357
 
        for record in records:
2358
 
            self.assertSubset([record.key], (key_basis, key_missing))
2359
 
            if record.key == key_missing:
2360
 
                self.assertIsInstance(record, AbsentContentFactory)
2361
 
            else:
2362
 
                reference = list(basis.get_record_stream([key_basis],
2363
 
                    'unordered', False))[0]
2364
 
                self.assertEqual(reference.key, record.key)
2365
 
                self.assertEqual(reference.sha1, record.sha1)
2366
 
                self.assertEqual(reference.storage_kind, record.storage_kind)
2367
 
                self.assertEqual(reference.get_bytes_as(reference.storage_kind),
2368
 
                    record.get_bytes_as(record.storage_kind))
2369
 
        # It's not strictly minimal, but it seems reasonable for now for it to
2370
 
        # ask which fallbacks have which parents.
2371
 
        self.assertEqual([
2372
 
            ("get_parent_map", set([key_basis, key_missing])),
2373
 
            ("get_record_stream", [key_basis], 'unordered', False)],
2374
 
            calls)
2375
 
 
2376
 
    def test_get_record_stream_ordered_deltas(self):
2377
 
        # ordering is preserved down into the fallback store.
2378
 
        basis, test = self.get_basis_and_test_knit()
2379
 
        key = ('foo',)
2380
 
        key_basis = ('bar',)
2381
 
        key_basis_2 = ('quux',)
2382
 
        key_missing = ('missing',)
2383
 
        test.add_lines(key, (key_basis,), ['foo\n'])
2384
 
        # Missing (from test knit) objects are retrieved from the basis:
2385
 
        basis.add_lines(key_basis, (key_basis_2,), ['foo\n', 'bar\n'])
2386
 
        basis.add_lines(key_basis_2, (), ['quux\n'])
2387
 
        basis.calls = []
2388
 
        # ask for in non-topological order
2389
 
        records = list(test.get_record_stream(
2390
 
            [key, key_basis, key_missing, key_basis_2], 'topological', False))
2391
 
        self.assertEqual(4, len(records))
2392
 
        results = []
2393
 
        for record in records:
2394
 
            self.assertSubset([record.key],
2395
 
                (key_basis, key_missing, key_basis_2, key))
2396
 
            if record.key == key_missing:
2397
 
                self.assertIsInstance(record, AbsentContentFactory)
2398
 
            else:
2399
 
                results.append((record.key, record.sha1, record.storage_kind,
2400
 
                    record.get_bytes_as(record.storage_kind)))
2401
 
        calls = list(basis.calls)
2402
 
        order = [record[0] for record in results]
2403
 
        self.assertEqual([key_basis_2, key_basis, key], order)
2404
 
        for result in results:
2405
 
            if result[0] == key:
2406
 
                source = test
2407
 
            else:
2408
 
                source = basis
2409
 
            record = source.get_record_stream([result[0]], 'unordered',
2410
 
                False).next()
2411
 
            self.assertEqual(record.key, result[0])
2412
 
            self.assertEqual(record.sha1, result[1])
2413
 
            self.assertEqual(record.storage_kind, result[2])
2414
 
            self.assertEqual(record.get_bytes_as(record.storage_kind), result[3])
2415
 
        # It's not strictly minimal, but it seems reasonable for now for it to
2416
 
        # ask which fallbacks have which parents.
2417
 
        self.assertEqual([
2418
 
            ("get_parent_map", set([key_basis, key_basis_2, key_missing])),
2419
 
            ("get_record_stream", [key_basis_2, key_basis], 'topological', False)],
2420
 
            calls)
2421
 
 
2422
 
    def test_get_sha1s(self):
2423
 
        # sha1's in the test knit are answered without asking the basis
2424
 
        basis, test = self.get_basis_and_test_knit()
2425
 
        key = ('foo',)
2426
 
        key_basis = ('bar',)
2427
 
        key_missing = ('missing',)
2428
 
        test.add_lines(key, (), ['foo\n'])
2429
 
        key_sha1sum = osutils.sha('foo\n').hexdigest()
2430
 
        sha1s = test.get_sha1s([key])
2431
 
        self.assertEqual({key: key_sha1sum}, sha1s)
2432
 
        self.assertEqual([], basis.calls)
2433
 
        # But texts that are not in the test knit are looked for in the basis
2434
 
        # directly (rather than via text reconstruction) so that remote servers
2435
 
        # etc don't have to answer with full content.
2436
 
        basis.add_lines(key_basis, (), ['foo\n', 'bar\n'])
2437
 
        basis_sha1sum = osutils.sha('foo\nbar\n').hexdigest()
2438
 
        basis.calls = []
2439
 
        sha1s = test.get_sha1s([key, key_missing, key_basis])
2440
 
        self.assertEqual({key: key_sha1sum,
2441
 
            key_basis: basis_sha1sum}, sha1s)
2442
 
        self.assertEqual([("get_sha1s", set([key_basis, key_missing]))],
2443
 
            basis.calls)
2444
 
 
2445
 
    def test_insert_record_stream(self):
2446
 
        # records are inserted as normal; insert_record_stream builds on
2447
 
        # add_lines, so a smoke test should be all that's needed:
2448
 
        key = ('foo',)
2449
 
        key_basis = ('bar',)
2450
 
        key_delta = ('zaphod',)
2451
 
        basis, test = self.get_basis_and_test_knit()
2452
 
        source = self.make_test_knit(name='source')
2453
 
        basis.add_lines(key_basis, (), ['foo\n'])
2454
 
        basis.calls = []
2455
 
        source.add_lines(key_basis, (), ['foo\n'])
2456
 
        source.add_lines(key_delta, (key_basis,), ['bar\n'])
2457
 
        stream = source.get_record_stream([key_delta], 'unordered', False)
2458
 
        test.insert_record_stream(stream)
2459
 
        # XXX: this does somewhat too many calls in making sure of whether it
2460
 
        # has to recreate the full text.
2461
 
        self.assertEqual([("get_parent_map", set([key_basis])),
2462
 
             ('get_parent_map', set([key_basis])),
2463
 
             ('get_record_stream', [key_basis], 'unordered', True)],
2464
 
            basis.calls)
2465
 
        self.assertEqual({key_delta:(key_basis,)},
2466
 
            test.get_parent_map([key_delta]))
2467
 
        self.assertEqual('bar\n', test.get_record_stream([key_delta],
2468
 
            'unordered', True).next().get_bytes_as('fulltext'))
2469
 
 
2470
 
    def test_iter_lines_added_or_present_in_keys(self):
2471
 
        # Lines from the basis are returned, and lines for a given key are only
2472
 
        # returned once.
2473
 
        key1 = ('foo1',)
2474
 
        key2 = ('foo2',)
2475
 
        # all sources are asked for keys:
2476
 
        basis, test = self.get_basis_and_test_knit()
2477
 
        basis.add_lines(key1, (), ["foo"])
2478
 
        basis.calls = []
2479
 
        lines = list(test.iter_lines_added_or_present_in_keys([key1]))
2480
 
        self.assertEqual([("foo\n", key1)], lines)
2481
 
        self.assertEqual([("iter_lines_added_or_present_in_keys", set([key1]))],
2482
 
            basis.calls)
2483
 
        # keys in both are not duplicated:
2484
 
        test.add_lines(key2, (), ["bar\n"])
2485
 
        basis.add_lines(key2, (), ["bar\n"])
2486
 
        basis.calls = []
2487
 
        lines = list(test.iter_lines_added_or_present_in_keys([key2]))
2488
 
        self.assertEqual([("bar\n", key2)], lines)
2489
 
        self.assertEqual([], basis.calls)
2490
 
 
2491
 
    def test_keys(self):
2492
 
        key1 = ('foo1',)
2493
 
        key2 = ('foo2',)
2494
 
        # all sources are asked for keys:
2495
 
        basis, test = self.get_basis_and_test_knit()
2496
 
        keys = test.keys()
2497
 
        self.assertEqual(set(), set(keys))
2498
 
        self.assertEqual([("keys",)], basis.calls)
2499
 
        # keys from a basis are returned:
2500
 
        basis.add_lines(key1, (), [])
2501
 
        basis.calls = []
2502
 
        keys = test.keys()
2503
 
        self.assertEqual(set([key1]), set(keys))
2504
 
        self.assertEqual([("keys",)], basis.calls)
2505
 
        # keys in both are not duplicated:
2506
 
        test.add_lines(key2, (), [])
2507
 
        basis.add_lines(key2, (), [])
2508
 
        basis.calls = []
2509
 
        keys = test.keys()
2510
 
        self.assertEqual(2, len(keys))
2511
 
        self.assertEqual(set([key1, key2]), set(keys))
2512
 
        self.assertEqual([("keys",)], basis.calls)
2513
 
 
2514
 
    def test_add_mpdiffs(self):
2515
 
        # records are inserted as normal; add_mpdiff builds on
2516
 
        # add_lines, so a smoke test should be all that's needed:
2517
 
        key = ('foo',)
2518
 
        key_basis = ('bar',)
2519
 
        key_delta = ('zaphod',)
2520
 
        basis, test = self.get_basis_and_test_knit()
2521
 
        source = self.make_test_knit(name='source')
2522
 
        basis.add_lines(key_basis, (), ['foo\n'])
2523
 
        basis.calls = []
2524
 
        source.add_lines(key_basis, (), ['foo\n'])
2525
 
        source.add_lines(key_delta, (key_basis,), ['bar\n'])
2526
 
        diffs = source.make_mpdiffs([key_delta])
2527
 
        test.add_mpdiffs([(key_delta, (key_basis,),
2528
 
            source.get_sha1s([key_delta])[key_delta], diffs[0])])
2529
 
        self.assertEqual([("get_parent_map", set([key_basis])),
2530
 
            ('get_record_stream', [key_basis], 'unordered', True),],
2531
 
            basis.calls)
2532
 
        self.assertEqual({key_delta:(key_basis,)},
2533
 
            test.get_parent_map([key_delta]))
2534
 
        self.assertEqual('bar\n', test.get_record_stream([key_delta],
2535
 
            'unordered', True).next().get_bytes_as('fulltext'))
2536
 
 
2537
 
    def test_make_mpdiffs(self):
2538
 
        # Generating an mpdiff across a stacking boundary should detect parent
2539
 
        # texts regions.
2540
 
        key = ('foo',)
2541
 
        key_left = ('bar',)
2542
 
        key_right = ('zaphod',)
2543
 
        basis, test = self.get_basis_and_test_knit()
2544
 
        basis.add_lines(key_left, (), ['bar\n'])
2545
 
        basis.add_lines(key_right, (), ['zaphod\n'])
2546
 
        basis.calls = []
2547
 
        test.add_lines(key, (key_left, key_right),
2548
 
            ['bar\n', 'foo\n', 'zaphod\n'])
2549
 
        diffs = test.make_mpdiffs([key])
2550
 
        self.assertEqual([
2551
 
            multiparent.MultiParent([multiparent.ParentText(0, 0, 0, 1),
2552
 
                multiparent.NewText(['foo\n']),
2553
 
                multiparent.ParentText(1, 0, 2, 1)])],
2554
 
            diffs)
2555
 
        self.assertEqual(3, len(basis.calls))
2556
 
        self.assertEqual([
2557
 
            ("get_parent_map", set([key_left, key_right])),
2558
 
            ("get_parent_map", set([key_left, key_right])),
2559
 
            ],
2560
 
            basis.calls[:-1])
2561
 
        last_call = basis.calls[-1]
2562
 
        self.assertEqual('get_record_stream', last_call[0])
2563
 
        self.assertEqual(set([key_left, key_right]), set(last_call[1]))
2564
 
        self.assertEqual('topological', last_call[2])
2565
 
        self.assertEqual(True, last_call[3])
2566
 
 
2567
 
 
2568
 
class TestNetworkBehaviour(KnitTests):
2569
 
    """Tests for getting data out of/into knits over the network."""
2570
 
 
2571
 
    def test_include_delta_closure_generates_a_knit_delta_closure(self):
2572
 
        vf = self.make_test_knit(name='test')
2573
 
        # put in three texts, giving ft, delta, delta
2574
 
        vf.add_lines(('base',), (), ['base\n', 'content\n'])
2575
 
        vf.add_lines(('d1',), (('base',),), ['d1\n'])
2576
 
        vf.add_lines(('d2',), (('d1',),), ['d2\n'])
2577
 
        # But heuristics could interfere, so check what happened:
2578
 
        self.assertEqual(['knit-ft-gz', 'knit-delta-gz', 'knit-delta-gz'],
2579
 
            [record.storage_kind for record in
2580
 
             vf.get_record_stream([('base',), ('d1',), ('d2',)],
2581
 
                'topological', False)])
2582
 
        # generate a stream of just the deltas include_delta_closure=True,
2583
 
        # serialise to the network, and check that we get a delta closure on the wire.
2584
 
        stream = vf.get_record_stream([('d1',), ('d2',)], 'topological', True)
2585
 
        netb = [record.get_bytes_as(record.storage_kind) for record in stream]
2586
 
        # The first bytes should be a memo from _ContentMapGenerator, and the
2587
 
        # second bytes should be empty (because its a API proxy not something
2588
 
        # for wire serialisation.
2589
 
        self.assertEqual('', netb[1])
2590
 
        bytes = netb[0]
2591
 
        kind, line_end = network_bytes_to_kind_and_offset(bytes)
2592
 
        self.assertEqual('knit-delta-closure', kind)
2593
 
 
2594
 
 
2595
 
class TestContentMapGenerator(KnitTests):
2596
 
    """Tests for ContentMapGenerator"""
2597
 
 
2598
 
    def test_get_record_stream_gives_records(self):
2599
 
        vf = self.make_test_knit(name='test')
2600
 
        # put in three texts, giving ft, delta, delta
2601
 
        vf.add_lines(('base',), (), ['base\n', 'content\n'])
2602
 
        vf.add_lines(('d1',), (('base',),), ['d1\n'])
2603
 
        vf.add_lines(('d2',), (('d1',),), ['d2\n'])
2604
 
        keys = [('d1',), ('d2',)]
2605
 
        generator = _VFContentMapGenerator(vf, keys,
2606
 
            global_map=vf.get_parent_map(keys))
2607
 
        for record in generator.get_record_stream():
2608
 
            if record.key == ('d1',):
2609
 
                self.assertEqual('d1\n', record.get_bytes_as('fulltext'))
2610
 
            else:
2611
 
                self.assertEqual('d2\n', record.get_bytes_as('fulltext'))
2612
 
 
2613
 
    def test_get_record_stream_kinds_are_raw(self):
2614
 
        vf = self.make_test_knit(name='test')
2615
 
        # put in three texts, giving ft, delta, delta
2616
 
        vf.add_lines(('base',), (), ['base\n', 'content\n'])
2617
 
        vf.add_lines(('d1',), (('base',),), ['d1\n'])
2618
 
        vf.add_lines(('d2',), (('d1',),), ['d2\n'])
2619
 
        keys = [('base',), ('d1',), ('d2',)]
2620
 
        generator = _VFContentMapGenerator(vf, keys,
2621
 
            global_map=vf.get_parent_map(keys))
2622
 
        kinds = {('base',): 'knit-delta-closure',
2623
 
            ('d1',): 'knit-delta-closure-ref',
2624
 
            ('d2',): 'knit-delta-closure-ref',
2625
 
            }
2626
 
        for record in generator.get_record_stream():
2627
 
            self.assertEqual(kinds[record.key], record.storage_kind)
 
2140
    def test_iter_parents(self):
 
2141
        index = self.two_graph_index()
 
2142
        self.assertEqual(set([
 
2143
            ('tip', ()), ('tail', ()), ('parent', ()), ('separate', ())
 
2144
            ]),
 
2145
            set(index.iter_parents(['tip', 'tail', 'ghost', 'parent', 'separate'])))
 
2146
        self.assertEqual(set([('tip', ())]),
 
2147
            set(index.iter_parents(['tip'])))
 
2148
        self.assertEqual(set(),
 
2149
            set(index.iter_parents([])))