~bzr-pqm/bzr/bzr.dev

« back to all changes in this revision

Viewing changes to bzrlib/groupcompress.py

  • Committer: Canonical.com Patch Queue Manager
  • Date: 2011-04-29 11:07:14 UTC
  • mfrom: (5813.1.1 realname-can-be-empty)
  • Revision ID: pqm@pqm.ubuntu.com-20110429110714-wr9f71ea9600lvb6
(jelmer) Allow realname to be empty in tests. (Jelmer Vernooij)

Show diffs side-by-side

added added

removed removed

Lines of Context:
1
 
# Copyright (C) 2008, 2009, 2010 Canonical Ltd
 
1
# Copyright (C) 2008-2011 Canonical Ltd
2
2
#
3
3
# This program is free software; you can redistribute it and/or modify
4
4
# it under the terms of the GNU General Public License as published by
23
23
except ImportError:
24
24
    pylzma = None
25
25
 
 
26
from bzrlib.lazy_import import lazy_import
 
27
lazy_import(globals(), """
26
28
from bzrlib import (
27
29
    annotate,
28
30
    debug,
29
31
    errors,
30
32
    graph as _mod_graph,
31
 
    knit,
32
33
    osutils,
33
34
    pack,
34
35
    static_tuple,
35
36
    trace,
 
37
    tsort,
36
38
    )
 
39
 
 
40
from bzrlib.repofmt import pack_repo
 
41
""")
 
42
 
37
43
from bzrlib.btree_index import BTreeBuilder
38
44
from bzrlib.lru_cache import LRUSizeCache
39
 
from bzrlib.tsort import topo_sort
40
45
from bzrlib.versionedfile import (
 
46
    _KeyRefs,
41
47
    adapter_registry,
42
48
    AbsentContentFactory,
43
49
    ChunkedContentFactory,
77
83
 
78
84
    present_keys = []
79
85
    for prefix in sorted(per_prefix_map):
80
 
        present_keys.extend(reversed(topo_sort(per_prefix_map[prefix])))
 
86
        present_keys.extend(reversed(tsort.topo_sort(per_prefix_map[prefix])))
81
87
    return present_keys
82
88
 
83
89
 
1046
1052
        index = _GCGraphIndex(graph_index, lambda:True, parents=parents,
1047
1053
            add_callback=graph_index.add_nodes,
1048
1054
            inconsistency_fatal=inconsistency_fatal)
1049
 
        access = knit._DirectPackAccess({})
 
1055
        access = pack_repo._DirectPackAccess({})
1050
1056
        access.set_writer(writer, graph_index, (transport, 'newpack'))
1051
1057
        result = GroupCompressVersionedFiles(index, access, delta)
1052
1058
        result.stream = stream
1186
1192
            _unadded_refs = {}
1187
1193
        self._unadded_refs = _unadded_refs
1188
1194
        self._group_cache = LRUSizeCache(max_size=50*1024*1024)
1189
 
        self._fallback_vfs = []
 
1195
        self._immediate_fallback_vfs = []
1190
1196
 
1191
1197
    def without_fallbacks(self):
1192
1198
        """Return a clone of this object without any fallbacks configured."""
1266
1272
 
1267
1273
        :param a_versioned_files: A VersionedFiles object.
1268
1274
        """
1269
 
        self._fallback_vfs.append(a_versioned_files)
 
1275
        self._immediate_fallback_vfs.append(a_versioned_files)
1270
1276
 
1271
1277
    def annotate(self, key):
1272
1278
        """See VersionedFiles.annotate."""
1312
1318
        # KnitVersionedFiles.get_known_graph_ancestry, but they don't share
1313
1319
        # ancestry.
1314
1320
        parent_map, missing_keys = self._index.find_ancestry(keys)
1315
 
        for fallback in self._fallback_vfs:
 
1321
        for fallback in self._transitive_fallbacks():
1316
1322
            if not missing_keys:
1317
1323
                break
1318
1324
            (f_parent_map, f_missing_keys) = fallback._index.find_ancestry(
1342
1348
            and so on.
1343
1349
        """
1344
1350
        result = {}
1345
 
        sources = [self._index] + self._fallback_vfs
 
1351
        sources = [self._index] + self._immediate_fallback_vfs
1346
1352
        source_results = []
1347
1353
        missing = set(keys)
1348
1354
        for source in sources:
1449
1455
        parent_map = {}
1450
1456
        key_to_source_map = {}
1451
1457
        source_results = []
1452
 
        for source in self._fallback_vfs:
 
1458
        for source in self._immediate_fallback_vfs:
1453
1459
            if not missing:
1454
1460
                break
1455
1461
            source_parents = source.get_parent_map(missing)
1470
1476
            the defined order, regardless of source.
1471
1477
        """
1472
1478
        if ordering == 'topological':
1473
 
            present_keys = topo_sort(parent_map)
 
1479
            present_keys = tsort.topo_sort(parent_map)
1474
1480
        else:
1475
1481
            # ordering == 'groupcompress'
1476
1482
            # XXX: This only optimizes for the target ordering. We may need
1832
1838
        """See VersionedFiles.keys."""
1833
1839
        if 'evil' in debug.debug_flags:
1834
1840
            trace.mutter_callsite(2, "keys scales with size of history")
1835
 
        sources = [self._index] + self._fallback_vfs
 
1841
        sources = [self._index] + self._immediate_fallback_vfs
1836
1842
        result = set()
1837
1843
        for source in sources:
1838
1844
            result.update(source.keys())
1921
1927
        # repeated over and over, this creates a surplus of ints
1922
1928
        self._int_cache = {}
1923
1929
        if track_external_parent_refs:
1924
 
            self._key_dependencies = knit._KeyRefs(
 
1930
            self._key_dependencies = _KeyRefs(
1925
1931
                track_new_keys=track_new_keys)
1926
1932
        else:
1927
1933
            self._key_dependencies = None