~bzr-pqm/bzr/bzr.dev

« back to all changes in this revision

Viewing changes to bzrlib/graph.py

  • Committer: Andrew Bennetts
  • Date: 2011-05-18 15:45:07 UTC
  • mto: This revision was merged to the branch mainline in revision 5895.
  • Revision ID: andrew.bennetts@canonical.com-20110518154507-t3qudgcb7fj3omjc
Use pydoctor in api-docs make target.

Show diffs side-by-side

added added

removed removed

Lines of Context:
250
250
        common ancestor of all border ancestors, because this shows that it
251
251
        cannot be a descendant of any border ancestor.
252
252
 
253
 
        The scaling of this operation should be proportional to:
254
 
 
 
253
        The scaling of this operation should be proportional to
255
254
        1. The number of uncommon ancestors
256
255
        2. The number of border ancestors
257
256
        3. The length of the shortest path between a border ancestor and an
389
388
 
390
389
        :param unique_revision: The revision_id whose ancestry we are
391
390
            interested in.
392
 
            (XXX: Would this API be better if we allowed multiple revisions on
393
 
            to be searched here?)
 
391
            XXX: Would this API be better if we allowed multiple revisions on
 
392
                 to be searched here?
394
393
        :param common_revisions: Revision_ids of ancestries to exclude.
395
394
        :return: A set of revisions in the ancestry of unique_revision
396
395
        """
1564
1563
 
1565
1564
        The recipe allows reconstruction of the same results at a later date.
1566
1565
 
1567
 
        :return: A tuple of `(search_kind_str, *details)`.  The details vary by
 
1566
        :return: A tuple of (search_kind_str, *details).  The details vary by
1568
1567
            kind of search result.
1569
1568
        """
1570
1569
        raise NotImplementedError(self.get_recipe)
1849
1848
    """Find all revisions missing in one repo for a some specific heads."""
1850
1849
 
1851
1850
    def __init__(self, to_repo, from_repo, required_ids, if_present_ids=None,
1852
 
            find_ghosts=False, limit=None):
 
1851
            find_ghosts=False):
1853
1852
        """Constructor.
1854
1853
 
1855
1854
        :param required_ids: revision IDs of heads that must be found, or else
1859
1858
        :param if_present_ids: revision IDs of heads that may be absent in the
1860
1859
            source repository.  If present, then their ancestry not already
1861
1860
            found in other will be included in the search result.
1862
 
        :param limit: maximum number of revisions to fetch
1863
1861
        """
1864
1862
        self.to_repo = to_repo
1865
1863
        self.from_repo = from_repo
1866
1864
        self.find_ghosts = find_ghosts
1867
1865
        self.required_ids = required_ids
1868
1866
        self.if_present_ids = if_present_ids
1869
 
        self.limit = limit
1870
1867
 
1871
1868
    def __repr__(self):
1872
1869
        if len(self.required_ids) > 5:
1878
1875
        else:
1879
1876
            ifp_revs_repr = repr(self.if_present_ids)
1880
1877
 
1881
 
        return ("<%s from:%r to:%r find_ghosts:%r req'd:%r if-present:%r"
1882
 
                "limit:%r>") % (
1883
 
                self.__class__.__name__, self.from_repo, self.to_repo,
1884
 
                self.find_ghosts, reqd_revs_repr, ifp_revs_repr,
1885
 
                self.limit)
 
1878
        return "<%s from:%r to:%r find_ghosts:%r req'd:%r if-present:%r>" % (
 
1879
            self.__class__.__name__, self.from_repo, self.to_repo,
 
1880
            self.find_ghosts, reqd_revs_repr, ifp_revs_repr)
1886
1881
 
1887
1882
    def execute(self):
1888
1883
        return self.to_repo.search_missing_revision_ids(
1889
1884
            self.from_repo, revision_ids=self.required_ids,
1890
 
            if_present_ids=self.if_present_ids, find_ghosts=self.find_ghosts,
1891
 
            limit=self.limit)
 
1885
            if_present_ids=self.if_present_ids, find_ghosts=self.find_ghosts)
1892
1886
 
1893
1887
 
1894
1888
def collapse_linear_regions(parent_map):
1980
1974
        return set([h[0] for h in head_keys])
1981
1975
 
1982
1976
    def merge_sort(self, tip_revision):
1983
 
        nodes = self._graph.merge_sort((tip_revision,))
1984
 
        for node in nodes:
1985
 
            node.key = node.key[0]
1986
 
        return nodes
 
1977
        return self._graph.merge_sort((tip_revision,))
1987
1978
 
1988
1979
    def add_node(self, revision, parents):
1989
1980
        self._graph.add_node((revision,), [(p,) for p in parents])