39
36
# At the moment they may alias the inventory and have old copies of it in
40
37
# memory. (Now done? -- mbp 20060309)
42
from binascii import hexlify
39
from cStringIO import StringIO
42
from bzrlib.lazy_import import lazy_import
43
lazy_import(globals(), """
44
from bisect import bisect_left
44
46
from copy import deepcopy
45
from cStringIO import StringIO
51
51
from time import time
53
from bzrlib.atomicfile import AtomicFile
54
from bzrlib.branch import (Branch,
56
from bzrlib.conflicts import Conflict, ConflictList, CONFLICT_SUFFIXES
57
import bzrlib.bzrdir as bzrdir
59
conflicts as _mod_conflicts,
78
from bzrlib.transport import get_transport
80
from bzrlib.workingtree_4 import WorkingTreeFormat4
83
from bzrlib import symbol_versioning
58
84
from bzrlib.decorators import needs_read_lock, needs_write_lock
59
import bzrlib.errors as errors
60
from bzrlib.errors import (BzrCheckError,
64
WeaveRevisionNotPresent,
68
MergeModifiedFormatError,
71
from bzrlib.inventory import InventoryEntry, Inventory
85
from bzrlib.inventory import InventoryEntry, Inventory, ROOT_ID, TreeReference
72
86
from bzrlib.lockable_files import LockableFiles, TransportLock
73
87
from bzrlib.lockdir import LockDir
74
from bzrlib.merge import merge_inner, transform_tree
88
import bzrlib.mutabletree
89
from bzrlib.mutabletree import needs_tree_write_lock
75
90
from bzrlib.osutils import (
102
from bzrlib.trace import mutter, note
103
from bzrlib.transport.local import LocalTransport
92
104
from bzrlib.progress import DummyProgress, ProgressPhase
93
from bzrlib.revision import NULL_REVISION
105
from bzrlib.revision import NULL_REVISION, CURRENT_REVISION
94
106
from bzrlib.rio import RioReader, rio_file, Stanza
95
from bzrlib.symbol_versioning import *
96
from bzrlib.textui import show_status
98
from bzrlib.transform import build_tree
99
from bzrlib.trace import mutter, note
100
from bzrlib.transport import get_transport
101
from bzrlib.transport.local import LocalTransport
106
# the regex here does the following:
107
# 1) remove any weird characters; we don't escape them but rather
109
# 2) match leading '.'s to make it not hidden
110
_gen_file_id_re = re.compile(r'[^\w.]|(^\.*)')
111
_gen_id_suffix = None
115
def _next_id_suffix():
116
"""Create a new file id suffix that is reasonably unique.
118
On the first call we combine the current time with 64 bits of randomness
119
to give a highly probably globally unique number. Then each call in the same
120
process adds 1 to a serial number we append to that unique value.
122
# XXX TODO: change bzrlib.add.smart_add to call workingtree.add() rather
123
# than having to move the id randomness out of the inner loop like this.
124
# XXX TODO: for the global randomness this uses we should add the thread-id
125
# before the serial #.
126
global _gen_id_suffix, _gen_id_serial
127
if _gen_id_suffix is None:
128
_gen_id_suffix = "-%s-%s-" % (compact_date(time()), rand_chars(16))
130
return _gen_id_suffix + str(_gen_id_serial)
107
from bzrlib.symbol_versioning import (deprecated_passed,
110
DEPRECATED_PARAMETER,
117
MERGE_MODIFIED_HEADER_1 = "BZR merge-modified list format 1"
118
CONFLICT_HEADER_1 = "BZR conflict list format 1"
121
@deprecated_function(zero_thirteen)
133
122
def gen_file_id(name):
134
123
"""Return new file id for the basename 'name'.
136
The uniqueness is supplied from _next_id_suffix.
125
Use bzrlib.generate_ids.gen_file_id() instead
138
# XXX TODO: squash the filename to lowercase.
139
# XXX TODO: truncate the filename to something like 20 or 30 chars.
140
# XXX TODO: consider what to do with ids that look like illegal filepaths
141
# on platforms we support.
142
return _gen_file_id_re.sub('', name) + _next_id_suffix()
127
return generate_ids.gen_file_id(name)
130
@deprecated_function(zero_thirteen)
145
131
def gen_root_id():
146
"""Return a new tree-root file id."""
147
return gen_file_id('TREE_ROOT')
132
"""Return a new tree-root file id.
134
This has been deprecated in favor of bzrlib.generate_ids.gen_root_id()
136
return generate_ids.gen_root_id()
150
139
class TreeEntry(object):
151
"""An entry that implements the minium interface used by commands.
140
"""An entry that implements the minimum interface used by commands.
153
142
This needs further inspection, it may be better to have
154
143
InventoryEntries without ids - though that seems wrong. For now,
527
600
__contains__ = has_id
529
602
def get_file_size(self, file_id):
603
file_id = osutils.safe_file_id(file_id)
530
604
return os.path.getsize(self.id2abspath(file_id))
533
def get_file_sha1(self, file_id, path=None):
607
def get_file_sha1(self, file_id, path=None, stat_value=None):
608
file_id = osutils.safe_file_id(file_id)
535
610
path = self._inventory.id2path(file_id)
536
return self._hashcache.get_sha1(path)
611
return self._hashcache.get_sha1(path, stat_value)
613
def get_file_mtime(self, file_id, path=None):
614
file_id = osutils.safe_file_id(file_id)
616
path = self.inventory.id2path(file_id)
617
return os.lstat(self.abspath(path)).st_mtime
538
619
if not supports_executable():
539
620
def is_executable(self, file_id, path=None):
621
file_id = osutils.safe_file_id(file_id)
540
622
return self._inventory[file_id].executable
542
624
def is_executable(self, file_id, path=None):
544
path = self._inventory.id2path(file_id)
626
file_id = osutils.safe_file_id(file_id)
627
path = self.id2path(file_id)
545
628
mode = os.lstat(self.abspath(path)).st_mode
546
return bool(stat.S_ISREG(mode) and stat.S_IEXEC&mode)
549
def add(self, files, ids=None):
550
"""Make files versioned.
552
Note that the command line normally calls smart_add instead,
553
which can automatically recurse.
555
This adds the files to the inventory, so that they will be
556
recorded by the next commit.
559
List of paths to add, relative to the base of the tree.
562
If set, use these instead of automatically generated ids.
563
Must be the same length as the list of files, but may
564
contain None for ids that are to be autogenerated.
566
TODO: Perhaps have an option to add the ids even if the files do
569
TODO: Perhaps callback with the ids and paths as they're added.
629
return bool(stat.S_ISREG(mode) and stat.S_IEXEC & mode)
631
@needs_tree_write_lock
632
def _add(self, files, ids, kinds):
633
"""See MutableTree._add."""
571
634
# TODO: Re-adding a file that is removed in the working copy
572
635
# should probably put it back with the previous ID.
573
if isinstance(files, basestring):
574
assert(ids is None or isinstance(ids, basestring))
580
ids = [None] * len(files)
582
assert(len(ids) == len(files))
636
# the read and write working inventory should not occur in this
637
# function - they should be part of lock_write and unlock.
584
638
inv = self.read_working_inventory()
585
for f,file_id in zip(files, ids):
586
if self.is_control_filename(f):
587
raise BzrError("cannot add control file %s" % quotefn(f))
592
raise BzrError("cannot add top-level %r" % f)
594
fullpath = normpath(self.abspath(f))
597
kind = file_kind(fullpath)
599
if e.errno == errno.ENOENT:
600
raise NoSuchFile(fullpath)
601
# maybe something better?
602
raise BzrError('cannot add: not a regular file, symlink or directory: %s' % quotefn(f))
604
if not InventoryEntry.versionable_kind(kind):
605
raise BzrError('cannot add: not a versionable file ('
606
'i.e. regular file, symlink or directory): %s' % quotefn(f))
639
for f, file_id, kind in zip(files, ids, kinds):
640
assert kind is not None
608
641
if file_id is None:
609
642
inv.add_path(f, kind=kind)
644
file_id = osutils.safe_file_id(file_id)
611
645
inv.add_path(f, kind=kind, file_id=file_id)
613
646
self._write_inventory(inv)
648
@needs_tree_write_lock
649
def _gather_kinds(self, files, kinds):
650
"""See MutableTree._gather_kinds."""
651
for pos, f in enumerate(files):
652
if kinds[pos] is None:
653
fullpath = normpath(self.abspath(f))
655
kinds[pos] = file_kind(fullpath)
657
if e.errno == errno.ENOENT:
658
raise errors.NoSuchFile(fullpath)
615
660
@needs_write_lock
661
def add_parent_tree_id(self, revision_id, allow_leftmost_as_ghost=False):
662
"""Add revision_id as a parent.
664
This is equivalent to retrieving the current list of parent ids
665
and setting the list to its value plus revision_id.
667
:param revision_id: The revision id to add to the parent list. It may
668
be a ghost revision as long as its not the first parent to be added,
669
or the allow_leftmost_as_ghost parameter is set True.
670
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
672
parents = self.get_parent_ids() + [revision_id]
673
self.set_parent_ids(parents, allow_leftmost_as_ghost=len(parents) > 1
674
or allow_leftmost_as_ghost)
676
@needs_tree_write_lock
677
def add_parent_tree(self, parent_tuple, allow_leftmost_as_ghost=False):
678
"""Add revision_id, tree tuple as a parent.
680
This is equivalent to retrieving the current list of parent trees
681
and setting the list to its value plus parent_tuple. See also
682
add_parent_tree_id - if you only have a parent id available it will be
683
simpler to use that api. If you have the parent already available, using
684
this api is preferred.
686
:param parent_tuple: The (revision id, tree) to add to the parent list.
687
If the revision_id is a ghost, pass None for the tree.
688
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
690
parent_ids = self.get_parent_ids() + [parent_tuple[0]]
691
if len(parent_ids) > 1:
692
# the leftmost may have already been a ghost, preserve that if it
694
allow_leftmost_as_ghost = True
695
self.set_parent_ids(parent_ids,
696
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
698
@needs_tree_write_lock
616
699
def add_pending_merge(self, *revision_ids):
617
700
# TODO: Perhaps should check at this point that the
618
701
# history of the revision is actually present?
619
p = self.pending_merges()
702
parents = self.get_parent_ids()
621
704
for rev_id in revision_ids:
705
if rev_id in parents:
707
parents.append(rev_id)
627
self.set_pending_merges(p)
710
self.set_parent_ids(parents, allow_leftmost_as_ghost=True)
712
@deprecated_method(zero_eleven)
630
714
def pending_merges(self):
631
715
"""Return a list of pending merges.
633
717
These are revisions that have been merged into the working
634
718
directory but not yet committed.
637
merges_file = self._control_files.get_utf8('pending-merges')
639
if e.errno != errno.ENOENT:
643
for l in merges_file.readlines():
644
p.append(l.rstrip('\n'))
720
As of 0.11 this is deprecated. Please see WorkingTree.get_parent_ids()
721
instead - which is available on all tree objects.
723
return self.get_parent_ids()[1:]
725
def _check_parents_for_ghosts(self, revision_ids, allow_leftmost_as_ghost):
726
"""Common ghost checking functionality from set_parent_*.
728
This checks that the left hand-parent exists if there are any
731
if len(revision_ids) > 0:
732
leftmost_id = revision_ids[0]
733
if (not allow_leftmost_as_ghost and not
734
self.branch.repository.has_revision(leftmost_id)):
735
raise errors.GhostRevisionUnusableHere(leftmost_id)
737
def _set_merges_from_parent_ids(self, parent_ids):
738
merges = parent_ids[1:]
739
self._control_files.put_bytes('pending-merges', '\n'.join(merges))
741
@needs_tree_write_lock
742
def set_parent_ids(self, revision_ids, allow_leftmost_as_ghost=False):
743
"""Set the parent ids to revision_ids.
745
See also set_parent_trees. This api will try to retrieve the tree data
746
for each element of revision_ids from the trees repository. If you have
747
tree data already available, it is more efficient to use
748
set_parent_trees rather than set_parent_ids. set_parent_ids is however
749
an easier API to use.
751
:param revision_ids: The revision_ids to set as the parent ids of this
752
working tree. Any of these may be ghosts.
754
revision_ids = [osutils.safe_revision_id(r) for r in revision_ids]
755
self._check_parents_for_ghosts(revision_ids,
756
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
758
if len(revision_ids) > 0:
759
self.set_last_revision(revision_ids[0])
761
self.set_last_revision(None)
763
self._set_merges_from_parent_ids(revision_ids)
765
@needs_tree_write_lock
766
def set_parent_trees(self, parents_list, allow_leftmost_as_ghost=False):
767
"""See MutableTree.set_parent_trees."""
768
parent_ids = [osutils.safe_revision_id(rev) for (rev, tree) in parents_list]
770
self._check_parents_for_ghosts(parent_ids,
771
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
773
if len(parent_ids) == 0:
774
leftmost_parent_id = None
775
leftmost_parent_tree = None
777
leftmost_parent_id, leftmost_parent_tree = parents_list[0]
779
if self._change_last_revision(leftmost_parent_id):
780
if leftmost_parent_tree is None:
781
# If we don't have a tree, fall back to reading the
782
# parent tree from the repository.
783
self._cache_basis_inventory(leftmost_parent_id)
785
inv = leftmost_parent_tree.inventory
786
xml = self._create_basis_xml_from_inventory(
787
leftmost_parent_id, inv)
788
self._write_basis_inventory(xml)
789
self._set_merges_from_parent_ids(parent_ids)
791
@needs_tree_write_lock
648
792
def set_pending_merges(self, rev_list):
649
self._control_files.put_utf8('pending-merges', '\n'.join(rev_list))
793
parents = self.get_parent_ids()
794
leftmost = parents[:1]
795
new_parents = leftmost + rev_list
796
self.set_parent_ids(new_parents)
798
@needs_tree_write_lock
652
799
def set_merge_modified(self, modified_hashes):
653
800
def iter_stanzas():
654
801
for file_id, hash in modified_hashes.iteritems():
655
yield Stanza(file_id=file_id, hash=hash)
802
yield Stanza(file_id=file_id.decode('utf8'), hash=hash)
656
803
self._put_rio('merge-hashes', iter_stanzas(), MERGE_MODIFIED_HEADER_1)
659
805
def _put_rio(self, filename, stanzas, header):
806
self._must_be_locked()
660
807
my_file = rio_file(stanzas, header)
661
808
self._control_files.put(filename, my_file)
810
@needs_write_lock # because merge pulls data into the branch.
811
def merge_from_branch(self, branch, to_revision=None):
812
"""Merge from a branch into this working tree.
814
:param branch: The branch to merge from.
815
:param to_revision: If non-None, the merge will merge to to_revision,
816
but not beyond it. to_revision does not need to be in the history
817
of the branch when it is supplied. If None, to_revision defaults to
818
branch.last_revision().
820
from bzrlib.merge import Merger, Merge3Merger
821
pb = bzrlib.ui.ui_factory.nested_progress_bar()
823
merger = Merger(self.branch, this_tree=self, pb=pb)
824
merger.pp = ProgressPhase("Merge phase", 5, pb)
825
merger.pp.next_phase()
826
# check that there are no
828
merger.check_basis(check_clean=True, require_commits=False)
829
if to_revision is None:
830
to_revision = branch.last_revision()
832
to_revision = osutils.safe_revision_id(to_revision)
833
merger.other_rev_id = to_revision
834
if merger.other_rev_id is None:
835
raise error.NoCommits(branch)
836
self.branch.fetch(branch, last_revision=merger.other_rev_id)
837
merger.other_basis = merger.other_rev_id
838
merger.other_tree = self.branch.repository.revision_tree(
840
merger.other_branch = branch
841
merger.pp.next_phase()
843
if merger.base_rev_id == merger.other_rev_id:
844
raise errors.PointlessMerge
845
merger.backup_files = False
846
merger.merge_type = Merge3Merger
847
merger.set_interesting_files(None)
848
merger.show_base = False
849
merger.reprocess = False
850
conflicts = merger.do_merge()
664
857
def merge_modified(self):
858
"""Return a dictionary of files modified by a merge.
860
The list is initialized by WorkingTree.set_merge_modified, which is
861
typically called after we make some automatic updates to the tree
864
This returns a map of file_id->sha1, containing only files which are
865
still in the working inventory and have that text hash.
666
868
hashfile = self._control_files.get('merge-hashes')
869
except errors.NoSuchFile:
669
871
merge_hashes = {}
671
873
if hashfile.next() != MERGE_MODIFIED_HEADER_1 + '\n':
672
raise MergeModifiedFormatError()
874
raise errors.MergeModifiedFormatError()
673
875
except StopIteration:
674
raise MergeModifiedFormatError()
876
raise errors.MergeModifiedFormatError()
675
877
for s in RioReader(hashfile):
676
file_id = s.get("file_id")
878
# RioReader reads in Unicode, so convert file_ids back to utf8
879
file_id = osutils.safe_file_id(s.get("file_id"), warn=False)
677
880
if file_id not in self.inventory:
680
if hash == self.get_file_sha1(file_id):
681
merge_hashes[file_id] = hash
882
text_hash = s.get("hash")
883
if text_hash == self.get_file_sha1(file_id):
884
merge_hashes[file_id] = text_hash
682
885
return merge_hashes
888
def mkdir(self, path, file_id=None):
889
"""See MutableTree.mkdir()."""
891
file_id = generate_ids.gen_file_id(os.path.basename(path))
892
os.mkdir(self.abspath(path))
893
self.add(path, file_id, 'directory')
684
896
def get_symlink_target(self, file_id):
897
file_id = osutils.safe_file_id(file_id)
685
898
return os.readlink(self.id2abspath(file_id))
687
def file_class(self, filename):
688
if self.path2id(filename):
690
elif self.is_ignored(filename):
695
def list_files(self):
901
def subsume(self, other_tree):
902
def add_children(inventory, entry):
903
for child_entry in entry.children.values():
904
inventory._byid[child_entry.file_id] = child_entry
905
if child_entry.kind == 'directory':
906
add_children(inventory, child_entry)
907
if other_tree.get_root_id() == self.get_root_id():
908
raise errors.BadSubsumeSource(self, other_tree,
909
'Trees have the same root')
911
other_tree_path = self.relpath(other_tree.basedir)
912
except errors.PathNotChild:
913
raise errors.BadSubsumeSource(self, other_tree,
914
'Tree is not contained by the other')
915
new_root_parent = self.path2id(osutils.dirname(other_tree_path))
916
if new_root_parent is None:
917
raise errors.BadSubsumeSource(self, other_tree,
918
'Parent directory is not versioned.')
919
# We need to ensure that the result of a fetch will have a
920
# versionedfile for the other_tree root, and only fetching into
921
# RepositoryKnit2 guarantees that.
922
if not self.branch.repository.supports_rich_root():
923
raise errors.SubsumeTargetNeedsUpgrade(other_tree)
924
other_tree.lock_tree_write()
926
new_parents = other_tree.get_parent_ids()
927
other_root = other_tree.inventory.root
928
other_root.parent_id = new_root_parent
929
other_root.name = osutils.basename(other_tree_path)
930
self.inventory.add(other_root)
931
add_children(self.inventory, other_root)
932
self._write_inventory(self.inventory)
933
# normally we don't want to fetch whole repositories, but i think
934
# here we really do want to consolidate the whole thing.
935
for parent_id in other_tree.get_parent_ids():
936
self.branch.fetch(other_tree.branch, parent_id)
937
self.add_parent_tree_id(parent_id)
940
other_tree.bzrdir.retire_bzrdir()
942
@needs_tree_write_lock
943
def extract(self, file_id, format=None):
944
"""Extract a subtree from this tree.
946
A new branch will be created, relative to the path for this tree.
949
segments = osutils.splitpath(path)
950
transport = self.branch.bzrdir.root_transport
951
for name in segments:
952
transport = transport.clone(name)
955
except errors.FileExists:
959
sub_path = self.id2path(file_id)
960
branch_transport = mkdirs(sub_path)
962
format = bzrdir.format_registry.make_bzrdir('dirstate-with-subtree')
964
branch_transport.mkdir('.')
965
except errors.FileExists:
967
branch_bzrdir = format.initialize_on_transport(branch_transport)
969
repo = branch_bzrdir.find_repository()
970
except errors.NoRepositoryPresent:
971
repo = branch_bzrdir.create_repository()
972
assert repo.supports_rich_root()
974
if not repo.supports_rich_root():
975
raise errors.RootNotRich()
976
new_branch = branch_bzrdir.create_branch()
977
new_branch.pull(self.branch)
978
for parent_id in self.get_parent_ids():
979
new_branch.fetch(self.branch, parent_id)
980
tree_transport = self.bzrdir.root_transport.clone(sub_path)
981
if tree_transport.base != branch_transport.base:
982
tree_bzrdir = format.initialize_on_transport(tree_transport)
983
branch.BranchReferenceFormat().initialize(tree_bzrdir, new_branch)
985
tree_bzrdir = branch_bzrdir
986
wt = tree_bzrdir.create_workingtree(NULL_REVISION)
987
wt.set_parent_ids(self.get_parent_ids())
988
my_inv = self.inventory
989
child_inv = Inventory(root_id=None)
990
new_root = my_inv[file_id]
991
my_inv.remove_recursive_id(file_id)
992
new_root.parent_id = None
993
child_inv.add(new_root)
994
self._write_inventory(my_inv)
995
wt._write_inventory(child_inv)
998
def _serialize(self, inventory, out_file):
999
xml5.serializer_v5.write_inventory(self._inventory, out_file)
1001
def _deserialize(selt, in_file):
1002
return xml5.serializer_v5.read_inventory(in_file)
1005
"""Write the in memory inventory to disk."""
1006
# TODO: Maybe this should only write on dirty ?
1007
if self._control_files._lock_mode != 'w':
1008
raise errors.NotWriteLocked(self)
1010
self._serialize(self._inventory, sio)
1012
self._control_files.put('inventory', sio)
1013
self._inventory_is_modified = False
1015
def list_files(self, include_root=False):
696
1016
"""Recursively list all files as (path, class, kind, id, entry).
698
1018
Lists, but does not descend into unversioned directories.
777
1122
new_children.sort()
778
1123
new_children = collections.deque(new_children)
779
1124
stack.append((f_ie.file_id, fp, fap, new_children))
780
# Break out of inner loop, so that we start outer loop with child
1125
# Break out of inner loop,
1126
# so that we start outer loop with child
783
1129
# if we finished all children, pop it off the stack
788
def move(self, from_paths, to_name):
1132
@needs_tree_write_lock
1133
def move(self, from_paths, to_dir=None, after=False, **kwargs):
789
1134
"""Rename files.
791
to_name must exist in the inventory.
1136
to_dir must exist in the inventory.
793
If to_name exists and is a directory, the files are moved into
1138
If to_dir exists and is a directory, the files are moved into
794
1139
it, keeping their old names.
796
Note that to_name is only the last component of the new name;
1141
Note that to_dir is only the last component of the new name;
797
1142
this doesn't change the directory.
1144
For each entry in from_paths the move mode will be determined
1147
The first mode moves the file in the filesystem and updates the
1148
inventory. The second mode only updates the inventory without
1149
touching the file on the filesystem. This is the new mode introduced
1152
move uses the second mode if 'after == True' and the target is not
1153
versioned but present in the working tree.
1155
move uses the second mode if 'after == False' and the source is
1156
versioned but no longer in the working tree, and the target is not
1157
versioned but present in the working tree.
1159
move uses the first mode if 'after == False' and the source is
1160
versioned and present in the working tree, and the target is not
1161
versioned and not present in the working tree.
1163
Everything else results in an error.
799
1165
This returns a list of (from_path, to_path) pairs for each
800
1166
entry that is moved.
803
## TODO: Option to move IDs only
1171
# check for deprecated use of signature
1173
to_dir = kwargs.get('to_name', None)
1175
raise TypeError('You must supply a target directory')
1177
symbol_versioning.warn('The parameter to_name was deprecated'
1178
' in version 0.13. Use to_dir instead',
1181
# check destination directory
804
1182
assert not isinstance(from_paths, basestring)
805
1183
inv = self.inventory
806
to_abs = self.abspath(to_name)
1184
to_abs = self.abspath(to_dir)
807
1185
if not isdir(to_abs):
808
raise BzrError("destination %r is not a directory" % to_abs)
809
if not self.has_filename(to_name):
810
raise BzrError("destination %r not in working directory" % to_abs)
811
to_dir_id = inv.path2id(to_name)
812
if to_dir_id == None and to_name != '':
813
raise BzrError("destination %r is not a versioned directory" % to_name)
1186
raise errors.BzrMoveFailedError('',to_dir,
1187
errors.NotADirectory(to_abs))
1188
if not self.has_filename(to_dir):
1189
raise errors.BzrMoveFailedError('',to_dir,
1190
errors.NotInWorkingDirectory(to_dir))
1191
to_dir_id = inv.path2id(to_dir)
1192
if to_dir_id is None:
1193
raise errors.BzrMoveFailedError('',to_dir,
1194
errors.NotVersionedError(path=str(to_dir)))
814
1196
to_dir_ie = inv[to_dir_id]
815
if to_dir_ie.kind not in ('directory', 'root_directory'):
816
raise BzrError("destination %r is not a directory" % to_abs)
818
to_idpath = inv.get_idpath(to_dir_id)
821
if not self.has_filename(f):
822
raise BzrError("%r does not exist in working tree" % f)
823
f_id = inv.path2id(f)
825
raise BzrError("%r is not versioned" % f)
826
name_tail = splitpath(f)[-1]
827
dest_path = pathjoin(to_name, name_tail)
828
if self.has_filename(dest_path):
829
raise BzrError("destination %r already exists" % dest_path)
830
if f_id in to_idpath:
831
raise BzrError("can't move %r to a subdirectory of itself" % f)
833
# OK, so there's a race here, it's possible that someone will
834
# create a file in this interval and then the rename might be
835
# left half-done. But we should have caught most problems.
836
orig_inv = deepcopy(self.inventory)
1197
if to_dir_ie.kind != 'directory':
1198
raise errors.BzrMoveFailedError('',to_dir,
1199
errors.NotADirectory(to_abs))
1201
# create rename entries and tuples
1202
for from_rel in from_paths:
1203
from_tail = splitpath(from_rel)[-1]
1204
from_id = inv.path2id(from_rel)
1206
raise errors.BzrMoveFailedError(from_rel,to_dir,
1207
errors.NotVersionedError(path=str(from_rel)))
1209
from_entry = inv[from_id]
1210
from_parent_id = from_entry.parent_id
1211
to_rel = pathjoin(to_dir, from_tail)
1212
rename_entry = WorkingTree._RenameEntry(from_rel=from_rel,
1214
from_tail=from_tail,
1215
from_parent_id=from_parent_id,
1216
to_rel=to_rel, to_tail=from_tail,
1217
to_parent_id=to_dir_id)
1218
rename_entries.append(rename_entry)
1219
rename_tuples.append((from_rel, to_rel))
1221
# determine which move mode to use. checks also for movability
1222
rename_entries = self._determine_mv_mode(rename_entries, after)
1224
original_modified = self._inventory_is_modified
839
name_tail = splitpath(f)[-1]
840
dest_path = pathjoin(to_name, name_tail)
841
result.append((f, dest_path))
842
inv.rename(inv.path2id(f), to_dir_id, name_tail)
844
rename(self.abspath(f), self.abspath(dest_path))
846
raise BzrError("failed to rename %r to %r: %s" %
847
(f, dest_path, e[1]),
848
["rename rolled back"])
1227
self._inventory_is_modified = True
1228
self._move(rename_entries)
850
1230
# restore the inventory on error
851
self._set_inventory(orig_inv)
1231
self._inventory_is_modified = original_modified
853
1233
self._write_inventory(inv)
857
def rename_one(self, from_rel, to_rel):
1234
return rename_tuples
1236
def _determine_mv_mode(self, rename_entries, after=False):
1237
"""Determines for each from-to pair if both inventory and working tree
1238
or only the inventory has to be changed.
1240
Also does basic plausability tests.
1242
inv = self.inventory
1244
for rename_entry in rename_entries:
1245
# store to local variables for easier reference
1246
from_rel = rename_entry.from_rel
1247
from_id = rename_entry.from_id
1248
to_rel = rename_entry.to_rel
1249
to_id = inv.path2id(to_rel)
1250
only_change_inv = False
1252
# check the inventory for source and destination
1254
raise errors.BzrMoveFailedError(from_rel,to_rel,
1255
errors.NotVersionedError(path=str(from_rel)))
1256
if to_id is not None:
1257
raise errors.BzrMoveFailedError(from_rel,to_rel,
1258
errors.AlreadyVersionedError(path=str(to_rel)))
1260
# try to determine the mode for rename (only change inv or change
1261
# inv and file system)
1263
if not self.has_filename(to_rel):
1264
raise errors.BzrMoveFailedError(from_id,to_rel,
1265
errors.NoSuchFile(path=str(to_rel),
1266
extra="New file has not been created yet"))
1267
only_change_inv = True
1268
elif not self.has_filename(from_rel) and self.has_filename(to_rel):
1269
only_change_inv = True
1270
elif self.has_filename(from_rel) and not self.has_filename(to_rel):
1271
only_change_inv = False
1273
# something is wrong, so lets determine what exactly
1274
if not self.has_filename(from_rel) and \
1275
not self.has_filename(to_rel):
1276
raise errors.BzrRenameFailedError(from_rel,to_rel,
1277
errors.PathsDoNotExist(paths=(str(from_rel),
1280
raise errors.RenameFailedFilesExist(from_rel, to_rel,
1281
extra="(Use --after to update the Bazaar id)")
1282
rename_entry.only_change_inv = only_change_inv
1283
return rename_entries
1285
def _move(self, rename_entries):
1286
"""Moves a list of files.
1288
Depending on the value of the flag 'only_change_inv', the
1289
file will be moved on the file system or not.
1291
inv = self.inventory
1294
for entry in rename_entries:
1296
self._move_entry(entry)
1298
self._rollback_move(moved)
1302
def _rollback_move(self, moved):
1303
"""Try to rollback a previous move in case of an filesystem error."""
1304
inv = self.inventory
1307
self._move_entry(_RenameEntry(entry.to_rel, entry.from_id,
1308
entry.to_tail, entry.to_parent_id, entry.from_rel,
1309
entry.from_tail, entry.from_parent_id,
1310
entry.only_change_inv))
1311
except errors.BzrMoveFailedError, e:
1312
raise errors.BzrMoveFailedError( '', '', "Rollback failed."
1313
" The working tree is in an inconsistent state."
1314
" Please consider doing a 'bzr revert'."
1315
" Error message is: %s" % e)
1317
def _move_entry(self, entry):
1318
inv = self.inventory
1319
from_rel_abs = self.abspath(entry.from_rel)
1320
to_rel_abs = self.abspath(entry.to_rel)
1321
if from_rel_abs == to_rel_abs:
1322
raise errors.BzrMoveFailedError(entry.from_rel, entry.to_rel,
1323
"Source and target are identical.")
1325
if not entry.only_change_inv:
1327
osutils.rename(from_rel_abs, to_rel_abs)
1329
raise errors.BzrMoveFailedError(entry.from_rel,
1331
inv.rename(entry.from_id, entry.to_parent_id, entry.to_tail)
1333
@needs_tree_write_lock
1334
def rename_one(self, from_rel, to_rel, after=False):
858
1335
"""Rename one file.
860
1337
This can change the directory or the filename or both.
1339
rename_one has several 'modes' to work. First, it can rename a physical
1340
file and change the file_id. That is the normal mode. Second, it can
1341
only change the file_id without touching any physical file. This is
1342
the new mode introduced in version 0.15.
1344
rename_one uses the second mode if 'after == True' and 'to_rel' is not
1345
versioned but present in the working tree.
1347
rename_one uses the second mode if 'after == False' and 'from_rel' is
1348
versioned but no longer in the working tree, and 'to_rel' is not
1349
versioned but present in the working tree.
1351
rename_one uses the first mode if 'after == False' and 'from_rel' is
1352
versioned and present in the working tree, and 'to_rel' is not
1353
versioned and not present in the working tree.
1355
Everything else results in an error.
862
1357
inv = self.inventory
863
if not self.has_filename(from_rel):
864
raise BzrError("can't rename: old working file %r does not exist" % from_rel)
865
if self.has_filename(to_rel):
866
raise BzrError("can't rename: new working file %r already exists" % to_rel)
868
file_id = inv.path2id(from_rel)
870
raise BzrError("can't rename: old name %r is not versioned" % from_rel)
873
from_parent = entry.parent_id
874
from_name = entry.name
876
if inv.path2id(to_rel):
877
raise BzrError("can't rename: new name %r is already versioned" % to_rel)
1360
# create rename entries and tuples
1361
from_tail = splitpath(from_rel)[-1]
1362
from_id = inv.path2id(from_rel)
1364
raise errors.BzrRenameFailedError(from_rel,to_rel,
1365
errors.NotVersionedError(path=str(from_rel)))
1366
from_entry = inv[from_id]
1367
from_parent_id = from_entry.parent_id
879
1368
to_dir, to_tail = os.path.split(to_rel)
880
1369
to_dir_id = inv.path2id(to_dir)
881
if to_dir_id == None and to_dir != '':
882
raise BzrError("can't determine destination directory id for %r" % to_dir)
884
mutter("rename_one:")
885
mutter(" file_id {%s}" % file_id)
886
mutter(" from_rel %r" % from_rel)
887
mutter(" to_rel %r" % to_rel)
888
mutter(" to_dir %r" % to_dir)
889
mutter(" to_dir_id {%s}" % to_dir_id)
891
inv.rename(file_id, to_dir_id, to_tail)
893
from_abs = self.abspath(from_rel)
894
to_abs = self.abspath(to_rel)
896
rename(from_abs, to_abs)
898
inv.rename(file_id, from_parent, from_name)
899
raise BzrError("failed to rename %r to %r: %s"
900
% (from_abs, to_abs, e[1]),
901
["rename rolled back"])
1370
rename_entry = WorkingTree._RenameEntry(from_rel=from_rel,
1372
from_tail=from_tail,
1373
from_parent_id=from_parent_id,
1374
to_rel=to_rel, to_tail=to_tail,
1375
to_parent_id=to_dir_id)
1376
rename_entries.append(rename_entry)
1378
# determine which move mode to use. checks also for movability
1379
rename_entries = self._determine_mv_mode(rename_entries, after)
1381
# check if the target changed directory and if the target directory is
1383
if to_dir_id is None:
1384
raise errors.BzrMoveFailedError(from_rel,to_rel,
1385
errors.NotVersionedError(path=str(to_dir)))
1387
# all checks done. now we can continue with our actual work
1388
mutter('rename_one:\n'
1393
' to_dir_id {%s}\n',
1394
from_id, from_rel, to_rel, to_dir, to_dir_id)
1396
self._move(rename_entries)
902
1397
self._write_inventory(inv)
1399
class _RenameEntry(object):
1400
def __init__(self, from_rel, from_id, from_tail, from_parent_id,
1401
to_rel, to_tail, to_parent_id, only_change_inv=False):
1402
self.from_rel = from_rel
1403
self.from_id = from_id
1404
self.from_tail = from_tail
1405
self.from_parent_id = from_parent_id
1406
self.to_rel = to_rel
1407
self.to_tail = to_tail
1408
self.to_parent_id = to_parent_id
1409
self.only_change_inv = only_change_inv
904
1411
@needs_read_lock
905
1412
def unknowns(self):
906
1413
"""Return all unknown files.
908
1415
These are files in the working directory that are not versioned or
909
1416
control files or ignored.
911
>>> from bzrlib.bzrdir import ScratchDir
912
>>> d = ScratchDir(files=['foo', 'foo~'])
913
>>> b = d.open_branch()
914
>>> tree = d.open_workingtree()
915
>>> map(str, tree.unknowns())
918
>>> list(b.unknowns())
920
>>> tree.remove('foo')
921
>>> list(b.unknowns())
924
for subp in self.extras():
925
if not self.is_ignored(subp):
1418
# force the extras method to be fully executed before returning, to
1419
# prevent race conditions with the lock
1421
[subp for subp in self.extras() if not self.is_ignored(subp)])
1423
@needs_tree_write_lock
1424
def unversion(self, file_ids):
1425
"""Remove the file ids in file_ids from the current versioned set.
1427
When a file_id is unversioned, all of its children are automatically
1430
:param file_ids: The file ids to stop versioning.
1431
:raises: NoSuchId if any fileid is not currently versioned.
1433
for file_id in file_ids:
1434
file_id = osutils.safe_file_id(file_id)
1435
if self._inventory.has_id(file_id):
1436
self._inventory.remove_recursive_id(file_id)
1438
raise errors.NoSuchId(self, file_id)
1440
# in the future this should just set a dirty bit to wait for the
1441
# final unlock. However, until all methods of workingtree start
1442
# with the current in -memory inventory rather than triggering
1443
# a read, it is more complex - we need to teach read_inventory
1444
# to know when to read, and when to not read first... and possibly
1445
# to save first when the in memory one may be corrupted.
1446
# so for now, we just only write it if it is indeed dirty.
1448
self._write_inventory(self._inventory)
928
1450
@deprecated_method(zero_eight)
929
1451
def iter_conflicts(self):
930
1452
"""List all files in the tree that have text or content conflicts.
1177
1716
if new_revision is None:
1178
1717
self.branch.set_revision_history([])
1180
# current format is locked in with the branch
1181
revision_history = self.branch.revision_history()
1183
position = revision_history.index(new_revision)
1185
raise errors.NoSuchRevision(self.branch, new_revision)
1186
self.branch.set_revision_history(revision_history[:position + 1])
1720
self.branch.generate_revision_history(new_revision)
1721
except errors.NoSuchRevision:
1722
# not present in the repo - dont try to set it deeper than the tip
1723
self.branch.set_revision_history([new_revision])
1726
def _write_basis_inventory(self, xml):
1727
"""Write the basis inventory XML to the basis-inventory file"""
1728
assert isinstance(xml, str), 'serialised xml must be bytestring.'
1729
path = self._basis_inventory_name()
1731
self._control_files.put(path, sio)
1733
def _create_basis_xml_from_inventory(self, revision_id, inventory):
1734
"""Create the text that will be saved in basis-inventory"""
1735
# TODO: jam 20070209 This should be redundant, as the revision_id
1736
# as all callers should have already converted the revision_id to
1738
inventory.revision_id = osutils.safe_revision_id(revision_id)
1739
return xml7.serializer_v7.write_inventory_to_string(inventory)
1189
1741
def _cache_basis_inventory(self, new_revision):
1190
1742
"""Cache new_revision as the basis inventory."""
1743
# TODO: this should allow the ready-to-use inventory to be passed in,
1744
# as commit already has that ready-to-use [while the format is the
1192
1747
# this double handles the inventory - unpack and repack -
1193
1748
# but is easier to understand. We can/should put a conditional
1194
1749
# in here based on whether the inventory is in the latest format
1195
1750
# - perhaps we should repack all inventories on a repository
1197
inv = self.branch.repository.get_inventory(new_revision)
1198
inv.revision_id = new_revision
1199
xml = bzrlib.xml5.serializer_v5.write_inventory_to_string(inv)
1201
path = self._basis_inventory_name()
1202
self._control_files.put_utf8(path, xml)
1203
except WeaveRevisionNotPresent:
1752
# the fast path is to copy the raw xml from the repository. If the
1753
# xml contains 'revision_id="', then we assume the right
1754
# revision_id is set. We must check for this full string, because a
1755
# root node id can legitimately look like 'revision_id' but cannot
1757
xml = self.branch.repository.get_inventory_xml(new_revision)
1758
firstline = xml.split('\n', 1)[0]
1759
if (not 'revision_id="' in firstline or
1760
'format="7"' not in firstline):
1761
inv = self.branch.repository.deserialise_inventory(
1763
xml = self._create_basis_xml_from_inventory(new_revision, inv)
1764
self._write_basis_inventory(xml)
1765
except (errors.NoSuchRevision, errors.RevisionNotPresent):
1206
1768
def read_basis_inventory(self):
1207
1769
"""Read the cached basis inventory."""
1208
1770
path = self._basis_inventory_name()
1209
return self._control_files.get_utf8(path).read()
1771
return self._control_files.get(path).read()
1211
1773
@needs_read_lock
1212
1774
def read_working_inventory(self):
1213
"""Read the working inventory."""
1775
"""Read the working inventory.
1777
:raises errors.InventoryModified: read_working_inventory will fail
1778
when the current in memory inventory has been modified.
1780
# conceptually this should be an implementation detail of the tree.
1781
# XXX: Deprecate this.
1214
1782
# ElementTree does its own conversion from UTF-8, so open in
1216
result = bzrlib.xml5.serializer_v5.read_inventory(
1217
self._control_files.get('inventory'))
1218
self._set_inventory(result)
1784
if self._inventory_is_modified:
1785
raise errors.InventoryModified(self)
1786
result = self._deserialize(self._control_files.get('inventory'))
1787
self._set_inventory(result, dirty=False)
1222
def remove(self, files, verbose=False):
1790
@needs_tree_write_lock
1791
def remove(self, files, verbose=False, to_file=None):
1223
1792
"""Remove nominated files from the working inventory..
1225
1794
This does not remove their text. This does not run on XXX on what? RBC
1322
1940
between multiple working trees, i.e. via shared storage, then we
1323
1941
would probably want to lock both the local tree, and the branch.
1325
# FIXME: We want to write out the hashcache only when the last lock on
1326
# this working copy is released. Peeking at the lock count is a bit
1327
# of a nasty hack; probably it's better to have a transaction object,
1328
# which can do some finalization when it's either successfully or
1329
# unsuccessfully completed. (Denys's original patch did that.)
1330
# RBC 20060206 hookinhg into transaction will couple lock and transaction
1331
# wrongly. Hookinh into unllock on the control files object is fine though.
1333
# TODO: split this per format so there is no ugly if block
1334
if self._hashcache.needs_write and (
1335
# dedicated lock files
1336
self._control_files._lock_count==1 or
1338
(self._control_files is self.branch.control_files and
1339
self._control_files._lock_count==3)):
1340
self._hashcache.write()
1341
# reverse order of locking.
1343
return self._control_files.unlock()
1345
self.branch.unlock()
1943
raise NotImplementedError(self.unlock)
1348
1945
def update(self):
1349
1946
"""Update a working tree along its branch.
1351
This will update the branch if its bound too, which means we have multiple trees involved:
1352
The new basis tree of the master.
1353
The old basis tree of the branch.
1354
The old basis tree of the working tree.
1355
The current working tree state.
1356
pathologically all three may be different, and non ancestors of each other.
1357
Conceptually we want to:
1358
Preserve the wt.basis->wt.state changes
1359
Transform the wt.basis to the new master basis.
1360
Apply a merge of the old branch basis to get any 'local' changes from it into the tree.
1361
Restore the wt.basis->wt.state changes.
1948
This will update the branch if its bound too, which means we have
1949
multiple trees involved:
1951
- The new basis tree of the master.
1952
- The old basis tree of the branch.
1953
- The old basis tree of the working tree.
1954
- The current working tree state.
1956
Pathologically, all three may be different, and non-ancestors of each
1957
other. Conceptually we want to:
1959
- Preserve the wt.basis->wt.state changes
1960
- Transform the wt.basis to the new master basis.
1961
- Apply a merge of the old branch basis to get any 'local' changes from
1963
- Restore the wt.basis->wt.state changes.
1363
1965
There isn't a single operation at the moment to do that, so we:
1364
Merge current state -> basis tree of the master w.r.t. the old tree basis.
1365
Do a 'normal' merge of the old branch basis if it is relevant.
1367
old_tip = self.branch.update()
1368
if old_tip is not None:
1369
self.add_pending_merge(old_tip)
1370
self.branch.lock_read()
1373
if self.last_revision() != self.branch.last_revision():
1374
# merge tree state up to new branch tip.
1375
basis = self.basis_tree()
1966
- Merge current state -> basis tree of the master w.r.t. the old tree
1968
- Do a 'normal' merge of the old branch basis if it is relevant.
1970
if self.branch.get_master_branch() is not None:
1972
update_branch = True
1974
self.lock_tree_write()
1975
update_branch = False
1978
old_tip = self.branch.update()
1981
return self._update_tree(old_tip)
1985
@needs_tree_write_lock
1986
def _update_tree(self, old_tip=None):
1987
"""Update a tree to the master branch.
1989
:param old_tip: if supplied, the previous tip revision the branch,
1990
before it was changed to the master branch's tip.
1992
# here if old_tip is not None, it is the old tip of the branch before
1993
# it was updated from the master branch. This should become a pending
1994
# merge in the working tree to preserve the user existing work. we
1995
# cant set that until we update the working trees last revision to be
1996
# one from the new branch, because it will just get absorbed by the
1997
# parent de-duplication logic.
1999
# We MUST save it even if an error occurs, because otherwise the users
2000
# local work is unreferenced and will appear to have been lost.
2004
last_rev = self.get_parent_ids()[0]
2007
if last_rev != self.branch.last_revision():
2008
# merge tree state up to new branch tip.
2009
basis = self.basis_tree()
1376
2012
to_tree = self.branch.basis_tree()
1377
result += merge_inner(self.branch,
2013
if basis.inventory.root is None:
2014
self.set_root_id(to_tree.inventory.root.file_id)
2016
result += merge.merge_inner(
1380
2020
this_tree=self)
1381
self.set_last_revision(self.branch.last_revision())
1382
if old_tip and old_tip != self.last_revision():
1383
# our last revision was not the prior branch last reivison
1384
# and we have converted that last revision to a pending merge.
1385
# base is somewhere between the branch tip now
1386
# and the now pending merge
1387
from bzrlib.revision import common_ancestor
1389
base_rev_id = common_ancestor(self.branch.last_revision(),
1391
self.branch.repository)
1392
except errors.NoCommonAncestor:
1394
base_tree = self.branch.repository.revision_tree(base_rev_id)
1395
other_tree = self.branch.repository.revision_tree(old_tip)
1396
result += merge_inner(self.branch,
1402
self.branch.unlock()
2023
# TODO - dedup parents list with things merged by pull ?
2024
# reuse the tree we've updated to to set the basis:
2025
parent_trees = [(self.branch.last_revision(), to_tree)]
2026
merges = self.get_parent_ids()[1:]
2027
# Ideally we ask the tree for the trees here, that way the working
2028
# tree can decide whether to give us teh entire tree or give us a
2029
# lazy initialised tree. dirstate for instance will have the trees
2030
# in ram already, whereas a last-revision + basis-inventory tree
2031
# will not, but also does not need them when setting parents.
2032
for parent in merges:
2033
parent_trees.append(
2034
(parent, self.branch.repository.revision_tree(parent)))
2035
if old_tip is not None:
2036
parent_trees.append(
2037
(old_tip, self.branch.repository.revision_tree(old_tip)))
2038
self.set_parent_trees(parent_trees)
2039
last_rev = parent_trees[0][0]
2041
# the working tree had the same last-revision as the master
2042
# branch did. We may still have pivot local work from the local
2043
# branch into old_tip:
2044
if old_tip is not None:
2045
self.add_parent_tree_id(old_tip)
2046
if old_tip and old_tip != last_rev:
2047
# our last revision was not the prior branch last revision
2048
# and we have converted that last revision to a pending merge.
2049
# base is somewhere between the branch tip now
2050
# and the now pending merge
2052
# Since we just modified the working tree and inventory, flush out
2053
# the current state, before we modify it again.
2054
# TODO: jam 20070214 WorkingTree3 doesn't require this, dirstate
2055
# requires it only because TreeTransform directly munges the
2056
# inventory and calls tree._write_inventory(). Ultimately we
2057
# should be able to remove this extra flush.
2059
from bzrlib.revision import common_ancestor
2061
base_rev_id = common_ancestor(self.branch.last_revision(),
2063
self.branch.repository)
2064
except errors.NoCommonAncestor:
2066
base_tree = self.branch.repository.revision_tree(base_rev_id)
2067
other_tree = self.branch.repository.revision_tree(old_tip)
2068
result += merge.merge_inner(
2075
def _write_hashcache_if_dirty(self):
2076
"""Write out the hashcache if it is dirty."""
2077
if self._hashcache.needs_write:
2079
self._hashcache.write()
2081
if e.errno not in (errno.EPERM, errno.EACCES):
2083
# TODO: jam 20061219 Should this be a warning? A single line
2084
# warning might be sufficient to let the user know what
2086
mutter('Could not write hashcache for %s\nError: %s',
2087
self._hashcache.cache_file_name(), e)
2089
@needs_tree_write_lock
1405
2090
def _write_inventory(self, inv):
1406
2091
"""Write inventory as the current inventory."""
1408
bzrlib.xml5.serializer_v5.write_inventory(inv, sio)
1410
self._control_files.put('inventory', sio)
1411
self._set_inventory(inv)
1412
mutter('wrote working inventory')
2092
self._set_inventory(inv, dirty=True)
1414
2095
def set_conflicts(self, arg):
1415
raise UnsupportedOperation(self.set_conflicts, self)
2096
raise errors.UnsupportedOperation(self.set_conflicts, self)
2098
def add_conflicts(self, arg):
2099
raise errors.UnsupportedOperation(self.add_conflicts, self)
1417
2101
@needs_read_lock
1418
2102
def conflicts(self):
1419
conflicts = ConflictList()
2103
conflicts = _mod_conflicts.ConflictList()
1420
2104
for conflicted in self._iter_conflicts():
1423
2107
if file_kind(self.abspath(conflicted)) != "file":
1426
if e.errno == errno.ENOENT:
2109
except errors.NoSuchFile:
1430
2111
if text is True:
1431
2112
for suffix in ('.THIS', '.OTHER'):
1433
2114
kind = file_kind(self.abspath(conflicted+suffix))
1435
if e.errno == errno.ENOENT:
2117
except errors.NoSuchFile:
1443
2121
ctype = {True: 'text conflict', False: 'contents conflict'}[text]
1444
conflicts.append(Conflict.factory(ctype, path=conflicted,
2122
conflicts.append(_mod_conflicts.Conflict.factory(ctype,
1445
2124
file_id=self.path2id(conflicted)))
1446
2125
return conflicts
2127
def walkdirs(self, prefix=""):
2128
"""Walk the directories of this tree.
2130
This API returns a generator, which is only valid during the current
2131
tree transaction - within a single lock_read or lock_write duration.
2133
If the tree is not locked, it may cause an error to be raised, depending
2134
on the tree implementation.
2136
disk_top = self.abspath(prefix)
2137
if disk_top.endswith('/'):
2138
disk_top = disk_top[:-1]
2139
top_strip_len = len(disk_top) + 1
2140
inventory_iterator = self._walkdirs(prefix)
2141
disk_iterator = osutils.walkdirs(disk_top, prefix)
2143
current_disk = disk_iterator.next()
2144
disk_finished = False
2146
if e.errno != errno.ENOENT:
2149
disk_finished = True
2151
current_inv = inventory_iterator.next()
2152
inv_finished = False
2153
except StopIteration:
2156
while not inv_finished or not disk_finished:
2157
if not disk_finished:
2158
# strip out .bzr dirs
2159
if current_disk[0][1][top_strip_len:] == '':
2160
# osutils.walkdirs can be made nicer -
2161
# yield the path-from-prefix rather than the pathjoined
2163
bzrdir_loc = bisect_left(current_disk[1], ('.bzr', '.bzr'))
2164
if current_disk[1][bzrdir_loc][0] == '.bzr':
2165
# we dont yield the contents of, or, .bzr itself.
2166
del current_disk[1][bzrdir_loc]
2168
# everything is unknown
2171
# everything is missing
2174
direction = cmp(current_inv[0][0], current_disk[0][0])
2176
# disk is before inventory - unknown
2177
dirblock = [(relpath, basename, kind, stat, None, None) for
2178
relpath, basename, kind, stat, top_path in current_disk[1]]
2179
yield (current_disk[0][0], None), dirblock
2181
current_disk = disk_iterator.next()
2182
except StopIteration:
2183
disk_finished = True
2185
# inventory is before disk - missing.
2186
dirblock = [(relpath, basename, 'unknown', None, fileid, kind)
2187
for relpath, basename, dkind, stat, fileid, kind in
2189
yield (current_inv[0][0], current_inv[0][1]), dirblock
2191
current_inv = inventory_iterator.next()
2192
except StopIteration:
2195
# versioned present directory
2196
# merge the inventory and disk data together
2198
for relpath, subiterator in itertools.groupby(sorted(
2199
current_inv[1] + current_disk[1], key=operator.itemgetter(0)), operator.itemgetter(1)):
2200
path_elements = list(subiterator)
2201
if len(path_elements) == 2:
2202
inv_row, disk_row = path_elements
2203
# versioned, present file
2204
dirblock.append((inv_row[0],
2205
inv_row[1], disk_row[2],
2206
disk_row[3], inv_row[4],
2208
elif len(path_elements[0]) == 5:
2210
dirblock.append((path_elements[0][0],
2211
path_elements[0][1], path_elements[0][2],
2212
path_elements[0][3], None, None))
2213
elif len(path_elements[0]) == 6:
2214
# versioned, absent file.
2215
dirblock.append((path_elements[0][0],
2216
path_elements[0][1], 'unknown', None,
2217
path_elements[0][4], path_elements[0][5]))
2219
raise NotImplementedError('unreachable code')
2220
yield current_inv[0], dirblock
2222
current_inv = inventory_iterator.next()
2223
except StopIteration:
2226
current_disk = disk_iterator.next()
2227
except StopIteration:
2228
disk_finished = True
2230
def _walkdirs(self, prefix=""):
2231
_directory = 'directory'
2232
# get the root in the inventory
2233
inv = self.inventory
2234
top_id = inv.path2id(prefix)
2238
pending = [(prefix, '', _directory, None, top_id, None)]
2241
currentdir = pending.pop()
2242
# 0 - relpath, 1- basename, 2- kind, 3- stat, 4-id, 5-kind
2243
top_id = currentdir[4]
2245
relroot = currentdir[0] + '/'
2248
# FIXME: stash the node in pending
2250
for name, child in entry.sorted_children():
2251
dirblock.append((relroot + name, name, child.kind, None,
2252
child.file_id, child.kind
2254
yield (currentdir[0], entry.file_id), dirblock
2255
# push the user specified dirs from dirblock
2256
for dir in reversed(dirblock):
2257
if dir[2] == _directory:
2260
@needs_tree_write_lock
2261
def auto_resolve(self):
2262
"""Automatically resolve text conflicts according to contents.
2264
Only text conflicts are auto_resolvable. Files with no conflict markers
2265
are considered 'resolved', because bzr always puts conflict markers
2266
into files that have text conflicts. The corresponding .THIS .BASE and
2267
.OTHER files are deleted, as per 'resolve'.
2268
:return: a tuple of ConflictLists: (un_resolved, resolved).
2270
un_resolved = _mod_conflicts.ConflictList()
2271
resolved = _mod_conflicts.ConflictList()
2272
conflict_re = re.compile('^(<{7}|={7}|>{7})')
2273
for conflict in self.conflicts():
2274
if (conflict.typestring != 'text conflict' or
2275
self.kind(conflict.file_id) != 'file'):
2276
un_resolved.append(conflict)
2278
my_file = open(self.id2abspath(conflict.file_id), 'rb')
2280
for line in my_file:
2281
if conflict_re.search(line):
2282
un_resolved.append(conflict)
2285
resolved.append(conflict)
2288
resolved.remove_files(self)
2289
self.set_conflicts(un_resolved)
2290
return un_resolved, resolved
2293
class WorkingTree2(WorkingTree):
2294
"""This is the Format 2 working tree.
2296
This was the first weave based working tree.
2297
- uses os locks for locking.
2298
- uses the branch last-revision.
2301
def lock_tree_write(self):
2302
"""See WorkingTree.lock_tree_write().
2304
In Format2 WorkingTrees we have a single lock for the branch and tree
2305
so lock_tree_write() degrades to lock_write().
2307
self.branch.lock_write()
2309
return self._control_files.lock_write()
2311
self.branch.unlock()
2315
# we share control files:
2316
if self._control_files._lock_count == 3:
2317
# _inventory_is_modified is always False during a read lock.
2318
if self._inventory_is_modified:
2320
self._write_hashcache_if_dirty()
2322
# reverse order of locking.
2324
return self._control_files.unlock()
2326
self.branch.unlock()
1449
2329
class WorkingTree3(WorkingTree):
1450
2330
"""This is the Format 3 working tree.