39
36
# At the moment they may alias the inventory and have old copies of it in
40
37
# memory. (Now done? -- mbp 20060309)
42
from binascii import hexlify
39
from cStringIO import StringIO
42
from bzrlib.lazy_import import lazy_import
43
lazy_import(globals(), """
44
from bisect import bisect_left
43
46
from copy import deepcopy
44
from cStringIO import StringIO
50
51
from time import time
52
from bzrlib.atomicfile import AtomicFile
53
from bzrlib.branch import (Branch,
55
from bzrlib.conflicts import Conflict, ConflictList, CONFLICT_SUFFIXES
56
import bzrlib.bzrdir as bzrdir
59
conflicts as _mod_conflicts,
78
from bzrlib.transport import get_transport
80
from bzrlib.workingtree_4 import WorkingTreeFormat4
83
from bzrlib import symbol_versioning
57
84
from bzrlib.decorators import needs_read_lock, needs_write_lock
58
import bzrlib.errors as errors
59
from bzrlib.errors import (BzrCheckError,
63
WeaveRevisionNotPresent,
67
MergeModifiedFormatError,
70
from bzrlib.inventory import InventoryEntry, Inventory
85
from bzrlib.inventory import InventoryEntry, Inventory, ROOT_ID, TreeReference
71
86
from bzrlib.lockable_files import LockableFiles, TransportLock
72
87
from bzrlib.lockdir import LockDir
73
from bzrlib.merge import merge_inner, transform_tree
88
import bzrlib.mutabletree
89
from bzrlib.mutabletree import needs_tree_write_lock
74
90
from bzrlib.osutils import (
102
from bzrlib.trace import mutter, note
103
from bzrlib.transport.local import LocalTransport
92
104
from bzrlib.progress import DummyProgress, ProgressPhase
93
from bzrlib.revision import NULL_REVISION
105
from bzrlib.revision import NULL_REVISION, CURRENT_REVISION
94
106
from bzrlib.rio import RioReader, rio_file, Stanza
95
from bzrlib.symbol_versioning import *
96
from bzrlib.textui import show_status
98
from bzrlib.transform import build_tree
99
from bzrlib.trace import mutter, note
100
from bzrlib.transport import get_transport
101
from bzrlib.transport.local import LocalTransport
106
# the regex here does the following:
107
# 1) remove any weird characters; we don't escape them but rather
109
# 2) match leading '.'s to make it not hidden
110
_gen_file_id_re = re.compile(r'[^\w.]|(^\.*)')
111
_gen_id_suffix = None
115
def _next_id_suffix():
116
"""Create a new file id suffix that is reasonably unique.
118
On the first call we combine the current time with 64 bits of randomness
119
to give a highly probably globally unique number. Then each call in the same
120
process adds 1 to a serial number we append to that unique value.
122
# XXX TODO: change bzrlib.add.smart_add to call workingtree.add() rather
123
# than having to move the id randomness out of the inner loop like this.
124
# XXX TODO: for the global randomness this uses we should add the thread-id
125
# before the serial #.
126
global _gen_id_suffix, _gen_id_serial
127
if _gen_id_suffix is None:
128
_gen_id_suffix = "-%s-%s-" % (compact_date(time()), rand_chars(16))
130
return _gen_id_suffix + str(_gen_id_serial)
107
from bzrlib.symbol_versioning import (deprecated_passed,
110
DEPRECATED_PARAMETER,
117
MERGE_MODIFIED_HEADER_1 = "BZR merge-modified list format 1"
118
CONFLICT_HEADER_1 = "BZR conflict list format 1"
121
@deprecated_function(zero_thirteen)
133
122
def gen_file_id(name):
134
123
"""Return new file id for the basename 'name'.
136
The uniqueness is supplied from _next_id_suffix.
125
Use bzrlib.generate_ids.gen_file_id() instead
138
# XXX TODO: squash the filename to lowercase.
139
# XXX TODO: truncate the filename to something like 20 or 30 chars.
140
# XXX TODO: consider what to do with ids that look like illegal filepaths
141
# on platforms we support.
142
return _gen_file_id_re.sub('', name) + _next_id_suffix()
127
return generate_ids.gen_file_id(name)
130
@deprecated_function(zero_thirteen)
145
131
def gen_root_id():
146
"""Return a new tree-root file id."""
147
return gen_file_id('TREE_ROOT')
132
"""Return a new tree-root file id.
134
This has been deprecated in favor of bzrlib.generate_ids.gen_root_id()
136
return generate_ids.gen_root_id()
150
139
class TreeEntry(object):
151
"""An entry that implements the minium interface used by commands.
140
"""An entry that implements the minimum interface used by commands.
153
142
This needs further inspection, it may be better to have
154
143
InventoryEntries without ids - though that seems wrong. For now,
527
600
__contains__ = has_id
529
602
def get_file_size(self, file_id):
603
file_id = osutils.safe_file_id(file_id)
530
604
return os.path.getsize(self.id2abspath(file_id))
533
def get_file_sha1(self, file_id):
534
path = self._inventory.id2path(file_id)
535
return self._hashcache.get_sha1(path)
537
def is_executable(self, file_id):
538
if not supports_executable():
607
def get_file_sha1(self, file_id, path=None, stat_value=None):
608
file_id = osutils.safe_file_id(file_id)
610
path = self._inventory.id2path(file_id)
611
return self._hashcache.get_sha1(path, stat_value)
613
def get_file_mtime(self, file_id, path=None):
614
file_id = osutils.safe_file_id(file_id)
616
path = self.inventory.id2path(file_id)
617
return os.lstat(self.abspath(path)).st_mtime
619
if not supports_executable():
620
def is_executable(self, file_id, path=None):
621
file_id = osutils.safe_file_id(file_id)
539
622
return self._inventory[file_id].executable
541
path = self._inventory.id2path(file_id)
624
def is_executable(self, file_id, path=None):
626
file_id = osutils.safe_file_id(file_id)
627
path = self.id2path(file_id)
542
628
mode = os.lstat(self.abspath(path)).st_mode
543
return bool(stat.S_ISREG(mode) and stat.S_IEXEC&mode)
546
def add(self, files, ids=None):
547
"""Make files versioned.
549
Note that the command line normally calls smart_add instead,
550
which can automatically recurse.
552
This adds the files to the inventory, so that they will be
553
recorded by the next commit.
556
List of paths to add, relative to the base of the tree.
559
If set, use these instead of automatically generated ids.
560
Must be the same length as the list of files, but may
561
contain None for ids that are to be autogenerated.
563
TODO: Perhaps have an option to add the ids even if the files do
566
TODO: Perhaps callback with the ids and paths as they're added.
629
return bool(stat.S_ISREG(mode) and stat.S_IEXEC & mode)
631
@needs_tree_write_lock
632
def _add(self, files, ids, kinds):
633
"""See MutableTree._add."""
568
634
# TODO: Re-adding a file that is removed in the working copy
569
635
# should probably put it back with the previous ID.
570
if isinstance(files, basestring):
571
assert(ids is None or isinstance(ids, basestring))
577
ids = [None] * len(files)
579
assert(len(ids) == len(files))
636
# the read and write working inventory should not occur in this
637
# function - they should be part of lock_write and unlock.
581
638
inv = self.read_working_inventory()
582
for f,file_id in zip(files, ids):
583
if self.is_control_filename(f):
584
raise BzrError("cannot add control file %s" % quotefn(f))
589
raise BzrError("cannot add top-level %r" % f)
591
fullpath = normpath(self.abspath(f))
594
kind = file_kind(fullpath)
596
if e.errno == errno.ENOENT:
597
raise NoSuchFile(fullpath)
598
# maybe something better?
599
raise BzrError('cannot add: not a regular file, symlink or directory: %s' % quotefn(f))
601
if not InventoryEntry.versionable_kind(kind):
602
raise BzrError('cannot add: not a versionable file ('
603
'i.e. regular file, symlink or directory): %s' % quotefn(f))
639
for f, file_id, kind in zip(files, ids, kinds):
640
assert kind is not None
605
641
if file_id is None:
606
642
inv.add_path(f, kind=kind)
644
file_id = osutils.safe_file_id(file_id)
608
645
inv.add_path(f, kind=kind, file_id=file_id)
610
646
self._write_inventory(inv)
648
@needs_tree_write_lock
649
def _gather_kinds(self, files, kinds):
650
"""See MutableTree._gather_kinds."""
651
for pos, f in enumerate(files):
652
if kinds[pos] is None:
653
fullpath = normpath(self.abspath(f))
655
kinds[pos] = file_kind(fullpath)
657
if e.errno == errno.ENOENT:
658
raise errors.NoSuchFile(fullpath)
612
660
@needs_write_lock
661
def add_parent_tree_id(self, revision_id, allow_leftmost_as_ghost=False):
662
"""Add revision_id as a parent.
664
This is equivalent to retrieving the current list of parent ids
665
and setting the list to its value plus revision_id.
667
:param revision_id: The revision id to add to the parent list. It may
668
be a ghost revision as long as its not the first parent to be added,
669
or the allow_leftmost_as_ghost parameter is set True.
670
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
672
parents = self.get_parent_ids() + [revision_id]
673
self.set_parent_ids(parents, allow_leftmost_as_ghost=len(parents) > 1
674
or allow_leftmost_as_ghost)
676
@needs_tree_write_lock
677
def add_parent_tree(self, parent_tuple, allow_leftmost_as_ghost=False):
678
"""Add revision_id, tree tuple as a parent.
680
This is equivalent to retrieving the current list of parent trees
681
and setting the list to its value plus parent_tuple. See also
682
add_parent_tree_id - if you only have a parent id available it will be
683
simpler to use that api. If you have the parent already available, using
684
this api is preferred.
686
:param parent_tuple: The (revision id, tree) to add to the parent list.
687
If the revision_id is a ghost, pass None for the tree.
688
:param allow_leftmost_as_ghost: Allow the first parent to be a ghost.
690
parent_ids = self.get_parent_ids() + [parent_tuple[0]]
691
if len(parent_ids) > 1:
692
# the leftmost may have already been a ghost, preserve that if it
694
allow_leftmost_as_ghost = True
695
self.set_parent_ids(parent_ids,
696
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
698
@needs_tree_write_lock
613
699
def add_pending_merge(self, *revision_ids):
614
700
# TODO: Perhaps should check at this point that the
615
701
# history of the revision is actually present?
616
p = self.pending_merges()
702
parents = self.get_parent_ids()
618
704
for rev_id in revision_ids:
705
if rev_id in parents:
707
parents.append(rev_id)
624
self.set_pending_merges(p)
710
self.set_parent_ids(parents, allow_leftmost_as_ghost=True)
712
@deprecated_method(zero_eleven)
627
714
def pending_merges(self):
628
715
"""Return a list of pending merges.
630
717
These are revisions that have been merged into the working
631
718
directory but not yet committed.
634
merges_file = self._control_files.get_utf8('pending-merges')
636
if e.errno != errno.ENOENT:
640
for l in merges_file.readlines():
641
p.append(l.rstrip('\n'))
720
As of 0.11 this is deprecated. Please see WorkingTree.get_parent_ids()
721
instead - which is available on all tree objects.
723
return self.get_parent_ids()[1:]
725
def _check_parents_for_ghosts(self, revision_ids, allow_leftmost_as_ghost):
726
"""Common ghost checking functionality from set_parent_*.
728
This checks that the left hand-parent exists if there are any
731
if len(revision_ids) > 0:
732
leftmost_id = revision_ids[0]
733
if (not allow_leftmost_as_ghost and not
734
self.branch.repository.has_revision(leftmost_id)):
735
raise errors.GhostRevisionUnusableHere(leftmost_id)
737
def _set_merges_from_parent_ids(self, parent_ids):
738
merges = parent_ids[1:]
739
self._control_files.put_bytes('pending-merges', '\n'.join(merges))
741
@needs_tree_write_lock
742
def set_parent_ids(self, revision_ids, allow_leftmost_as_ghost=False):
743
"""Set the parent ids to revision_ids.
745
See also set_parent_trees. This api will try to retrieve the tree data
746
for each element of revision_ids from the trees repository. If you have
747
tree data already available, it is more efficient to use
748
set_parent_trees rather than set_parent_ids. set_parent_ids is however
749
an easier API to use.
751
:param revision_ids: The revision_ids to set as the parent ids of this
752
working tree. Any of these may be ghosts.
754
revision_ids = [osutils.safe_revision_id(r) for r in revision_ids]
755
self._check_parents_for_ghosts(revision_ids,
756
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
758
if len(revision_ids) > 0:
759
self.set_last_revision(revision_ids[0])
761
self.set_last_revision(None)
763
self._set_merges_from_parent_ids(revision_ids)
765
@needs_tree_write_lock
766
def set_parent_trees(self, parents_list, allow_leftmost_as_ghost=False):
767
"""See MutableTree.set_parent_trees."""
768
parent_ids = [osutils.safe_revision_id(rev) for (rev, tree) in parents_list]
770
self._check_parents_for_ghosts(parent_ids,
771
allow_leftmost_as_ghost=allow_leftmost_as_ghost)
773
if len(parent_ids) == 0:
774
leftmost_parent_id = None
775
leftmost_parent_tree = None
777
leftmost_parent_id, leftmost_parent_tree = parents_list[0]
779
if self._change_last_revision(leftmost_parent_id):
780
if leftmost_parent_tree is None:
781
# If we don't have a tree, fall back to reading the
782
# parent tree from the repository.
783
self._cache_basis_inventory(leftmost_parent_id)
785
inv = leftmost_parent_tree.inventory
786
xml = self._create_basis_xml_from_inventory(
787
leftmost_parent_id, inv)
788
self._write_basis_inventory(xml)
789
self._set_merges_from_parent_ids(parent_ids)
791
@needs_tree_write_lock
645
792
def set_pending_merges(self, rev_list):
646
self._control_files.put_utf8('pending-merges', '\n'.join(rev_list))
793
parents = self.get_parent_ids()
794
leftmost = parents[:1]
795
new_parents = leftmost + rev_list
796
self.set_parent_ids(new_parents)
798
@needs_tree_write_lock
649
799
def set_merge_modified(self, modified_hashes):
650
800
def iter_stanzas():
651
801
for file_id, hash in modified_hashes.iteritems():
652
yield Stanza(file_id=file_id, hash=hash)
802
yield Stanza(file_id=file_id.decode('utf8'), hash=hash)
653
803
self._put_rio('merge-hashes', iter_stanzas(), MERGE_MODIFIED_HEADER_1)
656
805
def _put_rio(self, filename, stanzas, header):
806
self._must_be_locked()
657
807
my_file = rio_file(stanzas, header)
658
808
self._control_files.put(filename, my_file)
810
@needs_write_lock # because merge pulls data into the branch.
811
def merge_from_branch(self, branch, to_revision=None):
812
"""Merge from a branch into this working tree.
814
:param branch: The branch to merge from.
815
:param to_revision: If non-None, the merge will merge to to_revision,
816
but not beyond it. to_revision does not need to be in the history
817
of the branch when it is supplied. If None, to_revision defaults to
818
branch.last_revision().
820
from bzrlib.merge import Merger, Merge3Merger
821
pb = bzrlib.ui.ui_factory.nested_progress_bar()
823
merger = Merger(self.branch, this_tree=self, pb=pb)
824
merger.pp = ProgressPhase("Merge phase", 5, pb)
825
merger.pp.next_phase()
826
# check that there are no
828
merger.check_basis(check_clean=True, require_commits=False)
829
if to_revision is None:
830
to_revision = branch.last_revision()
832
to_revision = osutils.safe_revision_id(to_revision)
833
merger.other_rev_id = to_revision
834
if merger.other_rev_id is None:
835
raise error.NoCommits(branch)
836
self.branch.fetch(branch, last_revision=merger.other_rev_id)
837
merger.other_basis = merger.other_rev_id
838
merger.other_tree = self.branch.repository.revision_tree(
840
merger.other_branch = branch
841
merger.pp.next_phase()
843
if merger.base_rev_id == merger.other_rev_id:
844
raise errors.PointlessMerge
845
merger.backup_files = False
846
merger.merge_type = Merge3Merger
847
merger.set_interesting_files(None)
848
merger.show_base = False
849
merger.reprocess = False
850
conflicts = merger.do_merge()
661
857
def merge_modified(self):
858
"""Return a dictionary of files modified by a merge.
860
The list is initialized by WorkingTree.set_merge_modified, which is
861
typically called after we make some automatic updates to the tree
864
This returns a map of file_id->sha1, containing only files which are
865
still in the working inventory and have that text hash.
663
868
hashfile = self._control_files.get('merge-hashes')
869
except errors.NoSuchFile:
666
871
merge_hashes = {}
668
873
if hashfile.next() != MERGE_MODIFIED_HEADER_1 + '\n':
669
raise MergeModifiedFormatError()
874
raise errors.MergeModifiedFormatError()
670
875
except StopIteration:
671
raise MergeModifiedFormatError()
876
raise errors.MergeModifiedFormatError()
672
877
for s in RioReader(hashfile):
673
file_id = s.get("file_id")
878
# RioReader reads in Unicode, so convert file_ids back to utf8
879
file_id = osutils.safe_file_id(s.get("file_id"), warn=False)
674
880
if file_id not in self.inventory:
677
if hash == self.get_file_sha1(file_id):
678
merge_hashes[file_id] = hash
882
text_hash = s.get("hash")
883
if text_hash == self.get_file_sha1(file_id):
884
merge_hashes[file_id] = text_hash
679
885
return merge_hashes
888
def mkdir(self, path, file_id=None):
889
"""See MutableTree.mkdir()."""
891
file_id = generate_ids.gen_file_id(os.path.basename(path))
892
os.mkdir(self.abspath(path))
893
self.add(path, file_id, 'directory')
681
896
def get_symlink_target(self, file_id):
897
file_id = osutils.safe_file_id(file_id)
682
898
return os.readlink(self.id2abspath(file_id))
684
def file_class(self, filename):
685
if self.path2id(filename):
687
elif self.is_ignored(filename):
692
def list_files(self):
693
"""Recursively list all files as (path, class, kind, id).
901
def subsume(self, other_tree):
902
def add_children(inventory, entry):
903
for child_entry in entry.children.values():
904
inventory._byid[child_entry.file_id] = child_entry
905
if child_entry.kind == 'directory':
906
add_children(inventory, child_entry)
907
if other_tree.get_root_id() == self.get_root_id():
908
raise errors.BadSubsumeSource(self, other_tree,
909
'Trees have the same root')
911
other_tree_path = self.relpath(other_tree.basedir)
912
except errors.PathNotChild:
913
raise errors.BadSubsumeSource(self, other_tree,
914
'Tree is not contained by the other')
915
new_root_parent = self.path2id(osutils.dirname(other_tree_path))
916
if new_root_parent is None:
917
raise errors.BadSubsumeSource(self, other_tree,
918
'Parent directory is not versioned.')
919
# We need to ensure that the result of a fetch will have a
920
# versionedfile for the other_tree root, and only fetching into
921
# RepositoryKnit2 guarantees that.
922
if not self.branch.repository.supports_rich_root():
923
raise errors.SubsumeTargetNeedsUpgrade(other_tree)
924
other_tree.lock_tree_write()
926
new_parents = other_tree.get_parent_ids()
927
other_root = other_tree.inventory.root
928
other_root.parent_id = new_root_parent
929
other_root.name = osutils.basename(other_tree_path)
930
self.inventory.add(other_root)
931
add_children(self.inventory, other_root)
932
self._write_inventory(self.inventory)
933
# normally we don't want to fetch whole repositories, but i think
934
# here we really do want to consolidate the whole thing.
935
for parent_id in other_tree.get_parent_ids():
936
self.branch.fetch(other_tree.branch, parent_id)
937
self.add_parent_tree_id(parent_id)
940
other_tree.bzrdir.retire_bzrdir()
942
@needs_tree_write_lock
943
def extract(self, file_id, format=None):
944
"""Extract a subtree from this tree.
946
A new branch will be created, relative to the path for this tree.
949
segments = osutils.splitpath(path)
950
transport = self.branch.bzrdir.root_transport
951
for name in segments:
952
transport = transport.clone(name)
955
except errors.FileExists:
959
sub_path = self.id2path(file_id)
960
branch_transport = mkdirs(sub_path)
962
format = bzrdir.format_registry.make_bzrdir('dirstate-with-subtree')
964
branch_transport.mkdir('.')
965
except errors.FileExists:
967
branch_bzrdir = format.initialize_on_transport(branch_transport)
969
repo = branch_bzrdir.find_repository()
970
except errors.NoRepositoryPresent:
971
repo = branch_bzrdir.create_repository()
972
assert repo.supports_rich_root()
974
if not repo.supports_rich_root():
975
raise errors.RootNotRich()
976
new_branch = branch_bzrdir.create_branch()
977
new_branch.pull(self.branch)
978
for parent_id in self.get_parent_ids():
979
new_branch.fetch(self.branch, parent_id)
980
tree_transport = self.bzrdir.root_transport.clone(sub_path)
981
if tree_transport.base != branch_transport.base:
982
tree_bzrdir = format.initialize_on_transport(tree_transport)
983
branch.BranchReferenceFormat().initialize(tree_bzrdir, new_branch)
985
tree_bzrdir = branch_bzrdir
986
wt = tree_bzrdir.create_workingtree(NULL_REVISION)
987
wt.set_parent_ids(self.get_parent_ids())
988
my_inv = self.inventory
989
child_inv = Inventory(root_id=None)
990
new_root = my_inv[file_id]
991
my_inv.remove_recursive_id(file_id)
992
new_root.parent_id = None
993
child_inv.add(new_root)
994
self._write_inventory(my_inv)
995
wt._write_inventory(child_inv)
998
def _serialize(self, inventory, out_file):
999
xml5.serializer_v5.write_inventory(self._inventory, out_file)
1001
def _deserialize(selt, in_file):
1002
return xml5.serializer_v5.read_inventory(in_file)
1005
"""Write the in memory inventory to disk."""
1006
# TODO: Maybe this should only write on dirty ?
1007
if self._control_files._lock_mode != 'w':
1008
raise errors.NotWriteLocked(self)
1010
self._serialize(self._inventory, sio)
1012
self._control_files.put('inventory', sio)
1013
self._inventory_is_modified = False
1015
def list_files(self, include_root=False):
1016
"""Recursively list all files as (path, class, kind, id, entry).
695
1018
Lists, but does not descend into unversioned directories.
700
1023
Skips the control directory.
702
inv = self._inventory
704
def descend(from_dir_relpath, from_dir_id, dp):
1025
# list_files is an iterator, so @needs_read_lock doesn't work properly
1026
# with it. So callers should be careful to always read_lock the tree.
1027
if not self.is_locked():
1028
raise errors.ObjectNotLocked(self)
1030
inv = self.inventory
1031
if include_root is True:
1032
yield ('', 'V', 'directory', inv.root.file_id, inv.root)
1033
# Convert these into local objects to save lookup times
1034
pathjoin = osutils.pathjoin
1035
file_kind = osutils.file_kind
1037
# transport.base ends in a slash, we want the piece
1038
# between the last two slashes
1039
transport_base_dir = self.bzrdir.transport.base.rsplit('/', 2)[1]
1041
fk_entries = {'directory':TreeDirectory, 'file':TreeFile, 'symlink':TreeLink}
1043
# directory file_id, relative path, absolute path, reverse sorted children
1044
children = os.listdir(self.basedir)
1046
# jam 20060527 The kernel sized tree seems equivalent whether we
1047
# use a deque and popleft to keep them sorted, or if we use a plain
1048
# list and just reverse() them.
1049
children = collections.deque(children)
1050
stack = [(inv.root.file_id, u'', self.basedir, children)]
1052
from_dir_id, from_dir_relpath, from_dir_abspath, children = stack[-1]
1055
f = children.popleft()
708
1056
## TODO: If we find a subdirectory with its own .bzr
709
1057
## directory, then that is a separate tree and we
710
1058
## should exclude it.
712
1060
# the bzrdir for this tree
713
if self.bzrdir.transport.base.endswith(f + '/'):
1061
if transport_base_dir == f:
717
fp = appendpath(from_dir_relpath, f)
1064
# we know that from_dir_relpath and from_dir_abspath never end in a slash
1065
# and 'f' doesn't begin with one, we can do a string op, rather
1066
# than the checks of pathjoin(), all relative paths will have an extra slash
1068
fp = from_dir_relpath + '/' + f
720
fap = appendpath(dp, f)
1071
fap = from_dir_abspath + '/' + f
722
1073
f_ie = inv.get_child(from_dir_id, f)
725
elif self.is_ignored(fp):
1076
elif self.is_ignored(fp[1:]):
1079
# we may not have found this file, because of a unicode issue
1080
f_norm, can_access = osutils.normalized_filename(f)
1081
if f == f_norm or not can_access:
1082
# No change, so treat this file normally
1085
# this file can be accessed by a normalized path
1086
# check again if it is versioned
1087
# these lines are repeated here for performance
1089
fp = from_dir_relpath + '/' + f
1090
fap = from_dir_abspath + '/' + f
1091
f_ie = inv.get_child(from_dir_id, f)
1094
elif self.is_ignored(fp[1:]):
730
1099
fk = file_kind(fap)
733
1102
if f_ie.kind != fk:
734
raise BzrCheckError("file %r entered as kind %r id %r, "
736
% (fap, f_ie.kind, f_ie.file_id, fk))
1103
raise errors.BzrCheckError(
1104
"file %r entered as kind %r id %r, now of kind %r"
1105
% (fap, f_ie.kind, f_ie.file_id, fk))
738
1107
# make a last minute entry
1109
yield fp[1:], c, fk, f_ie.file_id, f_ie
742
if fk == 'directory':
743
entry = TreeDirectory()
746
elif fk == 'symlink':
1112
yield fp[1:], c, fk, None, fk_entries[fk]()
1114
yield fp[1:], c, fk, None, TreeEntry()
751
yield fp, c, fk, (f_ie and f_ie.file_id), entry
753
1117
if fk != 'directory':
757
# don't descend unversioned directories
760
for ff in descend(fp, f_ie.file_id, fap):
763
for f in descend(u'', inv.root.file_id, self.basedir):
767
def move(self, from_paths, to_name):
1120
# But do this child first
1121
new_children = os.listdir(fap)
1123
new_children = collections.deque(new_children)
1124
stack.append((f_ie.file_id, fp, fap, new_children))
1125
# Break out of inner loop,
1126
# so that we start outer loop with child
1129
# if we finished all children, pop it off the stack
1132
@needs_tree_write_lock
1133
def move(self, from_paths, to_dir=None, after=False, **kwargs):
768
1134
"""Rename files.
770
to_name must exist in the inventory.
1136
to_dir must exist in the inventory.
772
If to_name exists and is a directory, the files are moved into
1138
If to_dir exists and is a directory, the files are moved into
773
1139
it, keeping their old names.
775
Note that to_name is only the last component of the new name;
1141
Note that to_dir is only the last component of the new name;
776
1142
this doesn't change the directory.
1144
For each entry in from_paths the move mode will be determined
1147
The first mode moves the file in the filesystem and updates the
1148
inventory. The second mode only updates the inventory without
1149
touching the file on the filesystem. This is the new mode introduced
1152
move uses the second mode if 'after == True' and the target is not
1153
versioned but present in the working tree.
1155
move uses the second mode if 'after == False' and the source is
1156
versioned but no longer in the working tree, and the target is not
1157
versioned but present in the working tree.
1159
move uses the first mode if 'after == False' and the source is
1160
versioned and present in the working tree, and the target is not
1161
versioned and not present in the working tree.
1163
Everything else results in an error.
778
1165
This returns a list of (from_path, to_path) pairs for each
779
1166
entry that is moved.
782
## TODO: Option to move IDs only
1171
# check for deprecated use of signature
1173
to_dir = kwargs.get('to_name', None)
1175
raise TypeError('You must supply a target directory')
1177
symbol_versioning.warn('The parameter to_name was deprecated'
1178
' in version 0.13. Use to_dir instead',
1181
# check destination directory
783
1182
assert not isinstance(from_paths, basestring)
784
1183
inv = self.inventory
785
to_abs = self.abspath(to_name)
1184
to_abs = self.abspath(to_dir)
786
1185
if not isdir(to_abs):
787
raise BzrError("destination %r is not a directory" % to_abs)
788
if not self.has_filename(to_name):
789
raise BzrError("destination %r not in working directory" % to_abs)
790
to_dir_id = inv.path2id(to_name)
791
if to_dir_id == None and to_name != '':
792
raise BzrError("destination %r is not a versioned directory" % to_name)
1186
raise errors.BzrMoveFailedError('',to_dir,
1187
errors.NotADirectory(to_abs))
1188
if not self.has_filename(to_dir):
1189
raise errors.BzrMoveFailedError('',to_dir,
1190
errors.NotInWorkingDirectory(to_dir))
1191
to_dir_id = inv.path2id(to_dir)
1192
if to_dir_id is None:
1193
raise errors.BzrMoveFailedError('',to_dir,
1194
errors.NotVersionedError(path=str(to_dir)))
793
1196
to_dir_ie = inv[to_dir_id]
794
if to_dir_ie.kind not in ('directory', 'root_directory'):
795
raise BzrError("destination %r is not a directory" % to_abs)
797
to_idpath = inv.get_idpath(to_dir_id)
800
if not self.has_filename(f):
801
raise BzrError("%r does not exist in working tree" % f)
802
f_id = inv.path2id(f)
804
raise BzrError("%r is not versioned" % f)
805
name_tail = splitpath(f)[-1]
806
dest_path = appendpath(to_name, name_tail)
807
if self.has_filename(dest_path):
808
raise BzrError("destination %r already exists" % dest_path)
809
if f_id in to_idpath:
810
raise BzrError("can't move %r to a subdirectory of itself" % f)
812
# OK, so there's a race here, it's possible that someone will
813
# create a file in this interval and then the rename might be
814
# left half-done. But we should have caught most problems.
815
orig_inv = deepcopy(self.inventory)
1197
if to_dir_ie.kind != 'directory':
1198
raise errors.BzrMoveFailedError('',to_dir,
1199
errors.NotADirectory(to_abs))
1201
# create rename entries and tuples
1202
for from_rel in from_paths:
1203
from_tail = splitpath(from_rel)[-1]
1204
from_id = inv.path2id(from_rel)
1206
raise errors.BzrMoveFailedError(from_rel,to_dir,
1207
errors.NotVersionedError(path=str(from_rel)))
1209
from_entry = inv[from_id]
1210
from_parent_id = from_entry.parent_id
1211
to_rel = pathjoin(to_dir, from_tail)
1212
rename_entry = WorkingTree._RenameEntry(from_rel=from_rel,
1214
from_tail=from_tail,
1215
from_parent_id=from_parent_id,
1216
to_rel=to_rel, to_tail=from_tail,
1217
to_parent_id=to_dir_id)
1218
rename_entries.append(rename_entry)
1219
rename_tuples.append((from_rel, to_rel))
1221
# determine which move mode to use. checks also for movability
1222
rename_entries = self._determine_mv_mode(rename_entries, after)
1224
original_modified = self._inventory_is_modified
818
name_tail = splitpath(f)[-1]
819
dest_path = appendpath(to_name, name_tail)
820
result.append((f, dest_path))
821
inv.rename(inv.path2id(f), to_dir_id, name_tail)
823
rename(self.abspath(f), self.abspath(dest_path))
825
raise BzrError("failed to rename %r to %r: %s" %
826
(f, dest_path, e[1]),
827
["rename rolled back"])
1227
self._inventory_is_modified = True
1228
self._move(rename_entries)
829
1230
# restore the inventory on error
830
self._set_inventory(orig_inv)
1231
self._inventory_is_modified = original_modified
832
1233
self._write_inventory(inv)
836
def rename_one(self, from_rel, to_rel):
1234
return rename_tuples
1236
def _determine_mv_mode(self, rename_entries, after=False):
1237
"""Determines for each from-to pair if both inventory and working tree
1238
or only the inventory has to be changed.
1240
Also does basic plausability tests.
1242
inv = self.inventory
1244
for rename_entry in rename_entries:
1245
# store to local variables for easier reference
1246
from_rel = rename_entry.from_rel
1247
from_id = rename_entry.from_id
1248
to_rel = rename_entry.to_rel
1249
to_id = inv.path2id(to_rel)
1250
only_change_inv = False
1252
# check the inventory for source and destination
1254
raise errors.BzrMoveFailedError(from_rel,to_rel,
1255
errors.NotVersionedError(path=str(from_rel)))
1256
if to_id is not None:
1257
raise errors.BzrMoveFailedError(from_rel,to_rel,
1258
errors.AlreadyVersionedError(path=str(to_rel)))
1260
# try to determine the mode for rename (only change inv or change
1261
# inv and file system)
1263
if not self.has_filename(to_rel):
1264
raise errors.BzrMoveFailedError(from_id,to_rel,
1265
errors.NoSuchFile(path=str(to_rel),
1266
extra="New file has not been created yet"))
1267
only_change_inv = True
1268
elif not self.has_filename(from_rel) and self.has_filename(to_rel):
1269
only_change_inv = True
1270
elif self.has_filename(from_rel) and not self.has_filename(to_rel):
1271
only_change_inv = False
1273
# something is wrong, so lets determine what exactly
1274
if not self.has_filename(from_rel) and \
1275
not self.has_filename(to_rel):
1276
raise errors.BzrRenameFailedError(from_rel,to_rel,
1277
errors.PathsDoNotExist(paths=(str(from_rel),
1280
raise errors.RenameFailedFilesExist(from_rel, to_rel,
1281
extra="(Use --after to update the Bazaar id)")
1282
rename_entry.only_change_inv = only_change_inv
1283
return rename_entries
1285
def _move(self, rename_entries):
1286
"""Moves a list of files.
1288
Depending on the value of the flag 'only_change_inv', the
1289
file will be moved on the file system or not.
1291
inv = self.inventory
1294
for entry in rename_entries:
1296
self._move_entry(entry)
1298
self._rollback_move(moved)
1302
def _rollback_move(self, moved):
1303
"""Try to rollback a previous move in case of an filesystem error."""
1304
inv = self.inventory
1307
self._move_entry(_RenameEntry(entry.to_rel, entry.from_id,
1308
entry.to_tail, entry.to_parent_id, entry.from_rel,
1309
entry.from_tail, entry.from_parent_id,
1310
entry.only_change_inv))
1311
except errors.BzrMoveFailedError, e:
1312
raise errors.BzrMoveFailedError( '', '', "Rollback failed."
1313
" The working tree is in an inconsistent state."
1314
" Please consider doing a 'bzr revert'."
1315
" Error message is: %s" % e)
1317
def _move_entry(self, entry):
1318
inv = self.inventory
1319
from_rel_abs = self.abspath(entry.from_rel)
1320
to_rel_abs = self.abspath(entry.to_rel)
1321
if from_rel_abs == to_rel_abs:
1322
raise errors.BzrMoveFailedError(entry.from_rel, entry.to_rel,
1323
"Source and target are identical.")
1325
if not entry.only_change_inv:
1327
osutils.rename(from_rel_abs, to_rel_abs)
1329
raise errors.BzrMoveFailedError(entry.from_rel,
1331
inv.rename(entry.from_id, entry.to_parent_id, entry.to_tail)
1333
@needs_tree_write_lock
1334
def rename_one(self, from_rel, to_rel, after=False):
837
1335
"""Rename one file.
839
1337
This can change the directory or the filename or both.
1339
rename_one has several 'modes' to work. First, it can rename a physical
1340
file and change the file_id. That is the normal mode. Second, it can
1341
only change the file_id without touching any physical file. This is
1342
the new mode introduced in version 0.15.
1344
rename_one uses the second mode if 'after == True' and 'to_rel' is not
1345
versioned but present in the working tree.
1347
rename_one uses the second mode if 'after == False' and 'from_rel' is
1348
versioned but no longer in the working tree, and 'to_rel' is not
1349
versioned but present in the working tree.
1351
rename_one uses the first mode if 'after == False' and 'from_rel' is
1352
versioned and present in the working tree, and 'to_rel' is not
1353
versioned and not present in the working tree.
1355
Everything else results in an error.
841
1357
inv = self.inventory
842
if not self.has_filename(from_rel):
843
raise BzrError("can't rename: old working file %r does not exist" % from_rel)
844
if self.has_filename(to_rel):
845
raise BzrError("can't rename: new working file %r already exists" % to_rel)
847
file_id = inv.path2id(from_rel)
849
raise BzrError("can't rename: old name %r is not versioned" % from_rel)
852
from_parent = entry.parent_id
853
from_name = entry.name
855
if inv.path2id(to_rel):
856
raise BzrError("can't rename: new name %r is already versioned" % to_rel)
1360
# create rename entries and tuples
1361
from_tail = splitpath(from_rel)[-1]
1362
from_id = inv.path2id(from_rel)
1364
raise errors.BzrRenameFailedError(from_rel,to_rel,
1365
errors.NotVersionedError(path=str(from_rel)))
1366
from_entry = inv[from_id]
1367
from_parent_id = from_entry.parent_id
858
1368
to_dir, to_tail = os.path.split(to_rel)
859
1369
to_dir_id = inv.path2id(to_dir)
860
if to_dir_id == None and to_dir != '':
861
raise BzrError("can't determine destination directory id for %r" % to_dir)
863
mutter("rename_one:")
864
mutter(" file_id {%s}" % file_id)
865
mutter(" from_rel %r" % from_rel)
866
mutter(" to_rel %r" % to_rel)
867
mutter(" to_dir %r" % to_dir)
868
mutter(" to_dir_id {%s}" % to_dir_id)
870
inv.rename(file_id, to_dir_id, to_tail)
872
from_abs = self.abspath(from_rel)
873
to_abs = self.abspath(to_rel)
875
rename(from_abs, to_abs)
877
inv.rename(file_id, from_parent, from_name)
878
raise BzrError("failed to rename %r to %r: %s"
879
% (from_abs, to_abs, e[1]),
880
["rename rolled back"])
1370
rename_entry = WorkingTree._RenameEntry(from_rel=from_rel,
1372
from_tail=from_tail,
1373
from_parent_id=from_parent_id,
1374
to_rel=to_rel, to_tail=to_tail,
1375
to_parent_id=to_dir_id)
1376
rename_entries.append(rename_entry)
1378
# determine which move mode to use. checks also for movability
1379
rename_entries = self._determine_mv_mode(rename_entries, after)
1381
# check if the target changed directory and if the target directory is
1383
if to_dir_id is None:
1384
raise errors.BzrMoveFailedError(from_rel,to_rel,
1385
errors.NotVersionedError(path=str(to_dir)))
1387
# all checks done. now we can continue with our actual work
1388
mutter('rename_one:\n'
1393
' to_dir_id {%s}\n',
1394
from_id, from_rel, to_rel, to_dir, to_dir_id)
1396
self._move(rename_entries)
881
1397
self._write_inventory(inv)
1399
class _RenameEntry(object):
1400
def __init__(self, from_rel, from_id, from_tail, from_parent_id,
1401
to_rel, to_tail, to_parent_id, only_change_inv=False):
1402
self.from_rel = from_rel
1403
self.from_id = from_id
1404
self.from_tail = from_tail
1405
self.from_parent_id = from_parent_id
1406
self.to_rel = to_rel
1407
self.to_tail = to_tail
1408
self.to_parent_id = to_parent_id
1409
self.only_change_inv = only_change_inv
883
1411
@needs_read_lock
884
1412
def unknowns(self):
885
1413
"""Return all unknown files.
887
1415
These are files in the working directory that are not versioned or
888
1416
control files or ignored.
890
>>> from bzrlib.bzrdir import ScratchDir
891
>>> d = ScratchDir(files=['foo', 'foo~'])
892
>>> b = d.open_branch()
893
>>> tree = d.open_workingtree()
894
>>> map(str, tree.unknowns())
897
>>> list(b.unknowns())
899
>>> tree.remove('foo')
900
>>> list(b.unknowns())
903
for subp in self.extras():
904
if not self.is_ignored(subp):
1418
# force the extras method to be fully executed before returning, to
1419
# prevent race conditions with the lock
1421
[subp for subp in self.extras() if not self.is_ignored(subp)])
1423
@needs_tree_write_lock
1424
def unversion(self, file_ids):
1425
"""Remove the file ids in file_ids from the current versioned set.
1427
When a file_id is unversioned, all of its children are automatically
1430
:param file_ids: The file ids to stop versioning.
1431
:raises: NoSuchId if any fileid is not currently versioned.
1433
for file_id in file_ids:
1434
file_id = osutils.safe_file_id(file_id)
1435
if self._inventory.has_id(file_id):
1436
self._inventory.remove_recursive_id(file_id)
1438
raise errors.NoSuchId(self, file_id)
1440
# in the future this should just set a dirty bit to wait for the
1441
# final unlock. However, until all methods of workingtree start
1442
# with the current in -memory inventory rather than triggering
1443
# a read, it is more complex - we need to teach read_inventory
1444
# to know when to read, and when to not read first... and possibly
1445
# to save first when the in memory one may be corrupted.
1446
# so for now, we just only write it if it is indeed dirty.
1448
self._write_inventory(self._inventory)
907
1450
@deprecated_method(zero_eight)
908
1451
def iter_conflicts(self):
909
1452
"""List all files in the tree that have text or content conflicts.
1155
1716
if new_revision is None:
1156
1717
self.branch.set_revision_history([])
1158
# current format is locked in with the branch
1159
revision_history = self.branch.revision_history()
1161
position = revision_history.index(new_revision)
1163
raise errors.NoSuchRevision(self.branch, new_revision)
1164
self.branch.set_revision_history(revision_history[:position + 1])
1720
self.branch.generate_revision_history(new_revision)
1721
except errors.NoSuchRevision:
1722
# not present in the repo - dont try to set it deeper than the tip
1723
self.branch.set_revision_history([new_revision])
1726
def _write_basis_inventory(self, xml):
1727
"""Write the basis inventory XML to the basis-inventory file"""
1728
assert isinstance(xml, str), 'serialised xml must be bytestring.'
1729
path = self._basis_inventory_name()
1731
self._control_files.put(path, sio)
1733
def _create_basis_xml_from_inventory(self, revision_id, inventory):
1734
"""Create the text that will be saved in basis-inventory"""
1735
# TODO: jam 20070209 This should be redundant, as the revision_id
1736
# as all callers should have already converted the revision_id to
1738
inventory.revision_id = osutils.safe_revision_id(revision_id)
1739
return xml7.serializer_v7.write_inventory_to_string(inventory)
1167
1741
def _cache_basis_inventory(self, new_revision):
1168
1742
"""Cache new_revision as the basis inventory."""
1743
# TODO: this should allow the ready-to-use inventory to be passed in,
1744
# as commit already has that ready-to-use [while the format is the
1170
1747
# this double handles the inventory - unpack and repack -
1171
1748
# but is easier to understand. We can/should put a conditional
1172
1749
# in here based on whether the inventory is in the latest format
1173
1750
# - perhaps we should repack all inventories on a repository
1175
inv = self.branch.repository.get_inventory(new_revision)
1176
inv.revision_id = new_revision
1177
xml = bzrlib.xml5.serializer_v5.write_inventory_to_string(inv)
1179
path = self._basis_inventory_name()
1180
self._control_files.put_utf8(path, xml)
1181
except WeaveRevisionNotPresent:
1752
# the fast path is to copy the raw xml from the repository. If the
1753
# xml contains 'revision_id="', then we assume the right
1754
# revision_id is set. We must check for this full string, because a
1755
# root node id can legitimately look like 'revision_id' but cannot
1757
xml = self.branch.repository.get_inventory_xml(new_revision)
1758
firstline = xml.split('\n', 1)[0]
1759
if (not 'revision_id="' in firstline or
1760
'format="7"' not in firstline):
1761
inv = self.branch.repository.deserialise_inventory(
1763
xml = self._create_basis_xml_from_inventory(new_revision, inv)
1764
self._write_basis_inventory(xml)
1765
except (errors.NoSuchRevision, errors.RevisionNotPresent):
1184
1768
def read_basis_inventory(self):
1185
1769
"""Read the cached basis inventory."""
1186
1770
path = self._basis_inventory_name()
1187
return self._control_files.get_utf8(path).read()
1771
return self._control_files.get(path).read()
1189
1773
@needs_read_lock
1190
1774
def read_working_inventory(self):
1191
"""Read the working inventory."""
1775
"""Read the working inventory.
1777
:raises errors.InventoryModified: read_working_inventory will fail
1778
when the current in memory inventory has been modified.
1780
# conceptually this should be an implementation detail of the tree.
1781
# XXX: Deprecate this.
1192
1782
# ElementTree does its own conversion from UTF-8, so open in
1194
result = bzrlib.xml5.serializer_v5.read_inventory(
1195
self._control_files.get('inventory'))
1196
self._set_inventory(result)
1784
if self._inventory_is_modified:
1785
raise errors.InventoryModified(self)
1786
result = self._deserialize(self._control_files.get('inventory'))
1787
self._set_inventory(result, dirty=False)
1200
def remove(self, files, verbose=False):
1790
@needs_tree_write_lock
1791
def remove(self, files, verbose=False, to_file=None):
1201
1792
"""Remove nominated files from the working inventory..
1203
1794
This does not remove their text. This does not run on XXX on what? RBC
1300
1940
between multiple working trees, i.e. via shared storage, then we
1301
1941
would probably want to lock both the local tree, and the branch.
1303
# FIXME: We want to write out the hashcache only when the last lock on
1304
# this working copy is released. Peeking at the lock count is a bit
1305
# of a nasty hack; probably it's better to have a transaction object,
1306
# which can do some finalization when it's either successfully or
1307
# unsuccessfully completed. (Denys's original patch did that.)
1308
# RBC 20060206 hookinhg into transaction will couple lock and transaction
1309
# wrongly. Hookinh into unllock on the control files object is fine though.
1311
# TODO: split this per format so there is no ugly if block
1312
if self._hashcache.needs_write and (
1313
# dedicated lock files
1314
self._control_files._lock_count==1 or
1316
(self._control_files is self.branch.control_files and
1317
self._control_files._lock_count==3)):
1318
self._hashcache.write()
1319
# reverse order of locking.
1321
return self._control_files.unlock()
1323
self.branch.unlock()
1943
raise NotImplementedError(self.unlock)
1326
1945
def update(self):
1327
1946
"""Update a working tree along its branch.
1329
This will update the branch if its bound too, which means we have multiple trees involved:
1330
The new basis tree of the master.
1331
The old basis tree of the branch.
1332
The old basis tree of the working tree.
1333
The current working tree state.
1334
pathologically all three may be different, and non ancestors of each other.
1335
Conceptually we want to:
1336
Preserve the wt.basis->wt.state changes
1337
Transform the wt.basis to the new master basis.
1338
Apply a merge of the old branch basis to get any 'local' changes from it into the tree.
1339
Restore the wt.basis->wt.state changes.
1948
This will update the branch if its bound too, which means we have
1949
multiple trees involved:
1951
- The new basis tree of the master.
1952
- The old basis tree of the branch.
1953
- The old basis tree of the working tree.
1954
- The current working tree state.
1956
Pathologically, all three may be different, and non-ancestors of each
1957
other. Conceptually we want to:
1959
- Preserve the wt.basis->wt.state changes
1960
- Transform the wt.basis to the new master basis.
1961
- Apply a merge of the old branch basis to get any 'local' changes from
1963
- Restore the wt.basis->wt.state changes.
1341
1965
There isn't a single operation at the moment to do that, so we:
1342
Merge current state -> basis tree of the master w.r.t. the old tree basis.
1343
Do a 'normal' merge of the old branch basis if it is relevant.
1345
old_tip = self.branch.update()
1346
if old_tip is not None:
1347
self.add_pending_merge(old_tip)
1348
self.branch.lock_read()
1351
if self.last_revision() != self.branch.last_revision():
1352
# merge tree state up to new branch tip.
1353
basis = self.basis_tree()
1966
- Merge current state -> basis tree of the master w.r.t. the old tree
1968
- Do a 'normal' merge of the old branch basis if it is relevant.
1970
if self.branch.get_master_branch() is not None:
1972
update_branch = True
1974
self.lock_tree_write()
1975
update_branch = False
1978
old_tip = self.branch.update()
1981
return self._update_tree(old_tip)
1985
@needs_tree_write_lock
1986
def _update_tree(self, old_tip=None):
1987
"""Update a tree to the master branch.
1989
:param old_tip: if supplied, the previous tip revision the branch,
1990
before it was changed to the master branch's tip.
1992
# here if old_tip is not None, it is the old tip of the branch before
1993
# it was updated from the master branch. This should become a pending
1994
# merge in the working tree to preserve the user existing work. we
1995
# cant set that until we update the working trees last revision to be
1996
# one from the new branch, because it will just get absorbed by the
1997
# parent de-duplication logic.
1999
# We MUST save it even if an error occurs, because otherwise the users
2000
# local work is unreferenced and will appear to have been lost.
2004
last_rev = self.get_parent_ids()[0]
2007
if last_rev != self.branch.last_revision():
2008
# merge tree state up to new branch tip.
2009
basis = self.basis_tree()
1354
2012
to_tree = self.branch.basis_tree()
1355
result += merge_inner(self.branch,
2013
if basis.inventory.root is None:
2014
self.set_root_id(to_tree.inventory.root.file_id)
2016
result += merge.merge_inner(
1358
2020
this_tree=self)
1359
self.set_last_revision(self.branch.last_revision())
1360
if old_tip and old_tip != self.last_revision():
1361
# our last revision was not the prior branch last reivison
1362
# and we have converted that last revision to a pending merge.
1363
# base is somewhere between the branch tip now
1364
# and the now pending merge
1365
from bzrlib.revision import common_ancestor
1367
base_rev_id = common_ancestor(self.branch.last_revision(),
1369
self.branch.repository)
1370
except errors.NoCommonAncestor:
1372
base_tree = self.branch.repository.revision_tree(base_rev_id)
1373
other_tree = self.branch.repository.revision_tree(old_tip)
1374
result += merge_inner(self.branch,
1380
self.branch.unlock()
2023
# TODO - dedup parents list with things merged by pull ?
2024
# reuse the tree we've updated to to set the basis:
2025
parent_trees = [(self.branch.last_revision(), to_tree)]
2026
merges = self.get_parent_ids()[1:]
2027
# Ideally we ask the tree for the trees here, that way the working
2028
# tree can decide whether to give us teh entire tree or give us a
2029
# lazy initialised tree. dirstate for instance will have the trees
2030
# in ram already, whereas a last-revision + basis-inventory tree
2031
# will not, but also does not need them when setting parents.
2032
for parent in merges:
2033
parent_trees.append(
2034
(parent, self.branch.repository.revision_tree(parent)))
2035
if old_tip is not None:
2036
parent_trees.append(
2037
(old_tip, self.branch.repository.revision_tree(old_tip)))
2038
self.set_parent_trees(parent_trees)
2039
last_rev = parent_trees[0][0]
2041
# the working tree had the same last-revision as the master
2042
# branch did. We may still have pivot local work from the local
2043
# branch into old_tip:
2044
if old_tip is not None:
2045
self.add_parent_tree_id(old_tip)
2046
if old_tip and old_tip != last_rev:
2047
# our last revision was not the prior branch last revision
2048
# and we have converted that last revision to a pending merge.
2049
# base is somewhere between the branch tip now
2050
# and the now pending merge
2052
# Since we just modified the working tree and inventory, flush out
2053
# the current state, before we modify it again.
2054
# TODO: jam 20070214 WorkingTree3 doesn't require this, dirstate
2055
# requires it only because TreeTransform directly munges the
2056
# inventory and calls tree._write_inventory(). Ultimately we
2057
# should be able to remove this extra flush.
2059
from bzrlib.revision import common_ancestor
2061
base_rev_id = common_ancestor(self.branch.last_revision(),
2063
self.branch.repository)
2064
except errors.NoCommonAncestor:
2066
base_tree = self.branch.repository.revision_tree(base_rev_id)
2067
other_tree = self.branch.repository.revision_tree(old_tip)
2068
result += merge.merge_inner(
2075
def _write_hashcache_if_dirty(self):
2076
"""Write out the hashcache if it is dirty."""
2077
if self._hashcache.needs_write:
2079
self._hashcache.write()
2081
if e.errno not in (errno.EPERM, errno.EACCES):
2083
# TODO: jam 20061219 Should this be a warning? A single line
2084
# warning might be sufficient to let the user know what
2086
mutter('Could not write hashcache for %s\nError: %s',
2087
self._hashcache.cache_file_name(), e)
2089
@needs_tree_write_lock
1383
2090
def _write_inventory(self, inv):
1384
2091
"""Write inventory as the current inventory."""
1386
bzrlib.xml5.serializer_v5.write_inventory(inv, sio)
1388
self._control_files.put('inventory', sio)
1389
self._set_inventory(inv)
1390
mutter('wrote working inventory')
2092
self._set_inventory(inv, dirty=True)
1392
2095
def set_conflicts(self, arg):
1393
raise UnsupportedOperation(self.set_conflicts, self)
2096
raise errors.UnsupportedOperation(self.set_conflicts, self)
2098
def add_conflicts(self, arg):
2099
raise errors.UnsupportedOperation(self.add_conflicts, self)
1395
2101
@needs_read_lock
1396
2102
def conflicts(self):
1397
conflicts = ConflictList()
2103
conflicts = _mod_conflicts.ConflictList()
1398
2104
for conflicted in self._iter_conflicts():
1401
2107
if file_kind(self.abspath(conflicted)) != "file":
1404
if e.errno == errno.ENOENT:
2109
except errors.NoSuchFile:
1408
2111
if text is True:
1409
2112
for suffix in ('.THIS', '.OTHER'):
1411
2114
kind = file_kind(self.abspath(conflicted+suffix))
1413
if e.errno == errno.ENOENT:
2117
except errors.NoSuchFile:
1421
2121
ctype = {True: 'text conflict', False: 'contents conflict'}[text]
1422
conflicts.append(Conflict.factory(ctype, path=conflicted,
2122
conflicts.append(_mod_conflicts.Conflict.factory(ctype,
1423
2124
file_id=self.path2id(conflicted)))
1424
2125
return conflicts
2127
def walkdirs(self, prefix=""):
2128
"""Walk the directories of this tree.
2130
This API returns a generator, which is only valid during the current
2131
tree transaction - within a single lock_read or lock_write duration.
2133
If the tree is not locked, it may cause an error to be raised, depending
2134
on the tree implementation.
2136
disk_top = self.abspath(prefix)
2137
if disk_top.endswith('/'):
2138
disk_top = disk_top[:-1]
2139
top_strip_len = len(disk_top) + 1
2140
inventory_iterator = self._walkdirs(prefix)
2141
disk_iterator = osutils.walkdirs(disk_top, prefix)
2143
current_disk = disk_iterator.next()
2144
disk_finished = False
2146
if e.errno != errno.ENOENT:
2149
disk_finished = True
2151
current_inv = inventory_iterator.next()
2152
inv_finished = False
2153
except StopIteration:
2156
while not inv_finished or not disk_finished:
2157
if not disk_finished:
2158
# strip out .bzr dirs
2159
if current_disk[0][1][top_strip_len:] == '':
2160
# osutils.walkdirs can be made nicer -
2161
# yield the path-from-prefix rather than the pathjoined
2163
bzrdir_loc = bisect_left(current_disk[1], ('.bzr', '.bzr'))
2164
if current_disk[1][bzrdir_loc][0] == '.bzr':
2165
# we dont yield the contents of, or, .bzr itself.
2166
del current_disk[1][bzrdir_loc]
2168
# everything is unknown
2171
# everything is missing
2174
direction = cmp(current_inv[0][0], current_disk[0][0])
2176
# disk is before inventory - unknown
2177
dirblock = [(relpath, basename, kind, stat, None, None) for
2178
relpath, basename, kind, stat, top_path in current_disk[1]]
2179
yield (current_disk[0][0], None), dirblock
2181
current_disk = disk_iterator.next()
2182
except StopIteration:
2183
disk_finished = True
2185
# inventory is before disk - missing.
2186
dirblock = [(relpath, basename, 'unknown', None, fileid, kind)
2187
for relpath, basename, dkind, stat, fileid, kind in
2189
yield (current_inv[0][0], current_inv[0][1]), dirblock
2191
current_inv = inventory_iterator.next()
2192
except StopIteration:
2195
# versioned present directory
2196
# merge the inventory and disk data together
2198
for relpath, subiterator in itertools.groupby(sorted(
2199
current_inv[1] + current_disk[1], key=operator.itemgetter(0)), operator.itemgetter(1)):
2200
path_elements = list(subiterator)
2201
if len(path_elements) == 2:
2202
inv_row, disk_row = path_elements
2203
# versioned, present file
2204
dirblock.append((inv_row[0],
2205
inv_row[1], disk_row[2],
2206
disk_row[3], inv_row[4],
2208
elif len(path_elements[0]) == 5:
2210
dirblock.append((path_elements[0][0],
2211
path_elements[0][1], path_elements[0][2],
2212
path_elements[0][3], None, None))
2213
elif len(path_elements[0]) == 6:
2214
# versioned, absent file.
2215
dirblock.append((path_elements[0][0],
2216
path_elements[0][1], 'unknown', None,
2217
path_elements[0][4], path_elements[0][5]))
2219
raise NotImplementedError('unreachable code')
2220
yield current_inv[0], dirblock
2222
current_inv = inventory_iterator.next()
2223
except StopIteration:
2226
current_disk = disk_iterator.next()
2227
except StopIteration:
2228
disk_finished = True
2230
def _walkdirs(self, prefix=""):
2231
_directory = 'directory'
2232
# get the root in the inventory
2233
inv = self.inventory
2234
top_id = inv.path2id(prefix)
2238
pending = [(prefix, '', _directory, None, top_id, None)]
2241
currentdir = pending.pop()
2242
# 0 - relpath, 1- basename, 2- kind, 3- stat, 4-id, 5-kind
2243
top_id = currentdir[4]
2245
relroot = currentdir[0] + '/'
2248
# FIXME: stash the node in pending
2250
for name, child in entry.sorted_children():
2251
dirblock.append((relroot + name, name, child.kind, None,
2252
child.file_id, child.kind
2254
yield (currentdir[0], entry.file_id), dirblock
2255
# push the user specified dirs from dirblock
2256
for dir in reversed(dirblock):
2257
if dir[2] == _directory:
2260
@needs_tree_write_lock
2261
def auto_resolve(self):
2262
"""Automatically resolve text conflicts according to contents.
2264
Only text conflicts are auto_resolvable. Files with no conflict markers
2265
are considered 'resolved', because bzr always puts conflict markers
2266
into files that have text conflicts. The corresponding .THIS .BASE and
2267
.OTHER files are deleted, as per 'resolve'.
2268
:return: a tuple of ConflictLists: (un_resolved, resolved).
2270
un_resolved = _mod_conflicts.ConflictList()
2271
resolved = _mod_conflicts.ConflictList()
2272
conflict_re = re.compile('^(<{7}|={7}|>{7})')
2273
for conflict in self.conflicts():
2274
if (conflict.typestring != 'text conflict' or
2275
self.kind(conflict.file_id) != 'file'):
2276
un_resolved.append(conflict)
2278
my_file = open(self.id2abspath(conflict.file_id), 'rb')
2280
for line in my_file:
2281
if conflict_re.search(line):
2282
un_resolved.append(conflict)
2285
resolved.append(conflict)
2288
resolved.remove_files(self)
2289
self.set_conflicts(un_resolved)
2290
return un_resolved, resolved
2293
class WorkingTree2(WorkingTree):
2294
"""This is the Format 2 working tree.
2296
This was the first weave based working tree.
2297
- uses os locks for locking.
2298
- uses the branch last-revision.
2301
def lock_tree_write(self):
2302
"""See WorkingTree.lock_tree_write().
2304
In Format2 WorkingTrees we have a single lock for the branch and tree
2305
so lock_tree_write() degrades to lock_write().
2307
self.branch.lock_write()
2309
return self._control_files.lock_write()
2311
self.branch.unlock()
2315
# we share control files:
2316
if self._control_files._lock_count == 3:
2317
# _inventory_is_modified is always False during a read lock.
2318
if self._inventory_is_modified:
2320
self._write_hashcache_if_dirty()
2322
# reverse order of locking.
2324
return self._control_files.unlock()
2326
self.branch.unlock()
1427
2329
class WorkingTree3(WorkingTree):
1428
2330
"""This is the Format 3 working tree.