25
23
from inventory import Inventory
26
24
from trace import mutter, note
27
from tree import Tree, EmptyTree, RevisionTree, WorkingTree
25
from tree import Tree, EmptyTree, RevisionTree
28
26
from inventory import InventoryEntry, Inventory
29
27
from osutils import isdir, quotefn, isfile, uuid, sha_file, username, \
30
28
format_date, compact_date, pumpfile, user_email, rand_bytes, splitpath, \
31
joinpath, sha_string, file_kind, local_time_offset, appendpath
29
joinpath, sha_file, sha_string, file_kind, local_time_offset, appendpath
32
30
from store import ImmutableStore
33
31
from revision import Revision
34
from errors import bailout, BzrError
32
from errors import BzrError
35
33
from textui import show_status
36
from diff import diff_trees
38
35
BZR_BRANCH_FORMAT = "Bazaar-NG branch, format 0.0.4\n"
39
36
## TODO: Maybe include checks for common corruption of newlines, etc?
43
40
def find_branch(f, **args):
44
if f.startswith('http://') or f.startswith('https://'):
41
if f and (f.startswith('http://') or f.startswith('https://')):
45
42
import remotebranch
46
43
return remotebranch.RemoteBranch(f, **args)
48
45
return Branch(f, **args)
48
def find_cached_branch(f, cache_root, **args):
49
from remotebranch import RemoteBranch
50
br = find_branch(f, **args)
51
def cacheify(br, store_name):
52
from meta_store import CachedStore
53
cache_path = os.path.join(cache_root, store_name)
55
new_store = CachedStore(getattr(br, store_name), cache_path)
56
setattr(br, store_name, new_store)
58
if isinstance(br, RemoteBranch):
59
cacheify(br, 'inventory_store')
60
cacheify(br, 'text_store')
61
cacheify(br, 'revision_store')
65
def _relpath(base, path):
66
"""Return path relative to base, or raise exception.
68
The path may be either an absolute path or a path relative to the
69
current working directory.
71
Lifted out of Branch.relpath for ease of testing.
73
os.path.commonprefix (python2.4) has a bad bug that it works just
74
on string prefixes, assuming that '/u' is a prefix of '/u2'. This
75
avoids that problem."""
76
rp = os.path.abspath(path)
80
while len(head) >= len(base):
83
head, tail = os.path.split(head)
87
from errors import NotBranchError
88
raise NotBranchError("path %r is not within branch %r" % (rp, base))
51
93
def find_branch_root(f=None):
132
200
__repr__ = __str__
136
def lock(self, mode='w'):
137
"""Lock the on-disk branch, excluding other processes."""
143
om = os.O_WRONLY | os.O_CREAT
148
raise BzrError("invalid locking mode %r" % mode)
151
lockfile = os.open(self.controlfilename('branch-lock'), om)
153
if e.errno == errno.ENOENT:
154
# might not exist on branches from <0.0.4
155
self.controlfile('branch-lock', 'w').close()
156
lockfile = os.open(self.controlfilename('branch-lock'), om)
204
if self._lock_mode or self._lock:
205
from warnings import warn
206
warn("branch %r was not explicitly unlocked" % self)
211
def lock_write(self):
213
if self._lock_mode != 'w':
214
from errors import LockError
215
raise LockError("can't upgrade to a write lock from %r" %
217
self._lock_count += 1
219
from bzrlib.lock import WriteLock
221
self._lock = WriteLock(self.controlfilename('branch-lock'))
222
self._lock_mode = 'w'
229
assert self._lock_mode in ('r', 'w'), \
230
"invalid lock mode %r" % self._lock_mode
231
self._lock_count += 1
233
from bzrlib.lock import ReadLock
235
self._lock = ReadLock(self.controlfilename('branch-lock'))
236
self._lock_mode = 'r'
160
fcntl.lockf(lockfile, lm)
162
fcntl.lockf(lockfile, fcntl.LOCK_UN)
164
self._lockmode = None
166
self._lockmode = mode
168
warning("please write a locking method for platform %r" % sys.platform)
170
self._lockmode = None
172
self._lockmode = mode
175
def _need_readlock(self):
176
if self._lockmode not in ['r', 'w']:
177
raise BzrError('need read lock on branch, only have %r' % self._lockmode)
179
def _need_writelock(self):
180
if self._lockmode not in ['w']:
181
raise BzrError('need write lock on branch, only have %r' % self._lockmode)
242
if not self._lock_mode:
243
from errors import LockError
244
raise LockError('branch %r is not locked' % (self))
246
if self._lock_count > 1:
247
self._lock_count -= 1
251
self._lock_mode = self._lock_count = None
184
254
def abspath(self, name):
312
382
This puts the files in the Added state, so that they will be
313
383
recorded by the next commit.
386
List of paths to add, relative to the base of the tree.
389
If set, use these instead of automatically generated ids.
390
Must be the same length as the list of files, but may
391
contain None for ids that are to be autogenerated.
315
393
TODO: Perhaps have an option to add the ids even if the files do
318
396
TODO: Perhaps return the ids of the files? But then again it
319
is easy to retrieve them if they're needed.
321
TODO: Option to specify file id.
397
is easy to retrieve them if they're needed.
323
399
TODO: Adding a directory should optionally recurse down and
324
add all non-ignored children. Perhaps do that in a
327
>>> b = ScratchBranch(files=['foo'])
328
>>> 'foo' in b.unknowns()
333
>>> 'foo' in b.unknowns()
335
>>> bool(b.inventory.path2id('foo'))
341
Traceback (most recent call last):
343
BzrError: ('foo is already versioned', [])
345
>>> b.add(['nothere'])
346
Traceback (most recent call last):
347
BzrError: ('cannot add: not a regular file or directory: nothere', [])
400
add all non-ignored children. Perhaps do that in a
349
self._need_writelock()
351
403
# TODO: Re-adding a file that is removed in the working copy
352
404
# should probably put it back with the previous ID.
353
405
if isinstance(files, types.StringTypes):
406
assert(ids is None or isinstance(ids, types.StringTypes))
356
inv = self.read_working_inventory()
358
if is_control_file(f):
359
bailout("cannot add control file %s" % quotefn(f))
364
bailout("cannot add top-level %r" % f)
366
fullpath = os.path.normpath(self.abspath(f))
369
kind = file_kind(fullpath)
371
# maybe something better?
372
bailout('cannot add: not a regular file or directory: %s' % quotefn(f))
374
if kind != 'file' and kind != 'directory':
375
bailout('cannot add: not a regular file or directory: %s' % quotefn(f))
377
file_id = gen_file_id(f)
378
inv.add_path(f, kind=kind, file_id=file_id)
381
show_status('A', kind, quotefn(f))
383
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
385
self._write_inventory(inv)
412
ids = [None] * len(files)
414
assert(len(ids) == len(files))
418
inv = self.read_working_inventory()
419
for f,file_id in zip(files, ids):
420
if is_control_file(f):
421
raise BzrError("cannot add control file %s" % quotefn(f))
426
raise BzrError("cannot add top-level %r" % f)
428
fullpath = os.path.normpath(self.abspath(f))
431
kind = file_kind(fullpath)
433
# maybe something better?
434
raise BzrError('cannot add: not a regular file or directory: %s' % quotefn(f))
436
if kind != 'file' and kind != 'directory':
437
raise BzrError('cannot add: not a regular file or directory: %s' % quotefn(f))
440
file_id = gen_file_id(f)
441
inv.add_path(f, kind=kind, file_id=file_id)
444
print 'added', quotefn(f)
446
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
448
self._write_inventory(inv)
388
453
def print_file(self, file, revno):
389
454
"""Print `file` to stdout."""
390
self._need_readlock()
391
tree = self.revision_tree(self.lookup_revision(revno))
392
# use inventory as it was in that revision
393
file_id = tree.inventory.path2id(file)
395
bailout("%r is not present in revision %d" % (file, revno))
396
tree.print_file(file_id)
457
tree = self.revision_tree(self.lookup_revision(revno))
458
# use inventory as it was in that revision
459
file_id = tree.inventory.path2id(file)
461
raise BzrError("%r is not present in revision %d" % (file, revno))
462
tree.print_file(file_id)
399
467
def remove(self, files, verbose=False):
400
468
"""Mark nominated files for removal from the inventory.
435
481
## TODO: Normalize names
436
482
## TODO: Remove nested loops; better scalability
437
self._need_writelock()
439
483
if isinstance(files, types.StringTypes):
442
tree = self.working_tree()
445
# do this before any modifications
449
bailout("cannot remove unversioned file %s" % quotefn(f))
450
mutter("remove inventory entry %s {%s}" % (quotefn(f), fid))
452
# having remove it, it must be either ignored or unknown
453
if tree.is_ignored(f):
457
show_status(new_status, inv[fid].kind, quotefn(f))
489
tree = self.working_tree()
492
# do this before any modifications
496
raise BzrError("cannot remove unversioned file %s" % quotefn(f))
497
mutter("remove inventory entry %s {%s}" % (quotefn(f), fid))
499
# having remove it, it must be either ignored or unknown
500
if tree.is_ignored(f):
504
show_status(new_status, inv[fid].kind, quotefn(f))
507
self._write_inventory(inv)
512
# FIXME: this doesn't need to be a branch method
513
def set_inventory(self, new_inventory_list):
515
for path, file_id, parent, kind in new_inventory_list:
516
name = os.path.basename(path)
519
inv.add(InventoryEntry(file_id, name, kind, parent))
460
520
self._write_inventory(inv)
479
539
return self.working_tree().unknowns()
482
def commit(self, message, timestamp=None, timezone=None,
485
"""Commit working copy as a new revision.
487
The basic approach is to add all the file texts into the
488
store, then the inventory, then make a new revision pointing
489
to that inventory and store that.
491
This is not quite safe if the working copy changes during the
492
commit; for the moment that is simply not allowed. A better
493
approach is to make a temporary copy of the files before
494
computing their hashes, and then add those hashes in turn to
495
the inventory. This should mean at least that there are no
496
broken hash pointers. There is no way we can get a snapshot
497
of the whole directory at an instant. This would also have to
498
be robust against files disappearing, moving, etc. So the
499
whole thing is a bit hard.
501
timestamp -- if not None, seconds-since-epoch for a
502
postdated/predated commit.
504
self._need_writelock()
506
## TODO: Show branch names
508
# TODO: Don't commit if there are no changes, unless forced?
510
# First walk over the working inventory; and both update that
511
# and also build a new revision inventory. The revision
512
# inventory needs to hold the text-id, sha1 and size of the
513
# actual file versions committed in the revision. (These are
514
# not present in the working inventory.) We also need to
515
# detect missing/deleted files, and remove them from the
518
work_inv = self.read_working_inventory()
520
basis = self.basis_tree()
521
basis_inv = basis.inventory
523
for path, entry in work_inv.iter_entries():
524
## TODO: Cope with files that have gone missing.
526
## TODO: Check that the file kind has not changed from the previous
527
## revision of this file (if any).
531
p = self.abspath(path)
532
file_id = entry.file_id
533
mutter('commit prep file %s, id %r ' % (p, file_id))
535
if not os.path.exists(p):
536
mutter(" file is missing, removing from inventory")
538
show_status('D', entry.kind, quotefn(path))
539
missing_ids.append(file_id)
542
# TODO: Handle files that have been deleted
544
# TODO: Maybe a special case for empty files? Seems a
545
# waste to store them many times.
549
if basis_inv.has_id(file_id):
550
old_kind = basis_inv[file_id].kind
551
if old_kind != entry.kind:
552
bailout("entry %r changed kind from %r to %r"
553
% (file_id, old_kind, entry.kind))
555
if entry.kind == 'directory':
557
bailout("%s is entered as directory but not a directory" % quotefn(p))
558
elif entry.kind == 'file':
560
bailout("%s is entered as file but is not a file" % quotefn(p))
562
content = file(p, 'rb').read()
564
entry.text_sha1 = sha_string(content)
565
entry.text_size = len(content)
567
old_ie = basis_inv.has_id(file_id) and basis_inv[file_id]
569
and (old_ie.text_size == entry.text_size)
570
and (old_ie.text_sha1 == entry.text_sha1)):
571
## assert content == basis.get_file(file_id).read()
572
entry.text_id = basis_inv[file_id].text_id
573
mutter(' unchanged from previous text_id {%s}' %
577
entry.text_id = gen_file_id(entry.name)
578
self.text_store.add(content, entry.text_id)
579
mutter(' stored with text_id {%s}' % entry.text_id)
583
elif (old_ie.name == entry.name
584
and old_ie.parent_id == entry.parent_id):
589
show_status(state, entry.kind, quotefn(path))
591
for file_id in missing_ids:
592
# have to do this later so we don't mess up the iterator.
593
# since parents may be removed before their children we
596
# FIXME: There's probably a better way to do this; perhaps
597
# the workingtree should know how to filter itself.
598
if work_inv.has_id(file_id):
599
del work_inv[file_id]
602
inv_id = rev_id = _gen_revision_id(time.time())
604
inv_tmp = tempfile.TemporaryFile()
605
inv.write_xml(inv_tmp)
607
self.inventory_store.add(inv_tmp, inv_id)
608
mutter('new inventory_id is {%s}' % inv_id)
610
self._write_inventory(work_inv)
612
if timestamp == None:
613
timestamp = time.time()
615
if committer == None:
616
committer = username()
619
timezone = local_time_offset()
621
mutter("building commit log message")
622
rev = Revision(timestamp=timestamp,
625
precursor = self.last_patch(),
630
rev_tmp = tempfile.TemporaryFile()
631
rev.write_xml(rev_tmp)
633
self.revision_store.add(rev_tmp, rev_id)
634
mutter("new revision_id is {%s}" % rev_id)
636
## XXX: Everything up to here can simply be orphaned if we abort
637
## the commit; it will leave junk files behind but that doesn't
640
## TODO: Read back the just-generated changeset, and make sure it
641
## applies and recreates the right state.
643
## TODO: Also calculate and store the inventory SHA1
644
mutter("committing patch r%d" % (self.revno() + 1))
647
self.append_revision(rev_id)
650
note("commited r%d" % self.revno())
653
542
def append_revision(self, revision_id):
543
from bzrlib.atomicfile import AtomicFile
654
545
mutter("add {%s} to revision-history" % revision_id)
655
rev_history = self.revision_history()
657
tmprhname = self.controlfilename('revision-history.tmp')
658
rhname = self.controlfilename('revision-history')
660
f = file(tmprhname, 'wt')
661
rev_history.append(revision_id)
662
f.write('\n'.join(rev_history))
666
if sys.platform == 'win32':
668
os.rename(tmprhname, rhname)
546
rev_history = self.revision_history() + [revision_id]
548
f = AtomicFile(self.controlfilename('revision-history'))
550
for rev_id in rev_history:
672
557
def get_revision(self, revision_id):
673
558
"""Return the Revision object for a named revision"""
674
self._need_readlock()
559
if not revision_id or not isinstance(revision_id, basestring):
560
raise ValueError('invalid revision-id: %r' % revision_id)
675
561
r = Revision.read_xml(self.revision_store[revision_id])
676
562
assert r.revision_id == revision_id
565
def get_revision_sha1(self, revision_id):
566
"""Hash the stored value of a revision, and return it."""
567
# In the future, revision entries will be signed. At that
568
# point, it is probably best *not* to include the signature
569
# in the revision hash. Because that lets you re-sign
570
# the revision, (add signatures/remove signatures) and still
571
# have all hash pointers stay consistent.
572
# But for now, just hash the contents.
573
return sha_file(self.revision_store[revision_id])
680
576
def get_inventory(self, inventory_id):
681
577
"""Get Inventory object by hash.
703
602
>>> ScratchBranch().revision_history()
706
self._need_readlock()
707
return [l.rstrip('\r\n') for l in self.controlfile('revision-history', 'r').readlines()]
607
return [l.rstrip('\r\n') for l in
608
self.controlfile('revision-history', 'r').readlines()]
613
def common_ancestor(self, other, self_revno=None, other_revno=None):
616
>>> sb = ScratchBranch(files=['foo', 'foo~'])
617
>>> sb.common_ancestor(sb) == (None, None)
619
>>> commit.commit(sb, "Committing first revision", verbose=False)
620
>>> sb.common_ancestor(sb)[0]
622
>>> clone = sb.clone()
623
>>> commit.commit(sb, "Committing second revision", verbose=False)
624
>>> sb.common_ancestor(sb)[0]
626
>>> sb.common_ancestor(clone)[0]
628
>>> commit.commit(clone, "Committing divergent second revision",
630
>>> sb.common_ancestor(clone)[0]
632
>>> sb.common_ancestor(clone) == clone.common_ancestor(sb)
634
>>> sb.common_ancestor(sb) != clone.common_ancestor(clone)
636
>>> clone2 = sb.clone()
637
>>> sb.common_ancestor(clone2)[0]
639
>>> sb.common_ancestor(clone2, self_revno=1)[0]
641
>>> sb.common_ancestor(clone2, other_revno=1)[0]
644
my_history = self.revision_history()
645
other_history = other.revision_history()
646
if self_revno is None:
647
self_revno = len(my_history)
648
if other_revno is None:
649
other_revno = len(other_history)
650
indices = range(min((self_revno, other_revno)))
653
if my_history[r] == other_history[r]:
654
return r+1, my_history[r]
710
657
def enum_history(self, direction):
711
658
"""Return (revno, revision_id) for history of branch.
735
682
That is equivalent to the number of revisions committed to
738
>>> b = ScratchBranch()
741
>>> b.commit('no foo')
745
685
return len(self.revision_history())
748
688
def last_patch(self):
749
689
"""Return last patch hash, or None if no history.
751
>>> ScratchBranch().last_patch() == None
754
691
ph = self.revision_history()
698
def missing_revisions(self, other, stop_revision=None):
700
If self and other have not diverged, return a list of the revisions
701
present in other, but missing from self.
703
>>> from bzrlib.commit import commit
704
>>> bzrlib.trace.silent = True
705
>>> br1 = ScratchBranch()
706
>>> br2 = ScratchBranch()
707
>>> br1.missing_revisions(br2)
709
>>> commit(br2, "lala!", rev_id="REVISION-ID-1")
710
>>> br1.missing_revisions(br2)
712
>>> br2.missing_revisions(br1)
714
>>> commit(br1, "lala!", rev_id="REVISION-ID-1")
715
>>> br1.missing_revisions(br2)
717
>>> commit(br2, "lala!", rev_id="REVISION-ID-2A")
718
>>> br1.missing_revisions(br2)
720
>>> commit(br1, "lala!", rev_id="REVISION-ID-2B")
721
>>> br1.missing_revisions(br2)
722
Traceback (most recent call last):
723
DivergedBranches: These branches have diverged.
725
self_history = self.revision_history()
726
self_len = len(self_history)
727
other_history = other.revision_history()
728
other_len = len(other_history)
729
common_index = min(self_len, other_len) -1
730
if common_index >= 0 and \
731
self_history[common_index] != other_history[common_index]:
732
raise DivergedBranches(self, other)
734
if stop_revision is None:
735
stop_revision = other_len
736
elif stop_revision > other_len:
737
raise NoSuchRevision(self, stop_revision)
739
return other_history[self_len:stop_revision]
742
def update_revisions(self, other, stop_revision=None):
743
"""Pull in all new revisions from other branch.
745
>>> from bzrlib.commit import commit
746
>>> bzrlib.trace.silent = True
747
>>> br1 = ScratchBranch(files=['foo', 'bar'])
750
>>> commit(br1, "lala!", rev_id="REVISION-ID-1", verbose=False)
751
>>> br2 = ScratchBranch()
752
>>> br2.update_revisions(br1)
756
>>> br2.revision_history()
758
>>> br2.update_revisions(br1)
762
>>> br1.text_store.total_size() == br2.text_store.total_size()
765
from bzrlib.progress import ProgressBar
769
pb.update('comparing histories')
770
revision_ids = self.missing_revisions(other, stop_revision)
772
if hasattr(other.revision_store, "prefetch"):
773
other.revision_store.prefetch(revision_ids)
774
if hasattr(other.inventory_store, "prefetch"):
775
inventory_ids = [other.get_revision(r).inventory_id
776
for r in revision_ids]
777
other.inventory_store.prefetch(inventory_ids)
780
needed_texts = sets.Set()
782
for rev_id in revision_ids:
784
pb.update('fetching revision', i, len(revision_ids))
785
rev = other.get_revision(rev_id)
786
revisions.append(rev)
787
inv = other.get_inventory(str(rev.inventory_id))
788
for key, entry in inv.iter_entries():
789
if entry.text_id is None:
791
if entry.text_id not in self.text_store:
792
needed_texts.add(entry.text_id)
796
count = self.text_store.copy_multi(other.text_store, needed_texts)
797
print "Added %d texts." % count
798
inventory_ids = [ f.inventory_id for f in revisions ]
799
count = self.inventory_store.copy_multi(other.inventory_store,
801
print "Added %d inventories." % count
802
revision_ids = [ f.revision_id for f in revisions]
803
count = self.revision_store.copy_multi(other.revision_store,
805
for revision_id in revision_ids:
806
self.append_revision(revision_id)
807
print "Added %d revisions." % count
810
def commit(self, *args, **kw):
811
from bzrlib.commit import commit
812
commit(self, *args, **kw)
761
815
def lookup_revision(self, revno):
817
863
This can change the directory or the filename or both.
819
self._need_writelock()
820
tree = self.working_tree()
822
if not tree.has_filename(from_rel):
823
bailout("can't rename: old working file %r does not exist" % from_rel)
824
if tree.has_filename(to_rel):
825
bailout("can't rename: new working file %r already exists" % to_rel)
827
file_id = inv.path2id(from_rel)
829
bailout("can't rename: old name %r is not versioned" % from_rel)
831
if inv.path2id(to_rel):
832
bailout("can't rename: new name %r is already versioned" % to_rel)
834
to_dir, to_tail = os.path.split(to_rel)
835
to_dir_id = inv.path2id(to_dir)
836
if to_dir_id == None and to_dir != '':
837
bailout("can't determine destination directory id for %r" % to_dir)
839
mutter("rename_one:")
840
mutter(" file_id {%s}" % file_id)
841
mutter(" from_rel %r" % from_rel)
842
mutter(" to_rel %r" % to_rel)
843
mutter(" to_dir %r" % to_dir)
844
mutter(" to_dir_id {%s}" % to_dir_id)
846
inv.rename(file_id, to_dir_id, to_tail)
848
print "%s => %s" % (from_rel, to_rel)
850
from_abs = self.abspath(from_rel)
851
to_abs = self.abspath(to_rel)
853
os.rename(from_abs, to_abs)
855
bailout("failed to rename %r to %r: %s"
856
% (from_abs, to_abs, e[1]),
857
["rename rolled back"])
859
self._write_inventory(inv)
867
tree = self.working_tree()
869
if not tree.has_filename(from_rel):
870
raise BzrError("can't rename: old working file %r does not exist" % from_rel)
871
if tree.has_filename(to_rel):
872
raise BzrError("can't rename: new working file %r already exists" % to_rel)
874
file_id = inv.path2id(from_rel)
876
raise BzrError("can't rename: old name %r is not versioned" % from_rel)
878
if inv.path2id(to_rel):
879
raise BzrError("can't rename: new name %r is already versioned" % to_rel)
881
to_dir, to_tail = os.path.split(to_rel)
882
to_dir_id = inv.path2id(to_dir)
883
if to_dir_id == None and to_dir != '':
884
raise BzrError("can't determine destination directory id for %r" % to_dir)
886
mutter("rename_one:")
887
mutter(" file_id {%s}" % file_id)
888
mutter(" from_rel %r" % from_rel)
889
mutter(" to_rel %r" % to_rel)
890
mutter(" to_dir %r" % to_dir)
891
mutter(" to_dir_id {%s}" % to_dir_id)
893
inv.rename(file_id, to_dir_id, to_tail)
895
print "%s => %s" % (from_rel, to_rel)
897
from_abs = self.abspath(from_rel)
898
to_abs = self.abspath(to_rel)
900
os.rename(from_abs, to_abs)
902
raise BzrError("failed to rename %r to %r: %s"
903
% (from_abs, to_abs, e[1]),
904
["rename rolled back"])
906
self._write_inventory(inv)
863
911
def move(self, from_paths, to_name):
871
919
Note that to_name is only the last component of the new name;
872
920
this doesn't change the directory.
874
self._need_writelock()
875
## TODO: Option to move IDs only
876
assert not isinstance(from_paths, basestring)
877
tree = self.working_tree()
879
to_abs = self.abspath(to_name)
880
if not isdir(to_abs):
881
bailout("destination %r is not a directory" % to_abs)
882
if not tree.has_filename(to_name):
883
bailout("destination %r not in working directory" % to_abs)
884
to_dir_id = inv.path2id(to_name)
885
if to_dir_id == None and to_name != '':
886
bailout("destination %r is not a versioned directory" % to_name)
887
to_dir_ie = inv[to_dir_id]
888
if to_dir_ie.kind not in ('directory', 'root_directory'):
889
bailout("destination %r is not a directory" % to_abs)
891
to_idpath = Set(inv.get_idpath(to_dir_id))
894
if not tree.has_filename(f):
895
bailout("%r does not exist in working tree" % f)
896
f_id = inv.path2id(f)
898
bailout("%r is not versioned" % f)
899
name_tail = splitpath(f)[-1]
900
dest_path = appendpath(to_name, name_tail)
901
if tree.has_filename(dest_path):
902
bailout("destination %r already exists" % dest_path)
903
if f_id in to_idpath:
904
bailout("can't move %r to a subdirectory of itself" % f)
906
# OK, so there's a race here, it's possible that someone will
907
# create a file in this interval and then the rename might be
908
# left half-done. But we should have caught most problems.
911
name_tail = splitpath(f)[-1]
912
dest_path = appendpath(to_name, name_tail)
913
print "%s => %s" % (f, dest_path)
914
inv.rename(inv.path2id(f), to_dir_id, name_tail)
924
## TODO: Option to move IDs only
925
assert not isinstance(from_paths, basestring)
926
tree = self.working_tree()
928
to_abs = self.abspath(to_name)
929
if not isdir(to_abs):
930
raise BzrError("destination %r is not a directory" % to_abs)
931
if not tree.has_filename(to_name):
932
raise BzrError("destination %r not in working directory" % to_abs)
933
to_dir_id = inv.path2id(to_name)
934
if to_dir_id == None and to_name != '':
935
raise BzrError("destination %r is not a versioned directory" % to_name)
936
to_dir_ie = inv[to_dir_id]
937
if to_dir_ie.kind not in ('directory', 'root_directory'):
938
raise BzrError("destination %r is not a directory" % to_abs)
940
to_idpath = inv.get_idpath(to_dir_id)
943
if not tree.has_filename(f):
944
raise BzrError("%r does not exist in working tree" % f)
945
f_id = inv.path2id(f)
947
raise BzrError("%r is not versioned" % f)
948
name_tail = splitpath(f)[-1]
949
dest_path = appendpath(to_name, name_tail)
950
if tree.has_filename(dest_path):
951
raise BzrError("destination %r already exists" % dest_path)
952
if f_id in to_idpath:
953
raise BzrError("can't move %r to a subdirectory of itself" % f)
955
# OK, so there's a race here, it's possible that someone will
956
# create a file in this interval and then the rename might be
957
# left half-done. But we should have caught most problems.
960
name_tail = splitpath(f)[-1]
961
dest_path = appendpath(to_name, name_tail)
962
print "%s => %s" % (f, dest_path)
963
inv.rename(inv.path2id(f), to_dir_id, name_tail)
965
os.rename(self.abspath(f), self.abspath(dest_path))
967
raise BzrError("failed to rename %r to %r: %s" % (f, dest_path, e[1]),
968
["rename rolled back"])
970
self._write_inventory(inv)
975
def revert(self, filenames, old_tree=None, backups=True):
976
"""Restore selected files to the versions from a previous tree.
979
If true (default) backups are made of files before
982
from bzrlib.errors import NotVersionedError, BzrError
983
from bzrlib.atomicfile import AtomicFile
984
from bzrlib.osutils import backup_file
986
inv = self.read_working_inventory()
988
old_tree = self.basis_tree()
989
old_inv = old_tree.inventory
993
file_id = inv.path2id(fn)
995
raise NotVersionedError("not a versioned file", fn)
996
if not old_inv.has_id(file_id):
997
raise BzrError("file not present in old tree", fn, file_id)
998
nids.append((fn, file_id))
1000
# TODO: Rename back if it was previously at a different location
1002
# TODO: If given a directory, restore the entire contents from
1003
# the previous version.
1005
# TODO: Make a backup to a temporary file.
1007
# TODO: If the file previously didn't exist, delete it?
1008
for fn, file_id in nids:
1011
f = AtomicFile(fn, 'wb')
916
os.rename(self.abspath(f), self.abspath(dest_path))
918
bailout("failed to rename %r to %r: %s" % (f, dest_path, e[1]),
919
["rename rolled back"])
921
self._write_inventory(inv)
925
def show_status(self, show_all=False, file_list=None):
926
"""Display single-line status for non-ignored working files.
928
The list is show sorted in order by file name.
930
>>> b = ScratchBranch(files=['foo', 'foo~'])
936
>>> b.commit("add foo")
938
>>> os.unlink(b.abspath('foo'))
942
self._need_readlock()
944
# We have to build everything into a list first so that it can
945
# sorted by name, incorporating all the different sources.
947
# FIXME: Rather than getting things in random order and then sorting,
948
# just step through in order.
950
# Interesting case: the old ID for a file has been removed,
951
# but a new file has been created under that name.
953
old = self.basis_tree()
954
new = self.working_tree()
956
items = diff_trees(old, new)
957
# We want to filter out only if any file was provided in the file_list.
958
if isinstance(file_list, list) and len(file_list):
959
items = [item for item in items if item[3] in file_list]
961
for fs, fid, oldname, newname, kind in items:
963
show_status(fs, kind,
964
oldname + ' => ' + newname)
965
elif fs == 'A' or fs == 'M':
966
show_status(fs, kind, newname)
968
show_status(fs, kind, oldname)
971
show_status(fs, kind, newname)
974
show_status(fs, kind, newname)
976
show_status(fs, kind, newname)
978
bailout("weird file state %r" % ((fs, fid),))
1013
f.write(old_tree.get_file(file_id).read())
982
1020
class ScratchBranch(Branch):