15
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20
import sys, os, os.path, random, time, sha, sets, types, re, shutil, tempfile
21
import traceback, socket, fnmatch, difflib, time
22
from binascii import hexlify
25
from inventory import Inventory
26
from trace import mutter, note
27
from tree import Tree, EmptyTree, RevisionTree, WorkingTree
28
from inventory import InventoryEntry, Inventory
29
from osutils import isdir, quotefn, isfile, uuid, sha_file, username, \
30
format_date, compact_date, pumpfile, user_email, rand_bytes, splitpath, \
31
joinpath, sha_string, file_kind, local_time_offset, appendpath
32
from store import ImmutableStore
33
from revision import Revision
34
from errors import bailout, BzrError
35
from textui import show_status
36
from diff import diff_trees
21
from bzrlib.trace import mutter, note
22
from bzrlib.osutils import isdir, quotefn, compact_date, rand_bytes, splitpath, \
23
sha_file, appendpath, file_kind
24
from bzrlib.errors import BzrError
38
26
BZR_BRANCH_FORMAT = "Bazaar-NG branch, format 0.0.4\n"
39
27
## TODO: Maybe include checks for common corruption of newlines, etc?
31
def find_branch(f, **args):
32
if f and (f.startswith('http://') or f.startswith('https://')):
34
return remotebranch.RemoteBranch(f, **args)
36
return Branch(f, **args)
39
def find_cached_branch(f, cache_root, **args):
40
from remotebranch import RemoteBranch
41
br = find_branch(f, **args)
42
def cacheify(br, store_name):
43
from meta_store import CachedStore
44
cache_path = os.path.join(cache_root, store_name)
46
new_store = CachedStore(getattr(br, store_name), cache_path)
47
setattr(br, store_name, new_store)
49
if isinstance(br, RemoteBranch):
50
cacheify(br, 'inventory_store')
51
cacheify(br, 'text_store')
52
cacheify(br, 'revision_store')
56
def _relpath(base, path):
57
"""Return path relative to base, or raise exception.
59
The path may be either an absolute path or a path relative to the
60
current working directory.
62
Lifted out of Branch.relpath for ease of testing.
64
os.path.commonprefix (python2.4) has a bad bug that it works just
65
on string prefixes, assuming that '/u' is a prefix of '/u2'. This
66
avoids that problem."""
67
rp = os.path.abspath(path)
71
while len(head) >= len(base):
74
head, tail = os.path.split(head)
78
from errors import NotBranchError
79
raise NotBranchError("path %r is not within branch %r" % (rp, base))
43
84
def find_branch_root(f=None):
44
85
"""Find the branch root enclosing f, or pwd.
87
f may be a filename or a URL.
46
89
It is not necessary that f exists.
48
91
Basically we keep looking up until we find the control directory or
65
111
raise BzrError('%r is not in a branch' % orig_f)
114
class DivergedBranches(Exception):
115
def __init__(self, branch1, branch2):
116
self.branch1 = branch1
117
self.branch2 = branch2
118
Exception.__init__(self, "These branches have diverged.")
121
class NoSuchRevision(BzrError):
122
def __init__(self, branch, revision):
124
self.revision = revision
125
msg = "Branch %s has no revision %d" % (branch, revision)
126
BzrError.__init__(self, msg)
70
129
######################################################################
132
class Branch(object):
74
133
"""Branch holding a history of revisions.
76
TODO: Perhaps use different stores for different classes of object,
77
so that we can keep track of how much space each one uses,
78
or garbage-collect them.
80
TODO: Add a RemoteBranch subclass. For the basic case of read-only
81
HTTP access this should be very easy by,
82
just redirecting controlfile access into HTTP requests.
83
We would need a RemoteStore working similarly.
85
TODO: Keep the on-disk branch locked while the object exists.
136
Base directory of the branch.
142
If _lock_mode is true, a positive count of the number of times the
146
Lock object from bzrlib.lock.
89
153
def __init__(self, base, init=False, find_root=True):
90
154
"""Create new branch object at a particular location.
204
324
fmt = self.controlfile('branch-format', 'r').read()
205
325
fmt.replace('\r\n', '')
206
326
if fmt != BZR_BRANCH_FORMAT:
207
bailout('sorry, branch format %r not supported' % fmt,
208
['use a different bzr version',
209
'or remove the .bzr directory and "bzr init" again'])
327
raise BzrError('sorry, branch format %r not supported' % fmt,
328
['use a different bzr version',
329
'or remove the .bzr directory and "bzr init" again'])
212
333
def read_working_inventory(self):
213
334
"""Read the working inventory."""
215
# ElementTree does its own conversion from UTF-8, so open in
217
inv = Inventory.read_xml(self.controlfile('inventory', 'rb'))
218
mutter("loaded inventory of %d items in %f"
219
% (len(inv), time.time() - before))
335
from bzrlib.inventory import Inventory
336
from bzrlib.xml import unpack_xml
337
from time import time
341
# ElementTree does its own conversion from UTF-8, so open in
343
inv = unpack_xml(Inventory,
344
self.controlfile('inventory', 'rb'))
345
mutter("loaded inventory of %d items in %f"
346
% (len(inv), time() - before))
223
352
def _write_inventory(self, inv):
224
353
"""Update the working inventory.
251
384
This puts the files in the Added state, so that they will be
252
385
recorded by the next commit.
388
List of paths to add, relative to the base of the tree.
391
If set, use these instead of automatically generated ids.
392
Must be the same length as the list of files, but may
393
contain None for ids that are to be autogenerated.
254
395
TODO: Perhaps have an option to add the ids even if the files do
257
398
TODO: Perhaps return the ids of the files? But then again it
258
is easy to retrieve them if they're needed.
260
TODO: Option to specify file id.
399
is easy to retrieve them if they're needed.
262
401
TODO: Adding a directory should optionally recurse down and
263
add all non-ignored children. Perhaps do that in a
266
>>> b = ScratchBranch(files=['foo'])
267
>>> 'foo' in b.unknowns()
272
>>> 'foo' in b.unknowns()
274
>>> bool(b.inventory.path2id('foo'))
280
Traceback (most recent call last):
282
BzrError: ('foo is already versioned', [])
284
>>> b.add(['nothere'])
285
Traceback (most recent call last):
286
BzrError: ('cannot add: not a regular file or directory: nothere', [])
402
add all non-ignored children. Perhaps do that in a
405
from bzrlib.textui import show_status
289
406
# TODO: Re-adding a file that is removed in the working copy
290
407
# should probably put it back with the previous ID.
291
if isinstance(files, types.StringTypes):
408
if isinstance(files, basestring):
409
assert(ids is None or isinstance(ids, basestring))
294
inv = self.read_working_inventory()
296
if is_control_file(f):
297
bailout("cannot add control file %s" % quotefn(f))
302
bailout("cannot add top-level %r" % f)
304
fullpath = os.path.normpath(self.abspath(f))
307
kind = file_kind(fullpath)
309
# maybe something better?
310
bailout('cannot add: not a regular file or directory: %s' % quotefn(f))
312
if kind != 'file' and kind != 'directory':
313
bailout('cannot add: not a regular file or directory: %s' % quotefn(f))
315
file_id = gen_file_id(f)
316
inv.add_path(f, kind=kind, file_id=file_id)
319
show_status('A', kind, quotefn(f))
321
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
323
self._write_inventory(inv)
415
ids = [None] * len(files)
417
assert(len(ids) == len(files))
421
inv = self.read_working_inventory()
422
for f,file_id in zip(files, ids):
423
if is_control_file(f):
424
raise BzrError("cannot add control file %s" % quotefn(f))
429
raise BzrError("cannot add top-level %r" % f)
431
fullpath = os.path.normpath(self.abspath(f))
434
kind = file_kind(fullpath)
436
# maybe something better?
437
raise BzrError('cannot add: not a regular file or directory: %s' % quotefn(f))
439
if kind != 'file' and kind != 'directory':
440
raise BzrError('cannot add: not a regular file or directory: %s' % quotefn(f))
443
file_id = gen_file_id(f)
444
inv.add_path(f, kind=kind, file_id=file_id)
447
print 'added', quotefn(f)
449
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
451
self._write_inventory(inv)
326
456
def print_file(self, file, revno):
327
457
"""Print `file` to stdout."""
328
tree = self.revision_tree(self.lookup_revision(revno))
329
# use inventory as it was in that revision
330
file_id = tree.inventory.path2id(file)
332
bailout("%r is not present in revision %d" % (file, revno))
333
tree.print_file(file_id)
460
tree = self.revision_tree(self.lookup_revision(revno))
461
# use inventory as it was in that revision
462
file_id = tree.inventory.path2id(file)
464
raise BzrError("%r is not present in revision %d" % (file, revno))
465
tree.print_file(file_id)
336
470
def remove(self, files, verbose=False):
337
471
"""Mark nominated files for removal from the inventory.
369
481
is the opposite of add. Removing it is consistent with most
370
482
other tools. Maybe an option.
484
from bzrlib.textui import show_status
372
485
## TODO: Normalize names
373
486
## TODO: Remove nested loops; better scalability
375
if isinstance(files, types.StringTypes):
487
if isinstance(files, basestring):
378
tree = self.working_tree()
381
# do this before any modifications
385
bailout("cannot remove unversioned file %s" % quotefn(f))
386
mutter("remove inventory entry %s {%s}" % (quotefn(f), fid))
388
# having remove it, it must be either ignored or unknown
389
if tree.is_ignored(f):
393
show_status(new_status, inv[fid].kind, quotefn(f))
493
tree = self.working_tree()
496
# do this before any modifications
500
raise BzrError("cannot remove unversioned file %s" % quotefn(f))
501
mutter("remove inventory entry %s {%s}" % (quotefn(f), fid))
503
# having remove it, it must be either ignored or unknown
504
if tree.is_ignored(f):
508
show_status(new_status, inv[fid].kind, quotefn(f))
511
self._write_inventory(inv)
516
# FIXME: this doesn't need to be a branch method
517
def set_inventory(self, new_inventory_list):
518
from bzrlib.inventory import Inventory, InventoryEntry
520
for path, file_id, parent, kind in new_inventory_list:
521
name = os.path.basename(path)
524
inv.add(InventoryEntry(file_id, name, kind, parent))
396
525
self._write_inventory(inv)
415
544
return self.working_tree().unknowns()
418
def commit(self, message, timestamp=None, timezone=None,
421
"""Commit working copy as a new revision.
423
The basic approach is to add all the file texts into the
424
store, then the inventory, then make a new revision pointing
425
to that inventory and store that.
427
This is not quite safe if the working copy changes during the
428
commit; for the moment that is simply not allowed. A better
429
approach is to make a temporary copy of the files before
430
computing their hashes, and then add those hashes in turn to
431
the inventory. This should mean at least that there are no
432
broken hash pointers. There is no way we can get a snapshot
433
of the whole directory at an instant. This would also have to
434
be robust against files disappearing, moving, etc. So the
435
whole thing is a bit hard.
437
timestamp -- if not None, seconds-since-epoch for a
438
postdated/predated commit.
441
## TODO: Show branch names
443
# TODO: Don't commit if there are no changes, unless forced?
445
# First walk over the working inventory; and both update that
446
# and also build a new revision inventory. The revision
447
# inventory needs to hold the text-id, sha1 and size of the
448
# actual file versions committed in the revision. (These are
449
# not present in the working inventory.) We also need to
450
# detect missing/deleted files, and remove them from the
453
work_inv = self.read_working_inventory()
455
basis = self.basis_tree()
456
basis_inv = basis.inventory
458
for path, entry in work_inv.iter_entries():
459
## TODO: Cope with files that have gone missing.
461
## TODO: Check that the file kind has not changed from the previous
462
## revision of this file (if any).
466
p = self.abspath(path)
467
file_id = entry.file_id
468
mutter('commit prep file %s, id %r ' % (p, file_id))
470
if not os.path.exists(p):
471
mutter(" file is missing, removing from inventory")
473
show_status('D', entry.kind, quotefn(path))
474
missing_ids.append(file_id)
477
# TODO: Handle files that have been deleted
479
# TODO: Maybe a special case for empty files? Seems a
480
# waste to store them many times.
484
if basis_inv.has_id(file_id):
485
old_kind = basis_inv[file_id].kind
486
if old_kind != entry.kind:
487
bailout("entry %r changed kind from %r to %r"
488
% (file_id, old_kind, entry.kind))
490
if entry.kind == 'directory':
492
bailout("%s is entered as directory but not a directory" % quotefn(p))
493
elif entry.kind == 'file':
495
bailout("%s is entered as file but is not a file" % quotefn(p))
497
content = file(p, 'rb').read()
499
entry.text_sha1 = sha_string(content)
500
entry.text_size = len(content)
502
old_ie = basis_inv.has_id(file_id) and basis_inv[file_id]
504
and (old_ie.text_size == entry.text_size)
505
and (old_ie.text_sha1 == entry.text_sha1)):
506
## assert content == basis.get_file(file_id).read()
507
entry.text_id = basis_inv[file_id].text_id
508
mutter(' unchanged from previous text_id {%s}' %
512
entry.text_id = gen_file_id(entry.name)
513
self.text_store.add(content, entry.text_id)
514
mutter(' stored with text_id {%s}' % entry.text_id)
518
elif (old_ie.name == entry.name
519
and old_ie.parent_id == entry.parent_id):
524
show_status(state, entry.kind, quotefn(path))
526
for file_id in missing_ids:
527
# have to do this later so we don't mess up the iterator.
528
# since parents may be removed before their children we
531
# FIXME: There's probably a better way to do this; perhaps
532
# the workingtree should know how to filter itself.
533
if work_inv.has_id(file_id):
534
del work_inv[file_id]
537
inv_id = rev_id = _gen_revision_id(time.time())
539
inv_tmp = tempfile.TemporaryFile()
540
inv.write_xml(inv_tmp)
542
self.inventory_store.add(inv_tmp, inv_id)
543
mutter('new inventory_id is {%s}' % inv_id)
545
self._write_inventory(work_inv)
547
if timestamp == None:
548
timestamp = time.time()
550
if committer == None:
551
committer = username()
554
timezone = local_time_offset()
556
mutter("building commit log message")
557
rev = Revision(timestamp=timestamp,
560
precursor = self.last_patch(),
565
rev_tmp = tempfile.TemporaryFile()
566
rev.write_xml(rev_tmp)
568
self.revision_store.add(rev_tmp, rev_id)
569
mutter("new revision_id is {%s}" % rev_id)
571
## XXX: Everything up to here can simply be orphaned if we abort
572
## the commit; it will leave junk files behind but that doesn't
575
## TODO: Read back the just-generated changeset, and make sure it
576
## applies and recreates the right state.
578
## TODO: Also calculate and store the inventory SHA1
579
mutter("committing patch r%d" % (self.revno() + 1))
582
self.append_revision(rev_id)
585
note("commited r%d" % self.revno())
588
547
def append_revision(self, revision_id):
548
from bzrlib.atomicfile import AtomicFile
589
550
mutter("add {%s} to revision-history" % revision_id)
590
rev_history = self.revision_history()
592
tmprhname = self.controlfilename('revision-history.tmp')
593
rhname = self.controlfilename('revision-history')
595
f = file(tmprhname, 'wt')
596
rev_history.append(revision_id)
597
f.write('\n'.join(rev_history))
601
if sys.platform == 'win32':
603
os.rename(tmprhname, rhname)
551
rev_history = self.revision_history() + [revision_id]
553
f = AtomicFile(self.controlfilename('revision-history'))
555
for rev_id in rev_history:
607
562
def get_revision(self, revision_id):
608
563
"""Return the Revision object for a named revision"""
609
r = Revision.read_xml(self.revision_store[revision_id])
564
from bzrlib.revision import Revision
565
from bzrlib.xml import unpack_xml
569
if not revision_id or not isinstance(revision_id, basestring):
570
raise ValueError('invalid revision-id: %r' % revision_id)
571
r = unpack_xml(Revision, self.revision_store[revision_id])
610
575
assert r.revision_id == revision_id
579
def get_revision_sha1(self, revision_id):
580
"""Hash the stored value of a revision, and return it."""
581
# In the future, revision entries will be signed. At that
582
# point, it is probably best *not* to include the signature
583
# in the revision hash. Because that lets you re-sign
584
# the revision, (add signatures/remove signatures) and still
585
# have all hash pointers stay consistent.
586
# But for now, just hash the contents.
587
return sha_file(self.revision_store[revision_id])
614
590
def get_inventory(self, inventory_id):
635
622
>>> ScratchBranch().revision_history()
638
return [l.rstrip('\r\n') for l in self.controlfile('revision-history', 'r').readlines()]
627
return [l.rstrip('\r\n') for l in
628
self.controlfile('revision-history', 'r').readlines()]
633
def common_ancestor(self, other, self_revno=None, other_revno=None):
636
>>> sb = ScratchBranch(files=['foo', 'foo~'])
637
>>> sb.common_ancestor(sb) == (None, None)
639
>>> commit.commit(sb, "Committing first revision", verbose=False)
640
>>> sb.common_ancestor(sb)[0]
642
>>> clone = sb.clone()
643
>>> commit.commit(sb, "Committing second revision", verbose=False)
644
>>> sb.common_ancestor(sb)[0]
646
>>> sb.common_ancestor(clone)[0]
648
>>> commit.commit(clone, "Committing divergent second revision",
650
>>> sb.common_ancestor(clone)[0]
652
>>> sb.common_ancestor(clone) == clone.common_ancestor(sb)
654
>>> sb.common_ancestor(sb) != clone.common_ancestor(clone)
656
>>> clone2 = sb.clone()
657
>>> sb.common_ancestor(clone2)[0]
659
>>> sb.common_ancestor(clone2, self_revno=1)[0]
661
>>> sb.common_ancestor(clone2, other_revno=1)[0]
664
my_history = self.revision_history()
665
other_history = other.revision_history()
666
if self_revno is None:
667
self_revno = len(my_history)
668
if other_revno is None:
669
other_revno = len(other_history)
670
indices = range(min((self_revno, other_revno)))
673
if my_history[r] == other_history[r]:
674
return r+1, my_history[r]
677
def enum_history(self, direction):
678
"""Return (revno, revision_id) for history of branch.
681
'forward' is from earliest to latest
682
'reverse' is from latest to earliest
684
rh = self.revision_history()
685
if direction == 'forward':
690
elif direction == 'reverse':
696
raise ValueError('invalid history direction', direction)
644
702
That is equivalent to the number of revisions committed to
647
>>> b = ScratchBranch()
650
>>> b.commit('no foo')
654
705
return len(self.revision_history())
657
708
def last_patch(self):
658
709
"""Return last patch hash, or None if no history.
660
>>> ScratchBranch().last_patch() == None
663
711
ph = self.revision_history()
718
def missing_revisions(self, other, stop_revision=None):
720
If self and other have not diverged, return a list of the revisions
721
present in other, but missing from self.
723
>>> from bzrlib.commit import commit
724
>>> bzrlib.trace.silent = True
725
>>> br1 = ScratchBranch()
726
>>> br2 = ScratchBranch()
727
>>> br1.missing_revisions(br2)
729
>>> commit(br2, "lala!", rev_id="REVISION-ID-1")
730
>>> br1.missing_revisions(br2)
732
>>> br2.missing_revisions(br1)
734
>>> commit(br1, "lala!", rev_id="REVISION-ID-1")
735
>>> br1.missing_revisions(br2)
737
>>> commit(br2, "lala!", rev_id="REVISION-ID-2A")
738
>>> br1.missing_revisions(br2)
740
>>> commit(br1, "lala!", rev_id="REVISION-ID-2B")
741
>>> br1.missing_revisions(br2)
742
Traceback (most recent call last):
743
DivergedBranches: These branches have diverged.
745
self_history = self.revision_history()
746
self_len = len(self_history)
747
other_history = other.revision_history()
748
other_len = len(other_history)
749
common_index = min(self_len, other_len) -1
750
if common_index >= 0 and \
751
self_history[common_index] != other_history[common_index]:
752
raise DivergedBranches(self, other)
754
if stop_revision is None:
755
stop_revision = other_len
756
elif stop_revision > other_len:
757
raise NoSuchRevision(self, stop_revision)
759
return other_history[self_len:stop_revision]
762
def update_revisions(self, other, stop_revision=None):
763
"""Pull in all new revisions from other branch.
765
>>> from bzrlib.commit import commit
766
>>> bzrlib.trace.silent = True
767
>>> br1 = ScratchBranch(files=['foo', 'bar'])
770
>>> commit(br1, "lala!", rev_id="REVISION-ID-1", verbose=False)
771
>>> br2 = ScratchBranch()
772
>>> br2.update_revisions(br1)
776
>>> br2.revision_history()
778
>>> br2.update_revisions(br1)
782
>>> br1.text_store.total_size() == br2.text_store.total_size()
785
from bzrlib.progress import ProgressBar
789
from sets import Set as set
793
pb.update('comparing histories')
794
revision_ids = self.missing_revisions(other, stop_revision)
796
if hasattr(other.revision_store, "prefetch"):
797
other.revision_store.prefetch(revision_ids)
798
if hasattr(other.inventory_store, "prefetch"):
799
inventory_ids = [other.get_revision(r).inventory_id
800
for r in revision_ids]
801
other.inventory_store.prefetch(inventory_ids)
806
for rev_id in revision_ids:
808
pb.update('fetching revision', i, len(revision_ids))
809
rev = other.get_revision(rev_id)
810
revisions.append(rev)
811
inv = other.get_inventory(str(rev.inventory_id))
812
for key, entry in inv.iter_entries():
813
if entry.text_id is None:
815
if entry.text_id not in self.text_store:
816
needed_texts.add(entry.text_id)
820
count = self.text_store.copy_multi(other.text_store, needed_texts)
821
print "Added %d texts." % count
822
inventory_ids = [ f.inventory_id for f in revisions ]
823
count = self.inventory_store.copy_multi(other.inventory_store,
825
print "Added %d inventories." % count
826
revision_ids = [ f.revision_id for f in revisions]
827
count = self.revision_store.copy_multi(other.revision_store,
829
for revision_id in revision_ids:
830
self.append_revision(revision_id)
831
print "Added %d revisions." % count
834
def commit(self, *args, **kw):
835
from bzrlib.commit import commit
836
commit(self, *args, **kw)
670
839
def lookup_revision(self, revno):
723
def write_log(self, show_timezone='original', verbose=False):
724
"""Write out human-readable log of commits to this branch
726
utc -- If true, show dates in universal time, not local time."""
727
## TODO: Option to choose either original, utc or local timezone
730
for p in self.revision_history():
732
print 'revno:', revno
733
## TODO: Show hash if --id is given.
734
##print 'revision-hash:', p
735
rev = self.get_revision(p)
736
print 'committer:', rev.committer
737
print 'timestamp: %s' % (format_date(rev.timestamp, rev.timezone or 0,
740
## opportunistic consistency check, same as check_patch_chaining
741
if rev.precursor != precursor:
742
bailout("mismatched precursor!")
746
print ' (no message)'
748
for l in rev.message.split('\n'):
751
if verbose == True and precursor != None:
752
print 'changed files:'
753
tree = self.revision_tree(p)
754
prevtree = self.revision_tree(precursor)
756
for file_state, fid, old_name, new_name, kind in \
757
diff_trees(prevtree, tree, ):
758
if file_state == 'A' or file_state == 'M':
759
show_status(file_state, kind, new_name)
760
elif file_state == 'D':
761
show_status(file_state, kind, old_name)
762
elif file_state == 'R':
763
show_status(file_state, kind,
764
old_name + ' => ' + new_name)
770
886
def rename_one(self, from_rel, to_rel):
771
887
"""Rename one file.
773
889
This can change the directory or the filename or both.
775
tree = self.working_tree()
777
if not tree.has_filename(from_rel):
778
bailout("can't rename: old working file %r does not exist" % from_rel)
779
if tree.has_filename(to_rel):
780
bailout("can't rename: new working file %r already exists" % to_rel)
782
file_id = inv.path2id(from_rel)
784
bailout("can't rename: old name %r is not versioned" % from_rel)
786
if inv.path2id(to_rel):
787
bailout("can't rename: new name %r is already versioned" % to_rel)
789
to_dir, to_tail = os.path.split(to_rel)
790
to_dir_id = inv.path2id(to_dir)
791
if to_dir_id == None and to_dir != '':
792
bailout("can't determine destination directory id for %r" % to_dir)
794
mutter("rename_one:")
795
mutter(" file_id {%s}" % file_id)
796
mutter(" from_rel %r" % from_rel)
797
mutter(" to_rel %r" % to_rel)
798
mutter(" to_dir %r" % to_dir)
799
mutter(" to_dir_id {%s}" % to_dir_id)
801
inv.rename(file_id, to_dir_id, to_tail)
803
print "%s => %s" % (from_rel, to_rel)
805
from_abs = self.abspath(from_rel)
806
to_abs = self.abspath(to_rel)
808
os.rename(from_abs, to_abs)
810
bailout("failed to rename %r to %r: %s"
811
% (from_abs, to_abs, e[1]),
812
["rename rolled back"])
814
self._write_inventory(inv)
893
tree = self.working_tree()
895
if not tree.has_filename(from_rel):
896
raise BzrError("can't rename: old working file %r does not exist" % from_rel)
897
if tree.has_filename(to_rel):
898
raise BzrError("can't rename: new working file %r already exists" % to_rel)
900
file_id = inv.path2id(from_rel)
902
raise BzrError("can't rename: old name %r is not versioned" % from_rel)
904
if inv.path2id(to_rel):
905
raise BzrError("can't rename: new name %r is already versioned" % to_rel)
907
to_dir, to_tail = os.path.split(to_rel)
908
to_dir_id = inv.path2id(to_dir)
909
if to_dir_id == None and to_dir != '':
910
raise BzrError("can't determine destination directory id for %r" % to_dir)
912
mutter("rename_one:")
913
mutter(" file_id {%s}" % file_id)
914
mutter(" from_rel %r" % from_rel)
915
mutter(" to_rel %r" % to_rel)
916
mutter(" to_dir %r" % to_dir)
917
mutter(" to_dir_id {%s}" % to_dir_id)
919
inv.rename(file_id, to_dir_id, to_tail)
921
print "%s => %s" % (from_rel, to_rel)
923
from_abs = self.abspath(from_rel)
924
to_abs = self.abspath(to_rel)
926
os.rename(from_abs, to_abs)
928
raise BzrError("failed to rename %r to %r: %s"
929
% (from_abs, to_abs, e[1]),
930
["rename rolled back"])
932
self._write_inventory(inv)
818
937
def move(self, from_paths, to_name):
826
945
Note that to_name is only the last component of the new name;
827
946
this doesn't change the directory.
829
## TODO: Option to move IDs only
830
assert not isinstance(from_paths, basestring)
831
tree = self.working_tree()
833
to_abs = self.abspath(to_name)
834
if not isdir(to_abs):
835
bailout("destination %r is not a directory" % to_abs)
836
if not tree.has_filename(to_name):
837
bailout("destination %r not in working directory" % to_abs)
838
to_dir_id = inv.path2id(to_name)
839
if to_dir_id == None and to_name != '':
840
bailout("destination %r is not a versioned directory" % to_name)
841
to_dir_ie = inv[to_dir_id]
842
if to_dir_ie.kind not in ('directory', 'root_directory'):
843
bailout("destination %r is not a directory" % to_abs)
845
to_idpath = Set(inv.get_idpath(to_dir_id))
848
if not tree.has_filename(f):
849
bailout("%r does not exist in working tree" % f)
850
f_id = inv.path2id(f)
852
bailout("%r is not versioned" % f)
853
name_tail = splitpath(f)[-1]
854
dest_path = appendpath(to_name, name_tail)
855
if tree.has_filename(dest_path):
856
bailout("destination %r already exists" % dest_path)
857
if f_id in to_idpath:
858
bailout("can't move %r to a subdirectory of itself" % f)
860
# OK, so there's a race here, it's possible that someone will
861
# create a file in this interval and then the rename might be
862
# left half-done. But we should have caught most problems.
865
name_tail = splitpath(f)[-1]
866
dest_path = appendpath(to_name, name_tail)
867
print "%s => %s" % (f, dest_path)
868
inv.rename(inv.path2id(f), to_dir_id, name_tail)
870
os.rename(self.abspath(f), self.abspath(dest_path))
872
bailout("failed to rename %r to %r: %s" % (f, dest_path, e[1]),
873
["rename rolled back"])
875
self._write_inventory(inv)
879
def show_status(self, show_all=False):
880
"""Display single-line status for non-ignored working files.
882
The list is show sorted in order by file name.
884
>>> b = ScratchBranch(files=['foo', 'foo~'])
890
>>> b.commit("add foo")
892
>>> os.unlink(b.abspath('foo'))
950
## TODO: Option to move IDs only
951
assert not isinstance(from_paths, basestring)
952
tree = self.working_tree()
954
to_abs = self.abspath(to_name)
955
if not isdir(to_abs):
956
raise BzrError("destination %r is not a directory" % to_abs)
957
if not tree.has_filename(to_name):
958
raise BzrError("destination %r not in working directory" % to_abs)
959
to_dir_id = inv.path2id(to_name)
960
if to_dir_id == None and to_name != '':
961
raise BzrError("destination %r is not a versioned directory" % to_name)
962
to_dir_ie = inv[to_dir_id]
963
if to_dir_ie.kind not in ('directory', 'root_directory'):
964
raise BzrError("destination %r is not a directory" % to_abs)
966
to_idpath = inv.get_idpath(to_dir_id)
969
if not tree.has_filename(f):
970
raise BzrError("%r does not exist in working tree" % f)
971
f_id = inv.path2id(f)
973
raise BzrError("%r is not versioned" % f)
974
name_tail = splitpath(f)[-1]
975
dest_path = appendpath(to_name, name_tail)
976
if tree.has_filename(dest_path):
977
raise BzrError("destination %r already exists" % dest_path)
978
if f_id in to_idpath:
979
raise BzrError("can't move %r to a subdirectory of itself" % f)
981
# OK, so there's a race here, it's possible that someone will
982
# create a file in this interval and then the rename might be
983
# left half-done. But we should have caught most problems.
986
name_tail = splitpath(f)[-1]
987
dest_path = appendpath(to_name, name_tail)
988
print "%s => %s" % (f, dest_path)
989
inv.rename(inv.path2id(f), to_dir_id, name_tail)
991
os.rename(self.abspath(f), self.abspath(dest_path))
993
raise BzrError("failed to rename %r to %r: %s" % (f, dest_path, e[1]),
994
["rename rolled back"])
996
self._write_inventory(inv)
1001
def revert(self, filenames, old_tree=None, backups=True):
1002
"""Restore selected files to the versions from a previous tree.
1005
If true (default) backups are made of files before
1008
from bzrlib.errors import NotVersionedError, BzrError
1009
from bzrlib.atomicfile import AtomicFile
1010
from bzrlib.osutils import backup_file
896
TODO: Get state for single files.
1012
inv = self.read_working_inventory()
1013
if old_tree is None:
1014
old_tree = self.basis_tree()
1015
old_inv = old_tree.inventory
1018
for fn in filenames:
1019
file_id = inv.path2id(fn)
1021
raise NotVersionedError("not a versioned file", fn)
1022
if not old_inv.has_id(file_id):
1023
raise BzrError("file not present in old tree", fn, file_id)
1024
nids.append((fn, file_id))
1026
# TODO: Rename back if it was previously at a different location
1028
# TODO: If given a directory, restore the entire contents from
1029
# the previous version.
1031
# TODO: Make a backup to a temporary file.
1033
# TODO: If the file previously didn't exist, delete it?
1034
for fn, file_id in nids:
1037
f = AtomicFile(fn, 'wb')
1039
f.write(old_tree.get_file(file_id).read())
1045
def pending_merges(self):
1046
"""Return a list of pending merges.
1048
These are revisions that have been merged into the working
1049
directory but not yet committed.
899
# We have to build everything into a list first so that it can
900
# sorted by name, incorporating all the different sources.
902
# FIXME: Rather than getting things in random order and then sorting,
903
# just step through in order.
905
# Interesting case: the old ID for a file has been removed,
906
# but a new file has been created under that name.
908
old = self.basis_tree()
909
new = self.working_tree()
911
for fs, fid, oldname, newname, kind in diff_trees(old, new):
913
show_status(fs, kind,
914
oldname + ' => ' + newname)
915
elif fs == 'A' or fs == 'M':
916
show_status(fs, kind, newname)
918
show_status(fs, kind, oldname)
921
show_status(fs, kind, newname)
924
show_status(fs, kind, newname)
926
show_status(fs, kind, newname)
928
bailout("weird file state %r" % ((fs, fid),))
1051
cfn = self.controlfilename('pending-merges')
1052
if not os.path.exists(cfn):
1055
for l in self.controlfile('pending-merges', 'r').readlines():
1056
p.append(l.rstrip('\n'))
1060
def add_pending_merge(self, revision_id):
1061
from bzrlib.revision import validate_revision_id
1063
validate_revision_id(revision_id)
1065
p = self.pending_merges()
1066
if revision_id in p:
1068
p.append(revision_id)
1069
self.set_pending_merges(p)
1072
def set_pending_merges(self, rev_list):
1073
from bzrlib.atomicfile import AtomicFile
1076
f = AtomicFile(self.controlfilename('pending-merges'))
932
1088
class ScratchBranch(Branch):