15
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20
import sys, os, os.path, random, time, sha, sets, types, re, shutil, tempfile
21
import traceback, socket, fnmatch, difflib, time
22
from binascii import hexlify
22
from bzrlib.trace import mutter, note
23
from bzrlib.osutils import isdir, quotefn, compact_date, rand_bytes, \
25
sha_file, appendpath, file_kind
27
from bzrlib.errors import BzrError, InvalidRevisionNumber, InvalidRevisionId, \
28
DivergedBranches, NotBranchError
29
from bzrlib.textui import show_status
30
from bzrlib.revision import Revision
31
from bzrlib.delta import compare_trees
32
from bzrlib.tree import EmptyTree, RevisionTree
25
from inventory import Inventory
26
from trace import mutter, note
27
from tree import Tree, EmptyTree, RevisionTree, WorkingTree
28
from inventory import InventoryEntry, Inventory
29
from osutils import isdir, quotefn, isfile, uuid, sha_file, username, \
30
format_date, compact_date, pumpfile, user_email, rand_bytes, splitpath, \
31
joinpath, sha_string, file_kind, local_time_offset, appendpath
32
from store import ImmutableStore
33
from revision import Revision
34
from errors import bailout, BzrError
35
from textui import show_status
36
from diff import diff_trees
38
38
BZR_BRANCH_FORMAT = "Bazaar-NG branch, format 0.0.4\n"
39
39
## TODO: Maybe include checks for common corruption of newlines, etc?
42
# TODO: Some operations like log might retrieve the same revisions
43
# repeatedly to calculate deltas. We could perhaps have a weakref
44
# cache in memory to make this faster.
46
# TODO: please move the revision-string syntax stuff out of the branch
47
# object; it's clutter
50
def find_branch(f, **args):
51
if f and (f.startswith('http://') or f.startswith('https://')):
52
from bzrlib.remotebranch import RemoteBranch
53
return RemoteBranch(f, **args)
55
return Branch(f, **args)
58
def find_cached_branch(f, cache_root, **args):
59
from bzrlib.remotebranch import RemoteBranch
60
br = find_branch(f, **args)
61
def cacheify(br, store_name):
62
from bzrlib.meta_store import CachedStore
63
cache_path = os.path.join(cache_root, store_name)
65
new_store = CachedStore(getattr(br, store_name), cache_path)
66
setattr(br, store_name, new_store)
68
if isinstance(br, RemoteBranch):
69
cacheify(br, 'inventory_store')
70
cacheify(br, 'text_store')
71
cacheify(br, 'revision_store')
75
def _relpath(base, path):
76
"""Return path relative to base, or raise exception.
78
The path may be either an absolute path or a path relative to the
79
current working directory.
81
Lifted out of Branch.relpath for ease of testing.
83
os.path.commonprefix (python2.4) has a bad bug that it works just
84
on string prefixes, assuming that '/u' is a prefix of '/u2'. This
85
avoids that problem."""
86
rp = os.path.abspath(path)
90
while len(head) >= len(base):
93
head, tail = os.path.split(head)
97
raise NotBranchError("path %r is not within branch %r" % (rp, base))
102
43
def find_branch_root(f=None):
103
44
"""Find the branch root enclosing f, or pwd.
105
f may be a filename or a URL.
107
46
It is not necessary that f exists.
109
48
Basically we keep looking up until we find the control directory or
110
run into the root. If there isn't one, raises NotBranchError.
114
52
elif hasattr(os.path, 'realpath'):
115
53
f = os.path.realpath(f)
117
55
f = os.path.abspath(f)
118
if not os.path.exists(f):
119
raise BzrError('%r does not exist' % f)
205
119
__repr__ = __str__
209
if self._lock_mode or self._lock:
210
from bzrlib.warnings import warn
211
warn("branch %r was not explicitly unlocked" % self)
214
def lock_write(self):
216
if self._lock_mode != 'w':
217
from bzrlib.errors import LockError
218
raise LockError("can't upgrade to a write lock from %r" %
220
self._lock_count += 1
222
from bzrlib.lock import WriteLock
224
self._lock = WriteLock(self.controlfilename('branch-lock'))
225
self._lock_mode = 'w'
231
assert self._lock_mode in ('r', 'w'), \
232
"invalid lock mode %r" % self._lock_mode
233
self._lock_count += 1
235
from bzrlib.lock import ReadLock
237
self._lock = ReadLock(self.controlfilename('branch-lock'))
238
self._lock_mode = 'r'
242
if not self._lock_mode:
243
from bzrlib.errors import LockError
244
raise LockError('branch %r is not locked' % (self))
246
if self._lock_count > 1:
247
self._lock_count -= 1
251
self._lock_mode = self._lock_count = None
123
def lock(self, mode='w'):
124
"""Lock the on-disk branch, excluding other processes."""
130
om = os.O_WRONLY | os.O_CREAT
135
raise BzrError("invalid locking mode %r" % mode)
138
lockfile = os.open(self.controlfilename('branch-lock'), om)
140
if e.errno == errno.ENOENT:
141
# might not exist on branches from <0.0.4
142
self.controlfile('branch-lock', 'w').close()
143
lockfile = os.open(self.controlfilename('branch-lock'), om)
147
fcntl.lockf(lockfile, lm)
149
fcntl.lockf(lockfile, fcntl.LOCK_UN)
151
self._lockmode = None
153
self._lockmode = mode
155
warning("please write a locking method for platform %r" % sys.platform)
157
self._lockmode = None
159
self._lockmode = mode
162
def _need_readlock(self):
163
if self._lockmode not in ['r', 'w']:
164
raise BzrError('need read lock on branch, only have %r' % self._lockmode)
166
def _need_writelock(self):
167
if self._lockmode not in ['w']:
168
raise BzrError('need write lock on branch, only have %r' % self._lockmode)
253
171
def abspath(self, name):
254
172
"""Return absolute filename for something in the branch"""
255
173
return os.path.join(self.base, name)
257
176
def relpath(self, path):
258
177
"""Return path relative to this branch of something inside it.
260
179
Raises an error if path is not in this branch."""
261
return _relpath(self.base, path)
180
rp = os.path.realpath(path)
182
if not rp.startswith(self.base):
183
bailout("path %r is not within branch %r" % (rp, self.base))
184
rp = rp[len(self.base):]
185
rp = rp.lstrip(os.sep)
263
189
def controlfilename(self, file_or_path):
264
190
"""Return location relative to branch."""
265
if isinstance(file_or_path, basestring):
191
if isinstance(file_or_path, types.StringTypes):
266
192
file_or_path = [file_or_path]
267
193
return os.path.join(self.base, bzrlib.BZRDIR, *file_or_path)
370
270
That is to say, the inventory describing changes underway, that
371
271
will be committed to the next revision.
373
from bzrlib.atomicfile import AtomicFile
377
f = AtomicFile(self.controlfilename('inventory'), 'wb')
379
bzrlib.xml.serializer_v4.write_inventory(inv, f)
273
self._need_writelock()
274
## TODO: factor out to atomicfile? is rename safe on windows?
275
## TODO: Maybe some kind of clean/dirty marker on inventory?
276
tmpfname = self.controlfilename('inventory.tmp')
277
tmpf = file(tmpfname, 'wb')
280
inv_fname = self.controlfilename('inventory')
281
if sys.platform == 'win32':
283
os.rename(tmpfname, inv_fname)
386
284
mutter('wrote working inventory')
389
287
inventory = property(read_working_inventory, _write_inventory, None,
390
288
"""Inventory for the working copy.""")
393
def add(self, files, ids=None):
291
def add(self, files, verbose=False):
394
292
"""Make files versioned.
396
Note that the command line normally calls smart_add instead,
397
which can automatically recurse.
294
Note that the command line normally calls smart_add instead.
399
296
This puts the files in the Added state, so that they will be
400
297
recorded by the next commit.
403
List of paths to add, relative to the base of the tree.
406
If set, use these instead of automatically generated ids.
407
Must be the same length as the list of files, but may
408
contain None for ids that are to be autogenerated.
410
299
TODO: Perhaps have an option to add the ids even if the files do
413
TODO: Perhaps yield the ids and paths as they're added.
302
TODO: Perhaps return the ids of the files? But then again it
303
is easy to retrieve them if they're needed.
305
TODO: Option to specify file id.
307
TODO: Adding a directory should optionally recurse down and
308
add all non-ignored children. Perhaps do that in a
311
>>> b = ScratchBranch(files=['foo'])
312
>>> 'foo' in b.unknowns()
317
>>> 'foo' in b.unknowns()
319
>>> bool(b.inventory.path2id('foo'))
325
Traceback (most recent call last):
327
BzrError: ('foo is already versioned', [])
329
>>> b.add(['nothere'])
330
Traceback (most recent call last):
331
BzrError: ('cannot add: not a regular file or directory: nothere', [])
333
self._need_writelock()
415
335
# TODO: Re-adding a file that is removed in the working copy
416
336
# should probably put it back with the previous ID.
417
if isinstance(files, basestring):
418
assert(ids is None or isinstance(ids, basestring))
337
if isinstance(files, types.StringTypes):
424
ids = [None] * len(files)
426
assert(len(ids) == len(files))
430
inv = self.read_working_inventory()
431
for f,file_id in zip(files, ids):
432
if is_control_file(f):
433
raise BzrError("cannot add control file %s" % quotefn(f))
438
raise BzrError("cannot add top-level %r" % f)
440
fullpath = os.path.normpath(self.abspath(f))
443
kind = file_kind(fullpath)
445
# maybe something better?
446
raise BzrError('cannot add: not a regular file or directory: %s' % quotefn(f))
448
if kind != 'file' and kind != 'directory':
449
raise BzrError('cannot add: not a regular file or directory: %s' % quotefn(f))
452
file_id = gen_file_id(f)
453
inv.add_path(f, kind=kind, file_id=file_id)
455
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
457
self._write_inventory(inv)
340
inv = self.read_working_inventory()
342
if is_control_file(f):
343
bailout("cannot add control file %s" % quotefn(f))
348
bailout("cannot add top-level %r" % f)
350
fullpath = os.path.normpath(self.abspath(f))
353
kind = file_kind(fullpath)
355
# maybe something better?
356
bailout('cannot add: not a regular file or directory: %s' % quotefn(f))
358
if kind != 'file' and kind != 'directory':
359
bailout('cannot add: not a regular file or directory: %s' % quotefn(f))
361
file_id = gen_file_id(f)
362
inv.add_path(f, kind=kind, file_id=file_id)
365
show_status('A', kind, quotefn(f))
367
mutter("add file %s file_id:{%s} kind=%r" % (f, file_id, kind))
369
self._write_inventory(inv)
462
372
def print_file(self, file, revno):
463
373
"""Print `file` to stdout."""
466
tree = self.revision_tree(self.lookup_revision(revno))
467
# use inventory as it was in that revision
468
file_id = tree.inventory.path2id(file)
470
raise BzrError("%r is not present in revision %s" % (file, revno))
471
tree.print_file(file_id)
374
self._need_readlock()
375
tree = self.revision_tree(self.lookup_revision(revno))
376
# use inventory as it was in that revision
377
file_id = tree.inventory.path2id(file)
379
bailout("%r is not present in revision %d" % (file, revno))
380
tree.print_file(file_id)
476
383
def remove(self, files, verbose=False):
477
384
"""Mark nominated files for removal from the inventory.
549
463
return self.working_tree().unknowns()
552
def append_revision(self, *revision_ids):
553
from bzrlib.atomicfile import AtomicFile
555
for revision_id in revision_ids:
556
mutter("add {%s} to revision-history" % revision_id)
466
def commit(self, message, timestamp=None, timezone=None,
469
"""Commit working copy as a new revision.
471
The basic approach is to add all the file texts into the
472
store, then the inventory, then make a new revision pointing
473
to that inventory and store that.
475
This is not quite safe if the working copy changes during the
476
commit; for the moment that is simply not allowed. A better
477
approach is to make a temporary copy of the files before
478
computing their hashes, and then add those hashes in turn to
479
the inventory. This should mean at least that there are no
480
broken hash pointers. There is no way we can get a snapshot
481
of the whole directory at an instant. This would also have to
482
be robust against files disappearing, moving, etc. So the
483
whole thing is a bit hard.
485
timestamp -- if not None, seconds-since-epoch for a
486
postdated/predated commit.
488
self._need_writelock()
490
## TODO: Show branch names
492
# TODO: Don't commit if there are no changes, unless forced?
494
# First walk over the working inventory; and both update that
495
# and also build a new revision inventory. The revision
496
# inventory needs to hold the text-id, sha1 and size of the
497
# actual file versions committed in the revision. (These are
498
# not present in the working inventory.) We also need to
499
# detect missing/deleted files, and remove them from the
502
work_inv = self.read_working_inventory()
504
basis = self.basis_tree()
505
basis_inv = basis.inventory
507
for path, entry in work_inv.iter_entries():
508
## TODO: Cope with files that have gone missing.
510
## TODO: Check that the file kind has not changed from the previous
511
## revision of this file (if any).
515
p = self.abspath(path)
516
file_id = entry.file_id
517
mutter('commit prep file %s, id %r ' % (p, file_id))
519
if not os.path.exists(p):
520
mutter(" file is missing, removing from inventory")
522
show_status('D', entry.kind, quotefn(path))
523
missing_ids.append(file_id)
526
# TODO: Handle files that have been deleted
528
# TODO: Maybe a special case for empty files? Seems a
529
# waste to store them many times.
533
if basis_inv.has_id(file_id):
534
old_kind = basis_inv[file_id].kind
535
if old_kind != entry.kind:
536
bailout("entry %r changed kind from %r to %r"
537
% (file_id, old_kind, entry.kind))
539
if entry.kind == 'directory':
541
bailout("%s is entered as directory but not a directory" % quotefn(p))
542
elif entry.kind == 'file':
544
bailout("%s is entered as file but is not a file" % quotefn(p))
546
content = file(p, 'rb').read()
548
entry.text_sha1 = sha_string(content)
549
entry.text_size = len(content)
551
old_ie = basis_inv.has_id(file_id) and basis_inv[file_id]
553
and (old_ie.text_size == entry.text_size)
554
and (old_ie.text_sha1 == entry.text_sha1)):
555
## assert content == basis.get_file(file_id).read()
556
entry.text_id = basis_inv[file_id].text_id
557
mutter(' unchanged from previous text_id {%s}' %
561
entry.text_id = gen_file_id(entry.name)
562
self.text_store.add(content, entry.text_id)
563
mutter(' stored with text_id {%s}' % entry.text_id)
567
elif (old_ie.name == entry.name
568
and old_ie.parent_id == entry.parent_id):
573
show_status(state, entry.kind, quotefn(path))
575
for file_id in missing_ids:
576
# have to do this later so we don't mess up the iterator.
577
# since parents may be removed before their children we
580
# FIXME: There's probably a better way to do this; perhaps
581
# the workingtree should know how to filter itself.
582
if work_inv.has_id(file_id):
583
del work_inv[file_id]
586
inv_id = rev_id = _gen_revision_id(time.time())
588
inv_tmp = tempfile.TemporaryFile()
589
inv.write_xml(inv_tmp)
591
self.inventory_store.add(inv_tmp, inv_id)
592
mutter('new inventory_id is {%s}' % inv_id)
594
self._write_inventory(work_inv)
596
if timestamp == None:
597
timestamp = time.time()
599
if committer == None:
600
committer = username()
603
timezone = local_time_offset()
605
mutter("building commit log message")
606
rev = Revision(timestamp=timestamp,
609
precursor = self.last_patch(),
614
rev_tmp = tempfile.TemporaryFile()
615
rev.write_xml(rev_tmp)
617
self.revision_store.add(rev_tmp, rev_id)
618
mutter("new revision_id is {%s}" % rev_id)
620
## XXX: Everything up to here can simply be orphaned if we abort
621
## the commit; it will leave junk files behind but that doesn't
624
## TODO: Read back the just-generated changeset, and make sure it
625
## applies and recreates the right state.
627
## TODO: Also calculate and store the inventory SHA1
628
mutter("committing patch r%d" % (self.revno() + 1))
631
self.append_revision(rev_id)
634
note("commited r%d" % self.revno())
637
def append_revision(self, revision_id):
638
mutter("add {%s} to revision-history" % revision_id)
558
639
rev_history = self.revision_history()
559
rev_history.extend(revision_ids)
561
f = AtomicFile(self.controlfilename('revision-history'))
563
for rev_id in rev_history:
570
def get_revision_xml_file(self, revision_id):
571
"""Return XML file object for revision object."""
572
if not revision_id or not isinstance(revision_id, basestring):
573
raise InvalidRevisionId(revision_id)
578
return self.revision_store[revision_id]
579
except (IndexError, KeyError):
580
raise bzrlib.errors.NoSuchRevision(self, revision_id)
586
get_revision_xml = get_revision_xml_file
641
tmprhname = self.controlfilename('revision-history.tmp')
642
rhname = self.controlfilename('revision-history')
644
f = file(tmprhname, 'wt')
645
rev_history.append(revision_id)
646
f.write('\n'.join(rev_history))
650
if sys.platform == 'win32':
652
os.rename(tmprhname, rhname)
589
656
def get_revision(self, revision_id):
590
657
"""Return the Revision object for a named revision"""
591
xml_file = self.get_revision_xml_file(revision_id)
594
r = bzrlib.xml.serializer_v4.read_revision(xml_file)
595
except SyntaxError, e:
596
raise bzrlib.errors.BzrError('failed to unpack revision_xml',
658
self._need_readlock()
659
r = Revision.read_xml(self.revision_store[revision_id])
600
660
assert r.revision_id == revision_id
604
def get_revision_delta(self, revno):
605
"""Return the delta for one revision.
607
The delta is relative to its mainline predecessor, or the
608
empty tree for revision 1.
610
assert isinstance(revno, int)
611
rh = self.revision_history()
612
if not (1 <= revno <= len(rh)):
613
raise InvalidRevisionNumber(revno)
615
# revno is 1-based; list is 0-based
617
new_tree = self.revision_tree(rh[revno-1])
619
old_tree = EmptyTree()
621
old_tree = self.revision_tree(rh[revno-2])
623
return compare_trees(old_tree, new_tree)
627
def get_revision_sha1(self, revision_id):
628
"""Hash the stored value of a revision, and return it."""
629
# In the future, revision entries will be signed. At that
630
# point, it is probably best *not* to include the signature
631
# in the revision hash. Because that lets you re-sign
632
# the revision, (add signatures/remove signatures) and still
633
# have all hash pointers stay consistent.
634
# But for now, just hash the contents.
635
return bzrlib.osutils.sha_file(self.get_revision_xml(revision_id))
638
664
def get_inventory(self, inventory_id):
639
665
"""Get Inventory object by hash.
641
667
TODO: Perhaps for this and similar methods, take a revision
642
668
parameter which can be either an integer revno or a
644
from bzrlib.inventory import Inventory
646
f = self.get_inventory_xml_file(inventory_id)
647
return bzrlib.xml.serializer_v4.read_inventory(f)
650
def get_inventory_xml(self, inventory_id):
651
"""Get inventory XML as a file object."""
652
return self.inventory_store[inventory_id]
654
get_inventory_xml_file = get_inventory_xml
657
def get_inventory_sha1(self, inventory_id):
658
"""Return the sha1 hash of the inventory entry
660
return sha_file(self.get_inventory_xml(inventory_id))
670
self._need_readlock()
671
i = Inventory.read_xml(self.inventory_store[inventory_id])
663
675
def get_revision_inventory(self, revision_id):
664
676
"""Return inventory of a past revision."""
665
# bzr 0.0.6 imposes the constraint that the inventory_id
666
# must be the same as its revision, so this is trivial.
677
self._need_readlock()
667
678
if revision_id == None:
668
from bzrlib.inventory import Inventory
669
return Inventory(self.get_root_id())
671
return self.get_inventory(revision_id)
681
return self.get_inventory(self.get_revision(revision_id).inventory_id)
674
684
def revision_history(self):
736
697
That is equivalent to the number of revisions committed to
700
>>> b = ScratchBranch()
703
>>> b.commit('no foo')
739
707
return len(self.revision_history())
742
710
def last_patch(self):
743
711
"""Return last patch hash, or None if no history.
713
>>> ScratchBranch().last_patch() == None
745
716
ph = self.revision_history()
752
def missing_revisions(self, other, stop_revision=None, diverged_ok=False):
754
If self and other have not diverged, return a list of the revisions
755
present in other, but missing from self.
757
>>> from bzrlib.commit import commit
758
>>> bzrlib.trace.silent = True
759
>>> br1 = ScratchBranch()
760
>>> br2 = ScratchBranch()
761
>>> br1.missing_revisions(br2)
763
>>> commit(br2, "lala!", rev_id="REVISION-ID-1")
764
>>> br1.missing_revisions(br2)
766
>>> br2.missing_revisions(br1)
768
>>> commit(br1, "lala!", rev_id="REVISION-ID-1")
769
>>> br1.missing_revisions(br2)
771
>>> commit(br2, "lala!", rev_id="REVISION-ID-2A")
772
>>> br1.missing_revisions(br2)
774
>>> commit(br1, "lala!", rev_id="REVISION-ID-2B")
775
>>> br1.missing_revisions(br2)
776
Traceback (most recent call last):
777
DivergedBranches: These branches have diverged.
779
self_history = self.revision_history()
780
self_len = len(self_history)
781
other_history = other.revision_history()
782
other_len = len(other_history)
783
common_index = min(self_len, other_len) -1
784
if common_index >= 0 and \
785
self_history[common_index] != other_history[common_index]:
786
raise DivergedBranches(self, other)
788
if stop_revision is None:
789
stop_revision = other_len
790
elif stop_revision > other_len:
791
raise bzrlib.errors.NoSuchRevision(self, stop_revision)
793
return other_history[self_len:stop_revision]
796
def update_revisions(self, other, stop_revision=None):
797
"""Pull in all new revisions from other branch.
799
from bzrlib.fetch import greedy_fetch
800
from bzrlib.revision import get_intervening_revisions
802
pb = bzrlib.ui.ui_factory.progress_bar()
803
pb.update('comparing histories')
804
if stop_revision is None:
805
other_revision = other.last_patch()
807
other_revision = other.lookup_revision(stop_revision)
808
count = greedy_fetch(self, other, other_revision, pb)[0]
810
revision_ids = self.missing_revisions(other, stop_revision)
811
except DivergedBranches, e:
813
revision_ids = get_intervening_revisions(self.last_patch(),
814
other_revision, self)
815
assert self.last_patch() not in revision_ids
816
except bzrlib.errors.NotAncestor:
819
self.append_revision(*revision_ids)
822
def install_revisions(self, other, revision_ids, pb):
823
if hasattr(other.revision_store, "prefetch"):
824
other.revision_store.prefetch(revision_ids)
825
if hasattr(other.inventory_store, "prefetch"):
827
for rev_id in revision_ids:
829
revision = other.get_revision(rev_id).inventory_id
830
inventory_ids.append(revision)
831
except bzrlib.errors.NoSuchRevision:
833
other.inventory_store.prefetch(inventory_ids)
836
pb = bzrlib.ui.ui_factory.progress_bar()
843
for i, rev_id in enumerate(revision_ids):
844
pb.update('fetching revision', i+1, len(revision_ids))
846
rev = other.get_revision(rev_id)
847
except bzrlib.errors.NoSuchRevision:
851
revisions.append(rev)
852
inv = other.get_inventory(str(rev.inventory_id))
853
for key, entry in inv.iter_entries():
854
if entry.text_id is None:
856
if entry.text_id not in self.text_store:
857
needed_texts.add(entry.text_id)
861
count, cp_fail = self.text_store.copy_multi(other.text_store,
863
#print "Added %d texts." % count
864
inventory_ids = [ f.inventory_id for f in revisions ]
865
count, cp_fail = self.inventory_store.copy_multi(other.inventory_store,
867
#print "Added %d inventories." % count
868
revision_ids = [ f.revision_id for f in revisions]
870
count, cp_fail = self.revision_store.copy_multi(other.revision_store,
873
assert len(cp_fail) == 0
874
return count, failures
877
def commit(self, *args, **kw):
878
from bzrlib.commit import commit
879
commit(self, *args, **kw)
882
def lookup_revision(self, revision):
883
"""Return the revision identifier for a given revision information."""
884
revno, info = self._get_revision_info(revision)
888
def revision_id_to_revno(self, revision_id):
889
"""Given a revision id, return its revno"""
890
history = self.revision_history()
892
return history.index(revision_id) + 1
894
raise bzrlib.errors.NoSuchRevision(self, revision_id)
897
def get_revision_info(self, revision):
898
"""Return (revno, revision id) for revision identifier.
900
revision can be an integer, in which case it is assumed to be revno (though
901
this will translate negative values into positive ones)
902
revision can also be a string, in which case it is parsed for something like
903
'date:' or 'revid:' etc.
905
revno, rev_id = self._get_revision_info(revision)
907
raise bzrlib.errors.NoSuchRevision(self, revision)
910
def get_rev_id(self, revno, history=None):
911
"""Find the revision id of the specified revno."""
723
def lookup_revision(self, revno):
724
"""Return revision hash for revision number."""
915
history = self.revision_history()
916
elif revno <= 0 or revno > len(history):
917
raise bzrlib.errors.NoSuchRevision(self, revno)
918
return history[revno - 1]
920
def _get_revision_info(self, revision):
921
"""Return (revno, revision id) for revision specifier.
923
revision can be an integer, in which case it is assumed to be revno
924
(though this will translate negative values into positive ones)
925
revision can also be a string, in which case it is parsed for something
926
like 'date:' or 'revid:' etc.
928
A revid is always returned. If it is None, the specifier referred to
929
the null revision. If the revid does not occur in the revision
930
history, revno will be None.
936
try:# Convert to int if possible
937
revision = int(revision)
940
revs = self.revision_history()
941
if isinstance(revision, int):
943
revno = len(revs) + revision + 1
946
rev_id = self.get_rev_id(revno, revs)
947
elif isinstance(revision, basestring):
948
for prefix, func in Branch.REVISION_NAMESPACES.iteritems():
949
if revision.startswith(prefix):
950
result = func(self, revs, revision)
952
revno, rev_id = result
955
rev_id = self.get_rev_id(revno, revs)
958
raise BzrError('No namespace registered for string: %r' %
961
raise TypeError('Unhandled revision type %s' % revision)
965
raise bzrlib.errors.NoSuchRevision(self, revision)
968
def _namespace_revno(self, revs, revision):
969
"""Lookup a revision by revision number"""
970
assert revision.startswith('revno:')
972
return (int(revision[6:]),)
975
REVISION_NAMESPACES['revno:'] = _namespace_revno
977
def _namespace_revid(self, revs, revision):
978
assert revision.startswith('revid:')
979
rev_id = revision[len('revid:'):]
981
return revs.index(rev_id) + 1, rev_id
984
REVISION_NAMESPACES['revid:'] = _namespace_revid
986
def _namespace_last(self, revs, revision):
987
assert revision.startswith('last:')
989
offset = int(revision[5:])
994
raise BzrError('You must supply a positive value for --revision last:XXX')
995
return (len(revs) - offset + 1,)
996
REVISION_NAMESPACES['last:'] = _namespace_last
998
def _namespace_tag(self, revs, revision):
999
assert revision.startswith('tag:')
1000
raise BzrError('tag: namespace registered, but not implemented.')
1001
REVISION_NAMESPACES['tag:'] = _namespace_tag
1003
def _namespace_date(self, revs, revision):
1004
assert revision.startswith('date:')
1006
# Spec for date revisions:
1008
# value can be 'yesterday', 'today', 'tomorrow' or a YYYY-MM-DD string.
1009
# it can also start with a '+/-/='. '+' says match the first
1010
# entry after the given date. '-' is match the first entry before the date
1011
# '=' is match the first entry after, but still on the given date.
1013
# +2005-05-12 says find the first matching entry after May 12th, 2005 at 0:00
1014
# -2005-05-12 says find the first matching entry before May 12th, 2005 at 0:00
1015
# =2005-05-12 says find the first match after May 12th, 2005 at 0:00 but before
1016
# May 13th, 2005 at 0:00
1018
# So the proper way of saying 'give me all entries for today' is:
1019
# -r {date:+today}:{date:-tomorrow}
1020
# The default is '=' when not supplied
1023
if val[:1] in ('+', '-', '='):
1024
match_style = val[:1]
1027
today = datetime.datetime.today().replace(hour=0,minute=0,second=0,microsecond=0)
1028
if val.lower() == 'yesterday':
1029
dt = today - datetime.timedelta(days=1)
1030
elif val.lower() == 'today':
1032
elif val.lower() == 'tomorrow':
1033
dt = today + datetime.timedelta(days=1)
1036
# This should be done outside the function to avoid recompiling it.
1037
_date_re = re.compile(
1038
r'(?P<date>(?P<year>\d\d\d\d)-(?P<month>\d\d)-(?P<day>\d\d))?'
1040
r'(?P<time>(?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d))?)?'
1042
m = _date_re.match(val)
1043
if not m or (not m.group('date') and not m.group('time')):
1044
raise BzrError('Invalid revision date %r' % revision)
1047
year, month, day = int(m.group('year')), int(m.group('month')), int(m.group('day'))
1049
year, month, day = today.year, today.month, today.day
1051
hour = int(m.group('hour'))
1052
minute = int(m.group('minute'))
1053
if m.group('second'):
1054
second = int(m.group('second'))
1058
hour, minute, second = 0,0,0
1060
dt = datetime.datetime(year=year, month=month, day=day,
1061
hour=hour, minute=minute, second=second)
1065
if match_style == '-':
1067
elif match_style == '=':
1068
last = dt + datetime.timedelta(days=1)
1071
for i in range(len(revs)-1, -1, -1):
1072
r = self.get_revision(revs[i])
1073
# TODO: Handle timezone.
1074
dt = datetime.datetime.fromtimestamp(r.timestamp)
1075
if first >= dt and (last is None or dt >= last):
1078
for i in range(len(revs)):
1079
r = self.get_revision(revs[i])
1080
# TODO: Handle timezone.
1081
dt = datetime.datetime.fromtimestamp(r.timestamp)
1082
if first <= dt and (last is None or dt <= last):
1084
REVISION_NAMESPACES['date:'] = _namespace_date
1087
def _namespace_ancestor(self, revs, revision):
1088
from revision import common_ancestor, MultipleRevisionSources
1089
other_branch = find_branch(_trim_namespace('ancestor', revision))
1090
revision_a = self.last_patch()
1091
revision_b = other_branch.last_patch()
1092
for r, b in ((revision_a, self), (revision_b, other_branch)):
1094
raise bzrlib.errors.NoCommits(b)
1095
revision_source = MultipleRevisionSources(self, other_branch)
1096
result = common_ancestor(revision_a, revision_b, revision_source)
1098
revno = self.revision_id_to_revno(result)
1099
except bzrlib.errors.NoSuchRevision:
1104
REVISION_NAMESPACES['ancestor:'] = _namespace_ancestor
729
# list is 0-based; revisions are 1-based
730
return self.revision_history()[revno-1]
732
raise BzrError("no such revision %s" % revno)
1106
735
def revision_tree(self, revision_id):
1107
736
"""Return Tree for a revision on this branch.
1109
738
`revision_id` may be None for the null revision, in which case
1110
739
an `EmptyTree` is returned."""
1111
# TODO: refactor this to use an existing revision object
1112
# so we don't need to read it in twice.
740
self._need_readlock()
1113
741
if revision_id == None:
1114
742
return EmptyTree()
1142
779
This can change the directory or the filename or both.
781
self._need_writelock()
782
tree = self.working_tree()
784
if not tree.has_filename(from_rel):
785
bailout("can't rename: old working file %r does not exist" % from_rel)
786
if tree.has_filename(to_rel):
787
bailout("can't rename: new working file %r already exists" % to_rel)
789
file_id = inv.path2id(from_rel)
791
bailout("can't rename: old name %r is not versioned" % from_rel)
793
if inv.path2id(to_rel):
794
bailout("can't rename: new name %r is already versioned" % to_rel)
796
to_dir, to_tail = os.path.split(to_rel)
797
to_dir_id = inv.path2id(to_dir)
798
if to_dir_id == None and to_dir != '':
799
bailout("can't determine destination directory id for %r" % to_dir)
801
mutter("rename_one:")
802
mutter(" file_id {%s}" % file_id)
803
mutter(" from_rel %r" % from_rel)
804
mutter(" to_rel %r" % to_rel)
805
mutter(" to_dir %r" % to_dir)
806
mutter(" to_dir_id {%s}" % to_dir_id)
808
inv.rename(file_id, to_dir_id, to_tail)
810
print "%s => %s" % (from_rel, to_rel)
812
from_abs = self.abspath(from_rel)
813
to_abs = self.abspath(to_rel)
1146
tree = self.working_tree()
1147
inv = tree.inventory
1148
if not tree.has_filename(from_rel):
1149
raise BzrError("can't rename: old working file %r does not exist" % from_rel)
1150
if tree.has_filename(to_rel):
1151
raise BzrError("can't rename: new working file %r already exists" % to_rel)
1153
file_id = inv.path2id(from_rel)
1155
raise BzrError("can't rename: old name %r is not versioned" % from_rel)
1157
if inv.path2id(to_rel):
1158
raise BzrError("can't rename: new name %r is already versioned" % to_rel)
1160
to_dir, to_tail = os.path.split(to_rel)
1161
to_dir_id = inv.path2id(to_dir)
1162
if to_dir_id == None and to_dir != '':
1163
raise BzrError("can't determine destination directory id for %r" % to_dir)
1165
mutter("rename_one:")
1166
mutter(" file_id {%s}" % file_id)
1167
mutter(" from_rel %r" % from_rel)
1168
mutter(" to_rel %r" % to_rel)
1169
mutter(" to_dir %r" % to_dir)
1170
mutter(" to_dir_id {%s}" % to_dir_id)
1172
inv.rename(file_id, to_dir_id, to_tail)
1174
from_abs = self.abspath(from_rel)
1175
to_abs = self.abspath(to_rel)
1177
os.rename(from_abs, to_abs)
1179
raise BzrError("failed to rename %r to %r: %s"
1180
% (from_abs, to_abs, e[1]),
1181
["rename rolled back"])
1183
self._write_inventory(inv)
815
os.rename(from_abs, to_abs)
817
bailout("failed to rename %r to %r: %s"
818
% (from_abs, to_abs, e[1]),
819
["rename rolled back"])
821
self._write_inventory(inv)
1188
825
def move(self, from_paths, to_name):
1196
833
Note that to_name is only the last component of the new name;
1197
834
this doesn't change the directory.
1199
This returns a list of (from_path, to_path) pairs for each
1200
entry that is moved.
1205
## TODO: Option to move IDs only
1206
assert not isinstance(from_paths, basestring)
1207
tree = self.working_tree()
1208
inv = tree.inventory
1209
to_abs = self.abspath(to_name)
1210
if not isdir(to_abs):
1211
raise BzrError("destination %r is not a directory" % to_abs)
1212
if not tree.has_filename(to_name):
1213
raise BzrError("destination %r not in working directory" % to_abs)
1214
to_dir_id = inv.path2id(to_name)
1215
if to_dir_id == None and to_name != '':
1216
raise BzrError("destination %r is not a versioned directory" % to_name)
1217
to_dir_ie = inv[to_dir_id]
1218
if to_dir_ie.kind not in ('directory', 'root_directory'):
1219
raise BzrError("destination %r is not a directory" % to_abs)
1221
to_idpath = inv.get_idpath(to_dir_id)
1223
for f in from_paths:
1224
if not tree.has_filename(f):
1225
raise BzrError("%r does not exist in working tree" % f)
1226
f_id = inv.path2id(f)
1228
raise BzrError("%r is not versioned" % f)
1229
name_tail = splitpath(f)[-1]
1230
dest_path = appendpath(to_name, name_tail)
1231
if tree.has_filename(dest_path):
1232
raise BzrError("destination %r already exists" % dest_path)
1233
if f_id in to_idpath:
1234
raise BzrError("can't move %r to a subdirectory of itself" % f)
1236
# OK, so there's a race here, it's possible that someone will
1237
# create a file in this interval and then the rename might be
1238
# left half-done. But we should have caught most problems.
1240
for f in from_paths:
1241
name_tail = splitpath(f)[-1]
1242
dest_path = appendpath(to_name, name_tail)
1243
result.append((f, dest_path))
1244
inv.rename(inv.path2id(f), to_dir_id, name_tail)
1246
os.rename(self.abspath(f), self.abspath(dest_path))
1248
raise BzrError("failed to rename %r to %r: %s" % (f, dest_path, e[1]),
1249
["rename rolled back"])
1251
self._write_inventory(inv)
1258
def revert(self, filenames, old_tree=None, backups=True):
1259
"""Restore selected files to the versions from a previous tree.
1262
If true (default) backups are made of files before
1265
from bzrlib.errors import NotVersionedError, BzrError
1266
from bzrlib.atomicfile import AtomicFile
1267
from bzrlib.osutils import backup_file
1269
inv = self.read_working_inventory()
1270
if old_tree is None:
1271
old_tree = self.basis_tree()
1272
old_inv = old_tree.inventory
1275
for fn in filenames:
1276
file_id = inv.path2id(fn)
1278
raise NotVersionedError("not a versioned file", fn)
1279
if not old_inv.has_id(file_id):
1280
raise BzrError("file not present in old tree", fn, file_id)
1281
nids.append((fn, file_id))
1283
# TODO: Rename back if it was previously at a different location
1285
# TODO: If given a directory, restore the entire contents from
1286
# the previous version.
1288
# TODO: Make a backup to a temporary file.
1290
# TODO: If the file previously didn't exist, delete it?
1291
for fn, file_id in nids:
1294
f = AtomicFile(fn, 'wb')
1296
f.write(old_tree.get_file(file_id).read())
1302
def pending_merges(self):
1303
"""Return a list of pending merges.
1305
These are revisions that have been merged into the working
1306
directory but not yet committed.
1308
cfn = self.controlfilename('pending-merges')
1309
if not os.path.exists(cfn):
1312
for l in self.controlfile('pending-merges', 'r').readlines():
1313
p.append(l.rstrip('\n'))
1317
def add_pending_merge(self, revision_id):
1318
from bzrlib.revision import validate_revision_id
1320
validate_revision_id(revision_id)
1322
p = self.pending_merges()
1323
if revision_id in p:
1325
p.append(revision_id)
1326
self.set_pending_merges(p)
1329
def set_pending_merges(self, rev_list):
1330
from bzrlib.atomicfile import AtomicFile
1333
f = AtomicFile(self.controlfilename('pending-merges'))
1344
def get_parent(self):
1345
"""Return the parent location of the branch.
1347
This is the default location for push/pull/missing. The usual
1348
pattern is that the user can override it by specifying a
1352
_locs = ['parent', 'pull', 'x-pull']
1355
return self.controlfile(l, 'r').read().strip('\n')
1357
if e.errno != errno.ENOENT:
1362
def set_parent(self, url):
1363
# TODO: Maybe delete old location files?
1364
from bzrlib.atomicfile import AtomicFile
1367
f = AtomicFile(self.controlfilename('parent'))
1376
def check_revno(self, revno):
1378
Check whether a revno corresponds to any revision.
1379
Zero (the NULL revision) is considered valid.
1382
self.check_real_revno(revno)
1384
def check_real_revno(self, revno):
1386
Check whether a revno corresponds to a real revision.
1387
Zero (the NULL revision) is considered invalid
1389
if revno < 1 or revno > self.revno():
1390
raise InvalidRevisionNumber(revno)
836
self._need_writelock()
837
## TODO: Option to move IDs only
838
assert not isinstance(from_paths, basestring)
839
tree = self.working_tree()
841
to_abs = self.abspath(to_name)
842
if not isdir(to_abs):
843
bailout("destination %r is not a directory" % to_abs)
844
if not tree.has_filename(to_name):
845
bailout("destination %r not in working directory" % to_abs)
846
to_dir_id = inv.path2id(to_name)
847
if to_dir_id == None and to_name != '':
848
bailout("destination %r is not a versioned directory" % to_name)
849
to_dir_ie = inv[to_dir_id]
850
if to_dir_ie.kind not in ('directory', 'root_directory'):
851
bailout("destination %r is not a directory" % to_abs)
853
to_idpath = Set(inv.get_idpath(to_dir_id))
856
if not tree.has_filename(f):
857
bailout("%r does not exist in working tree" % f)
858
f_id = inv.path2id(f)
860
bailout("%r is not versioned" % f)
861
name_tail = splitpath(f)[-1]
862
dest_path = appendpath(to_name, name_tail)
863
if tree.has_filename(dest_path):
864
bailout("destination %r already exists" % dest_path)
865
if f_id in to_idpath:
866
bailout("can't move %r to a subdirectory of itself" % f)
868
# OK, so there's a race here, it's possible that someone will
869
# create a file in this interval and then the rename might be
870
# left half-done. But we should have caught most problems.
873
name_tail = splitpath(f)[-1]
874
dest_path = appendpath(to_name, name_tail)
875
print "%s => %s" % (f, dest_path)
876
inv.rename(inv.path2id(f), to_dir_id, name_tail)
878
os.rename(self.abspath(f), self.abspath(dest_path))
880
bailout("failed to rename %r to %r: %s" % (f, dest_path, e[1]),
881
["rename rolled back"])
883
self._write_inventory(inv)
887
def show_status(self, show_all=False):
888
"""Display single-line status for non-ignored working files.
890
The list is show sorted in order by file name.
892
>>> b = ScratchBranch(files=['foo', 'foo~'])
898
>>> b.commit("add foo")
900
>>> os.unlink(b.abspath('foo'))
904
TODO: Get state for single files.
906
self._need_readlock()
908
# We have to build everything into a list first so that it can
909
# sorted by name, incorporating all the different sources.
911
# FIXME: Rather than getting things in random order and then sorting,
912
# just step through in order.
914
# Interesting case: the old ID for a file has been removed,
915
# but a new file has been created under that name.
917
old = self.basis_tree()
918
new = self.working_tree()
920
for fs, fid, oldname, newname, kind in diff_trees(old, new):
922
show_status(fs, kind,
923
oldname + ' => ' + newname)
924
elif fs == 'A' or fs == 'M':
925
show_status(fs, kind, newname)
927
show_status(fs, kind, oldname)
930
show_status(fs, kind, newname)
933
show_status(fs, kind, newname)
935
show_status(fs, kind, newname)
937
bailout("weird file state %r" % ((fs, fid),))
1395
941
class ScratchBranch(Branch):