3
# Copyright (C) 2005 Canonical Ltd
5
# This program is free software; you can redistribute it and/or modify
6
# it under the terms of the GNU General Public License as published by
7
# the Free Software Foundation; either version 2 of the License, or
8
# (at your option) any later version.
10
# This program is distributed in the hope that it will be useful,
11
# but WITHOUT ANY WARRANTY; without even the implied warranty of
12
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13
# GNU General Public License for more details.
15
# You should have received a copy of the GNU General Public License
16
# along with this program; if not, write to the Free Software
17
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20
"""Proxy object for access to remote branches.
22
At the moment remote branches are only for HTTP and only for read
28
from cStringIO import StringIO
31
from errors import BzrError, BzrCheckError
32
from branch import Branch, BZR_BRANCH_FORMAT
33
from trace import mutter
35
# velocitynet.com.au transparently proxies connections and thereby
36
# breaks keep-alive -- sucks!
39
ENABLE_URLGRABBER = True
41
from bzrlib.errors import BzrError
43
class GetFailed(BzrError):
44
def __init__(self, url, status):
45
BzrError.__init__(self, "Get %s failed with status %s" % (url, status))
50
import util.urlgrabber
51
import util.urlgrabber.keepalive
52
util.urlgrabber.keepalive.DEBUG = 0
53
def get_url(path, compressed=False):
58
mutter("grab url %s" % url)
59
url_f = util.urlgrabber.urlopen(url, keepalive=1, close_connection=0)
60
if url_f.status != 200:
61
raise GetFailed(url, url_f.status)
65
return gzip.GzipFile(fileobj=StringIO(url_f.read()))
66
except urllib2.URLError, e:
67
raise BzrError("remote fetch failed: %r: %s" % (url, e))
69
def get_url(url, compressed=False):
73
mutter("get_url %s" % url)
74
url_f = urllib2.urlopen(url)
76
return gzip.GzipFile(fileobj=StringIO(url_f.read()))
82
def _find_remote_root(url):
83
"""Return the prefix URL that corresponds to the branch root."""
87
ff = get_url(url + '/.bzr/branch-format')
92
fmt = fmt.rstrip('\r\n')
93
if fmt != BZR_BRANCH_FORMAT.rstrip('\r\n'):
94
raise BzrError("sorry, branch format %r not supported at url %s"
98
except urllib2.URLError:
102
idx = url.rindex('/')
104
raise BzrError('no branch root found for URL %s' % orig_url)
110
class RemoteBranch(Branch):
111
def __init__(self, baseurl, find_root=True):
112
"""Create new proxy for a remote branch."""
114
self.baseurl = _find_remote_root(baseurl)
116
self.baseurl = baseurl
119
self.inventory_store = RemoteStore(baseurl + '/.bzr/inventory-store/')
120
self.text_store = RemoteStore(baseurl + '/.bzr/text-store/')
121
self.revision_store = RemoteStore(baseurl + '/.bzr/revision-store/')
124
b = getattr(self, 'baseurl', 'undefined')
125
return '%s(%r)' % (self.__class__.__name__, b)
129
def controlfile(self, filename, mode):
130
if mode not in ('rb', 'rt', 'r'):
131
raise BzrError("file mode %r not supported for remote branches" % mode)
132
return get_url(self.baseurl + '/.bzr/' + filename, False)
136
# no locking for remote branches yet
139
def lock_write(self):
140
from errors import LockError
141
raise LockError("write lock not supported for remote branch %s"
148
def relpath(self, path):
149
if not path.startswith(self.baseurl):
150
raise BzrError('path %r is not under base URL %r'
151
% (path, self.baseurl))
152
pl = len(self.baseurl)
153
return path[pl:].lstrip('/')
156
def get_revision(self, revision_id):
157
from bzrlib.revision import Revision
158
from bzrlib.xml import unpack_xml
159
revf = self.revision_store[revision_id]
160
r = unpack_xml(Revision, revf)
161
if r.revision_id != revision_id:
162
raise BzrCheckError('revision stored as {%s} actually contains {%s}'
163
% (revision_id, r.revision_id))
167
class RemoteStore(object):
168
def __init__(self, baseurl):
169
self._baseurl = baseurl
172
def _path(self, name):
174
raise ValueError('invalid store id', name)
175
return self._baseurl + '/' + name
177
def __getitem__(self, fileid):
178
p = self._path(fileid)
180
return get_url(p, compressed=True)
182
raise KeyError(fileid)