1
# Copyright (C) 2005 Robey Pointer <robey@lag.net>, Canonical Ltd
1
# Copyright (C) 2005, 2006, 2007, 2008, 2009 Canonical Ltd
3
3
# This program is free software; you can redistribute it and/or modify
4
4
# it under the terms of the GNU General Public License as published by
5
5
# the Free Software Foundation; either version 2 of the License, or
6
6
# (at your option) any later version.
8
8
# This program is distributed in the hope that it will be useful,
9
9
# but WITHOUT ANY WARRANTY; without even the implied warranty of
10
10
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11
11
# GNU General Public License for more details.
13
13
# You should have received a copy of the GNU General Public License
14
14
# along with this program; if not, write to the Free Software
15
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
15
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
17
17
"""Implementation of Transport over SFTP, using paramiko."""
19
# TODO: Remove the transport-based lock_read and lock_write methods. They'll
20
# then raise TransportNotPossible, which will break remote access to any
21
# formats which rely on OS-level locks. That should be fine as those formats
22
# are pretty old, but these combinations may have to be removed from the test
23
# suite. Those formats all date back to 0.7; so we should be able to remove
24
# these methods when we officially drop support for those formats.
32
from bzrlib.config import config_dir, ensure_config_dir_exists
33
from bzrlib.errors import (ConnectionError,
35
TransportNotPossible, NoSuchFile, PathNotChild,
46
from bzrlib.errors import (FileExists,
47
NoSuchFile, PathNotChild,
37
LockError, ParamikoNotPresent
39
from bzrlib.osutils import pathjoin, fancy_rename
40
from bzrlib.trace import mutter, warning, error
41
from bzrlib.transport import Transport, Server, urlescape
53
from bzrlib.osutils import pathjoin, fancy_rename, getcwd
54
from bzrlib.symbol_versioning import (
57
from bzrlib.trace import mutter, warning
58
from bzrlib.transport import (
67
# Disable one particular warning that comes from paramiko in Python2.5; if
68
# this is emitted at the wrong time it tends to cause spurious test failures
69
# or at least noise in the test case::
71
# [1770/7639 in 86s, 1 known failures, 50 skipped, 2 missing features]
72
# test_permissions.TestSftpPermissions.test_new_files
73
# /var/lib/python-support/python2.5/paramiko/message.py:226: DeprecationWarning: integer argument expected, got float
74
# self.packet.write(struct.pack('>I', n))
75
warnings.filterwarnings('ignore',
76
'integer argument expected, got float',
77
category=DeprecationWarning,
78
module='paramiko.message')
51
87
CMD_HANDLE, CMD_OPEN)
52
88
from paramiko.sftp_attr import SFTPAttributes
53
89
from paramiko.sftp_file import SFTPFile
54
from paramiko.sftp_client import SFTPClient
56
if 'sftp' not in urlparse.uses_netloc:
57
urlparse.uses_netloc.append('sftp')
59
# don't use prefetch unless paramiko version >= 1.5.2 (there were bugs earlier)
60
_default_do_prefetch = False
61
if getattr(paramiko, '__version_info__', (0, 0, 0)) >= (1, 5, 2):
62
_default_do_prefetch = True
66
if sys.platform == 'win32':
67
# close_fds not supported on win32
72
def _get_ssh_vendor():
73
"""Find out what version of SSH is on the system."""
75
if _ssh_vendor is not None:
80
if 'BZR_SSH' in os.environ:
81
_ssh_vendor = os.environ['BZR_SSH']
82
if _ssh_vendor == 'paramiko':
87
p = subprocess.Popen(['ssh', '-V'],
89
stdin=subprocess.PIPE,
90
stdout=subprocess.PIPE,
91
stderr=subprocess.PIPE)
92
returncode = p.returncode
93
stdout, stderr = p.communicate()
97
if 'OpenSSH' in stderr:
98
mutter('ssh implementation is OpenSSH')
99
_ssh_vendor = 'openssh'
100
elif 'SSH Secure Shell' in stderr:
101
mutter('ssh implementation is SSH Corp.')
104
if _ssh_vendor != 'none':
107
# XXX: 20051123 jamesh
108
# A check for putty's plink or lsh would go here.
110
mutter('falling back to paramiko implementation')
114
class SFTPSubprocess:
115
"""A socket-like object that talks to an ssh subprocess via pipes."""
116
def __init__(self, hostname, vendor, port=None, user=None):
117
assert vendor in ['openssh', 'ssh']
118
if vendor == 'openssh':
120
'-oForwardX11=no', '-oForwardAgent=no',
121
'-oClearAllForwardings=yes', '-oProtocol=2',
122
'-oNoHostAuthenticationForLocalhost=yes']
124
args.extend(['-p', str(port)])
126
args.extend(['-l', user])
127
args.extend(['-s', hostname, 'sftp'])
128
elif vendor == 'ssh':
131
args.extend(['-p', str(port)])
133
args.extend(['-l', user])
134
args.extend(['-s', 'sftp', hostname])
136
self.proc = subprocess.Popen(args, close_fds=_close_fds,
137
stdin=subprocess.PIPE,
138
stdout=subprocess.PIPE)
140
def send(self, data):
141
return os.write(self.proc.stdin.fileno(), data)
143
def recv_ready(self):
144
# TODO: jam 20051215 this function is necessary to support the
145
# pipelined() function. In reality, it probably should use
146
# poll() or select() to actually return if there is data
147
# available, otherwise we probably don't get any benefit
150
def recv(self, count):
151
return os.read(self.proc.stdout.fileno(), count)
154
self.proc.stdin.close()
155
self.proc.stdout.close()
159
class LoopbackSFTP(object):
160
"""Simple wrapper for a socket that pretends to be a paramiko Channel."""
162
def __init__(self, sock):
165
def send(self, data):
166
return self.__socket.send(data)
169
return self.__socket.recv(n)
171
def recv_ready(self):
175
self.__socket.close()
181
# This is a weakref dictionary, so that we can reuse connections
182
# that are still active. Long term, it might be nice to have some
183
# sort of expiration policy, such as disconnect if inactive for
184
# X seconds. But that requires a lot more fanciness.
185
_connected_hosts = weakref.WeakValueDictionary()
188
def load_host_keys():
190
Load system host keys (probably doesn't work on windows) and any
191
"discovered" keys from previous sessions.
193
global SYSTEM_HOSTKEYS, BZR_HOSTKEYS
195
SYSTEM_HOSTKEYS = paramiko.util.load_host_keys(os.path.expanduser('~/.ssh/known_hosts'))
197
mutter('failed to load system host keys: ' + str(e))
198
bzr_hostkey_path = pathjoin(config_dir(), 'ssh_host_keys')
200
BZR_HOSTKEYS = paramiko.util.load_host_keys(bzr_hostkey_path)
202
mutter('failed to load bzr host keys: ' + str(e))
206
def save_host_keys():
208
Save "discovered" host keys in $(config)/ssh_host_keys/.
210
global SYSTEM_HOSTKEYS, BZR_HOSTKEYS
211
bzr_hostkey_path = pathjoin(config_dir(), 'ssh_host_keys')
212
ensure_config_dir_exists()
215
f = open(bzr_hostkey_path, 'w')
216
f.write('# SSH host keys collected by bzr\n')
217
for hostname, keys in BZR_HOSTKEYS.iteritems():
218
for keytype, key in keys.iteritems():
219
f.write('%s %s %s\n' % (hostname, keytype, key.get_base64()))
222
mutter('failed to save bzr host keys: ' + str(e))
92
_paramiko_version = getattr(paramiko, '__version_info__', (0, 0, 0))
93
# don't use prefetch unless paramiko version >= 1.5.5 (there were bugs earlier)
94
_default_do_prefetch = (_paramiko_version >= (1, 5, 5))
225
97
class SFTPLock(object):
226
"""This fakes a lock in a remote location."""
98
"""This fakes a lock in a remote location.
100
A present lock is indicated just by the existence of a file. This
101
doesn't work well on all transports and they are only used in
102
deprecated storage formats.
227
105
__slots__ = ['path', 'lock_path', 'lock_file', 'transport']
228
107
def __init__(self, path, transport):
229
assert isinstance(transport, SFTPTransport)
231
108
self.lock_file = None
233
110
self.lock_path = path + '.write-lock'
256
133
# What specific errors should we catch here?
259
class SFTPTransport (Transport):
261
Transport implementation for SFTP access.
137
class _SFTPReadvHelper(object):
138
"""A class to help with managing the state of a readv request."""
140
# See _get_requests for an explanation.
141
_max_request_size = 32768
143
def __init__(self, original_offsets, relpath, _report_activity):
144
"""Create a new readv helper.
146
:param original_offsets: The original requests given by the caller of
148
:param relpath: The name of the file (if known)
149
:param _report_activity: A Transport._report_activity bound method,
150
to be called as data arrives.
152
self.original_offsets = list(original_offsets)
153
self.relpath = relpath
154
self._report_activity = _report_activity
156
def _get_requests(self):
157
"""Break up the offsets into individual requests over sftp.
159
The SFTP spec only requires implementers to support 32kB requests. We
160
could try something larger (openssh supports 64kB), but then we have to
161
handle requests that fail.
162
So instead, we just break up our maximum chunks into 32kB chunks, and
163
asyncronously requests them.
164
Newer versions of paramiko would do the chunking for us, but we want to
165
start processing results right away, so we do it ourselves.
167
# TODO: Because we issue async requests, we don't 'fudge' any extra
168
# data. I'm not 100% sure that is the best choice.
170
# The first thing we do, is to collapse the individual requests as much
171
# as possible, so we don't issues requests <32kB
172
sorted_offsets = sorted(self.original_offsets)
173
coalesced = list(ConnectedTransport._coalesce_offsets(sorted_offsets,
174
limit=0, fudge_factor=0))
176
for c_offset in coalesced:
177
start = c_offset.start
178
size = c_offset.length
180
# Break this up into 32kB requests
182
next_size = min(size, self._max_request_size)
183
requests.append((start, next_size))
186
if 'sftp' in debug.debug_flags:
187
mutter('SFTP.readv(%s) %s offsets => %s coalesced => %s requests',
188
self.relpath, len(sorted_offsets), len(coalesced),
192
def request_and_yield_offsets(self, fp):
193
"""Request the data from the remote machine, yielding the results.
195
:param fp: A Paramiko SFTPFile object that supports readv.
196
:return: Yield the data requested by the original readv caller, one by
199
requests = self._get_requests()
200
offset_iter = iter(self.original_offsets)
201
cur_offset, cur_size = offset_iter.next()
202
# paramiko .readv() yields strings that are in the order of the requests
203
# So we track the current request to know where the next data is
204
# being returned from.
210
# This is used to buffer chunks which we couldn't process yet
211
# It is (start, end, data) tuples.
213
# Create an 'unlimited' data stream, so we stop based on requests,
214
# rather than just because the data stream ended. This lets us detect
216
data_stream = itertools.chain(fp.readv(requests),
217
itertools.repeat(None))
218
for (start, length), data in itertools.izip(requests, data_stream):
220
if cur_coalesced is not None:
221
raise errors.ShortReadvError(self.relpath,
222
start, length, len(data))
223
if len(data) != length:
224
raise errors.ShortReadvError(self.relpath,
225
start, length, len(data))
226
self._report_activity(length, 'read')
228
# This is the first request, just buffer it
229
buffered_data = [data]
230
buffered_len = length
232
elif start == last_end:
233
# The data we are reading fits neatly on the previous
234
# buffer, so this is all part of a larger coalesced range.
235
buffered_data.append(data)
236
buffered_len += length
238
# We have an 'interrupt' in the data stream. So we know we are
239
# at a request boundary.
241
# We haven't consumed the buffer so far, so put it into
242
# data_chunks, and continue.
243
buffered = ''.join(buffered_data)
244
data_chunks.append((input_start, buffered))
246
buffered_data = [data]
247
buffered_len = length
248
last_end = start + length
249
if input_start == cur_offset and cur_size <= buffered_len:
250
# Simplify the next steps a bit by transforming buffered_data
251
# into a single string. We also have the nice property that
252
# when there is only one string ''.join([x]) == x, so there is
254
buffered = ''.join(buffered_data)
255
# Clean out buffered data so that we keep memory
259
# TODO: We *could* also consider the case where cur_offset is in
260
# in the buffered range, even though it doesn't *start*
261
# the buffered range. But for packs we pretty much always
262
# read in order, so you won't get any extra data in the
264
while (input_start == cur_offset
265
and (buffered_offset + cur_size) <= buffered_len):
266
# We've buffered enough data to process this request, spit it
268
cur_data = buffered[buffered_offset:buffered_offset + cur_size]
269
# move the direct pointer into our buffered data
270
buffered_offset += cur_size
271
# Move the start-of-buffer pointer
272
input_start += cur_size
273
# Yield the requested data
274
yield cur_offset, cur_data
275
cur_offset, cur_size = offset_iter.next()
276
# at this point, we've consumed as much of buffered as we can,
277
# so break off the portion that we consumed
278
if buffered_offset == len(buffered_data):
279
# No tail to leave behind
283
buffered = buffered[buffered_offset:]
284
buffered_data = [buffered]
285
buffered_len = len(buffered)
287
buffered = ''.join(buffered_data)
289
data_chunks.append((input_start, buffered))
291
if 'sftp' in debug.debug_flags:
292
mutter('SFTP readv left with %d out-of-order bytes',
293
sum(map(lambda x: len(x[1]), data_chunks)))
294
# We've processed all the readv data, at this point, anything we
295
# couldn't process is in data_chunks. This doesn't happen often, so
296
# this code path isn't optimized
297
# We use an interesting process for data_chunks
298
# Specifically if we have "bisect_left([(start, len, entries)],
300
# If start == qstart, then we get the specific node. Otherwise we
301
# get the previous node
303
idx = bisect.bisect_left(data_chunks, (cur_offset,))
304
if idx < len(data_chunks) and data_chunks[idx][0] == cur_offset:
305
# The data starts here
306
data = data_chunks[idx][1][:cur_size]
308
# The data is in a portion of a previous page
310
sub_offset = cur_offset - data_chunks[idx][0]
311
data = data_chunks[idx][1]
312
data = data[sub_offset:sub_offset + cur_size]
314
# We are missing the page where the data should be found,
317
if len(data) != cur_size:
318
raise AssertionError('We must have miscalulated.'
319
' We expected %d bytes, but only found %d'
320
% (cur_size, len(data)))
321
yield cur_offset, data
322
cur_offset, cur_size = offset_iter.next()
325
class SFTPTransport(ConnectedTransport):
326
"""Transport implementation for SFTP access."""
263
328
_do_prefetch = _default_do_prefetch
265
def __init__(self, base, clone_from=None):
266
assert base.startswith('sftp://')
267
self._parse_url(base)
268
base = self._unparse_url()
271
super(SFTPTransport, self).__init__(base)
272
if clone_from is None:
275
# use the same ssh connection, etc
276
self._sftp = clone_from._sftp
277
# super saves 'self.base'
279
def should_cache(self):
281
Return True if the data pulled across should be cached locally.
285
def clone(self, offset=None):
287
Return a new SFTPTransport with root at self.base + offset.
288
We share the same SFTP session between such transports, because it's
289
fairly expensive to set them up.
292
return SFTPTransport(self.base, self)
294
return SFTPTransport(self.abspath(offset), self)
296
def abspath(self, relpath):
298
Return the full url to the given relative path.
300
@param relpath: the relative path or path components
301
@type relpath: str or list
303
return self._unparse_url(self._remote_path(relpath))
329
# TODO: jam 20060717 Conceivably these could be configurable, either
330
# by auto-tuning at run-time, or by a configuration (per host??)
331
# but the performance curve is pretty flat, so just going with
332
# reasonable defaults.
333
_max_readv_combine = 200
334
# Having to round trip to the server means waiting for a response,
335
# so it is better to download extra bytes.
336
# 8KiB had good performance for both local and remote network operations
337
_bytes_to_read_before_seek = 8192
339
# The sftp spec says that implementations SHOULD allow reads
340
# to be at least 32K. paramiko.readv() does an async request
341
# for the chunks. So we need to keep it within a single request
342
# size for paramiko <= 1.6.1. paramiko 1.6.2 will probably chop
343
# up the request itself, rather than us having to worry about it
344
_max_request_size = 32768
346
def __init__(self, base, _from_transport=None):
347
super(SFTPTransport, self).__init__(base,
348
_from_transport=_from_transport)
305
350
def _remote_path(self, relpath):
306
351
"""Return the path to be passed along the sftp protocol for relpath.
308
relpath is a urlencoded string.
310
# FIXME: share the common code across transports
311
assert isinstance(relpath, basestring)
312
relpath = urllib.unquote(relpath).split('/')
313
basepath = self._path.split('/')
314
if len(basepath) > 0 and basepath[-1] == '':
315
basepath = basepath[:-1]
319
if len(basepath) == 0:
320
# In most filesystems, a request for the parent
321
# of root, just returns root.
329
path = '/'.join(basepath)
332
def relpath(self, abspath):
333
username, password, host, port, path = self._split_url(abspath)
335
if (username != self._username):
336
error.append('username mismatch')
337
if (host != self._host):
338
error.append('host mismatch')
339
if (port != self._port):
340
error.append('port mismatch')
341
if (not path.startswith(self._path)):
342
error.append('path mismatch')
344
extra = ': ' + ', '.join(error)
345
raise PathNotChild(abspath, self.base, extra=extra)
347
return path[pl:].strip('/')
353
:param relpath: is a urlencoded string.
355
relative = urlutils.unescape(relpath).encode('utf-8')
356
remote_path = self._combine_paths(self._path, relative)
357
# the initial slash should be removed from the path, and treated as a
358
# homedir relative path (the path begins with a double slash if it is
359
# absolute). see draft-ietf-secsh-scp-sftp-ssh-uri-03.txt
360
# RBC 20060118 we are not using this as its too user hostile. instead
361
# we are following lftp and using /~/foo to mean '~/foo'
362
# vila--20070602 and leave absolute paths begin with a single slash.
363
if remote_path.startswith('/~/'):
364
remote_path = remote_path[3:]
365
elif remote_path == '/~':
369
def _create_connection(self, credentials=None):
370
"""Create a new connection with the provided credentials.
372
:param credentials: The credentials needed to establish the connection.
374
:return: The created connection and its associated credentials.
376
The credentials are only the password as it may have been entered
377
interactively by the user and may be different from the one provided
378
in base url at transport creation time.
380
if credentials is None:
381
password = self._password
383
password = credentials
385
vendor = ssh._get_ssh_vendor()
388
auth = config.AuthenticationConfig()
389
user = auth.get_user('ssh', self._host, self._port)
390
connection = vendor.connect_sftp(self._user, password,
391
self._host, self._port)
392
return connection, (user, password)
395
"""Ensures that a connection is established"""
396
connection = self._get_connection()
397
if connection is None:
398
# First connection ever
399
connection, credentials = self._create_connection()
400
self._set_connection(connection, credentials)
349
403
def has(self, relpath):
351
405
Does the target location exist?
354
self._sftp.stat(self._remote_path(relpath))
408
self._get_sftp().stat(self._remote_path(relpath))
409
# stat result is about 20 bytes, let's say
410
self._report_activity(20, 'read')
359
def get(self, relpath, decode=False):
361
Get the file at the given relative path.
415
def get(self, relpath):
416
"""Get the file at the given relative path.
363
418
:param relpath: The relative path to the file
421
# FIXME: by returning the file directly, we don't pass this
422
# through to report_activity. We could try wrapping the object
423
# before it's returned. For readv and get_bytes it's handled in
424
# the higher-level function.
366
426
path = self._remote_path(relpath)
367
f = self._sftp.file(path, mode='rb')
427
f = self._get_sftp().file(path, mode='rb')
368
428
if self._do_prefetch and (getattr(f, 'prefetch', None) is not None):
371
431
except (IOError, paramiko.SSHException), e:
432
self._translate_io_exception(e, path, ': error retrieving',
433
failure_exc=errors.ReadError)
435
def get_bytes(self, relpath):
436
# reimplement this here so that we can report how many bytes came back
437
f = self.get(relpath)
440
self._report_activity(len(bytes), 'read')
445
def _readv(self, relpath, offsets):
446
"""See Transport.readv()"""
447
# We overload the default readv() because we want to use a file
448
# that does not have prefetch enabled.
449
# Also, if we have a new paramiko, it implements an async readv()
454
path = self._remote_path(relpath)
455
fp = self._get_sftp().file(path, mode='rb')
456
readv = getattr(fp, 'readv', None)
458
return self._sftp_readv(fp, offsets, relpath)
459
if 'sftp' in debug.debug_flags:
460
mutter('seek and read %s offsets', len(offsets))
461
return self._seek_and_read(fp, offsets, relpath)
462
except (IOError, paramiko.SSHException), e:
372
463
self._translate_io_exception(e, path, ': error retrieving')
374
def get_partial(self, relpath, start, length=None):
376
Get just part of a file.
378
:param relpath: Path to the file, relative to base
379
:param start: The starting position to read from
380
:param length: The length to read. A length of None indicates
381
read to the end of the file.
382
:return: A file-like object containing at least the specified bytes.
383
Some implementations may return objects which can be read
384
past this length, but this is not guaranteed.
386
# TODO: implement get_partial_multi to help with knit support
387
f = self.get(relpath)
389
if self._do_prefetch and hasattr(f, 'prefetch'):
393
def put(self, relpath, f, mode=None):
395
Copy the file-like or string object into the location.
465
def recommended_page_size(self):
466
"""See Transport.recommended_page_size().
468
For SFTP we suggest a large page size to reduce the overhead
469
introduced by latency.
473
def _sftp_readv(self, fp, offsets, relpath):
474
"""Use the readv() member of fp to do async readv.
476
Then read them using paramiko.readv(). paramiko.readv()
477
does not support ranges > 64K, so it caps the request size, and
478
just reads until it gets all the stuff it wants.
480
helper = _SFTPReadvHelper(offsets, relpath, self._report_activity)
481
return helper.request_and_yield_offsets(fp)
483
def put_file(self, relpath, f, mode=None):
485
Copy the file-like object into the location.
397
487
:param relpath: Location to put the contents, relative to base.
398
:param f: File-like or string object.
488
:param f: File-like object.
399
489
:param mode: The final mode for the file
401
491
final_path = self._remote_path(relpath)
402
self._put(final_path, f, mode=mode)
492
return self._put(final_path, f, mode=mode)
404
494
def _put(self, abspath, f, mode=None):
405
495
"""Helper function so both put() and copy_abspaths can reuse the code"""
633
def _mkdir(self, abspath, mode=None):
639
self._report_activity(len(abspath), 'write')
640
self._get_sftp().mkdir(abspath, local_mode)
641
self._report_activity(1, 'read')
643
# chmod a dir through sftp will erase any sgid bit set
644
# on the server side. So, if the bit mode are already
645
# set, avoid the chmod. If the mode is not fine but
646
# the sgid bit is set, report a warning to the user
647
# with the umask fix.
648
stat = self._get_sftp().lstat(abspath)
649
mode = mode & 0777 # can't set special bits anyway
650
if mode != stat.st_mode & 0777:
651
if stat.st_mode & 06000:
652
warning('About to chmod %s over sftp, which will result'
653
' in its suid or sgid bits being cleared. If'
654
' you want to preserve those bits, change your '
655
' environment on the server to use umask 0%03o.'
656
% (abspath, 0777 - mode))
657
self._get_sftp().chmod(abspath, mode=mode)
658
except (paramiko.SSHException, IOError), e:
659
self._translate_io_exception(e, abspath, ': unable to mkdir',
660
failure_exc=FileExists)
451
662
def mkdir(self, relpath, mode=None):
452
663
"""Create a directory at the given path."""
664
self._mkdir(self._remote_path(relpath), mode=mode)
666
def open_write_stream(self, relpath, mode=None):
667
"""See Transport.open_write_stream."""
668
# initialise the file to zero-length
669
# this is three round trips, but we don't use this
670
# api more than once per write_group at the moment so
671
# it is a tolerable overhead. Better would be to truncate
672
# the file after opening. RBC 20070805
673
self.put_bytes_non_atomic(relpath, "", mode)
674
abspath = self._remote_path(relpath)
675
# TODO: jam 20060816 paramiko doesn't publicly expose a way to
676
# set the file mode at create time. If it does, use it.
677
# But for now, we just chmod later anyway.
454
path = self._remote_path(relpath)
455
# In the paramiko documentation, it says that passing a mode flag
456
# will filtered against the server umask.
457
# StubSFTPServer does not do this, which would be nice, because it is
458
# what we really want :)
459
# However, real servers do use umask, so we really should do it that way
460
self._sftp.mkdir(path)
462
self._sftp.chmod(path, mode=mode)
680
handle = self._get_sftp().file(abspath, mode='wb')
681
handle.set_pipelined(True)
463
682
except (paramiko.SSHException, IOError), e:
464
self._translate_io_exception(e, path, ': unable to mkdir',
465
failure_exc=FileExists)
683
self._translate_io_exception(e, abspath,
685
_file_streams[self.abspath(relpath)] = handle
686
return FileFileStream(self, relpath, handle)
467
def _translate_io_exception(self, e, path, more_info='', failure_exc=NoSuchFile):
688
def _translate_io_exception(self, e, path, more_info='',
689
failure_exc=PathError):
468
690
"""Translate a paramiko or IOError into a friendlier exception.
470
692
:param e: The original exception
597
842
# that we have taken the lock.
598
843
return SFTPLock(relpath, self)
600
def _unparse_url(self, path=None):
603
path = urllib.quote(path)
604
# handle homedir paths
605
if not path.startswith('/'):
607
netloc = urllib.quote(self._host)
608
if self._username is not None:
609
netloc = '%s@%s' % (urllib.quote(self._username), netloc)
610
if self._port is not None:
611
netloc = '%s:%d' % (netloc, self._port)
613
return urlparse.urlunparse(('sftp', netloc, path, '', '', ''))
615
def _split_url(self, url):
616
if isinstance(url, unicode):
617
url = url.encode('utf-8')
618
(scheme, netloc, path, params,
619
query, fragment) = urlparse.urlparse(url, allow_fragments=False)
620
assert scheme == 'sftp'
621
username = password = host = port = None
623
username, host = netloc.split('@', 1)
625
username, password = username.split(':', 1)
626
password = urllib.unquote(password)
627
username = urllib.unquote(username)
632
host, port = host.rsplit(':', 1)
636
# TODO: Should this be ConnectionError?
637
raise TransportError('%s: invalid port number' % port)
638
host = urllib.unquote(host)
640
path = urllib.unquote(path)
642
# the initial slash should be removed from the path, and treated
643
# as a homedir relative path (the path begins with a double slash
644
# if it is absolute).
645
# see draft-ietf-secsh-scp-sftp-ssh-uri-03.txt
646
# RBC 20060118 we are not using this as its too user hostile. instead
647
# we are following lftp and using /~/foo to mean '~/foo'.
648
# handle homedir paths
649
if path.startswith('/~/'):
653
return (username, password, host, port, path)
655
def _parse_url(self, url):
656
(self._username, self._password,
657
self._host, self._port, self._path) = self._split_url(url)
659
def _sftp_connect(self):
660
"""Connect to the remote sftp server.
661
After this, self._sftp should have a valid connection (or
662
we raise an TransportError 'could not connect').
664
TODO: Raise a more reasonable ConnectionFailed exception
666
global _connected_hosts
668
idx = (self._host, self._port, self._username)
670
self._sftp = _connected_hosts[idx]
675
vendor = _get_ssh_vendor()
676
if vendor == 'loopback':
677
sock = socket.socket()
678
sock.connect((self._host, self._port))
679
self._sftp = SFTPClient(LoopbackSFTP(sock))
680
elif vendor != 'none':
681
sock = SFTPSubprocess(self._host, vendor, self._port,
683
self._sftp = SFTPClient(sock)
685
self._paramiko_connect()
687
_connected_hosts[idx] = self._sftp
689
def _paramiko_connect(self):
690
global SYSTEM_HOSTKEYS, BZR_HOSTKEYS
695
t = paramiko.Transport((self._host, self._port or 22))
696
t.set_log_channel('bzr.paramiko')
698
except paramiko.SSHException, e:
699
raise ConnectionError('Unable to reach SSH host %s:%d' %
700
(self._host, self._port), e)
702
server_key = t.get_remote_server_key()
703
server_key_hex = paramiko.util.hexify(server_key.get_fingerprint())
704
keytype = server_key.get_name()
705
if SYSTEM_HOSTKEYS.has_key(self._host) and SYSTEM_HOSTKEYS[self._host].has_key(keytype):
706
our_server_key = SYSTEM_HOSTKEYS[self._host][keytype]
707
our_server_key_hex = paramiko.util.hexify(our_server_key.get_fingerprint())
708
elif BZR_HOSTKEYS.has_key(self._host) and BZR_HOSTKEYS[self._host].has_key(keytype):
709
our_server_key = BZR_HOSTKEYS[self._host][keytype]
710
our_server_key_hex = paramiko.util.hexify(our_server_key.get_fingerprint())
712
warning('Adding %s host key for %s: %s' % (keytype, self._host, server_key_hex))
713
if not BZR_HOSTKEYS.has_key(self._host):
714
BZR_HOSTKEYS[self._host] = {}
715
BZR_HOSTKEYS[self._host][keytype] = server_key
716
our_server_key = server_key
717
our_server_key_hex = paramiko.util.hexify(our_server_key.get_fingerprint())
719
if server_key != our_server_key:
720
filename1 = os.path.expanduser('~/.ssh/known_hosts')
721
filename2 = pathjoin(config_dir(), 'ssh_host_keys')
722
raise TransportError('Host keys for %s do not match! %s != %s' % \
723
(self._host, our_server_key_hex, server_key_hex),
724
['Try editing %s or %s' % (filename1, filename2)])
729
self._sftp = t.open_sftp_client()
730
except paramiko.SSHException, e:
731
raise ConnectionError('Unable to start sftp client %s:%d' %
732
(self._host, self._port), e)
734
def _sftp_auth(self, transport):
735
# paramiko requires a username, but it might be none if nothing was supplied
736
# use the local username, just in case.
737
# We don't override self._username, because if we aren't using paramiko,
738
# the username might be specified in ~/.ssh/config and we don't want to
739
# force it to something else
740
# Also, it would mess up the self.relpath() functionality
741
username = self._username or getpass.getuser()
743
# Paramiko tries to open a socket.AF_UNIX in order to connect
744
# to ssh-agent. That attribute doesn't exist on win32 (it does in cygwin)
745
# so we get an AttributeError exception. For now, just don't try to
746
# connect to an agent if we are on win32
747
if sys.platform != 'win32':
748
agent = paramiko.Agent()
749
for key in agent.get_keys():
750
mutter('Trying SSH agent key %s' % paramiko.util.hexify(key.get_fingerprint()))
752
transport.auth_publickey(username, key)
754
except paramiko.SSHException, e:
757
# okay, try finding id_rsa or id_dss? (posix only)
758
if self._try_pkey_auth(transport, paramiko.RSAKey, username, 'id_rsa'):
760
if self._try_pkey_auth(transport, paramiko.DSSKey, username, 'id_dsa'):
765
transport.auth_password(username, self._password)
767
except paramiko.SSHException, e:
770
# FIXME: Don't keep a password held in memory if you can help it
771
#self._password = None
773
# give up and ask for a password
774
password = bzrlib.ui.ui_factory.get_password(
775
prompt='SSH %(user)s@%(host)s password',
776
user=username, host=self._host)
778
transport.auth_password(username, password)
779
except paramiko.SSHException, e:
780
raise ConnectionError('Unable to authenticate to SSH host as %s@%s' %
781
(username, self._host), e)
783
def _try_pkey_auth(self, transport, pkey_class, username, filename):
784
filename = os.path.expanduser('~/.ssh/' + filename)
786
key = pkey_class.from_private_key_file(filename)
787
transport.auth_publickey(username, key)
789
except paramiko.PasswordRequiredException:
790
password = bzrlib.ui.ui_factory.get_password(
791
prompt='SSH %(filename)s password',
794
key = pkey_class.from_private_key_file(filename, password)
795
transport.auth_publickey(username, key)
797
except paramiko.SSHException:
798
mutter('SSH authentication via %s key failed.' % (os.path.basename(filename),))
799
except paramiko.SSHException:
800
mutter('SSH authentication via %s key failed.' % (os.path.basename(filename),))
805
845
def _sftp_open_exclusive(self, abspath, mode=None):
806
846
"""Open a remote path exclusively.
868
918
self._socket.bind(('localhost', 0))
869
919
self._socket.listen(1)
870
920
self.port = self._socket.getsockname()[1]
871
self.stop_event = threading.Event()
874
s, _ = self._socket.accept()
875
# now close the listen socket
878
self._callback(s, self.stop_event)
880
pass #Ignore socket errors
882
# probably a failed test
883
warning('Exception from within unit test server thread: %r' % x)
921
self._stop_event = threading.Event()
886
self.stop_event.set()
924
# called from outside this thread
925
self._stop_event.set()
887
926
# use a timeout here, because if the test fails, the server thread may
888
927
# never notice the stop_event.
933
readable, writable_unused, exception_unused = \
934
select.select([self._socket], [], [], 0.1)
935
if self._stop_event.isSet():
937
if len(readable) == 0:
940
s, addr_unused = self._socket.accept()
941
# because the loopback socket is inline, and transports are
942
# never explicitly closed, best to launch a new thread.
943
threading.Thread(target=self._callback, args=(s,)).start()
944
except socket.error, x:
945
sys.excepthook(*sys.exc_info())
946
warning('Socket error during accept() within unit test server'
949
# probably a failed test; unit test thread will log the
951
sys.excepthook(*sys.exc_info())
952
warning('Exception from within unit test server thread: %r' %
956
class SocketDelay(object):
957
"""A socket decorator to make TCP appear slower.
959
This changes recv, send, and sendall to add a fixed latency to each python
960
call if a new roundtrip is detected. That is, when a recv is called and the
961
flag new_roundtrip is set, latency is charged. Every send and send_all
964
In addition every send, sendall and recv sleeps a bit per character send to
967
Not all methods are implemented, this is deliberate as this class is not a
968
replacement for the builtin sockets layer. fileno is not implemented to
969
prevent the proxy being bypassed.
973
_proxied_arguments = dict.fromkeys([
974
"close", "getpeername", "getsockname", "getsockopt", "gettimeout",
975
"setblocking", "setsockopt", "settimeout", "shutdown"])
977
def __init__(self, sock, latency, bandwidth=1.0,
980
:param bandwith: simulated bandwith (MegaBit)
981
:param really_sleep: If set to false, the SocketDelay will just
982
increase a counter, instead of calling time.sleep. This is useful for
983
unittesting the SocketDelay.
986
self.latency = latency
987
self.really_sleep = really_sleep
988
self.time_per_byte = 1 / (bandwidth / 8.0 * 1024 * 1024)
989
self.new_roundtrip = False
992
if self.really_sleep:
995
SocketDelay.simulated_time += s
997
def __getattr__(self, attr):
998
if attr in SocketDelay._proxied_arguments:
999
return getattr(self.sock, attr)
1000
raise AttributeError("'SocketDelay' object has no attribute %r" %
1004
return SocketDelay(self.sock.dup(), self.latency, self.time_per_byte,
1007
def recv(self, *args):
1008
data = self.sock.recv(*args)
1009
if data and self.new_roundtrip:
1010
self.new_roundtrip = False
1011
self.sleep(self.latency)
1012
self.sleep(len(data) * self.time_per_byte)
1015
def sendall(self, data, flags=0):
1016
if not self.new_roundtrip:
1017
self.new_roundtrip = True
1018
self.sleep(self.latency)
1019
self.sleep(len(data) * self.time_per_byte)
1020
return self.sock.sendall(data, flags)
1022
def send(self, data, flags=0):
1023
if not self.new_roundtrip:
1024
self.new_roundtrip = True
1025
self.sleep(self.latency)
1026
bytes_sent = self.sock.send(data, flags)
1027
self.sleep(bytes_sent * self.time_per_byte)
892
1031
class SFTPServer(Server):
893
1032
"""Common code for SFTP server facilities."""
1034
def __init__(self, server_interface=StubServer):
896
1035
self._original_vendor = None
897
1036
self._homedir = None
898
1037
self._server_homedir = None
899
1038
self._listener = None
900
1039
self._root = None
901
self._vendor = 'none'
1040
self._vendor = ssh.ParamikoVendor()
1041
self._server_interface = server_interface
902
1042
# sftp server logs
1044
self.add_latency = 0
905
1046
def _get_sftp_url(self, path):
906
1047
"""Calculate an sftp url to this server for path."""