13
14
# You should have received a copy of the GNU General Public License
14
15
# along with this program; if not, write to the Free Software
15
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
16
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
17
18
"""Implementation of Transport over SFTP, using paramiko."""
19
# TODO: Remove the transport-based lock_read and lock_write methods. They'll
20
# then raise TransportNotPossible, which will break remote access to any
21
# formats which rely on OS-level locks. That should be fine as those formats
22
# are pretty old, but these combinations may have to be removed from the test
23
# suite. Those formats all date back to 0.7; so we should be able to remove
24
# these methods when we officially drop support for those formats.
44
from bzrlib.errors import (FileExists,
45
NoSuchFile, PathNotChild,
36
from bzrlib.config import config_dir, ensure_config_dir_exists
37
from bzrlib.errors import (ConnectionError,
39
TransportNotPossible, NoSuchFile, PathNotChild,
49
43
ParamikoNotPresent,
51
45
from bzrlib.osutils import pathjoin, fancy_rename, getcwd
52
from bzrlib.symbol_versioning import (
55
from bzrlib.trace import mutter, warning
46
from bzrlib.trace import mutter, warning, error
56
47
from bzrlib.transport import (
48
register_urlparse_netloc_protocol,
65
# Disable one particular warning that comes from paramiko in Python2.5; if
66
# this is emitted at the wrong time it tends to cause spurious test failures
67
# or at least noise in the test case::
69
# [1770/7639 in 86s, 1 known failures, 50 skipped, 2 missing features]
70
# test_permissions.TestSftpPermissions.test_new_files
71
# /var/lib/python-support/python2.5/paramiko/message.py:226: DeprecationWarning: integer argument expected, got float
72
# self.packet.write(struct.pack('>I', n))
73
warnings.filterwarnings('ignore',
74
'integer argument expected, got float',
75
category=DeprecationWarning,
76
module='paramiko.message')
54
import bzrlib.urlutils as urlutils
85
63
CMD_HANDLE, CMD_OPEN)
86
64
from paramiko.sftp_attr import SFTPAttributes
87
65
from paramiko.sftp_file import SFTPFile
66
from paramiko.sftp_client import SFTPClient
69
register_urlparse_netloc_protocol('sftp')
73
# TODO: This should possibly ignore SIGHUP as well, but bzr currently
74
# doesn't handle it itself.
75
# <https://launchpad.net/products/bzr/+bug/41433/+index>
77
signal.signal(signal.SIGINT, signal.SIG_IGN)
80
def os_specific_subprocess_params():
81
"""Get O/S specific subprocess parameters."""
82
if sys.platform == 'win32':
83
# setting the process group and closing fds is not supported on
87
# We close fds other than the pipes as the child process does not need
90
# We also set the child process to ignore SIGINT. Normally the signal
91
# would be sent to every process in the foreground process group, but
92
# this causes it to be seen only by bzr and not by ssh. Python will
93
# generate a KeyboardInterrupt in bzr, and we will then have a chance
94
# to release locks or do other cleanup over ssh before the connection
96
# <https://launchpad.net/products/bzr/+bug/5987>
98
# Running it in a separate process group is not good because then it
99
# can't get non-echoed input of a password or passphrase.
100
# <https://launchpad.net/products/bzr/+bug/40508>
101
return {'preexec_fn': _ignore_sigint,
90
106
_paramiko_version = getattr(paramiko, '__version_info__', (0, 0, 0))
91
107
# don't use prefetch unless paramiko version >= 1.5.5 (there were bugs earlier)
92
108
_default_do_prefetch = (_paramiko_version >= (1, 5, 5))
110
# Paramiko 1.5 tries to open a socket.AF_UNIX in order to connect
111
# to ssh-agent. That attribute doesn't exist on win32 (it does in cygwin)
112
# so we get an AttributeError exception. So we will not try to
113
# connect to an agent if we are on win32 and using Paramiko older than 1.6
114
_use_ssh_agent = (sys.platform != 'win32' or _paramiko_version >= (1, 6, 0))
118
def _get_ssh_vendor():
119
"""Find out what version of SSH is on the system."""
121
if _ssh_vendor is not None:
126
if 'BZR_SSH' in os.environ:
127
_ssh_vendor = os.environ['BZR_SSH']
128
if _ssh_vendor == 'paramiko':
133
p = subprocess.Popen(['ssh', '-V'],
134
stdin=subprocess.PIPE,
135
stdout=subprocess.PIPE,
136
stderr=subprocess.PIPE,
137
**os_specific_subprocess_params())
138
returncode = p.returncode
139
stdout, stderr = p.communicate()
143
if 'OpenSSH' in stderr:
144
mutter('ssh implementation is OpenSSH')
145
_ssh_vendor = 'openssh'
146
elif 'SSH Secure Shell' in stderr:
147
mutter('ssh implementation is SSH Corp.')
150
if _ssh_vendor != 'none':
153
# XXX: 20051123 jamesh
154
# A check for putty's plink or lsh would go here.
156
mutter('falling back to paramiko implementation')
160
class SFTPSubprocess:
161
"""A socket-like object that talks to an ssh subprocess via pipes."""
162
def __init__(self, hostname, vendor, port=None, user=None):
163
assert vendor in ['openssh', 'ssh']
164
if vendor == 'openssh':
166
'-oForwardX11=no', '-oForwardAgent=no',
167
'-oClearAllForwardings=yes', '-oProtocol=2',
168
'-oNoHostAuthenticationForLocalhost=yes']
170
args.extend(['-p', str(port)])
172
args.extend(['-l', user])
173
args.extend(['-s', hostname, 'sftp'])
174
elif vendor == 'ssh':
177
args.extend(['-p', str(port)])
179
args.extend(['-l', user])
180
args.extend(['-s', 'sftp', hostname])
182
self.proc = subprocess.Popen(args,
183
stdin=subprocess.PIPE,
184
stdout=subprocess.PIPE,
185
**os_specific_subprocess_params())
187
def send(self, data):
188
return os.write(self.proc.stdin.fileno(), data)
190
def recv_ready(self):
191
# TODO: jam 20051215 this function is necessary to support the
192
# pipelined() function. In reality, it probably should use
193
# poll() or select() to actually return if there is data
194
# available, otherwise we probably don't get any benefit
197
def recv(self, count):
198
return os.read(self.proc.stdout.fileno(), count)
201
self.proc.stdin.close()
202
self.proc.stdout.close()
206
class LoopbackSFTP(object):
207
"""Simple wrapper for a socket that pretends to be a paramiko Channel."""
209
def __init__(self, sock):
212
def send(self, data):
213
return self.__socket.send(data)
216
return self.__socket.recv(n)
218
def recv_ready(self):
222
self.__socket.close()
228
# This is a weakref dictionary, so that we can reuse connections
229
# that are still active. Long term, it might be nice to have some
230
# sort of expiration policy, such as disconnect if inactive for
231
# X seconds. But that requires a lot more fanciness.
232
_connected_hosts = weakref.WeakValueDictionary()
234
def clear_connection_cache():
235
"""Remove all hosts from the SFTP connection cache.
237
Primarily useful for test cases wanting to force garbage collection.
239
_connected_hosts.clear()
242
def load_host_keys():
244
Load system host keys (probably doesn't work on windows) and any
245
"discovered" keys from previous sessions.
247
global SYSTEM_HOSTKEYS, BZR_HOSTKEYS
249
SYSTEM_HOSTKEYS = paramiko.util.load_host_keys(os.path.expanduser('~/.ssh/known_hosts'))
251
mutter('failed to load system host keys: ' + str(e))
252
bzr_hostkey_path = pathjoin(config_dir(), 'ssh_host_keys')
254
BZR_HOSTKEYS = paramiko.util.load_host_keys(bzr_hostkey_path)
256
mutter('failed to load bzr host keys: ' + str(e))
260
def save_host_keys():
262
Save "discovered" host keys in $(config)/ssh_host_keys/.
264
global SYSTEM_HOSTKEYS, BZR_HOSTKEYS
265
bzr_hostkey_path = pathjoin(config_dir(), 'ssh_host_keys')
266
ensure_config_dir_exists()
269
f = open(bzr_hostkey_path, 'w')
270
f.write('# SSH host keys collected by bzr\n')
271
for hostname, keys in BZR_HOSTKEYS.iteritems():
272
for keytype, key in keys.iteritems():
273
f.write('%s %s %s\n' % (hostname, keytype, key.get_base64()))
276
mutter('failed to save bzr host keys: ' + str(e))
95
279
class SFTPLock(object):
96
"""This fakes a lock in a remote location.
98
A present lock is indicated just by the existence of a file. This
99
doesn't work well on all transports and they are only used in
100
deprecated storage formats.
280
"""This fakes a lock in a remote location."""
103
281
__slots__ = ['path', 'lock_path', 'lock_file', 'transport']
105
282
def __init__(self, path, transport):
283
assert isinstance(transport, SFTPTransport)
106
285
self.lock_file = None
108
287
self.lock_path = path + '.write-lock'
135
class _SFTPReadvHelper(object):
136
"""A class to help with managing the state of a readv request."""
138
# See _get_requests for an explanation.
139
_max_request_size = 32768
141
def __init__(self, original_offsets, relpath, _report_activity):
142
"""Create a new readv helper.
144
:param original_offsets: The original requests given by the caller of
146
:param relpath: The name of the file (if known)
147
:param _report_activity: A Transport._report_activity bound method,
148
to be called as data arrives.
150
self.original_offsets = list(original_offsets)
151
self.relpath = relpath
152
self._report_activity = _report_activity
154
def _get_requests(self):
155
"""Break up the offsets into individual requests over sftp.
157
The SFTP spec only requires implementers to support 32kB requests. We
158
could try something larger (openssh supports 64kB), but then we have to
159
handle requests that fail.
160
So instead, we just break up our maximum chunks into 32kB chunks, and
161
asyncronously requests them.
162
Newer versions of paramiko would do the chunking for us, but we want to
163
start processing results right away, so we do it ourselves.
165
# TODO: Because we issue async requests, we don't 'fudge' any extra
166
# data. I'm not 100% sure that is the best choice.
168
# The first thing we do, is to collapse the individual requests as much
169
# as possible, so we don't issues requests <32kB
170
sorted_offsets = sorted(self.original_offsets)
171
coalesced = list(ConnectedTransport._coalesce_offsets(sorted_offsets,
172
limit=0, fudge_factor=0))
174
for c_offset in coalesced:
175
start = c_offset.start
176
size = c_offset.length
178
# Break this up into 32kB requests
180
next_size = min(size, self._max_request_size)
181
requests.append((start, next_size))
184
if 'sftp' in debug.debug_flags:
185
mutter('SFTP.readv(%s) %s offsets => %s coalesced => %s requests',
186
self.relpath, len(sorted_offsets), len(coalesced),
190
def request_and_yield_offsets(self, fp):
191
"""Request the data from the remote machine, yielding the results.
193
:param fp: A Paramiko SFTPFile object that supports readv.
194
:return: Yield the data requested by the original readv caller, one by
197
requests = self._get_requests()
198
offset_iter = iter(self.original_offsets)
199
cur_offset, cur_size = offset_iter.next()
200
# paramiko .readv() yields strings that are in the order of the requests
201
# So we track the current request to know where the next data is
202
# being returned from.
208
# This is used to buffer chunks which we couldn't process yet
209
# It is (start, end, data) tuples.
211
# Create an 'unlimited' data stream, so we stop based on requests,
212
# rather than just because the data stream ended. This lets us detect
214
data_stream = itertools.chain(fp.readv(requests),
215
itertools.repeat(None))
216
for (start, length), data in itertools.izip(requests, data_stream):
218
if cur_coalesced is not None:
219
raise errors.ShortReadvError(self.relpath,
220
start, length, len(data))
221
if len(data) != length:
222
raise errors.ShortReadvError(self.relpath,
223
start, length, len(data))
224
self._report_activity(length, 'read')
226
# This is the first request, just buffer it
227
buffered_data = [data]
228
buffered_len = length
230
elif start == last_end:
231
# The data we are reading fits neatly on the previous
232
# buffer, so this is all part of a larger coalesced range.
233
buffered_data.append(data)
234
buffered_len += length
236
# We have an 'interrupt' in the data stream. So we know we are
237
# at a request boundary.
239
# We haven't consumed the buffer so far, so put it into
240
# data_chunks, and continue.
241
buffered = ''.join(buffered_data)
242
data_chunks.append((input_start, buffered))
244
buffered_data = [data]
245
buffered_len = length
246
last_end = start + length
247
if input_start == cur_offset and cur_size <= buffered_len:
248
# Simplify the next steps a bit by transforming buffered_data
249
# into a single string. We also have the nice property that
250
# when there is only one string ''.join([x]) == x, so there is
252
buffered = ''.join(buffered_data)
253
# Clean out buffered data so that we keep memory
257
# TODO: We *could* also consider the case where cur_offset is in
258
# in the buffered range, even though it doesn't *start*
259
# the buffered range. But for packs we pretty much always
260
# read in order, so you won't get any extra data in the
262
while (input_start == cur_offset
263
and (buffered_offset + cur_size) <= buffered_len):
264
# We've buffered enough data to process this request, spit it
266
cur_data = buffered[buffered_offset:buffered_offset + cur_size]
267
# move the direct pointer into our buffered data
268
buffered_offset += cur_size
269
# Move the start-of-buffer pointer
270
input_start += cur_size
271
# Yield the requested data
272
yield cur_offset, cur_data
273
cur_offset, cur_size = offset_iter.next()
274
# at this point, we've consumed as much of buffered as we can,
275
# so break off the portion that we consumed
276
if buffered_offset == len(buffered_data):
277
# No tail to leave behind
281
buffered = buffered[buffered_offset:]
282
buffered_data = [buffered]
283
buffered_len = len(buffered)
285
buffered = ''.join(buffered_data)
287
data_chunks.append((input_start, buffered))
289
if 'sftp' in debug.debug_flags:
290
mutter('SFTP readv left with %d out-of-order bytes',
291
sum(map(lambda x: len(x[1]), data_chunks)))
292
# We've processed all the readv data, at this point, anything we
293
# couldn't process is in data_chunks. This doesn't happen often, so
294
# this code path isn't optimized
295
# We use an interesting process for data_chunks
296
# Specifically if we have "bisect_left([(start, len, entries)],
298
# If start == qstart, then we get the specific node. Otherwise we
299
# get the previous node
301
idx = bisect.bisect_left(data_chunks, (cur_offset,))
302
if idx < len(data_chunks) and data_chunks[idx][0] == cur_offset:
303
# The data starts here
304
data = data_chunks[idx][1][:cur_size]
306
# The data is in a portion of a previous page
308
sub_offset = cur_offset - data_chunks[idx][0]
309
data = data_chunks[idx][1]
310
data = data[sub_offset:sub_offset + cur_size]
312
# We are missing the page where the data should be found,
315
if len(data) != cur_size:
316
raise AssertionError('We must have miscalulated.'
317
' We expected %d bytes, but only found %d'
318
% (cur_size, len(data)))
319
yield cur_offset, data
320
cur_offset, cur_size = offset_iter.next()
323
class SFTPTransport(ConnectedTransport):
314
class SFTPTransport(Transport):
324
315
"""Transport implementation for SFTP access."""
326
317
_do_prefetch = _default_do_prefetch
341
332
# up the request itself, rather than us having to worry about it
342
333
_max_request_size = 32768
344
def __init__(self, base, _from_transport=None):
345
super(SFTPTransport, self).__init__(base,
346
_from_transport=_from_transport)
335
def __init__(self, base, clone_from=None):
336
assert base.startswith('sftp://')
337
self._parse_url(base)
338
base = self._unparse_url()
341
super(SFTPTransport, self).__init__(base)
342
if clone_from is None:
345
# use the same ssh connection, etc
346
self._sftp = clone_from._sftp
347
# super saves 'self.base'
349
def should_cache(self):
351
Return True if the data pulled across should be cached locally.
355
def clone(self, offset=None):
357
Return a new SFTPTransport with root at self.base + offset.
358
We share the same SFTP session between such transports, because it's
359
fairly expensive to set them up.
362
return SFTPTransport(self.base, self)
364
return SFTPTransport(self.abspath(offset), self)
366
def abspath(self, relpath):
368
Return the full url to the given relative path.
370
@param relpath: the relative path or path components
371
@type relpath: str or list
373
return self._unparse_url(self._remote_path(relpath))
348
375
def _remote_path(self, relpath):
349
376
"""Return the path to be passed along the sftp protocol for relpath.
351
:param relpath: is a urlencoded string.
353
relative = urlutils.unescape(relpath).encode('utf-8')
354
remote_path = self._combine_paths(self._path, relative)
355
# the initial slash should be removed from the path, and treated as a
356
# homedir relative path (the path begins with a double slash if it is
357
# absolute). see draft-ietf-secsh-scp-sftp-ssh-uri-03.txt
358
# RBC 20060118 we are not using this as its too user hostile. instead
359
# we are following lftp and using /~/foo to mean '~/foo'
360
# vila--20070602 and leave absolute paths begin with a single slash.
361
if remote_path.startswith('/~/'):
362
remote_path = remote_path[3:]
363
elif remote_path == '/~':
367
def _create_connection(self, credentials=None):
368
"""Create a new connection with the provided credentials.
370
:param credentials: The credentials needed to establish the connection.
372
:return: The created connection and its associated credentials.
374
The credentials are only the password as it may have been entered
375
interactively by the user and may be different from the one provided
376
in base url at transport creation time.
378
if credentials is None:
379
password = self._password
381
password = credentials
383
vendor = ssh._get_ssh_vendor()
386
auth = config.AuthenticationConfig()
387
user = auth.get_user('ssh', self._host, self._port)
388
connection = vendor.connect_sftp(self._user, password,
389
self._host, self._port)
390
return connection, (user, password)
393
"""Ensures that a connection is established"""
394
connection = self._get_connection()
395
if connection is None:
396
# First connection ever
397
connection, credentials = self._create_connection()
398
self._set_connection(connection, credentials)
378
relpath is a urlencoded string.
380
# FIXME: share the common code across transports
381
assert isinstance(relpath, basestring)
382
relpath = urlutils.unescape(relpath).split('/')
383
basepath = self._path.split('/')
384
if len(basepath) > 0 and basepath[-1] == '':
385
basepath = basepath[:-1]
389
if len(basepath) == 0:
390
# In most filesystems, a request for the parent
391
# of root, just returns root.
399
path = '/'.join(basepath)
400
# mutter('relpath => remotepath %s => %s', relpath, path)
403
def relpath(self, abspath):
404
username, password, host, port, path = self._split_url(abspath)
406
if (username != self._username):
407
error.append('username mismatch')
408
if (host != self._host):
409
error.append('host mismatch')
410
if (port != self._port):
411
error.append('port mismatch')
412
if (not path.startswith(self._path)):
413
error.append('path mismatch')
415
extra = ': ' + ', '.join(error)
416
raise PathNotChild(abspath, self.base, extra=extra)
418
return path[pl:].strip('/')
401
420
def has(self, relpath):
403
422
Does the target location exist?
406
self._get_sftp().stat(self._remote_path(relpath))
407
# stat result is about 20 bytes, let's say
408
self._report_activity(20, 'read')
425
self._sftp.stat(self._remote_path(relpath))
413
430
def get(self, relpath):
414
"""Get the file at the given relative path.
432
Get the file at the given relative path.
416
434
:param relpath: The relative path to the file
419
# FIXME: by returning the file directly, we don't pass this
420
# through to report_activity. We could try wrapping the object
421
# before it's returned. For readv and get_bytes it's handled in
422
# the higher-level function.
424
437
path = self._remote_path(relpath)
425
f = self._get_sftp().file(path, mode='rb')
438
f = self._sftp.file(path, mode='rb')
426
439
if self._do_prefetch and (getattr(f, 'prefetch', None) is not None):
429
442
except (IOError, paramiko.SSHException), e:
430
self._translate_io_exception(e, path, ': error retrieving',
431
failure_exc=errors.ReadError)
433
def get_bytes(self, relpath):
434
# reimplement this here so that we can report how many bytes came back
435
f = self.get(relpath)
438
self._report_activity(len(bytes), 'read')
443
def _readv(self, relpath, offsets):
443
self._translate_io_exception(e, path, ': error retrieving')
445
def readv(self, relpath, offsets):
444
446
"""See Transport.readv()"""
445
447
# We overload the default readv() because we want to use a file
446
448
# that does not have prefetch enabled.
452
454
path = self._remote_path(relpath)
453
fp = self._get_sftp().file(path, mode='rb')
455
fp = self._sftp.file(path, mode='rb')
454
456
readv = getattr(fp, 'readv', None)
456
return self._sftp_readv(fp, offsets, relpath)
457
if 'sftp' in debug.debug_flags:
458
mutter('seek and read %s offsets', len(offsets))
459
return self._seek_and_read(fp, offsets, relpath)
458
return self._sftp_readv(fp, offsets)
459
mutter('seek and read %s offsets', len(offsets))
460
return self._seek_and_read(fp, offsets)
460
461
except (IOError, paramiko.SSHException), e:
461
462
self._translate_io_exception(e, path, ': error retrieving')
463
def recommended_page_size(self):
464
"""See Transport.recommended_page_size().
466
For SFTP we suggest a large page size to reduce the overhead
467
introduced by latency.
471
def _sftp_readv(self, fp, offsets, relpath):
464
def _sftp_readv(self, fp, offsets):
472
465
"""Use the readv() member of fp to do async readv.
474
Then read them using paramiko.readv(). paramiko.readv()
467
And then read them using paramiko.readv(). paramiko.readv()
475
468
does not support ranges > 64K, so it caps the request size, and
476
just reads until it gets all the stuff it wants.
478
helper = _SFTPReadvHelper(offsets, relpath, self._report_activity)
479
return helper.request_and_yield_offsets(fp)
481
def put_file(self, relpath, f, mode=None):
483
Copy the file-like object into the location.
469
just reads until it gets all the stuff it wants
471
offsets = list(offsets)
472
sorted_offsets = sorted(offsets)
474
# The algorithm works as follows:
475
# 1) Coalesce nearby reads into a single chunk
476
# This generates a list of combined regions, the total size
477
# and the size of the sub regions. This coalescing step is limited
478
# in the number of nearby chunks to combine, and is allowed to
479
# skip small breaks in the requests. Limiting it makes sure that
480
# we can start yielding some data earlier, and skipping means we
481
# make fewer requests. (Beneficial even when using async)
482
# 2) Break up this combined regions into chunks that are smaller
483
# than 64KiB. Technically the limit is 65536, but we are a
484
# little bit conservative. This is because sftp has a maximum
485
# return chunk size of 64KiB (max size of an unsigned short)
486
# 3) Issue a readv() to paramiko to create an async request for
488
# 4) Read in the data as it comes back, until we've read one
489
# continuous section as determined in step 1
490
# 5) Break up the full sections into hunks for the original requested
491
# offsets. And put them in a cache
492
# 6) Check if the next request is in the cache, and if it is, remove
493
# it from the cache, and yield its data. Continue until no more
494
# entries are in the cache.
495
# 7) loop back to step 4 until all data has been read
497
# TODO: jam 20060725 This could be optimized one step further, by
498
# attempting to yield whatever data we have read, even before
499
# the first coallesced section has been fully processed.
501
# When coalescing for use with readv(), we don't really need to
502
# use any fudge factor, because the requests are made asynchronously
503
coalesced = list(self._coalesce_offsets(sorted_offsets,
504
limit=self._max_readv_combine,
508
for c_offset in coalesced:
509
start = c_offset.start
510
size = c_offset.length
512
# We need to break this up into multiple requests
514
next_size = min(size, self._max_request_size)
515
requests.append((start, next_size))
519
mutter('SFTP.readv() %s offsets => %s coalesced => %s requests',
520
len(offsets), len(coalesced), len(requests))
522
# Queue the current read until we have read the full coalesced section
525
cur_coalesced_stack = iter(coalesced)
526
cur_coalesced = cur_coalesced_stack.next()
528
# Cache the results, but only until they have been fulfilled
530
# turn the list of offsets into a stack
531
offset_stack = iter(offsets)
532
cur_offset_and_size = offset_stack.next()
534
for data in fp.readv(requests):
536
cur_data_len += len(data)
538
if cur_data_len < cur_coalesced.length:
540
assert cur_data_len == cur_coalesced.length, \
541
"Somehow we read too much: %s != %s" % (cur_data_len,
542
cur_coalesced.length)
543
all_data = ''.join(cur_data)
547
for suboffset, subsize in cur_coalesced.ranges:
548
key = (cur_coalesced.start+suboffset, subsize)
549
data_map[key] = all_data[suboffset:suboffset+subsize]
551
# Now that we've read some data, see if we can yield anything back
552
while cur_offset_and_size in data_map:
553
this_data = data_map.pop(cur_offset_and_size)
554
yield cur_offset_and_size[0], this_data
555
cur_offset_and_size = offset_stack.next()
557
# Now that we've read all of the data for this coalesced section
559
cur_coalesced = cur_coalesced_stack.next()
561
def put(self, relpath, f, mode=None):
563
Copy the file-like or string object into the location.
485
565
:param relpath: Location to put the contents, relative to base.
486
:param f: File-like object.
566
:param f: File-like or string object.
487
567
:param mode: The final mode for the file
489
569
final_path = self._remote_path(relpath)
490
return self._put(final_path, f, mode=mode)
570
self._put(final_path, f, mode=mode)
492
572
def _put(self, abspath, f, mode=None):
493
573
"""Helper function so both put() and copy_abspaths can reuse the code"""
534
self._get_sftp().remove(tmp_abspath)
600
self._sftp.remove(tmp_abspath)
536
602
# raise the saved except
538
604
# raise the original with its traceback if we can.
541
def _put_non_atomic_helper(self, relpath, writer, mode=None,
542
create_parent_dir=False,
544
abspath = self._remote_path(relpath)
546
# TODO: jam 20060816 paramiko doesn't publicly expose a way to
547
# set the file mode at create time. If it does, use it.
548
# But for now, we just chmod later anyway.
550
def _open_and_write_file():
551
"""Try to open the target file, raise error on failure"""
555
fout = self._get_sftp().file(abspath, mode='wb')
556
fout.set_pipelined(True)
558
except (paramiko.SSHException, IOError), e:
559
self._translate_io_exception(e, abspath,
562
# This is designed to chmod() right before we close.
563
# Because we set_pipelined() earlier, theoretically we might
564
# avoid the round trip for fout.close()
566
self._get_sftp().chmod(abspath, mode)
571
if not create_parent_dir:
572
_open_and_write_file()
575
# Try error handling to create the parent directory if we need to
577
_open_and_write_file()
579
# Try to create the parent directory, and then go back to
581
parent_dir = os.path.dirname(abspath)
582
self._mkdir(parent_dir, dir_mode)
583
_open_and_write_file()
585
def put_file_non_atomic(self, relpath, f, mode=None,
586
create_parent_dir=False,
588
"""Copy the file-like object into the target location.
590
This function is not strictly safe to use. It is only meant to
591
be used when you already know that the target does not exist.
592
It is not safe, because it will open and truncate the remote
593
file. So there may be a time when the file has invalid contents.
595
:param relpath: The remote location to put the contents.
596
:param f: File-like object.
597
:param mode: Possible access permissions for new file.
598
None means do not set remote permissions.
599
:param create_parent_dir: If we cannot create the target file because
600
the parent directory does not exist, go ahead and
601
create it, and then try again.
605
self._put_non_atomic_helper(relpath, writer, mode=mode,
606
create_parent_dir=create_parent_dir,
609
def put_bytes_non_atomic(self, relpath, bytes, mode=None,
610
create_parent_dir=False,
614
self._put_non_atomic_helper(relpath, writer, mode=mode,
615
create_parent_dir=create_parent_dir,
618
607
def iter_files_recursive(self):
619
608
"""Walk the relative paths of all files in this transport."""
620
# progress is handled by list_dir
621
609
queue = list(self.list_dir('.'))
623
611
relpath = queue.pop(0)
631
def _mkdir(self, abspath, mode=None):
637
self._report_activity(len(abspath), 'write')
638
self._get_sftp().mkdir(abspath, local_mode)
639
self._report_activity(1, 'read')
641
# chmod a dir through sftp will erase any sgid bit set
642
# on the server side. So, if the bit mode are already
643
# set, avoid the chmod. If the mode is not fine but
644
# the sgid bit is set, report a warning to the user
645
# with the umask fix.
646
stat = self._get_sftp().lstat(abspath)
647
mode = mode & 0777 # can't set special bits anyway
648
if mode != stat.st_mode & 0777:
649
if stat.st_mode & 06000:
650
warning('About to chmod %s over sftp, which will result'
651
' in its suid or sgid bits being cleared. If'
652
' you want to preserve those bits, change your '
653
' environment on the server to use umask 0%03o.'
654
% (abspath, 0777 - mode))
655
self._get_sftp().chmod(abspath, mode=mode)
656
except (paramiko.SSHException, IOError), e:
657
self._translate_io_exception(e, abspath, ': unable to mkdir',
658
failure_exc=FileExists)
660
619
def mkdir(self, relpath, mode=None):
661
620
"""Create a directory at the given path."""
662
self._mkdir(self._remote_path(relpath), mode=mode)
664
def open_write_stream(self, relpath, mode=None):
665
"""See Transport.open_write_stream."""
666
# initialise the file to zero-length
667
# this is three round trips, but we don't use this
668
# api more than once per write_group at the moment so
669
# it is a tolerable overhead. Better would be to truncate
670
# the file after opening. RBC 20070805
671
self.put_bytes_non_atomic(relpath, "", mode)
672
abspath = self._remote_path(relpath)
673
# TODO: jam 20060816 paramiko doesn't publicly expose a way to
674
# set the file mode at create time. If it does, use it.
675
# But for now, we just chmod later anyway.
621
path = self._remote_path(relpath)
678
handle = self._get_sftp().file(abspath, mode='wb')
679
handle.set_pipelined(True)
623
# In the paramiko documentation, it says that passing a mode flag
624
# will filtered against the server umask.
625
# StubSFTPServer does not do this, which would be nice, because it is
626
# what we really want :)
627
# However, real servers do use umask, so we really should do it that way
628
self._sftp.mkdir(path)
630
self._sftp.chmod(path, mode=mode)
680
631
except (paramiko.SSHException, IOError), e:
681
self._translate_io_exception(e, abspath,
683
_file_streams[self.abspath(relpath)] = handle
684
return FileFileStream(self, relpath, handle)
632
self._translate_io_exception(e, path, ': unable to mkdir',
633
failure_exc=FileExists)
686
def _translate_io_exception(self, e, path, more_info='',
635
def _translate_io_exception(self, e, path, more_info='',
687
636
failure_exc=PathError):
688
637
"""Translate a paramiko or IOError into a friendlier exception.
694
643
:param failure_exc: Paramiko has the super fun ability to raise completely
695
644
opaque errors that just set "e.args = ('Failure',)" with
696
645
no more information.
697
If this parameter is set, it defines the exception
646
If this parameter is set, it defines the exception
698
647
to raise in these cases.
700
649
# paramiko seems to generate detailless errors.
701
650
self._translate_error(e, path, raise_generic=False)
702
if getattr(e, 'args', None) is not None:
651
if hasattr(e, 'args'):
703
652
if (e.args == ('No such file or directory',) or
704
653
e.args == ('No such file',)):
705
654
raise NoSuchFile(path, str(e) + more_info)
706
if (e.args == ('mkdir failed',) or
707
e.args[0].startswith('syserr: File exists')):
655
if (e.args == ('mkdir failed',)):
708
656
raise FileExists(path, str(e) + more_info)
709
657
# strange but true, for the paramiko server.
710
658
if (e.args == ('Failure',)):
711
659
raise failure_exc(path, str(e) + more_info)
712
# Can be something like args = ('Directory not empty:
713
# '/srv/bazaar.launchpad.net/blah...: '
714
# [Errno 39] Directory not empty',)
715
if (e.args[0].startswith('Directory not empty: ')
716
or getattr(e, 'errno', None) == errno.ENOTEMPTY):
717
raise errors.DirectoryNotEmpty(path, str(e))
718
660
mutter('Raising exception with args %s', e.args)
719
if getattr(e, 'errno', None) is not None:
661
if hasattr(e, 'errno'):
720
662
mutter('Raising exception with errno %s', e.errno)
723
def append_file(self, relpath, f, mode=None):
665
def append(self, relpath, f, mode=None):
725
667
Append the text in the file-like object into the final
729
671
path = self._remote_path(relpath)
730
fout = self._get_sftp().file(path, 'ab')
672
fout = self._sftp.file(path, 'ab')
731
673
if mode is not None:
732
self._get_sftp().chmod(path, mode)
674
self._sftp.chmod(path, mode)
733
675
result = fout.tell()
734
676
self._pump(f, fout)
840
774
# that we have taken the lock.
841
775
return SFTPLock(relpath, self)
777
def _unparse_url(self, path=None):
780
path = urllib.quote(path)
781
# handle homedir paths
782
if not path.startswith('/'):
784
netloc = urllib.quote(self._host)
785
if self._username is not None:
786
netloc = '%s@%s' % (urllib.quote(self._username), netloc)
787
if self._port is not None:
788
netloc = '%s:%d' % (netloc, self._port)
789
return urlparse.urlunparse(('sftp', netloc, path, '', '', ''))
791
def _split_url(self, url):
792
(scheme, username, password, host, port, path) = split_url(url)
793
assert scheme == 'sftp'
795
# the initial slash should be removed from the path, and treated
796
# as a homedir relative path (the path begins with a double slash
797
# if it is absolute).
798
# see draft-ietf-secsh-scp-sftp-ssh-uri-03.txt
799
# RBC 20060118 we are not using this as its too user hostile. instead
800
# we are following lftp and using /~/foo to mean '~/foo'.
801
# handle homedir paths
802
if path.startswith('/~/'):
806
return (username, password, host, port, path)
808
def _parse_url(self, url):
809
(self._username, self._password,
810
self._host, self._port, self._path) = self._split_url(url)
812
def _sftp_connect(self):
813
"""Connect to the remote sftp server.
814
After this, self._sftp should have a valid connection (or
815
we raise an TransportError 'could not connect').
817
TODO: Raise a more reasonable ConnectionFailed exception
819
global _connected_hosts
821
idx = (self._host, self._port, self._username)
823
self._sftp = _connected_hosts[idx]
828
vendor = _get_ssh_vendor()
829
if vendor == 'loopback':
830
sock = socket.socket()
832
sock.connect((self._host, self._port))
833
except socket.error, e:
834
raise ConnectionError('Unable to connect to SSH host %s:%s: %s'
835
% (self._host, self._port, e))
836
self._sftp = SFTPClient(LoopbackSFTP(sock))
837
elif vendor != 'none':
839
sock = SFTPSubprocess(self._host, vendor, self._port,
841
self._sftp = SFTPClient(sock)
842
except (EOFError, paramiko.SSHException), e:
843
raise ConnectionError('Unable to connect to SSH host %s:%s: %s'
844
% (self._host, self._port, e))
845
except (OSError, IOError), e:
846
# If the machine is fast enough, ssh can actually exit
847
# before we try and send it the sftp request, which
848
# raises a Broken Pipe
849
if e.errno not in (errno.EPIPE,):
851
raise ConnectionError('Unable to connect to SSH host %s:%s: %s'
852
% (self._host, self._port, e))
854
self._paramiko_connect()
856
_connected_hosts[idx] = self._sftp
858
def _paramiko_connect(self):
859
global SYSTEM_HOSTKEYS, BZR_HOSTKEYS
864
t = paramiko.Transport((self._host, self._port or 22))
865
t.set_log_channel('bzr.paramiko')
867
except (paramiko.SSHException, socket.error), e:
868
raise ConnectionError('Unable to reach SSH host %s:%s: %s'
869
% (self._host, self._port, e))
871
server_key = t.get_remote_server_key()
872
server_key_hex = paramiko.util.hexify(server_key.get_fingerprint())
873
keytype = server_key.get_name()
874
if SYSTEM_HOSTKEYS.has_key(self._host) and SYSTEM_HOSTKEYS[self._host].has_key(keytype):
875
our_server_key = SYSTEM_HOSTKEYS[self._host][keytype]
876
our_server_key_hex = paramiko.util.hexify(our_server_key.get_fingerprint())
877
elif BZR_HOSTKEYS.has_key(self._host) and BZR_HOSTKEYS[self._host].has_key(keytype):
878
our_server_key = BZR_HOSTKEYS[self._host][keytype]
879
our_server_key_hex = paramiko.util.hexify(our_server_key.get_fingerprint())
881
warning('Adding %s host key for %s: %s' % (keytype, self._host, server_key_hex))
882
if not BZR_HOSTKEYS.has_key(self._host):
883
BZR_HOSTKEYS[self._host] = {}
884
BZR_HOSTKEYS[self._host][keytype] = server_key
885
our_server_key = server_key
886
our_server_key_hex = paramiko.util.hexify(our_server_key.get_fingerprint())
888
if server_key != our_server_key:
889
filename1 = os.path.expanduser('~/.ssh/known_hosts')
890
filename2 = pathjoin(config_dir(), 'ssh_host_keys')
891
raise TransportError('Host keys for %s do not match! %s != %s' % \
892
(self._host, our_server_key_hex, server_key_hex),
893
['Try editing %s or %s' % (filename1, filename2)])
898
self._sftp = t.open_sftp_client()
899
except paramiko.SSHException, e:
900
raise ConnectionError('Unable to start sftp client %s:%d' %
901
(self._host, self._port), e)
903
def _sftp_auth(self, transport):
904
# paramiko requires a username, but it might be none if nothing was supplied
905
# use the local username, just in case.
906
# We don't override self._username, because if we aren't using paramiko,
907
# the username might be specified in ~/.ssh/config and we don't want to
908
# force it to something else
909
# Also, it would mess up the self.relpath() functionality
910
username = self._username or getpass.getuser()
913
agent = paramiko.Agent()
914
for key in agent.get_keys():
915
mutter('Trying SSH agent key %s' % paramiko.util.hexify(key.get_fingerprint()))
917
transport.auth_publickey(username, key)
919
except paramiko.SSHException, e:
922
# okay, try finding id_rsa or id_dss? (posix only)
923
if self._try_pkey_auth(transport, paramiko.RSAKey, username, 'id_rsa'):
925
if self._try_pkey_auth(transport, paramiko.DSSKey, username, 'id_dsa'):
930
transport.auth_password(username, self._password)
932
except paramiko.SSHException, e:
935
# FIXME: Don't keep a password held in memory if you can help it
936
#self._password = None
938
# give up and ask for a password
939
password = bzrlib.ui.ui_factory.get_password(
940
prompt='SSH %(user)s@%(host)s password',
941
user=username, host=self._host)
943
transport.auth_password(username, password)
944
except paramiko.SSHException, e:
945
raise ConnectionError('Unable to authenticate to SSH host as %s@%s' %
946
(username, self._host), e)
948
def _try_pkey_auth(self, transport, pkey_class, username, filename):
949
filename = os.path.expanduser('~/.ssh/' + filename)
951
key = pkey_class.from_private_key_file(filename)
952
transport.auth_publickey(username, key)
954
except paramiko.PasswordRequiredException:
955
password = bzrlib.ui.ui_factory.get_password(
956
prompt='SSH %(filename)s password',
959
key = pkey_class.from_private_key_file(filename, password)
960
transport.auth_publickey(username, key)
962
except paramiko.SSHException:
963
mutter('SSH authentication via %s key failed.' % (os.path.basename(filename),))
964
except paramiko.SSHException:
965
mutter('SSH authentication via %s key failed.' % (os.path.basename(filename),))
843
970
def _sftp_open_exclusive(self, abspath, mode=None):
844
971
"""Open a remote path exclusively.
854
981
:param abspath: The remote absolute path where the file should be opened
855
982
:param mode: The mode permissions bits for the new file
857
# TODO: jam 20060816 Paramiko >= 1.6.2 (probably earlier) supports
858
# using the 'x' flag to indicate SFTP_FLAG_EXCL.
859
# However, there is no way to set the permission mode at open
860
# time using the sftp_client.file() functionality.
861
path = self._get_sftp()._adjust_cwd(abspath)
984
path = self._sftp._adjust_cwd(abspath)
862
985
# mutter('sftp abspath %s => %s', abspath, path)
863
986
attr = SFTPAttributes()
864
987
if mode is not None:
865
988
attr.st_mode = mode
866
omode = (SFTP_FLAG_WRITE | SFTP_FLAG_CREATE
989
omode = (SFTP_FLAG_WRITE | SFTP_FLAG_CREATE
867
990
| SFTP_FLAG_TRUNC | SFTP_FLAG_EXCL)
869
t, msg = self._get_sftp()._request(CMD_OPEN, path, omode, attr)
992
t, msg = self._sftp._request(CMD_OPEN, path, omode, attr)
870
993
if t != CMD_HANDLE:
871
994
raise TransportError('Expected an SFTP handle')
872
995
handle = msg.get_string()
873
return SFTPFile(self._get_sftp(), handle, 'wb', -1)
996
return SFTPFile(self._sftp, handle, 'wb', -1)
874
997
except (paramiko.SSHException, IOError), e:
875
998
self._translate_io_exception(e, abspath, ': unable to open',
876
999
failure_exc=FileExists)
878
def _can_roundtrip_unix_modebits(self):
879
if sys.platform == 'win32':
1002
# ------------- server test implementation --------------
1006
from bzrlib.tests.stub_sftp import StubServer, StubSFTPServer
1008
STUB_SERVER_KEY = """
1009
-----BEGIN RSA PRIVATE KEY-----
1010
MIICWgIBAAKBgQDTj1bqB4WmayWNPB+8jVSYpZYk80Ujvj680pOTh2bORBjbIAyz
1011
oWGW+GUjzKxTiiPvVmxFgx5wdsFvF03v34lEVVhMpouqPAYQ15N37K/ir5XY+9m/
1012
d8ufMCkjeXsQkKqFbAlQcnWMCRnOoPHS3I4vi6hmnDDeeYTSRvfLbW0fhwIBIwKB
1013
gBIiOqZYaoqbeD9OS9z2K9KR2atlTxGxOJPXiP4ESqP3NVScWNwyZ3NXHpyrJLa0
1014
EbVtzsQhLn6rF+TzXnOlcipFvjsem3iYzCpuChfGQ6SovTcOjHV9z+hnpXvQ/fon
1015
soVRZY65wKnF7IAoUwTmJS9opqgrN6kRgCd3DASAMd1bAkEA96SBVWFt/fJBNJ9H
1016
tYnBKZGw0VeHOYmVYbvMSstssn8un+pQpUm9vlG/bp7Oxd/m+b9KWEh2xPfv6zqU
1017
avNwHwJBANqzGZa/EpzF4J8pGti7oIAPUIDGMtfIcmqNXVMckrmzQ2vTfqtkEZsA
1018
4rE1IERRyiJQx6EJsz21wJmGV9WJQ5kCQQDwkS0uXqVdFzgHO6S++tjmjYcxwr3g
1019
H0CoFYSgbddOT6miqRskOQF3DZVkJT3kyuBgU2zKygz52ukQZMqxCb1fAkASvuTv
1020
qfpH87Qq5kQhNKdbbwbmd2NxlNabazPijWuphGTdW0VfJdWfklyS2Kr+iqrs/5wV
1021
HhathJt636Eg7oIjAkA8ht3MQ+XSl9yIJIS8gVpbPxSw5OMfw0PjVE7tBdQruiSc
1022
nvuQES5C9BMHjF39LZiGH1iLQy7FgdHyoP+eodI7
1023
-----END RSA PRIVATE KEY-----
1027
class SocketListener(threading.Thread):
1029
def __init__(self, callback):
1030
threading.Thread.__init__(self)
1031
self._callback = callback
1032
self._socket = socket.socket()
1033
self._socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
1034
self._socket.bind(('localhost', 0))
1035
self._socket.listen(1)
1036
self.port = self._socket.getsockname()[1]
1037
self._stop_event = threading.Event()
1040
# called from outside this thread
1041
self._stop_event.set()
1042
# use a timeout here, because if the test fails, the server thread may
1043
# never notice the stop_event.
1045
self._socket.close()
1049
readable, writable_unused, exception_unused = \
1050
select.select([self._socket], [], [], 0.1)
1051
if self._stop_event.isSet():
1053
if len(readable) == 0:
1056
s, addr_unused = self._socket.accept()
1057
# because the loopback socket is inline, and transports are
1058
# never explicitly closed, best to launch a new thread.
1059
threading.Thread(target=self._callback, args=(s,)).start()
1060
except socket.error, x:
1061
sys.excepthook(*sys.exc_info())
1062
warning('Socket error during accept() within unit test server'
1064
except Exception, x:
1065
# probably a failed test; unit test thread will log the
1067
sys.excepthook(*sys.exc_info())
1068
warning('Exception from within unit test server thread: %r' %
1072
class SocketDelay(object):
1073
"""A socket decorator to make TCP appear slower.
1075
This changes recv, send, and sendall to add a fixed latency to each python
1076
call if a new roundtrip is detected. That is, when a recv is called and the
1077
flag new_roundtrip is set, latency is charged. Every send and send_all
1080
In addition every send, sendall and recv sleeps a bit per character send to
1083
Not all methods are implemented, this is deliberate as this class is not a
1084
replacement for the builtin sockets layer. fileno is not implemented to
1085
prevent the proxy being bypassed.
1089
_proxied_arguments = dict.fromkeys([
1090
"close", "getpeername", "getsockname", "getsockopt", "gettimeout",
1091
"setblocking", "setsockopt", "settimeout", "shutdown"])
1093
def __init__(self, sock, latency, bandwidth=1.0,
1096
:param bandwith: simulated bandwith (MegaBit)
1097
:param really_sleep: If set to false, the SocketDelay will just
1098
increase a counter, instead of calling time.sleep. This is useful for
1099
unittesting the SocketDelay.
1102
self.latency = latency
1103
self.really_sleep = really_sleep
1104
self.time_per_byte = 1 / (bandwidth / 8.0 * 1024 * 1024)
1105
self.new_roundtrip = False
1108
if self.really_sleep:
1111
SocketDelay.simulated_time += s
1113
def __getattr__(self, attr):
1114
if attr in SocketDelay._proxied_arguments:
1115
return getattr(self.sock, attr)
1116
raise AttributeError("'SocketDelay' object has no attribute %r" %
1120
return SocketDelay(self.sock.dup(), self.latency, self.time_per_byte,
1123
def recv(self, *args):
1124
data = self.sock.recv(*args)
1125
if data and self.new_roundtrip:
1126
self.new_roundtrip = False
1127
self.sleep(self.latency)
1128
self.sleep(len(data) * self.time_per_byte)
1131
def sendall(self, data, flags=0):
1132
if not self.new_roundtrip:
1133
self.new_roundtrip = True
1134
self.sleep(self.latency)
1135
self.sleep(len(data) * self.time_per_byte)
1136
return self.sock.sendall(data, flags)
1138
def send(self, data, flags=0):
1139
if not self.new_roundtrip:
1140
self.new_roundtrip = True
1141
self.sleep(self.latency)
1142
bytes_sent = self.sock.send(data, flags)
1143
self.sleep(bytes_sent * self.time_per_byte)
1147
class SFTPServer(Server):
1148
"""Common code for SFTP server facilities."""
1151
self._original_vendor = None
1152
self._homedir = None
1153
self._server_homedir = None
1154
self._listener = None
1156
self._vendor = 'none'
1159
self.add_latency = 0
1161
def _get_sftp_url(self, path):
1162
"""Calculate an sftp url to this server for path."""
1163
return 'sftp://foo:bar@localhost:%d/%s' % (self._listener.port, path)
1165
def log(self, message):
1166
"""StubServer uses this to log when a new server is created."""
1167
self.logs.append(message)
1169
def _run_server_entry(self, sock):
1170
"""Entry point for all implementations of _run_server.
1172
If self.add_latency is > 0.000001 then sock is given a latency adding
1175
if self.add_latency > 0.000001:
1176
sock = SocketDelay(sock, self.add_latency)
1177
return self._run_server(sock)
1179
def _run_server(self, s):
1180
ssh_server = paramiko.Transport(s)
1181
key_file = pathjoin(self._homedir, 'test_rsa.key')
1182
f = open(key_file, 'w')
1183
f.write(STUB_SERVER_KEY)
1185
host_key = paramiko.RSAKey.from_private_key_file(key_file)
1186
ssh_server.add_server_key(host_key)
1187
server = StubServer(self)
1188
ssh_server.set_subsystem_handler('sftp', paramiko.SFTPServer,
1189
StubSFTPServer, root=self._root,
1190
home=self._server_homedir)
1191
event = threading.Event()
1192
ssh_server.start_server(event, server)
1197
self._original_vendor = _ssh_vendor
1198
_ssh_vendor = self._vendor
1199
if sys.platform == 'win32':
1200
# Win32 needs to use the UNICODE api
1201
self._homedir = getcwd()
1203
# But Linux SFTP servers should just deal in bytestreams
1204
self._homedir = os.getcwd()
1205
if self._server_homedir is None:
1206
self._server_homedir = self._homedir
1208
if sys.platform == 'win32':
1210
self._listener = SocketListener(self._run_server_entry)
1211
self._listener.setDaemon(True)
1212
self._listener.start()
1215
"""See bzrlib.transport.Server.tearDown."""
1217
self._listener.stop()
1218
_ssh_vendor = self._original_vendor
1220
def get_bogus_url(self):
1221
"""See bzrlib.transport.Server.get_bogus_url."""
1222
# this is chosen to try to prevent trouble with proxies, wierd dns, etc
1223
# we bind a random socket, so that we get a guaranteed unused port
1224
# we just never listen on that port
1226
s.bind(('localhost', 0))
1227
return 'sftp://%s:%s/' % s.getsockname()
1230
class SFTPFullAbsoluteServer(SFTPServer):
1231
"""A test server for sftp transports, using absolute urls and ssh."""
1234
"""See bzrlib.transport.Server.get_url."""
1235
return self._get_sftp_url(urlutils.escape(self._homedir[1:]))
1238
class SFTPServerWithoutSSH(SFTPServer):
1239
"""An SFTP server that uses a simple TCP socket pair rather than SSH."""
1242
super(SFTPServerWithoutSSH, self).__init__()
1243
self._vendor = 'loopback'
1245
def _run_server(self, sock):
1246
class FakeChannel(object):
1247
def get_transport(self):
1249
def get_log_channel(self):
1253
def get_hexdump(self):
1258
server = paramiko.SFTPServer(FakeChannel(), 'sftp', StubServer(self), StubSFTPServer,
1259
root=self._root, home=self._server_homedir)
1261
server.start_subsystem('sftp', None, sock)
1262
except socket.error, e:
1263
if (len(e.args) > 0) and (e.args[0] == errno.EPIPE):
1264
# it's okay for the client to disconnect abruptly
1265
# (bug in paramiko 1.6: it should absorb this exception)
1269
except Exception, e:
1270
import sys; sys.stderr.write('\nEXCEPTION %r\n\n' % e.__class__)
1271
server.finish_subsystem()
1274
class SFTPAbsoluteServer(SFTPServerWithoutSSH):
1275
"""A test server for sftp transports, using absolute urls."""
1278
"""See bzrlib.transport.Server.get_url."""
1279
if sys.platform == 'win32':
1280
return self._get_sftp_url(urlutils.escape(self._homedir))
1282
return self._get_sftp_url(urlutils.escape(self._homedir[1:]))
1285
class SFTPHomeDirServer(SFTPServerWithoutSSH):
1286
"""A test server for sftp transports, using homedir relative urls."""
1289
"""See bzrlib.transport.Server.get_url."""
1290
return self._get_sftp_url("~/")
1293
class SFTPSiblingAbsoluteServer(SFTPAbsoluteServer):
1294
"""A test servere for sftp transports, using absolute urls to non-home."""
1297
self._server_homedir = '/dev/noone/runs/tests/here'
1298
super(SFTPSiblingAbsoluteServer, self).setUp()
886
1301
def get_test_permutations():
887
1302
"""Return the permutations to be used in testing."""
888
from bzrlib.tests import stub_sftp
889
return [(SFTPTransport, stub_sftp.SFTPAbsoluteServer),
890
(SFTPTransport, stub_sftp.SFTPHomeDirServer),
891
(SFTPTransport, stub_sftp.SFTPSiblingAbsoluteServer),
1303
return [(SFTPTransport, SFTPAbsoluteServer),
1304
(SFTPTransport, SFTPHomeDirServer),
1305
(SFTPTransport, SFTPSiblingAbsoluteServer),