2015-09-25 18:57:14 +00:00
|
|
|
""" Swift storage driver.
|
|
|
|
|
|
|
|
Uses: http://docs.openstack.org/developer/swift/overview_large_objects.html
|
|
|
|
"""
|
2016-01-14 20:30:01 +00:00
|
|
|
import os.path
|
|
|
|
import copy
|
|
|
|
import hmac
|
|
|
|
import string
|
|
|
|
import logging
|
|
|
|
from uuid import uuid4
|
|
|
|
|
2015-05-21 19:22:59 +00:00
|
|
|
from swiftclient.client import Connection, ClientException
|
2015-09-14 21:49:35 +00:00
|
|
|
from urlparse import urlparse
|
2015-05-21 19:22:59 +00:00
|
|
|
from random import SystemRandom
|
2015-09-14 21:49:35 +00:00
|
|
|
from hashlib import sha1
|
|
|
|
from time import time
|
2015-09-25 18:57:14 +00:00
|
|
|
from collections import namedtuple
|
2016-01-14 20:30:01 +00:00
|
|
|
|
2015-09-25 18:57:14 +00:00
|
|
|
from util.registry import filelike
|
2016-01-14 20:30:01 +00:00
|
|
|
from storage.basestorage import BaseStorage
|
|
|
|
from util.registry.generatorfile import GeneratorFile
|
2015-09-14 21:49:35 +00:00
|
|
|
|
2015-05-21 19:22:59 +00:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2015-09-25 18:57:14 +00:00
|
|
|
_PartUploadMetadata = namedtuple('_PartUploadMetadata', ['path', 'offset', 'length'])
|
|
|
|
_SEGMENTS_KEY = 'segments'
|
|
|
|
_SEGMENT_DIRECTORY = 'segments'
|
|
|
|
_MAXIMUM_SEGMENT_SIZE = 5000000000 # 5 GB
|
2015-07-02 14:52:43 +00:00
|
|
|
|
2015-05-21 19:22:59 +00:00
|
|
|
class SwiftStorage(BaseStorage):
|
2016-01-15 16:15:40 +00:00
|
|
|
def __init__(self, metric_queue, swift_container, storage_path, auth_url, swift_user,
|
2015-09-14 21:49:35 +00:00
|
|
|
swift_password, auth_version=None, os_options=None, ca_cert_path=None,
|
2016-01-14 20:30:01 +00:00
|
|
|
temp_url_key=None, simple_path_concat=False):
|
2015-08-26 21:08:42 +00:00
|
|
|
super(SwiftStorage, self).__init__()
|
2015-05-21 19:22:59 +00:00
|
|
|
self._swift_container = swift_container
|
2016-01-14 20:30:01 +00:00
|
|
|
|
|
|
|
self._storage_path = storage_path.lstrip('/')
|
|
|
|
self._simple_path_concat = simple_path_concat
|
2015-05-21 19:22:59 +00:00
|
|
|
|
|
|
|
self._auth_url = auth_url
|
|
|
|
self._ca_cert_path = ca_cert_path
|
|
|
|
|
|
|
|
self._swift_user = swift_user
|
|
|
|
self._swift_password = swift_password
|
|
|
|
|
2015-09-14 21:49:35 +00:00
|
|
|
self._temp_url_key = temp_url_key
|
|
|
|
|
2015-09-02 21:37:46 +00:00
|
|
|
try:
|
|
|
|
self._auth_version = int(auth_version or '2')
|
|
|
|
except ValueError:
|
|
|
|
self._auth_version = 2
|
|
|
|
|
2015-05-21 19:22:59 +00:00
|
|
|
self._os_options = os_options or {}
|
|
|
|
|
|
|
|
self._initialized = False
|
|
|
|
|
|
|
|
def _get_connection(self):
|
|
|
|
return Connection(
|
2016-01-14 20:30:01 +00:00
|
|
|
authurl=self._auth_url,
|
|
|
|
cacert=self._ca_cert_path,
|
2015-05-21 19:22:59 +00:00
|
|
|
|
2016-01-14 20:30:01 +00:00
|
|
|
user=self._swift_user,
|
|
|
|
key=self._swift_password,
|
2015-09-14 21:49:35 +00:00
|
|
|
|
2016-01-14 20:30:01 +00:00
|
|
|
auth_version=self._auth_version,
|
|
|
|
os_options=self._os_options
|
|
|
|
)
|
2015-05-21 19:22:59 +00:00
|
|
|
|
2016-01-14 20:30:01 +00:00
|
|
|
def _normalize_path(self, object_path):
|
|
|
|
""" No matter what inputs we get, we are going to return a path without a leading or trailing
|
|
|
|
'/'
|
|
|
|
"""
|
|
|
|
if self._simple_path_concat:
|
|
|
|
return (self._storage_path + object_path).rstrip('/')
|
|
|
|
else:
|
|
|
|
return os.path.join(self._storage_path, object_path).rstrip('/')
|
2015-05-21 19:22:59 +00:00
|
|
|
|
|
|
|
def _get_object(self, path, chunk_size=None):
|
|
|
|
path = self._normalize_path(path)
|
|
|
|
try:
|
2015-08-07 17:41:40 +00:00
|
|
|
_, obj = self._get_connection().get_object(self._swift_container, path,
|
2016-01-14 20:30:01 +00:00
|
|
|
resp_chunk_size=chunk_size)
|
2015-05-21 19:22:59 +00:00
|
|
|
return obj
|
|
|
|
except Exception:
|
|
|
|
logger.exception('Could not get object: %s', path)
|
|
|
|
raise IOError('Path %s not found' % path)
|
|
|
|
|
2015-09-25 18:57:14 +00:00
|
|
|
def _put_object(self, path, content, chunk=None, content_type=None, content_encoding=None,
|
|
|
|
headers=None):
|
2015-05-21 19:22:59 +00:00
|
|
|
path = self._normalize_path(path)
|
2015-09-25 18:57:14 +00:00
|
|
|
headers = headers or {}
|
2015-05-21 19:22:59 +00:00
|
|
|
|
|
|
|
if content_encoding is not None:
|
|
|
|
headers['Content-Encoding'] = content_encoding
|
|
|
|
|
|
|
|
try:
|
2015-08-07 17:41:40 +00:00
|
|
|
self._get_connection().put_object(self._swift_container, path, content,
|
2015-05-21 19:22:59 +00:00
|
|
|
chunk_size=chunk, content_type=content_type,
|
|
|
|
headers=headers)
|
|
|
|
except ClientException:
|
2015-05-26 20:35:12 +00:00
|
|
|
# We re-raise client exception here so that validation of config during setup can see
|
|
|
|
# the client exception messages.
|
2015-05-21 19:22:59 +00:00
|
|
|
raise
|
|
|
|
except Exception:
|
|
|
|
logger.exception('Could not put object: %s', path)
|
|
|
|
raise IOError("Could not put content: %s" % path)
|
|
|
|
|
|
|
|
def _head_object(self, path):
|
|
|
|
path = self._normalize_path(path)
|
|
|
|
try:
|
2015-08-07 17:41:40 +00:00
|
|
|
return self._get_connection().head_object(self._swift_container, path)
|
2015-05-21 19:22:59 +00:00
|
|
|
except Exception:
|
|
|
|
logger.exception('Could not head object: %s', path)
|
|
|
|
return None
|
|
|
|
|
2016-02-11 22:00:38 +00:00
|
|
|
def get_direct_download_url(self, object_path, expires_in=60, requires_cors=False, head=False):
|
2015-05-21 19:22:59 +00:00
|
|
|
if requires_cors:
|
|
|
|
return None
|
|
|
|
|
2015-09-14 21:49:35 +00:00
|
|
|
# Reference: http://docs.openstack.org/juno/config-reference/content/object-storage-tempurl.html
|
|
|
|
if not self._temp_url_key:
|
|
|
|
return None
|
|
|
|
|
|
|
|
# Retrieve the auth details for the connection.
|
|
|
|
try:
|
|
|
|
object_url_value, _ = self._get_connection().get_auth()
|
|
|
|
except ClientException:
|
|
|
|
logger.exception('Got client exception when trying to load Swift auth')
|
|
|
|
return None
|
|
|
|
|
|
|
|
object_url = urlparse(object_url_value)
|
|
|
|
scheme = object_url.scheme
|
2016-01-14 20:30:01 +00:00
|
|
|
path = object_url.path.rstrip('/')
|
2015-09-14 21:49:35 +00:00
|
|
|
hostname = object_url.netloc
|
|
|
|
|
|
|
|
object_path = self._normalize_path(object_path)
|
|
|
|
|
|
|
|
# Generate the signed HMAC body.
|
2016-02-11 22:00:38 +00:00
|
|
|
method = 'HEAD' if head else 'GET'
|
2015-09-14 21:49:35 +00:00
|
|
|
expires = int(time() + expires_in)
|
2016-01-14 20:30:01 +00:00
|
|
|
full_path = '%s/%s/%s' % (path, self._swift_container, object_path)
|
2015-09-14 21:49:35 +00:00
|
|
|
|
|
|
|
hmac_body = '%s\n%s\n%s' % (method, expires, full_path)
|
|
|
|
sig = hmac.new(self._temp_url_key.encode('utf-8'), hmac_body.encode('utf-8'), sha1).hexdigest()
|
|
|
|
|
|
|
|
surl = '{scheme}://{host}{full_path}?temp_url_sig={sig}&temp_url_expires={expires}'
|
|
|
|
return surl.format(scheme=scheme, host=hostname, full_path=full_path, sig=sig, expires=expires)
|
|
|
|
|
|
|
|
def validate(self, client):
|
|
|
|
if self._temp_url_key:
|
|
|
|
# Add a file to test direct download.
|
|
|
|
self.put_content('dd_path', 'testing 3456')
|
|
|
|
|
|
|
|
# Generate a direct download URL.
|
|
|
|
dd_url = self.get_direct_download_url('dd_path')
|
|
|
|
|
|
|
|
if not dd_url:
|
|
|
|
self.remove('dd_path')
|
|
|
|
raise Exception('Could not validate direct download URL; the token may be invalid.')
|
|
|
|
|
|
|
|
# Try to retrieve the direct download URL.
|
|
|
|
response = client.get(dd_url, timeout=2)
|
|
|
|
|
|
|
|
# Remove the test file.
|
|
|
|
self.remove('dd_path')
|
|
|
|
|
|
|
|
if response.status_code != 200:
|
|
|
|
logger.debug('Direct download failure: %s => %s with body %s', dd_url,
|
|
|
|
response.status_code, response.text)
|
|
|
|
|
|
|
|
msg = 'Direct download URL failed with status code %s. Please check your temp-url-key.'
|
|
|
|
raise Exception(msg % response.status_code)
|
2015-05-21 19:22:59 +00:00
|
|
|
|
|
|
|
def get_content(self, path):
|
|
|
|
return self._get_object(path)
|
|
|
|
|
|
|
|
def put_content(self, path, content):
|
|
|
|
self._put_object(path, content)
|
|
|
|
|
|
|
|
def stream_read(self, path):
|
|
|
|
for data in self._get_object(path, self.buffer_size):
|
|
|
|
yield data
|
|
|
|
|
|
|
|
def stream_read_file(self, path):
|
2015-07-02 14:52:43 +00:00
|
|
|
return GeneratorFile(self.stream_read(path))
|
2015-05-21 19:22:59 +00:00
|
|
|
|
|
|
|
def stream_write(self, path, fp, content_type=None, content_encoding=None):
|
|
|
|
self._put_object(path, fp, self.buffer_size, content_type=content_type,
|
|
|
|
content_encoding=content_encoding)
|
|
|
|
|
|
|
|
def exists(self, path):
|
|
|
|
return bool(self._head_object(path))
|
|
|
|
|
|
|
|
def remove(self, path):
|
|
|
|
path = self._normalize_path(path)
|
|
|
|
try:
|
2015-08-07 17:41:40 +00:00
|
|
|
self._get_connection().delete_object(self._swift_container, path)
|
2015-05-21 19:22:59 +00:00
|
|
|
except Exception:
|
|
|
|
raise IOError('Cannot delete path: %s' % path)
|
|
|
|
|
|
|
|
def _random_checksum(self, count):
|
|
|
|
chars = string.ascii_uppercase + string.digits
|
|
|
|
return ''.join(SystemRandom().choice(chars) for _ in range(count))
|
|
|
|
|
|
|
|
def get_checksum(self, path):
|
|
|
|
headers = self._head_object(path)
|
|
|
|
if not headers:
|
|
|
|
raise IOError('Cannot lookup path: %s' % path)
|
|
|
|
|
|
|
|
return headers.get('etag', '')[1:-1][:7] or self._random_checksum(7)
|
2015-09-25 18:57:14 +00:00
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _segment_list_from_metadata(storage_metadata):
|
|
|
|
return [_PartUploadMetadata(*segment_args) for segment_args in storage_metadata[_SEGMENTS_KEY]]
|
|
|
|
|
|
|
|
def initiate_chunked_upload(self):
|
|
|
|
random_uuid = str(uuid4())
|
|
|
|
|
|
|
|
metadata = {
|
|
|
|
_SEGMENTS_KEY: [],
|
|
|
|
}
|
|
|
|
|
|
|
|
return random_uuid, metadata
|
|
|
|
|
2015-11-30 20:45:45 +00:00
|
|
|
def stream_upload_chunk(self, uuid, offset, length, in_fp, storage_metadata, content_type=None):
|
2015-12-10 04:16:33 +00:00
|
|
|
error = None
|
|
|
|
|
2015-09-25 18:57:14 +00:00
|
|
|
if length == 0:
|
2015-12-10 04:16:33 +00:00
|
|
|
return 0, storage_metadata, error
|
2015-09-25 18:57:14 +00:00
|
|
|
|
|
|
|
# Note: Swift limits segments to a maximum of 5GB, so we keep writing segments until we
|
|
|
|
# are finished hitting the data limit.
|
|
|
|
total_bytes_written = 0
|
|
|
|
while True:
|
2015-12-10 04:16:33 +00:00
|
|
|
try:
|
|
|
|
bytes_written, storage_metadata = self._stream_upload_segment(uuid, offset, length, in_fp,
|
2016-01-14 20:30:01 +00:00
|
|
|
storage_metadata,
|
|
|
|
content_type)
|
2015-12-10 04:16:33 +00:00
|
|
|
except IOError as ex:
|
|
|
|
logger.warn('stream write error: %s', ex)
|
|
|
|
error = ex
|
|
|
|
break
|
2015-09-25 18:57:14 +00:00
|
|
|
|
|
|
|
if length != filelike.READ_UNTIL_END:
|
|
|
|
length = length - bytes_written
|
|
|
|
|
|
|
|
offset = offset + bytes_written
|
|
|
|
total_bytes_written = total_bytes_written + bytes_written
|
|
|
|
if bytes_written == 0 or length <= 0:
|
2015-12-10 04:16:33 +00:00
|
|
|
return total_bytes_written, storage_metadata, error
|
|
|
|
|
|
|
|
return total_bytes_written, storage_metadata, error
|
2015-09-25 18:57:14 +00:00
|
|
|
|
2015-11-30 20:45:45 +00:00
|
|
|
def _stream_upload_segment(self, uuid, offset, length, in_fp, storage_metadata, content_type):
|
2015-09-25 18:57:14 +00:00
|
|
|
updated_metadata = copy.deepcopy(storage_metadata)
|
|
|
|
segment_count = len(updated_metadata[_SEGMENTS_KEY])
|
|
|
|
segment_path = '%s/%s/%s' % (_SEGMENT_DIRECTORY, uuid, segment_count)
|
|
|
|
|
|
|
|
# Track the number of bytes read and if an explicit length is specified, limit the
|
|
|
|
# file stream to that length.
|
|
|
|
if length == filelike.READ_UNTIL_END:
|
|
|
|
length = _MAXIMUM_SEGMENT_SIZE
|
|
|
|
else:
|
|
|
|
length = min(_MAXIMUM_SEGMENT_SIZE, length)
|
|
|
|
|
|
|
|
limiting_fp = filelike.LimitingStream(in_fp, length)
|
|
|
|
|
|
|
|
# Write the segment to Swift.
|
2015-11-30 20:45:45 +00:00
|
|
|
self.stream_write(segment_path, limiting_fp, content_type)
|
2015-09-25 18:57:14 +00:00
|
|
|
|
|
|
|
# We are only going to track keys to which data was confirmed written.
|
2015-09-30 18:19:25 +00:00
|
|
|
bytes_written = limiting_fp.tell()
|
2015-09-25 18:57:14 +00:00
|
|
|
if bytes_written > 0:
|
|
|
|
updated_metadata[_SEGMENTS_KEY].append(_PartUploadMetadata(segment_path, offset,
|
|
|
|
bytes_written))
|
|
|
|
|
|
|
|
return bytes_written, updated_metadata
|
|
|
|
|
|
|
|
def complete_chunked_upload(self, uuid, final_path, storage_metadata):
|
|
|
|
""" Complete the chunked upload and store the final results in the path indicated.
|
|
|
|
Returns nothing.
|
|
|
|
"""
|
|
|
|
# Finally, we write an empty file at the proper location with a X-Object-Manifest
|
|
|
|
# header pointing to the prefix for the segments.
|
2016-01-14 20:30:01 +00:00
|
|
|
segments_prefix_path = self._normalize_path('%s/%s' % (_SEGMENT_DIRECTORY, uuid))
|
2015-09-25 18:57:14 +00:00
|
|
|
contained_segments_prefix_path = '%s/%s' % (self._swift_container, segments_prefix_path)
|
|
|
|
|
|
|
|
self._put_object(final_path, '', headers={'X-Object-Manifest': contained_segments_prefix_path})
|
|
|
|
|
|
|
|
def cancel_chunked_upload(self, uuid, storage_metadata):
|
|
|
|
""" Cancel the chunked upload and clean up any outstanding partially uploaded data.
|
|
|
|
Returns nothing.
|
|
|
|
"""
|
|
|
|
# Delete all the uploaded segments.
|
|
|
|
for segment in SwiftStorage._segment_list_from_metadata(storage_metadata):
|
|
|
|
self.remove(segment.path)
|