Add the registry code.
This commit is contained in:
parent
5151ea4412
commit
deee70d53b
7 changed files with 806 additions and 0 deletions
263
endpoints/registry.py
Normal file
263
endpoints/registry.py
Normal file
|
@ -0,0 +1,263 @@
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from flask import make_response, request, session, Response, abort
|
||||||
|
import simplejson as json
|
||||||
|
from functools import wraps
|
||||||
|
from datetime import datetime
|
||||||
|
from time import time
|
||||||
|
|
||||||
|
import storage
|
||||||
|
|
||||||
|
from app import app
|
||||||
|
from auth.auth import process_auth
|
||||||
|
from util import checksums
|
||||||
|
|
||||||
|
|
||||||
|
store = storage.load()
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class SocketReader(object):
|
||||||
|
def __init__(self, fp):
|
||||||
|
self._fp = fp
|
||||||
|
self.handlers = []
|
||||||
|
|
||||||
|
def add_handler(self, handler):
|
||||||
|
self.handlers.append(handler)
|
||||||
|
|
||||||
|
def read(self, n=-1):
|
||||||
|
buf = self._fp.read(n)
|
||||||
|
if not buf:
|
||||||
|
return ''
|
||||||
|
for handler in self.handlers:
|
||||||
|
handler(buf)
|
||||||
|
return buf
|
||||||
|
|
||||||
|
|
||||||
|
def require_completion(f):
|
||||||
|
"""This make sure that the image push correctly finished."""
|
||||||
|
@wraps(f)
|
||||||
|
def wrapper(*args, **kwargs):
|
||||||
|
if store.exists(store.image_mark_path(kwargs['image_id'])):
|
||||||
|
abort(400) #'Image is being uploaded, retry later')
|
||||||
|
return f(*args, **kwargs)
|
||||||
|
return wrapper
|
||||||
|
|
||||||
|
|
||||||
|
def set_cache_headers(f):
|
||||||
|
"""Returns HTTP headers suitable for caching."""
|
||||||
|
@wraps(f)
|
||||||
|
def wrapper(*args, **kwargs):
|
||||||
|
# Set TTL to 1 year by default
|
||||||
|
ttl = 31536000
|
||||||
|
expires = datetime.fromtimestamp(int(time()) + ttl)
|
||||||
|
expires = expires.strftime('%a, %d %b %Y %H:%M:%S GMT')
|
||||||
|
headers = {
|
||||||
|
'Cache-Control': 'public, max-age={0}'.format(ttl),
|
||||||
|
'Expires': expires,
|
||||||
|
'Last-Modified': 'Thu, 01 Jan 1970 00:00:00 GMT',
|
||||||
|
}
|
||||||
|
if 'If-Modified-Since' in request.headers:
|
||||||
|
response = make_response('Not modified', 304)
|
||||||
|
response.headers.extend(headers)
|
||||||
|
return response
|
||||||
|
kwargs['headers'] = headers
|
||||||
|
# Prevent the Cookie to be sent when the object is cacheable
|
||||||
|
session.modified = False
|
||||||
|
return f(*args, **kwargs)
|
||||||
|
return wrapper
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/images/<image_id>/layer', methods=['GET'])
|
||||||
|
@process_auth
|
||||||
|
@require_completion
|
||||||
|
@set_cache_headers
|
||||||
|
def get_image_layer(image_id, headers):
|
||||||
|
try:
|
||||||
|
return Response(store.stream_read(store.image_layer_path(
|
||||||
|
image_id)), headers=headers)
|
||||||
|
except IOError:
|
||||||
|
abort(404) #'Image not found', 404)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/images/<image_id>/layer', methods=['PUT'])
|
||||||
|
@process_auth
|
||||||
|
def put_image_layer(image_id):
|
||||||
|
try:
|
||||||
|
json_data = store.get_content(store.image_json_path(image_id))
|
||||||
|
except IOError:
|
||||||
|
abort(404) #'Image not found', 404)
|
||||||
|
layer_path = store.image_layer_path(image_id)
|
||||||
|
mark_path = store.image_mark_path(image_id)
|
||||||
|
if store.exists(layer_path) and not store.exists(mark_path):
|
||||||
|
abort(409) #'Image already exists', 409)
|
||||||
|
input_stream = request.stream
|
||||||
|
if request.headers.get('transfer-encoding') == 'chunked':
|
||||||
|
# Careful, might work only with WSGI servers supporting chunked
|
||||||
|
# encoding (Gunicorn)
|
||||||
|
input_stream = request.environ['wsgi.input']
|
||||||
|
# compute checksums
|
||||||
|
csums = []
|
||||||
|
sr = SocketReader(input_stream)
|
||||||
|
tmp, store_hndlr = storage.temp_store_handler()
|
||||||
|
sr.add_handler(store_hndlr)
|
||||||
|
h, sum_hndlr = checksums.simple_checksum_handler(json_data)
|
||||||
|
sr.add_handler(sum_hndlr)
|
||||||
|
store.stream_write(layer_path, sr)
|
||||||
|
csums.append('sha256:{0}'.format(h.hexdigest()))
|
||||||
|
try:
|
||||||
|
tmp.seek(0)
|
||||||
|
csums.append(checksums.compute_tarsum(tmp, json_data))
|
||||||
|
tmp.close()
|
||||||
|
except (IOError, checksums.TarError) as e:
|
||||||
|
logger.debug('put_image_layer: Error when computing tarsum '
|
||||||
|
'{0}'.format(e))
|
||||||
|
try:
|
||||||
|
checksum = store.get_content(store.image_checksum_path(image_id))
|
||||||
|
except IOError:
|
||||||
|
# We don't have a checksum stored yet, that's fine skipping the check.
|
||||||
|
# Not removing the mark though, image is not downloadable yet.
|
||||||
|
session['checksum'] = csums
|
||||||
|
return make_response('true', 200)
|
||||||
|
# We check if the checksums provided matches one the one we computed
|
||||||
|
if checksum not in csums:
|
||||||
|
logger.debug('put_image_layer: Wrong checksum')
|
||||||
|
abort(400) #'Checksum mismatch, ignoring the layer')
|
||||||
|
# Checksum is ok, we remove the marker
|
||||||
|
store.remove(mark_path)
|
||||||
|
return make_response('true', 200)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/images/<image_id>/checksum', methods=['PUT'])
|
||||||
|
@process_auth
|
||||||
|
def put_image_checksum(image_id):
|
||||||
|
checksum = request.headers.get('X-Docker-Checksum')
|
||||||
|
if not checksum:
|
||||||
|
abort(400) #'Missing Image\'s checksum')
|
||||||
|
if not session.get('checksum'):
|
||||||
|
abort(400) #'Checksum not found in Cookie')
|
||||||
|
if not store.exists(store.image_json_path(image_id)):
|
||||||
|
abort(404) #'Image not found', 404)
|
||||||
|
mark_path = store.image_mark_path(image_id)
|
||||||
|
if not store.exists(mark_path):
|
||||||
|
abort(409) #'Cannot set this image checksum', 409)
|
||||||
|
err = store_checksum(image_id, checksum)
|
||||||
|
if err:
|
||||||
|
abort(err)
|
||||||
|
if checksum not in session.get('checksum', []):
|
||||||
|
logger.debug('put_image_layer: Wrong checksum')
|
||||||
|
abort(400) #'Checksum mismatch')
|
||||||
|
# Checksum is ok, we remove the marker
|
||||||
|
store.remove(mark_path)
|
||||||
|
return make_response('true', 200)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/images/<image_id>/json', methods=['GET'])
|
||||||
|
@process_auth
|
||||||
|
@require_completion
|
||||||
|
@set_cache_headers
|
||||||
|
def get_image_json(image_id, headers):
|
||||||
|
try:
|
||||||
|
data = store.get_content(store.image_json_path(image_id))
|
||||||
|
except IOError:
|
||||||
|
abort(404) #'Image not found', 404)
|
||||||
|
try:
|
||||||
|
size = store.get_size(store.image_layer_path(image_id))
|
||||||
|
headers['X-Docker-Size'] = str(size)
|
||||||
|
except OSError:
|
||||||
|
pass
|
||||||
|
checksum_path = store.image_checksum_path(image_id)
|
||||||
|
if store.exists(checksum_path):
|
||||||
|
headers['X-Docker-Checksum'] = store.get_content(checksum_path)
|
||||||
|
response = make_response(data, 200)
|
||||||
|
response.headers.extend(headers)
|
||||||
|
return response
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/images/<image_id>/ancestry', methods=['GET'])
|
||||||
|
@process_auth
|
||||||
|
@require_completion
|
||||||
|
@set_cache_headers
|
||||||
|
def get_image_ancestry(image_id, headers):
|
||||||
|
try:
|
||||||
|
data = store.get_content(store.image_ancestry_path(image_id))
|
||||||
|
except IOError:
|
||||||
|
abort(404) #'Image not found', 404)
|
||||||
|
response = make_response(json.dumps(json.loads(data)), 200)
|
||||||
|
response.headers.extend(headers)
|
||||||
|
return response
|
||||||
|
|
||||||
|
|
||||||
|
def generate_ancestry(image_id, parent_id=None):
|
||||||
|
if not parent_id:
|
||||||
|
store.put_content(store.image_ancestry_path(image_id),
|
||||||
|
json.dumps([image_id]))
|
||||||
|
return
|
||||||
|
data = store.get_content(store.image_ancestry_path(parent_id))
|
||||||
|
data = json.loads(data)
|
||||||
|
data.insert(0, image_id)
|
||||||
|
store.put_content(store.image_ancestry_path(image_id), json.dumps(data))
|
||||||
|
|
||||||
|
|
||||||
|
def check_images_list(image_id):
|
||||||
|
full_repos_name = session.get('repository')
|
||||||
|
if not full_repos_name:
|
||||||
|
# We only enforce this check when there is a repos name in the session
|
||||||
|
# otherwise it means that the auth is disabled.
|
||||||
|
return True
|
||||||
|
try:
|
||||||
|
path = store.images_list_path(*full_repos_name.split('/'))
|
||||||
|
images_list = json.loads(store.get_content(path))
|
||||||
|
except IOError:
|
||||||
|
return False
|
||||||
|
return (image_id in images_list)
|
||||||
|
|
||||||
|
|
||||||
|
def store_checksum(image_id, checksum):
|
||||||
|
checksum_parts = checksum.split(':')
|
||||||
|
if len(checksum_parts) != 2:
|
||||||
|
return 'Invalid checksum format'
|
||||||
|
# We store the checksum
|
||||||
|
checksum_path = store.image_checksum_path(image_id)
|
||||||
|
store.put_content(checksum_path, checksum)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/images/<image_id>/json', methods=['PUT'])
|
||||||
|
@process_auth
|
||||||
|
def put_image_json(image_id):
|
||||||
|
try:
|
||||||
|
data = json.loads(request.data)
|
||||||
|
except json.JSONDecodeError:
|
||||||
|
pass
|
||||||
|
if not data or not isinstance(data, dict):
|
||||||
|
abort(400) #'Invalid JSON')
|
||||||
|
if 'id' not in data:
|
||||||
|
abort(400) #'Missing key `id\' in JSON')
|
||||||
|
# Read the checksum
|
||||||
|
checksum = request.headers.get('X-Docker-Checksum')
|
||||||
|
if checksum:
|
||||||
|
# Storing the checksum is optional at this stage
|
||||||
|
err = store_checksum(image_id, checksum)
|
||||||
|
if err:
|
||||||
|
abort(err)
|
||||||
|
else:
|
||||||
|
# We cleanup any old checksum in case it's a retry after a fail
|
||||||
|
store.remove(store.image_checksum_path(image_id))
|
||||||
|
if image_id != data['id']:
|
||||||
|
abort(400) #'JSON data contains invalid id')
|
||||||
|
if check_images_list(image_id) is False:
|
||||||
|
abort(400) #'This image does not belong to the repository')
|
||||||
|
parent_id = data.get('parent')
|
||||||
|
if parent_id and not store.exists(store.image_json_path(data['parent'])):
|
||||||
|
abort(400) #'Image depends on a non existing parent')
|
||||||
|
json_path = store.image_json_path(image_id)
|
||||||
|
mark_path = store.image_mark_path(image_id)
|
||||||
|
if store.exists(json_path) and not store.exists(mark_path):
|
||||||
|
abort(409) #'Image already exists', 409)
|
||||||
|
# If we reach that point, it means that this is a new image or a retry
|
||||||
|
# on a failed push
|
||||||
|
store.put_content(mark_path, 'true')
|
||||||
|
store.put_content(json_path, request.data)
|
||||||
|
generate_ancestry(image_id, parent_id)
|
||||||
|
return make_response('true', 200)
|
93
endpoints/tags.py
Normal file
93
endpoints/tags.py
Normal file
|
@ -0,0 +1,93 @@
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from flask import abort, request, jsonify, make_response
|
||||||
|
import simplejson as json
|
||||||
|
|
||||||
|
import storage
|
||||||
|
|
||||||
|
from app import app
|
||||||
|
from util.names import parse_repository_name
|
||||||
|
from auth.auth import process_auth
|
||||||
|
from auth.permissions import (ReadRepositoryPermission,
|
||||||
|
ModifyRepositoryPermission)
|
||||||
|
from data import model
|
||||||
|
|
||||||
|
|
||||||
|
store = storage.load()
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/repositories/<path:repository>/tags',
|
||||||
|
methods=['GET'])
|
||||||
|
@process_auth
|
||||||
|
@parse_repository_name
|
||||||
|
def get_tags(namespace, repository):
|
||||||
|
permission = ReadRepositoryPermission(namespace, repository)
|
||||||
|
|
||||||
|
if permission.can():
|
||||||
|
tags = model.list_repository_tags(namespace, repository)
|
||||||
|
tag_map = {tag.name: tag.image.image_id for tag in tags}
|
||||||
|
return jsonify(tag_map)
|
||||||
|
|
||||||
|
abort(403)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/repositories/<path:repository>/tags/<tag>',
|
||||||
|
methods=['GET'])
|
||||||
|
@process_auth
|
||||||
|
@parse_repository_name
|
||||||
|
def get_tag(namespace, repository, tag):
|
||||||
|
permission = ReadRepositoryPermission(namespace, repository)
|
||||||
|
|
||||||
|
if permission.can():
|
||||||
|
tag_image = model.get_tag_image(namespace, repository, tag)
|
||||||
|
response = make_response(tag_image.image_id, 200)
|
||||||
|
|
||||||
|
abort(403)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/repositories/<path:repository>/tags/<tag>',
|
||||||
|
methods=['PUT'])
|
||||||
|
@process_auth
|
||||||
|
@parse_repository_name
|
||||||
|
def put_tag(namespace, repository, tag):
|
||||||
|
permission = ModifyRepositoryPermission(namespace, repository)
|
||||||
|
|
||||||
|
if permission.can():
|
||||||
|
image_id = json.loads(request.data)
|
||||||
|
model.create_or_update_tag(namespace, repository, tag, image_id)
|
||||||
|
|
||||||
|
return make_response('Created', 200)
|
||||||
|
|
||||||
|
abort(403)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/repositories/<path:repository>/tags/<tag>',
|
||||||
|
methods=['DELETE'])
|
||||||
|
@process_auth
|
||||||
|
@parse_repository_name
|
||||||
|
def delete_tag(namespace, repository, tag):
|
||||||
|
permission = ModifyRepositoryPermission(namespace, repository)
|
||||||
|
|
||||||
|
if permission.can():
|
||||||
|
model.delete_tag(namespace, repository, tag_name)
|
||||||
|
|
||||||
|
return make_response('Deleted', 204)
|
||||||
|
|
||||||
|
abort(403)
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/v1/repositories/<path:repository>/tags',
|
||||||
|
methods=['DELETE'])
|
||||||
|
@process_auth
|
||||||
|
@parse_repository_name
|
||||||
|
def delete_repository(namespace, repository):
|
||||||
|
permission = ModifyRepositoryPermission(namespace, repository)
|
||||||
|
|
||||||
|
if permission.can():
|
||||||
|
model.delete_all_repository_tags(namespace, repository)
|
||||||
|
|
||||||
|
return make_response('Deleted', 204)
|
||||||
|
|
||||||
|
abort(403)
|
13
registry_wsgi.py
Normal file
13
registry_wsgi.py
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from app import app
|
||||||
|
|
||||||
|
import endpoints.registry
|
||||||
|
import endpoints.tags
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
FORMAT = '%(asctime)-15s - %(levelname)s - %(pathname)s - ' + \
|
||||||
|
'%(funcName)s - %(message)s'
|
||||||
|
logging.basicConfig(format=FORMAT, level=logging.DEBUG)
|
||||||
|
|
||||||
|
app.run(port=5000, debug=True)
|
138
storage/__init__.py
Normal file
138
storage/__init__.py
Normal file
|
@ -0,0 +1,138 @@
|
||||||
|
|
||||||
|
import contextlib
|
||||||
|
import tempfile
|
||||||
|
|
||||||
|
|
||||||
|
__all__ = ['load']
|
||||||
|
|
||||||
|
|
||||||
|
class Storage(object):
|
||||||
|
|
||||||
|
"""Storage is organized as follow:
|
||||||
|
$ROOT/images/<image_id>/json
|
||||||
|
$ROOT/images/<image_id>/layer
|
||||||
|
$ROOT/repositories/<namespace>/<repository_name>/<tag_name>
|
||||||
|
"""
|
||||||
|
|
||||||
|
# Useful if we want to change those locations later without rewriting
|
||||||
|
# the code which uses Storage
|
||||||
|
repositories = 'repositories'
|
||||||
|
images = 'images'
|
||||||
|
# Set the IO buffer to 64kB
|
||||||
|
buffer_size = 64 * 1024
|
||||||
|
|
||||||
|
#FIXME(samalba): Move all path resolver in each module (out of the base)
|
||||||
|
def images_list_path(self, namespace, repository):
|
||||||
|
return '{0}/{1}/{2}/_images_list'.format(self.repositories,
|
||||||
|
namespace,
|
||||||
|
repository)
|
||||||
|
|
||||||
|
def image_json_path(self, image_id):
|
||||||
|
return '{0}/{1}/json'.format(self.images, image_id)
|
||||||
|
|
||||||
|
def image_mark_path(self, image_id):
|
||||||
|
return '{0}/{1}/_inprogress'.format(self.images, image_id)
|
||||||
|
|
||||||
|
def image_checksum_path(self, image_id):
|
||||||
|
return '{0}/{1}/_checksum'.format(self.images, image_id)
|
||||||
|
|
||||||
|
def image_layer_path(self, image_id):
|
||||||
|
return '{0}/{1}/layer'.format(self.images, image_id)
|
||||||
|
|
||||||
|
def image_ancestry_path(self, image_id):
|
||||||
|
return '{0}/{1}/ancestry'.format(self.images, image_id)
|
||||||
|
|
||||||
|
def tag_path(self, namespace, repository, tagname=None):
|
||||||
|
if not tagname:
|
||||||
|
return '{0}/{1}/{2}'.format(self.repositories,
|
||||||
|
namespace,
|
||||||
|
repository)
|
||||||
|
return '{0}/{1}/{2}/tag_{3}'.format(self.repositories,
|
||||||
|
namespace,
|
||||||
|
repository,
|
||||||
|
tagname)
|
||||||
|
|
||||||
|
def index_images_path(self, namespace, repository):
|
||||||
|
return '{0}/{1}/{2}/_index_images'.format(self.repositories,
|
||||||
|
namespace,
|
||||||
|
repository)
|
||||||
|
|
||||||
|
def get_content(self, path):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def put_content(self, path, content):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def stream_read(self, path):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def stream_write(self, path, fp):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def list_directory(self, path=None):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def exists(self, path):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def remove(self, path):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def get_size(self, path):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
|
||||||
|
@contextlib.contextmanager
|
||||||
|
def store_stream(stream):
|
||||||
|
"""Stores the entire stream to a temporary file."""
|
||||||
|
tmpf = tempfile.TemporaryFile()
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
buf = stream.read(4096)
|
||||||
|
if not buf:
|
||||||
|
break
|
||||||
|
tmpf.write(buf)
|
||||||
|
except IOError:
|
||||||
|
break
|
||||||
|
tmpf.seek(0)
|
||||||
|
yield tmpf
|
||||||
|
tmpf.close()
|
||||||
|
|
||||||
|
|
||||||
|
def temp_store_handler():
|
||||||
|
tmpf = tempfile.TemporaryFile()
|
||||||
|
|
||||||
|
def fn(buf):
|
||||||
|
try:
|
||||||
|
tmpf.write(buf)
|
||||||
|
except IOError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
return tmpf, fn
|
||||||
|
|
||||||
|
|
||||||
|
from local import LocalStorage
|
||||||
|
from s3 import S3Storage
|
||||||
|
|
||||||
|
|
||||||
|
_storage = {}
|
||||||
|
|
||||||
|
|
||||||
|
def load(kind=None):
|
||||||
|
"""Returns the right storage class according to the configuration."""
|
||||||
|
global _storage
|
||||||
|
|
||||||
|
# TODO hard code to local for now
|
||||||
|
kind = 'local'
|
||||||
|
# if not kind:
|
||||||
|
# kind = cfg.storage.lower()
|
||||||
|
if kind in _storage:
|
||||||
|
return _storage[kind]
|
||||||
|
if kind == 's3':
|
||||||
|
store = S3Storage('/registry', 'access_key', 'secret_key', 'bucket')
|
||||||
|
elif kind == 'local':
|
||||||
|
store = LocalStorage('/tmp/registry')
|
||||||
|
else:
|
||||||
|
raise ValueError('Not supported storage \'{0}\''.format(kind))
|
||||||
|
_storage[kind] = store
|
||||||
|
return store
|
82
storage/local.py
Normal file
82
storage/local.py
Normal file
|
@ -0,0 +1,82 @@
|
||||||
|
|
||||||
|
import os
|
||||||
|
import shutil
|
||||||
|
|
||||||
|
from . import Storage
|
||||||
|
|
||||||
|
|
||||||
|
class LocalStorage(Storage):
|
||||||
|
|
||||||
|
def __init__(self, storage_path):
|
||||||
|
self._root_path = storage_path
|
||||||
|
|
||||||
|
def _init_path(self, path=None, create=False):
|
||||||
|
path = os.path.join(self._root_path, path) if path else self._root_path
|
||||||
|
if create is True:
|
||||||
|
dirname = os.path.dirname(path)
|
||||||
|
if not os.path.exists(dirname):
|
||||||
|
os.makedirs(dirname)
|
||||||
|
return path
|
||||||
|
|
||||||
|
def get_content(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
with open(path, mode='r') as f:
|
||||||
|
return f.read()
|
||||||
|
|
||||||
|
def put_content(self, path, content):
|
||||||
|
path = self._init_path(path, create=True)
|
||||||
|
with open(path, mode='w') as f:
|
||||||
|
f.write(content)
|
||||||
|
return path
|
||||||
|
|
||||||
|
def stream_read(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
with open(path, mode='rb') as f:
|
||||||
|
while True:
|
||||||
|
buf = f.read(self.buffer_size)
|
||||||
|
if not buf:
|
||||||
|
break
|
||||||
|
yield buf
|
||||||
|
|
||||||
|
def stream_write(self, path, fp):
|
||||||
|
# Size is mandatory
|
||||||
|
path = self._init_path(path, create=True)
|
||||||
|
with open(path, mode='wb') as f:
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
buf = fp.read(self.buffer_size)
|
||||||
|
if not buf:
|
||||||
|
break
|
||||||
|
f.write(buf)
|
||||||
|
except IOError:
|
||||||
|
break
|
||||||
|
|
||||||
|
def list_directory(self, path=None):
|
||||||
|
path = self._init_path(path)
|
||||||
|
prefix = path[len(self._root_path) + 1:] + '/'
|
||||||
|
exists = False
|
||||||
|
for d in os.listdir(path):
|
||||||
|
exists = True
|
||||||
|
yield prefix + d
|
||||||
|
if exists is False:
|
||||||
|
# Raises OSError even when the directory is empty
|
||||||
|
# (to be consistent with S3)
|
||||||
|
raise OSError('No such directory: \'{0}\''.format(path))
|
||||||
|
|
||||||
|
def exists(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
return os.path.exists(path)
|
||||||
|
|
||||||
|
def remove(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
if os.path.isdir(path):
|
||||||
|
shutil.rmtree(path)
|
||||||
|
return
|
||||||
|
try:
|
||||||
|
os.remove(path)
|
||||||
|
except OSError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
def get_size(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
return os.path.getsize(path)
|
128
storage/s3.py
Normal file
128
storage/s3.py
Normal file
|
@ -0,0 +1,128 @@
|
||||||
|
|
||||||
|
import cStringIO as StringIO
|
||||||
|
import os
|
||||||
|
|
||||||
|
import boto.s3.connection
|
||||||
|
import boto.s3.key
|
||||||
|
|
||||||
|
from . import Storage
|
||||||
|
|
||||||
|
|
||||||
|
class S3Storage(Storage):
|
||||||
|
|
||||||
|
def __init__(self, storage_path, s3_access_key, s3_secret_key, s3_bucket):
|
||||||
|
self._s3_conn = \
|
||||||
|
boto.s3.connection.S3Connection(s3_access_key,
|
||||||
|
s3_secret_key,
|
||||||
|
is_secure=False)
|
||||||
|
self._s3_bucket = self._s3_conn.get_bucket(s3_bucket)
|
||||||
|
self._root_path = storage_path
|
||||||
|
|
||||||
|
def _debug_key(self, key):
|
||||||
|
"""Used for debugging only."""
|
||||||
|
orig_meth = key.bucket.connection.make_request
|
||||||
|
|
||||||
|
def new_meth(*args, **kwargs):
|
||||||
|
print '#' * 16
|
||||||
|
print args
|
||||||
|
print kwargs
|
||||||
|
print '#' * 16
|
||||||
|
return orig_meth(*args, **kwargs)
|
||||||
|
key.bucket.connection.make_request = new_meth
|
||||||
|
|
||||||
|
def _init_path(self, path=None):
|
||||||
|
path = os.path.join(self._root_path, path) if path else self._root_path
|
||||||
|
if path and path[0] == '/':
|
||||||
|
return path[1:]
|
||||||
|
return path
|
||||||
|
|
||||||
|
def get_content(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
key = boto.s3.key.Key(self._s3_bucket, path)
|
||||||
|
if not key.exists():
|
||||||
|
raise IOError('No such key: \'{0}\''.format(path))
|
||||||
|
return key.get_contents_as_string()
|
||||||
|
|
||||||
|
def put_content(self, path, content):
|
||||||
|
path = self._init_path(path)
|
||||||
|
key = boto.s3.key.Key(self._s3_bucket, path)
|
||||||
|
key.set_contents_from_string(content)
|
||||||
|
return path
|
||||||
|
|
||||||
|
def stream_read(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
key = boto.s3.key.Key(self._s3_bucket, path)
|
||||||
|
if not key.exists():
|
||||||
|
raise IOError('No such key: \'{0}\''.format(path))
|
||||||
|
while True:
|
||||||
|
buf = key.read(self.buffer_size)
|
||||||
|
if not buf:
|
||||||
|
break
|
||||||
|
yield buf
|
||||||
|
|
||||||
|
def stream_write(self, path, fp):
|
||||||
|
# Minimum size of upload part size on S3 is 5MB
|
||||||
|
buffer_size = 5 * 1024 * 1024
|
||||||
|
if self.buffer_size > buffer_size:
|
||||||
|
buffer_size = self.buffer_size
|
||||||
|
path = self._init_path(path)
|
||||||
|
mp = self._s3_bucket.initiate_multipart_upload(path)
|
||||||
|
num_part = 1
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
buf = fp.read(buffer_size)
|
||||||
|
if not buf:
|
||||||
|
break
|
||||||
|
io = StringIO.StringIO(buf)
|
||||||
|
mp.upload_part_from_file(io, num_part)
|
||||||
|
num_part += 1
|
||||||
|
io.close()
|
||||||
|
except IOError:
|
||||||
|
break
|
||||||
|
mp.complete_upload()
|
||||||
|
|
||||||
|
def list_directory(self, path=None):
|
||||||
|
path = self._init_path(path)
|
||||||
|
if not path.endswith('/'):
|
||||||
|
path += '/'
|
||||||
|
ln = 0
|
||||||
|
if self._root_path != '/':
|
||||||
|
ln = len(self._root_path)
|
||||||
|
exists = False
|
||||||
|
for key in self._s3_bucket.list(prefix=path, delimiter='/'):
|
||||||
|
exists = True
|
||||||
|
name = key.name
|
||||||
|
if name.endswith('/'):
|
||||||
|
yield name[ln:-1]
|
||||||
|
else:
|
||||||
|
yield name[ln:]
|
||||||
|
if exists is False:
|
||||||
|
# In order to be compliant with the LocalStorage API. Even though
|
||||||
|
# S3 does not have a concept of folders.
|
||||||
|
raise OSError('No such directory: \'{0}\''.format(path))
|
||||||
|
|
||||||
|
def exists(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
key = boto.s3.key.Key(self._s3_bucket, path)
|
||||||
|
return key.exists()
|
||||||
|
|
||||||
|
def remove(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
key = boto.s3.key.Key(self._s3_bucket, path)
|
||||||
|
if key.exists():
|
||||||
|
# It's a file
|
||||||
|
key.delete()
|
||||||
|
return
|
||||||
|
# We assume it's a directory
|
||||||
|
if not path.endswith('/'):
|
||||||
|
path += '/'
|
||||||
|
for key in self._s3_bucket.list(prefix=path, delimiter='/'):
|
||||||
|
key.delete()
|
||||||
|
|
||||||
|
def get_size(self, path):
|
||||||
|
path = self._init_path(path)
|
||||||
|
# Lookup does a HEAD HTTP Request on the object
|
||||||
|
key = self._s3_bucket.lookup(path)
|
||||||
|
if not key:
|
||||||
|
raise OSError('No such key: \'{0}\''.format(path))
|
||||||
|
return key.size
|
89
util/checksums.py
Normal file
89
util/checksums.py
Normal file
|
@ -0,0 +1,89 @@
|
||||||
|
|
||||||
|
import hashlib
|
||||||
|
import logging
|
||||||
|
import tarfile
|
||||||
|
|
||||||
|
|
||||||
|
TarError = tarfile.TarError
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def sha256_file(fp, data=None):
|
||||||
|
h = hashlib.sha256(data or '')
|
||||||
|
if not fp:
|
||||||
|
return h.hexdigest()
|
||||||
|
while True:
|
||||||
|
buf = fp.read(4096)
|
||||||
|
if not buf:
|
||||||
|
break
|
||||||
|
h.update(buf)
|
||||||
|
return h.hexdigest()
|
||||||
|
|
||||||
|
|
||||||
|
def sha256_string(s):
|
||||||
|
return hashlib.sha256(s).hexdigest()
|
||||||
|
|
||||||
|
|
||||||
|
def compute_tarsum(fp, json_data):
|
||||||
|
header_fields = ('name', 'mode', 'uid', 'gid', 'size', 'mtime',
|
||||||
|
'type', 'linkname', 'uname', 'gname', 'devmajor',
|
||||||
|
'devminor')
|
||||||
|
tar = tarfile.open(mode='r|*', fileobj=fp)
|
||||||
|
hashes = []
|
||||||
|
for member in tar:
|
||||||
|
header = ''
|
||||||
|
for field in header_fields:
|
||||||
|
value = getattr(member, field)
|
||||||
|
if field == 'type':
|
||||||
|
field = 'typeflag'
|
||||||
|
elif field == 'name':
|
||||||
|
if member.isdir() and not value.endswith('/'):
|
||||||
|
value += '/'
|
||||||
|
header += '{0}{1}'.format(field, value)
|
||||||
|
h = None
|
||||||
|
try:
|
||||||
|
if member.size > 0:
|
||||||
|
f = tar.extractfile(member)
|
||||||
|
h = sha256_file(f, header)
|
||||||
|
else:
|
||||||
|
h = sha256_string(header)
|
||||||
|
except KeyError:
|
||||||
|
h = sha256_string(header)
|
||||||
|
hashes.append(h)
|
||||||
|
#log = '\n+ filename: {0}\n'.format(member.name)
|
||||||
|
#log += '+ header: {0}\n'.format(header)
|
||||||
|
#log += '+ hash: {0}\n'.format(h)
|
||||||
|
#log += '*' * 16
|
||||||
|
#logger.debug('checksums.compute_tarsum: \n{0}'.format(log))
|
||||||
|
tar.close()
|
||||||
|
hashes.sort()
|
||||||
|
data = json_data + ''.join(hashes)
|
||||||
|
#logger.debug('checksums.compute_tarsum: '
|
||||||
|
# 'Hashes: \n{0}\n{1}'.format('\n'.join(hashes), '-' * 16))
|
||||||
|
tarsum = 'tarsum+sha256:{0}'.format(sha256_string(data))
|
||||||
|
logger.debug('checksums.compute_tarsum: return {0}'.format(tarsum))
|
||||||
|
return tarsum
|
||||||
|
|
||||||
|
|
||||||
|
def simple_checksum_handler(json_data):
|
||||||
|
h = hashlib.sha256(json_data + '\n')
|
||||||
|
|
||||||
|
def fn(buf):
|
||||||
|
h.update(buf)
|
||||||
|
return h, fn
|
||||||
|
|
||||||
|
|
||||||
|
def compute_simple(fp, json_data):
|
||||||
|
data = json_data + '\n'
|
||||||
|
return 'sha256:{0}'.format(sha256_file(fp, data))
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
import sys
|
||||||
|
if len(sys.argv) < 3:
|
||||||
|
print 'Usage: {0} json_file layer_file'.format(sys.argv[0])
|
||||||
|
sys.exit(1)
|
||||||
|
json_data = file(sys.argv[1]).read()
|
||||||
|
fp = open(sys.argv[2])
|
||||||
|
print compute_simple(fp, json_data)
|
||||||
|
print compute_tarsum(fp, json_data)
|
Reference in a new issue