Change V1 to use the manifest builder and new registry data model
This commit is contained in:
parent
65d5be23c7
commit
4520f9e842
12 changed files with 291 additions and 689 deletions
|
@ -7,38 +7,37 @@ from time import time
|
|||
|
||||
from flask import make_response, request, session, Response, redirect, abort as flask_abort
|
||||
|
||||
from app import storage as store, app, metric_queue
|
||||
from app import storage as store, app
|
||||
from auth.auth_context import get_authenticated_user
|
||||
from auth.decorators import extract_namespace_repo_from_session, process_auth
|
||||
from auth.permissions import (ReadRepositoryPermission, ModifyRepositoryPermission)
|
||||
from data import model, database
|
||||
from data import database
|
||||
from data.registry_model import registry_model
|
||||
from data.registry_model.blobuploader import upload_blob, BlobUploadSettings, BlobUploadException
|
||||
from data.registry_model.manifestbuilder import lookup_manifest_builder
|
||||
from digest import checksums
|
||||
from endpoints.v1 import v1_bp
|
||||
from endpoints.v1.models_pre_oci import pre_oci_model as model
|
||||
from endpoints.v1.index import ensure_namespace_enabled
|
||||
from endpoints.decorators import anon_protect
|
||||
from util.http import abort, exact_abort
|
||||
from util.registry.filelike import SocketReader
|
||||
from util.registry import gzipstream
|
||||
from util.registry.replication import queue_storage_replication
|
||||
from util.registry.torrent import PieceHasher
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def require_completion(f):
|
||||
"""This make sure that the image push correctly finished."""
|
||||
|
||||
""" This make sure that the image push correctly finished. """
|
||||
@wraps(f)
|
||||
def wrapper(namespace, repository, *args, **kwargs):
|
||||
image_id = kwargs['image_id']
|
||||
if model.is_image_uploading(namespace, repository, image_id):
|
||||
abort(400, 'Image %(image_id)s is being uploaded, retry later', issue='upload-in-progress',
|
||||
image_id=image_id)
|
||||
return f(namespace, repository, *args, **kwargs)
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository)
|
||||
if repository_ref is not None:
|
||||
legacy_image = registry_model.get_legacy_image(repository_ref, image_id)
|
||||
if legacy_image is not None and legacy_image.uploading:
|
||||
abort(400, 'Image %(image_id)s is being uploaded, retry later', issue='upload-in-progress',
|
||||
image_id=image_id)
|
||||
|
||||
return f(namespace, repository, *args, **kwargs)
|
||||
return wrapper
|
||||
|
||||
|
||||
|
@ -76,24 +75,23 @@ def set_cache_headers(f):
|
|||
@anon_protect
|
||||
def head_image_layer(namespace, repository, image_id, headers):
|
||||
permission = ReadRepositoryPermission(namespace, repository)
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
|
||||
logger.debug('Checking repo permissions')
|
||||
if permission.can() or model.repository_is_public(namespace, repository):
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
if permission.can() or (repository_ref is not None and repository_ref.is_public):
|
||||
if repository_ref is None:
|
||||
abort(404)
|
||||
|
||||
logger.debug('Looking up placement locations')
|
||||
locations, _ = model.placement_locations_and_path_docker_v1(namespace, repository, image_id)
|
||||
if locations is None:
|
||||
legacy_image = registry_model.get_legacy_image(repository_ref, image_id, include_blob=True)
|
||||
if legacy_image is None:
|
||||
logger.debug('Could not find any blob placement locations')
|
||||
abort(404, 'Image %(image_id)s not found', issue='unknown-image', image_id=image_id)
|
||||
|
||||
# Add the Accept-Ranges header if the storage engine supports resumable
|
||||
# downloads.
|
||||
extra_headers = {}
|
||||
if store.get_supports_resumable_downloads(locations):
|
||||
if store.get_supports_resumable_downloads(legacy_image.blob.placements):
|
||||
logger.debug('Storage supports resumable downloads')
|
||||
extra_headers['Accept-Ranges'] = 'bytes'
|
||||
|
||||
|
@ -114,21 +112,23 @@ def head_image_layer(namespace, repository, image_id, headers):
|
|||
@anon_protect
|
||||
def get_image_layer(namespace, repository, image_id, headers):
|
||||
permission = ReadRepositoryPermission(namespace, repository)
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
|
||||
logger.debug('Checking repo permissions')
|
||||
if permission.can() or model.repository_is_public(namespace, repository):
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
if permission.can() or (repository_ref is not None and repository_ref.is_public):
|
||||
if repository_ref is None:
|
||||
abort(404)
|
||||
|
||||
logger.debug('Looking up placement locations and path')
|
||||
locations, path = model.placement_locations_and_path_docker_v1(namespace, repository, image_id)
|
||||
if not locations or not path:
|
||||
legacy_image = registry_model.get_legacy_image(repository_ref, image_id, include_blob=True)
|
||||
if legacy_image is None:
|
||||
abort(404, 'Image %(image_id)s not found', issue='unknown-image', image_id=image_id)
|
||||
|
||||
path = legacy_image.blob.storage_path
|
||||
|
||||
try:
|
||||
logger.debug('Looking up the direct download URL for path: %s', path)
|
||||
direct_download_url = store.get_direct_download_url(locations, path, request.remote_addr)
|
||||
direct_download_url = store.get_direct_download_url(legacy_image.blob.placements, path,
|
||||
request.remote_addr)
|
||||
if direct_download_url:
|
||||
logger.debug('Returning direct download URL')
|
||||
resp = redirect(direct_download_url)
|
||||
|
@ -137,7 +137,7 @@ def get_image_layer(namespace, repository, image_id, headers):
|
|||
# Close the database handle here for this process before we send the long download.
|
||||
database.close_db_filter(None)
|
||||
logger.debug('Streaming layer data')
|
||||
return Response(store.stream_read(locations, path), headers=headers)
|
||||
return Response(store.stream_read(legacy_image.blob.placements, path), headers=headers)
|
||||
except (IOError, AttributeError):
|
||||
logger.exception('Image layer data not found')
|
||||
abort(404, 'Image %(image_id)s not found', issue='unknown-image', image_id=image_id)
|
||||
|
@ -156,31 +156,31 @@ def put_image_layer(namespace, repository, image_id):
|
|||
if not permission.can():
|
||||
abort(403)
|
||||
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
if repository_ref is None:
|
||||
abort(403)
|
||||
|
||||
logger.debug('Retrieving image')
|
||||
if model.storage_exists(namespace, repository, image_id):
|
||||
legacy_image = registry_model.get_legacy_image(repository_ref, image_id)
|
||||
if legacy_image is not None and not legacy_image.uploading:
|
||||
exact_abort(409, 'Image already exists')
|
||||
|
||||
v1_metadata = model.docker_v1_metadata(namespace, repository, image_id)
|
||||
if v1_metadata is None:
|
||||
logger.debug('Checking for image in manifest builder')
|
||||
builder = lookup_manifest_builder(repository_ref, session.get('manifest_builder'))
|
||||
if builder is None:
|
||||
abort(400)
|
||||
|
||||
layer = builder.lookup_layer(image_id)
|
||||
if layer is None:
|
||||
abort(404)
|
||||
|
||||
logger.debug('Storing layer data')
|
||||
|
||||
input_stream = request.stream
|
||||
if request.headers.get('transfer-encoding') == 'chunked':
|
||||
# Careful, might work only with WSGI servers supporting chunked
|
||||
# encoding (Gunicorn)
|
||||
input_stream = request.environ['wsgi.input']
|
||||
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository)
|
||||
if repository_ref is None:
|
||||
abort(404)
|
||||
|
||||
expiration_sec = app.config['PUSH_TEMP_TAG_EXPIRATION_SEC']
|
||||
settings = BlobUploadSettings(maximum_blob_size=app.config['MAXIMUM_LAYER_SIZE'],
|
||||
bittorrent_piece_size=app.config['BITTORRENT_PIECE_SIZE'],
|
||||
|
@ -190,13 +190,13 @@ def put_image_layer(namespace, repository, image_id):
|
|||
|
||||
# Add a handler that copies the data into a temp file. This is used to calculate the tarsum,
|
||||
# which is only needed for older versions of Docker.
|
||||
requires_tarsum = session.get('checksum_format') == 'tarsum'
|
||||
requires_tarsum = bool(builder.get_layer_checksums(layer))
|
||||
if requires_tarsum:
|
||||
tmp, tmp_hndlr = store.temp_store_handler()
|
||||
extra_handlers.append(tmp_hndlr)
|
||||
|
||||
# Add a handler which computes the simple Docker V1 checksum.
|
||||
h, sum_hndlr = checksums.simple_checksum_handler(v1_metadata.compat_json)
|
||||
h, sum_hndlr = checksums.simple_checksum_handler(layer.v1_metadata_string)
|
||||
extra_handlers.append(sum_hndlr)
|
||||
|
||||
uploaded_blob = None
|
||||
|
@ -209,37 +209,34 @@ def put_image_layer(namespace, repository, image_id):
|
|||
logger.exception('Exception when writing image data')
|
||||
abort(520, 'Image %(image_id)s could not be written. Please try again.', image_id=image_id)
|
||||
|
||||
# Save the blob for the image.
|
||||
model.update_image_blob(namespace, repository, image_id, uploaded_blob)
|
||||
|
||||
# Send a job to the work queue to replicate the image layer.
|
||||
# TODO: move this into a better place.
|
||||
queue_storage_replication(namespace, uploaded_blob)
|
||||
|
||||
# Append the computed checksum.
|
||||
# Compute the final checksum
|
||||
csums = []
|
||||
csums.append('sha256:{0}'.format(h.hexdigest()))
|
||||
|
||||
try:
|
||||
if requires_tarsum:
|
||||
tmp.seek(0)
|
||||
csums.append(checksums.compute_tarsum(tmp, v1_metadata.compat_json))
|
||||
csums.append(checksums.compute_tarsum(tmp, layer.v1_metadata_string))
|
||||
tmp.close()
|
||||
except (IOError, checksums.TarError) as exc:
|
||||
logger.debug('put_image_layer: Error when computing tarsum %s', exc)
|
||||
|
||||
if v1_metadata.checksum is None:
|
||||
# We don't have a checksum stored yet, that's fine skipping the check.
|
||||
# Not removing the mark though, image is not downloadable yet.
|
||||
session['checksum'] = csums
|
||||
session['content_checksum'] = uploaded_blob.digest
|
||||
return make_response('true', 200)
|
||||
# If there was already a precomputed checksum, validate against it now.
|
||||
if builder.get_layer_checksums(layer):
|
||||
checksum = builder.get_layer_checksums(layer)[0]
|
||||
if not builder.validate_layer_checksum(layer, checksum):
|
||||
logger.debug('put_image_checksum: Wrong checksum. Given: %s and expected: %s', checksum,
|
||||
builder.get_layer_checksums(layer))
|
||||
abort(400, 'Checksum mismatch for image: %(image_id)s', issue='checksum-mismatch',
|
||||
image_id=image_id)
|
||||
|
||||
# We check if the checksums provided matches one the one we computed
|
||||
if v1_metadata.checksum not in csums:
|
||||
logger.warning('put_image_layer: Wrong checksum')
|
||||
abort(400, 'Checksum mismatch; ignoring the layer for image %(image_id)s',
|
||||
issue='checksum-mismatch', image_id=image_id)
|
||||
# Assign the blob to the layer in the manifest.
|
||||
if not builder.assign_layer_blob(layer, uploaded_blob, csums):
|
||||
abort(500, 'Something went wrong')
|
||||
|
||||
# Send a job to the work queue to replicate the image layer.
|
||||
# TODO: move this into a better place.
|
||||
queue_storage_replication(namespace, uploaded_blob)
|
||||
|
||||
return make_response('true', 200)
|
||||
|
||||
|
@ -255,10 +252,9 @@ def put_image_checksum(namespace, repository, image_id):
|
|||
if not permission.can():
|
||||
abort(403)
|
||||
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
if repository_ref is None:
|
||||
abort(403)
|
||||
|
||||
# Docker Version < 0.10 (tarsum+sha):
|
||||
old_checksum = request.headers.get('X-Docker-Checksum')
|
||||
|
@ -266,40 +262,27 @@ def put_image_checksum(namespace, repository, image_id):
|
|||
# Docker Version >= 0.10 (sha):
|
||||
new_checksum = request.headers.get('X-Docker-Checksum-Payload')
|
||||
|
||||
# Store whether we need to calculate the tarsum.
|
||||
if new_checksum:
|
||||
session['checksum_format'] = 'sha256'
|
||||
else:
|
||||
session['checksum_format'] = 'tarsum'
|
||||
|
||||
checksum = new_checksum or old_checksum
|
||||
if not checksum:
|
||||
abort(400, "Missing checksum for image %(image_id)s", issue='missing-checksum',
|
||||
image_id=image_id)
|
||||
|
||||
if not session.get('checksum'):
|
||||
abort(400, 'Checksum not found in Cookie for image %(image_id)s',
|
||||
issue='missing-checksum-cookie', image_id=image_id)
|
||||
logger.debug('Checking for image in manifest builder')
|
||||
builder = lookup_manifest_builder(repository_ref, session.get('manifest_builder'))
|
||||
if builder is None:
|
||||
abort(400)
|
||||
|
||||
logger.debug('Looking up repo image')
|
||||
v1_metadata = model.docker_v1_metadata(namespace, repository, image_id)
|
||||
if not v1_metadata:
|
||||
abort(404, 'Image not found: %(image_id)s', issue='unknown-image', image_id=image_id)
|
||||
layer = builder.lookup_layer(image_id)
|
||||
if layer is None:
|
||||
abort(404)
|
||||
|
||||
logger.debug('Looking up repo layer data')
|
||||
if not v1_metadata.compat_json:
|
||||
abort(404, 'Image not found: %(image_id)s', issue='unknown-image', image_id=image_id)
|
||||
if old_checksum:
|
||||
builder.save_precomputed_checksum(layer, checksum)
|
||||
return make_response('true', 200)
|
||||
|
||||
logger.debug('Storing image and checksum')
|
||||
content_checksum = session.get('content_checksum', None)
|
||||
checksum_parts = checksum.split(':')
|
||||
if len(checksum_parts) != 2:
|
||||
abort(400, 'Invalid checksum format')
|
||||
|
||||
if checksum not in session.get('checksum', []):
|
||||
logger.debug('session checksums: %s', session.get('checksum', []))
|
||||
logger.debug('client supplied checksum: %s', checksum)
|
||||
logger.debug('put_image_checksum: Wrong checksum')
|
||||
if not builder.validate_layer_checksum(layer, checksum):
|
||||
logger.debug('put_image_checksum: Wrong checksum. Given: %s and expected: %s', checksum,
|
||||
builder.get_layer_checksums(layer))
|
||||
abort(400, 'Checksum mismatch for image: %(image_id)s', issue='checksum-mismatch',
|
||||
image_id=image_id)
|
||||
|
||||
|
@ -316,27 +299,22 @@ def put_image_checksum(namespace, repository, image_id):
|
|||
def get_image_json(namespace, repository, image_id, headers):
|
||||
logger.debug('Checking repo permissions')
|
||||
permission = ReadRepositoryPermission(namespace, repository)
|
||||
if not permission.can() and not model.repository_is_public(namespace, repository):
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
if not permission.can() and not (repository_ref is not None and repository_ref.is_public):
|
||||
abort(403)
|
||||
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
|
||||
logger.debug('Looking up repo image')
|
||||
v1_metadata = model.docker_v1_metadata(namespace, repository, image_id)
|
||||
if v1_metadata is None:
|
||||
legacy_image = registry_model.get_legacy_image(repository_ref, image_id, include_blob=True)
|
||||
if legacy_image is None:
|
||||
flask_abort(404)
|
||||
|
||||
logger.debug('Looking up repo layer size')
|
||||
size = model.get_image_size(namespace, repository, image_id)
|
||||
size = legacy_image.blob.compressed_size
|
||||
if size is not None:
|
||||
# Note: X-Docker-Size is optional and we *can* end up with a NULL image_size,
|
||||
# so handle this case rather than failing.
|
||||
headers['X-Docker-Size'] = str(size)
|
||||
|
||||
response = make_response(v1_metadata.compat_json, 200)
|
||||
response = make_response(legacy_image.v1_metadata_string, 200)
|
||||
response.headers.extend(headers)
|
||||
return response
|
||||
|
||||
|
@ -351,20 +329,19 @@ def get_image_json(namespace, repository, image_id, headers):
|
|||
def get_image_ancestry(namespace, repository, image_id, headers):
|
||||
logger.debug('Checking repo permissions')
|
||||
permission = ReadRepositoryPermission(namespace, repository)
|
||||
if not permission.can() and not model.repository_is_public(namespace, repository):
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
if not permission.can() and not (repository_ref is not None and repository_ref.is_public):
|
||||
abort(403)
|
||||
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
|
||||
ancestry_docker_ids = model.image_ancestry(namespace, repository, image_id)
|
||||
if ancestry_docker_ids is None:
|
||||
logger.debug('Looking up repo image')
|
||||
legacy_image = registry_model.get_legacy_image(repository_ref, image_id, include_parents=True)
|
||||
if legacy_image is None:
|
||||
abort(404, 'Image %(image_id)s not found', issue='unknown-image', image_id=image_id)
|
||||
|
||||
# We can not use jsonify here because we are returning a list not an object
|
||||
response = make_response(json.dumps(ancestry_docker_ids), 200)
|
||||
# NOTE: We can not use jsonify here because we are returning a list not an object.
|
||||
ancestor_ids = ([legacy_image.docker_image_id] +
|
||||
[a.docker_image_id for a in legacy_image.parents])
|
||||
response = make_response(json.dumps(ancestor_ids), 200)
|
||||
response.headers.extend(headers)
|
||||
return response
|
||||
|
||||
|
@ -380,10 +357,13 @@ def put_image_json(namespace, repository, image_id):
|
|||
if not permission.can():
|
||||
abort(403)
|
||||
|
||||
repo = model.get_repository(namespace, repository)
|
||||
if repo.kind != 'image':
|
||||
msg = 'This repository is for managing %s resources and not container images.' % repo.kind
|
||||
abort(405, message=msg, image_id=image_id)
|
||||
repository_ref = registry_model.lookup_repository(namespace, repository, kind_filter='image')
|
||||
if repository_ref is None:
|
||||
abort(403)
|
||||
|
||||
builder = lookup_manifest_builder(repository_ref, session.get('manifest_builder'))
|
||||
if builder is None:
|
||||
abort(400)
|
||||
|
||||
logger.debug('Parsing image JSON')
|
||||
try:
|
||||
|
@ -405,46 +385,12 @@ def put_image_json(namespace, repository, image_id):
|
|||
image_id=image_id)
|
||||
|
||||
logger.debug('Looking up repo image')
|
||||
|
||||
if not model.repository_exists(namespace, repository):
|
||||
abort(404, 'Repository does not exist: %(namespace)s/%(repository)s', issue='no-repo',
|
||||
namespace=namespace, repository=repository)
|
||||
|
||||
parent_id = data.get('parent', None)
|
||||
if parent_id:
|
||||
logger.debug('Looking up parent image')
|
||||
if model.docker_v1_metadata(namespace, repository, parent_id) is None:
|
||||
abort(400, 'Image %(image_id)s depends on non existing parent image %(parent_id)s',
|
||||
issue='invalid-request', image_id=image_id, parent_id=parent_id)
|
||||
|
||||
v1_metadata = model.docker_v1_metadata(namespace, repository, image_id)
|
||||
if v1_metadata is None:
|
||||
username = get_authenticated_user() and get_authenticated_user().username
|
||||
logger.debug('Image not found, creating or linking image with initiating user context: %s',
|
||||
username)
|
||||
location_pref = store.preferred_locations[0]
|
||||
model.create_or_link_image(username, namespace, repository, image_id, location_pref)
|
||||
v1_metadata = model.docker_v1_metadata(namespace, repository, image_id)
|
||||
|
||||
# Create a temporary tag to prevent this image from getting garbage collected while the push
|
||||
# is in progress.
|
||||
model.create_temp_hidden_tag(namespace, repository, image_id,
|
||||
app.config['PUSH_TEMP_TAG_EXPIRATION_SEC'])
|
||||
|
||||
logger.debug('Checking if image already exists')
|
||||
if v1_metadata and not model.is_image_uploading(namespace, repository, image_id):
|
||||
exact_abort(409, 'Image already exists')
|
||||
|
||||
model.update_image_uploading(namespace, repository, image_id, True)
|
||||
|
||||
# If we reach that point, it means that this is a new image or a retry
|
||||
# on a failed push, save the metadata
|
||||
command_list = data.get('container_config', {}).get('Cmd', None)
|
||||
command = json.dumps(command_list) if command_list else None
|
||||
|
||||
logger.debug('Setting image metadata')
|
||||
model.update_docker_v1_metadata(namespace, repository, image_id,
|
||||
data.get('created'),
|
||||
data.get('comment'), command, uploaded_metadata, parent_id)
|
||||
location_pref = store.preferred_locations[0]
|
||||
username = get_authenticated_user() and get_authenticated_user().username
|
||||
layer = builder.start_layer(image_id, uploaded_metadata, location_pref, username,
|
||||
app.config['PUSH_TEMP_TAG_EXPIRATION_SEC'])
|
||||
if layer is None:
|
||||
abort(400, 'Image %(image_id)s has invalid metadata',
|
||||
issue='invalid-request', image_id=image_id)
|
||||
|
||||
return make_response('true', 200)
|
||||
|
|
Reference in a new issue