2016-11-22 19:32:10 +00:00
package copy
import (
"bytes"
"compress/gzip"
"fmt"
"io"
"io/ioutil"
"reflect"
2017-05-17 17:18:35 +00:00
"strings"
2017-03-13 16:33:17 +00:00
"time"
2016-11-22 19:32:10 +00:00
pb "gopkg.in/cheggaaa/pb.v1"
"github.com/Sirupsen/logrus"
"github.com/containers/image/image"
2017-03-13 16:33:17 +00:00
"github.com/containers/image/pkg/compression"
2016-11-22 19:32:10 +00:00
"github.com/containers/image/signature"
"github.com/containers/image/transports"
"github.com/containers/image/types"
2016-10-17 13:53:40 +00:00
"github.com/opencontainers/go-digest"
"github.com/pkg/errors"
2016-11-22 19:32:10 +00:00
)
type digestingReader struct {
source io . Reader
digester digest . Digester
expectedDigest digest . Digest
validationFailed bool
}
// imageCopier allows us to keep track of diffID values for blobs, and other
// data, that we're copying between images, and cache other information that
// might allow us to take some shortcuts
type imageCopier struct {
copiedBlobs map [ digest . Digest ] digest . Digest
cachedDiffIDs map [ digest . Digest ] digest . Digest
manifestUpdates * types . ManifestUpdateOptions
dest types . ImageDestination
src types . Image
rawSource types . ImageSource
diffIDsAreNeeded bool
canModifyManifest bool
reportWriter io . Writer
2017-03-13 16:33:17 +00:00
progressInterval time . Duration
progress chan types . ProgressProperties
2016-11-22 19:32:10 +00:00
}
// newDigestingReader returns an io.Reader implementation with contents of source, which will eventually return a non-EOF error
// and set validationFailed to true if the source stream does not match expectedDigest.
func newDigestingReader ( source io . Reader , expectedDigest digest . Digest ) ( * digestingReader , error ) {
if err := expectedDigest . Validate ( ) ; err != nil {
2016-10-17 13:53:40 +00:00
return nil , errors . Errorf ( "Invalid digest specification %s" , expectedDigest )
2016-11-22 19:32:10 +00:00
}
digestAlgorithm := expectedDigest . Algorithm ( )
if ! digestAlgorithm . Available ( ) {
2016-10-17 13:53:40 +00:00
return nil , errors . Errorf ( "Invalid digest specification %s: unsupported digest algorithm %s" , expectedDigest , digestAlgorithm )
2016-11-22 19:32:10 +00:00
}
return & digestingReader {
source : source ,
2016-10-17 13:53:40 +00:00
digester : digestAlgorithm . Digester ( ) ,
2016-11-22 19:32:10 +00:00
expectedDigest : expectedDigest ,
validationFailed : false ,
} , nil
}
func ( d * digestingReader ) Read ( p [ ] byte ) ( int , error ) {
n , err := d . source . Read ( p )
if n > 0 {
if n2 , err := d . digester . Hash ( ) . Write ( p [ : n ] ) ; n2 != n || err != nil {
// Coverage: This should not happen, the hash.Hash interface requires
// d.digest.Write to never return an error, and the io.Writer interface
// requires n2 == len(input) if no error is returned.
2016-10-17 13:53:40 +00:00
return 0 , errors . Wrapf ( err , "Error updating digest during verification: %d vs. %d" , n2 , n )
2016-11-22 19:32:10 +00:00
}
}
if err == io . EOF {
actualDigest := d . digester . Digest ( )
if actualDigest != d . expectedDigest {
d . validationFailed = true
2016-10-17 13:53:40 +00:00
return 0 , errors . Errorf ( "Digest did not match, expected %s, got %s" , d . expectedDigest , actualDigest )
2016-11-22 19:32:10 +00:00
}
}
return n , err
}
// Options allows supplying non-default configuration modifying the behavior of CopyImage.
type Options struct {
RemoveSignatures bool // Remove any pre-existing signatures. SignBy will still add a new signature.
SignBy string // If non-empty, asks for a signature to be added during the copy, and specifies a key ID, as accepted by signature.NewGPGSigningMechanism().SignDockerManifest(),
ReportWriter io . Writer
SourceCtx * types . SystemContext
DestinationCtx * types . SystemContext
2017-03-13 16:33:17 +00:00
ProgressInterval time . Duration // time to wait between reports to signal the progress channel
Progress chan types . ProgressProperties // Reported to when ProgressInterval has arrived for a single artifact+offset.
2016-11-22 19:32:10 +00:00
}
2017-03-13 16:33:17 +00:00
// Image copies image from srcRef to destRef, using policyContext to validate
// source image admissibility.
func Image ( policyContext * signature . PolicyContext , destRef , srcRef types . ImageReference , options * Options ) ( retErr error ) {
// NOTE this function uses an output parameter for the error return value.
// Setting this and returning is the ideal way to return an error.
//
// the defers in this routine will wrap the error return with its own errors
// which can be valuable context in the middle of a multi-streamed copy.
if options == nil {
options = & Options { }
}
2016-11-22 19:32:10 +00:00
reportWriter := ioutil . Discard
2017-03-13 16:33:17 +00:00
if options . ReportWriter != nil {
2016-11-22 19:32:10 +00:00
reportWriter = options . ReportWriter
}
2017-03-13 16:33:17 +00:00
2016-11-22 19:32:10 +00:00
writeReport := func ( f string , a ... interface { } ) {
fmt . Fprintf ( reportWriter , f , a ... )
}
dest , err := destRef . NewImageDestination ( options . DestinationCtx )
if err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrapf ( err , "Error initializing destination %s" , transports . ImageName ( destRef ) )
2016-11-22 19:32:10 +00:00
}
2017-03-13 16:33:17 +00:00
defer func ( ) {
if err := dest . Close ( ) ; err != nil {
retErr = errors . Wrapf ( retErr , " (dest: %v)" , err )
}
} ( )
2016-11-22 19:32:10 +00:00
destSupportedManifestMIMETypes := dest . SupportedManifestMIMETypes ( )
rawSource , err := srcRef . NewImageSource ( options . SourceCtx , destSupportedManifestMIMETypes )
if err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrapf ( err , "Error initializing source %s" , transports . ImageName ( srcRef ) )
2016-11-22 19:32:10 +00:00
}
unparsedImage := image . UnparsedFromSource ( rawSource )
defer func ( ) {
if unparsedImage != nil {
2017-03-13 16:33:17 +00:00
if err := unparsedImage . Close ( ) ; err != nil {
retErr = errors . Wrapf ( retErr , " (unparsed: %v)" , err )
}
2016-11-22 19:32:10 +00:00
}
} ( )
// Please keep this policy check BEFORE reading any other information about the image.
if allowed , err := policyContext . IsRunningImageAllowed ( unparsedImage ) ; ! allowed || err != nil { // Be paranoid and fail if either return value indicates so.
2016-10-17 13:53:40 +00:00
return errors . Wrap ( err , "Source image rejected" )
2016-11-22 19:32:10 +00:00
}
src , err := image . FromUnparsedImage ( unparsedImage )
if err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrapf ( err , "Error initializing image from source %s" , transports . ImageName ( srcRef ) )
2016-11-22 19:32:10 +00:00
}
unparsedImage = nil
2017-03-13 16:33:17 +00:00
defer func ( ) {
if err := src . Close ( ) ; err != nil {
retErr = errors . Wrapf ( retErr , " (source: %v)" , err )
}
} ( )
2016-11-22 19:32:10 +00:00
if src . IsMultiImage ( ) {
2016-10-17 13:53:40 +00:00
return errors . Errorf ( "can not copy %s: manifest contains multiple images" , transports . ImageName ( srcRef ) )
2016-11-22 19:32:10 +00:00
}
var sigs [ ] [ ] byte
2017-03-13 16:33:17 +00:00
if options . RemoveSignatures {
2016-11-22 19:32:10 +00:00
sigs = [ ] [ ] byte { }
} else {
writeReport ( "Getting image source signatures\n" )
s , err := src . Signatures ( )
if err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrap ( err , "Error reading signatures" )
2016-11-22 19:32:10 +00:00
}
sigs = s
}
if len ( sigs ) != 0 {
writeReport ( "Checking if image destination supports signatures\n" )
if err := dest . SupportsSignatures ( ) ; err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrap ( err , "Can not copy signatures" )
2016-11-22 19:32:10 +00:00
}
}
canModifyManifest := len ( sigs ) == 0
manifestUpdates := types . ManifestUpdateOptions { }
2017-05-17 17:18:35 +00:00
manifestUpdates . InformationOnly . Destination = dest
2016-11-22 19:32:10 +00:00
2017-05-17 17:18:35 +00:00
if err := updateEmbeddedDockerReference ( & manifestUpdates , dest , src , canModifyManifest ) ; err != nil {
return err
}
// We compute preferredManifestMIMEType only to show it in error messages.
// Without having to add this context in an error message, we would be happy enough to know only that no conversion is needed.
preferredManifestMIMEType , otherManifestMIMETypeCandidates , err := determineManifestConversion ( & manifestUpdates , src , destSupportedManifestMIMETypes , canModifyManifest )
if err != nil {
2016-11-22 19:32:10 +00:00
return err
}
// If src.UpdatedImageNeedsLayerDiffIDs(manifestUpdates) will be true, it needs to be true by the time we get here.
ic := imageCopier {
copiedBlobs : make ( map [ digest . Digest ] digest . Digest ) ,
cachedDiffIDs : make ( map [ digest . Digest ] digest . Digest ) ,
manifestUpdates : & manifestUpdates ,
dest : dest ,
src : src ,
rawSource : rawSource ,
diffIDsAreNeeded : src . UpdatedImageNeedsLayerDiffIDs ( manifestUpdates ) ,
canModifyManifest : canModifyManifest ,
reportWriter : reportWriter ,
2017-03-13 16:33:17 +00:00
progressInterval : options . ProgressInterval ,
progress : options . Progress ,
2016-11-22 19:32:10 +00:00
}
if err := ic . copyLayers ( ) ; err != nil {
return err
}
2017-05-17 17:18:35 +00:00
// With docker/distribution registries we do not know whether the registry accepts schema2 or schema1 only;
// and at least with the OpenShift registry "acceptschema2" option, there is no way to detect the support
// without actually trying to upload something and getting a types.ManifestTypeRejectedError.
// So, try the preferred manifest MIME type. If the process succeeds, fine…
manifest , err := ic . copyUpdatedConfigAndManifest ( )
if err != nil {
logrus . Debugf ( "Writing manifest using preferred type %s failed: %v" , preferredManifestMIMEType , err )
// … if it fails, _and_ the failure is because the manifest is rejected, we may have other options.
if _ , isManifestRejected := errors . Cause ( err ) . ( types . ManifestTypeRejectedError ) ; ! isManifestRejected || len ( otherManifestMIMETypeCandidates ) == 0 {
// We don’ t have other options.
// In principle the code below would handle this as well, but the resulting error message is fairly ugly.
// Don’ t bother the user with MIME types if we have no choice.
return err
2016-11-22 19:32:10 +00:00
}
2017-05-17 17:18:35 +00:00
// If the original MIME type is acceptable, determineManifestConversion always uses it as preferredManifestMIMEType.
// So if we are here, we will definitely be trying to convert the manifest.
// With !canModifyManifest, that would just be a string of repeated failures for the same reason,
// so let’ s bail out early and with a better error message.
if ! canModifyManifest {
return errors . Wrap ( err , "Writing manifest failed (and converting it is not possible)" )
2016-11-22 19:32:10 +00:00
}
2017-05-17 17:18:35 +00:00
// errs is a list of errors when trying various manifest types. Also serves as an "upload succeeded" flag when set to nil.
errs := [ ] string { fmt . Sprintf ( "%s(%v)" , preferredManifestMIMEType , err ) }
for _ , manifestMIMEType := range otherManifestMIMETypeCandidates {
logrus . Debugf ( "Trying to use manifest type %s…" , manifestMIMEType )
manifestUpdates . ManifestMIMEType = manifestMIMEType
attemptedManifest , err := ic . copyUpdatedConfigAndManifest ( )
if err != nil {
logrus . Debugf ( "Upload of manifest type %s failed: %v" , manifestMIMEType , err )
errs = append ( errs , fmt . Sprintf ( "%s(%v)" , manifestMIMEType , err ) )
continue
}
2016-11-22 19:32:10 +00:00
2017-05-17 17:18:35 +00:00
// We have successfully uploaded a manifest.
manifest = attemptedManifest
errs = nil // Mark this as a success so that we don't abort below.
break
2017-04-03 07:22:44 +00:00
}
2017-05-17 17:18:35 +00:00
if errs != nil {
return fmt . Errorf ( "Uploading manifest failed, attempted the following formats: %s" , strings . Join ( errs , ", " ) )
2016-11-22 19:32:10 +00:00
}
2017-05-17 17:18:35 +00:00
}
2016-11-22 19:32:10 +00:00
2017-05-17 17:18:35 +00:00
if options . SignBy != "" {
newSig , err := createSignature ( dest , manifest , options . SignBy , reportWriter )
2016-11-22 19:32:10 +00:00
if err != nil {
2017-05-17 17:18:35 +00:00
return err
2016-11-22 19:32:10 +00:00
}
sigs = append ( sigs , newSig )
}
writeReport ( "Storing signatures\n" )
if err := dest . PutSignatures ( sigs ) ; err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrap ( err , "Error writing signatures" )
2016-11-22 19:32:10 +00:00
}
if err := dest . Commit ( ) ; err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrap ( err , "Error committing the finished image" )
2016-11-22 19:32:10 +00:00
}
return nil
}
2017-05-17 17:18:35 +00:00
// updateEmbeddedDockerReference handles the Docker reference embedded in Docker schema1 manifests.
func updateEmbeddedDockerReference ( manifestUpdates * types . ManifestUpdateOptions , dest types . ImageDestination , src types . Image , canModifyManifest bool ) error {
destRef := dest . Reference ( ) . DockerReference ( )
if destRef == nil {
return nil // Destination does not care about Docker references
}
if ! src . EmbeddedDockerReferenceConflicts ( destRef ) {
return nil // No reference embedded in the manifest, or it matches destRef already.
}
if ! canModifyManifest {
return errors . Errorf ( " Copying a schema1 image with an embedded Docker reference to % s ( Docker reference % s ) would invalidate existing signatures . Explicitly enable signature removal to proceed anyway " ,
transports . ImageName ( dest . Reference ( ) ) , destRef . String ( ) )
}
manifestUpdates . EmbeddedDockerReference = destRef
return nil
}
2016-11-22 19:32:10 +00:00
// copyLayers copies layers from src/rawSource to dest, using and updating ic.manifestUpdates if necessary and ic.canModifyManifest.
func ( ic * imageCopier ) copyLayers ( ) error {
srcInfos := ic . src . LayerInfos ( )
destInfos := [ ] types . BlobInfo { }
diffIDs := [ ] digest . Digest { }
for _ , srcLayer := range srcInfos {
var (
destInfo types . BlobInfo
diffID digest . Digest
err error
)
if ic . dest . AcceptsForeignLayerURLs ( ) && len ( srcLayer . URLs ) != 0 {
// DiffIDs are, currently, needed only when converting from schema1.
// In which case src.LayerInfos will not have URLs because schema1
// does not support them.
if ic . diffIDsAreNeeded {
return errors . New ( "getting DiffID for foreign layers is unimplemented" )
}
destInfo = srcLayer
fmt . Fprintf ( ic . reportWriter , "Skipping foreign layer %q copy to %s\n" , destInfo . Digest , ic . dest . Reference ( ) . Transport ( ) . Name ( ) )
} else {
destInfo , diffID , err = ic . copyLayer ( srcLayer )
if err != nil {
return err
}
}
destInfos = append ( destInfos , destInfo )
diffIDs = append ( diffIDs , diffID )
}
ic . manifestUpdates . InformationOnly . LayerInfos = destInfos
if ic . diffIDsAreNeeded {
ic . manifestUpdates . InformationOnly . LayerDiffIDs = diffIDs
}
if layerDigestsDiffer ( srcInfos , destInfos ) {
ic . manifestUpdates . LayerInfos = destInfos
}
return nil
}
// layerDigestsDiffer return true iff the digests in a and b differ (ignoring sizes and possible other fields)
func layerDigestsDiffer ( a , b [ ] types . BlobInfo ) bool {
if len ( a ) != len ( b ) {
return true
}
for i := range a {
if a [ i ] . Digest != b [ i ] . Digest {
return true
}
}
return false
}
2017-05-17 17:18:35 +00:00
// copyUpdatedConfigAndManifest updates the image per ic.manifestUpdates, if necessary,
// stores the resulting config and manifest to the destination, and returns the stored manifest.
func ( ic * imageCopier ) copyUpdatedConfigAndManifest ( ) ( [ ] byte , error ) {
pendingImage := ic . src
if ! reflect . DeepEqual ( * ic . manifestUpdates , types . ManifestUpdateOptions { InformationOnly : ic . manifestUpdates . InformationOnly } ) {
if ! ic . canModifyManifest {
return nil , errors . Errorf ( "Internal error: copy needs an updated manifest but that was known to be forbidden" )
}
if ! ic . diffIDsAreNeeded && ic . src . UpdatedImageNeedsLayerDiffIDs ( * ic . manifestUpdates ) {
// We have set ic.diffIDsAreNeeded based on the preferred MIME type returned by determineManifestConversion.
// So, this can only happen if we are trying to upload using one of the other MIME type candidates.
// Because UpdatedImageNeedsLayerDiffIDs is true only when converting from s1 to s2, this case should only arise
// when ic.dest.SupportedManifestMIMETypes() includes both s1 and s2, the upload using s1 failed, and we are now trying s2.
// Supposedly s2-only registries do not exist or are extremely rare, so failing with this error message is good enough for now.
// If handling such registries turns out to be necessary, we could compute ic.diffIDsAreNeeded based on the full list of manifest MIME type candidates.
return nil , errors . Errorf ( "Can not convert image to %s, preparing DiffIDs for this case is not supported" , ic . manifestUpdates . ManifestMIMEType )
}
pi , err := ic . src . UpdatedImage ( * ic . manifestUpdates )
if err != nil {
return nil , errors . Wrap ( err , "Error creating an updated image manifest" )
}
pendingImage = pi
}
manifest , _ , err := pendingImage . Manifest ( )
if err != nil {
return nil , errors . Wrap ( err , "Error reading manifest" )
}
if err := ic . copyConfig ( pendingImage ) ; err != nil {
return nil , err
}
fmt . Fprintf ( ic . reportWriter , "Writing manifest to image destination\n" )
if err := ic . dest . PutManifest ( manifest ) ; err != nil {
return nil , errors . Wrap ( err , "Error writing manifest" )
}
return manifest , nil
}
2016-11-22 19:32:10 +00:00
// copyConfig copies config.json, if any, from src to dest.
func ( ic * imageCopier ) copyConfig ( src types . Image ) error {
srcInfo := src . ConfigInfo ( )
if srcInfo . Digest != "" {
fmt . Fprintf ( ic . reportWriter , "Copying config %s\n" , srcInfo . Digest )
configBlob , err := src . ConfigBlob ( )
if err != nil {
2016-10-17 13:53:40 +00:00
return errors . Wrapf ( err , "Error reading config blob %s" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
destInfo , err := ic . copyBlobFromStream ( bytes . NewReader ( configBlob ) , srcInfo , nil , false )
if err != nil {
return err
}
if destInfo . Digest != srcInfo . Digest {
2016-10-17 13:53:40 +00:00
return errors . Errorf ( "Internal error: copying uncompressed config blob %s changed digest to %s" , srcInfo . Digest , destInfo . Digest )
2016-11-22 19:32:10 +00:00
}
}
return nil
}
// diffIDResult contains both a digest value and an error from diffIDComputationGoroutine.
// We could also send the error through the pipeReader, but this more cleanly separates the copying of the layer and the DiffID computation.
type diffIDResult struct {
digest digest . Digest
err error
}
// copyLayer copies a layer with srcInfo (with known Digest and possibly known Size) in src to dest, perhaps compressing it if canCompress,
// and returns a complete blobInfo of the copied layer, and a value for LayerDiffIDs if diffIDIsNeeded
func ( ic * imageCopier ) copyLayer ( srcInfo types . BlobInfo ) ( types . BlobInfo , digest . Digest , error ) {
// Check if we already have a blob with this digest
haveBlob , extantBlobSize , err := ic . dest . HasBlob ( srcInfo )
2017-04-03 07:22:44 +00:00
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , "" , errors . Wrapf ( err , "Error checking for blob %s at destination" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
// If we already have a cached diffID for this blob, we don't need to compute it
diffIDIsNeeded := ic . diffIDsAreNeeded && ( ic . cachedDiffIDs [ srcInfo . Digest ] == "" )
// If we already have the blob, and we don't need to recompute the diffID, then we might be able to avoid reading it again
if haveBlob && ! diffIDIsNeeded {
// Check the blob sizes match, if we were given a size this time
if srcInfo . Size != - 1 && srcInfo . Size != extantBlobSize {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , "" , errors . Errorf ( "Error: blob %s is already present, but with size %d instead of %d" , srcInfo . Digest , extantBlobSize , srcInfo . Size )
2016-11-22 19:32:10 +00:00
}
srcInfo . Size = extantBlobSize
// Tell the image destination that this blob's delta is being applied again. For some image destinations, this can be faster than using GetBlob/PutBlob
blobinfo , err := ic . dest . ReapplyBlob ( srcInfo )
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , "" , errors . Wrapf ( err , "Error reapplying blob %s at destination" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
fmt . Fprintf ( ic . reportWriter , "Skipping fetch of repeat blob %s\n" , srcInfo . Digest )
return blobinfo , ic . cachedDiffIDs [ srcInfo . Digest ] , err
}
// Fallback: copy the layer, computing the diffID if we need to do so
fmt . Fprintf ( ic . reportWriter , "Copying blob %s\n" , srcInfo . Digest )
srcStream , srcBlobSize , err := ic . rawSource . GetBlob ( srcInfo )
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , "" , errors . Wrapf ( err , "Error reading blob %s" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
defer srcStream . Close ( )
blobInfo , diffIDChan , err := ic . copyLayerFromStream ( srcStream , types . BlobInfo { Digest : srcInfo . Digest , Size : srcBlobSize } ,
diffIDIsNeeded )
if err != nil {
return types . BlobInfo { } , "" , err
}
var diffIDResult diffIDResult // = {digest:""}
if diffIDIsNeeded {
diffIDResult = <- diffIDChan
if diffIDResult . err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , "" , errors . Wrap ( diffIDResult . err , "Error computing layer DiffID" )
2016-11-22 19:32:10 +00:00
}
logrus . Debugf ( "Computed DiffID %s for layer %s" , diffIDResult . digest , srcInfo . Digest )
ic . cachedDiffIDs [ srcInfo . Digest ] = diffIDResult . digest
}
return blobInfo , diffIDResult . digest , nil
}
// copyLayerFromStream is an implementation detail of copyLayer; mostly providing a separate “defer” scope.
// it copies a blob with srcInfo (with known Digest and possibly known Size) from srcStream to dest,
// perhaps compressing the stream if canCompress,
// and returns a complete blobInfo of the copied blob and perhaps a <-chan diffIDResult if diffIDIsNeeded, to be read by the caller.
func ( ic * imageCopier ) copyLayerFromStream ( srcStream io . Reader , srcInfo types . BlobInfo ,
diffIDIsNeeded bool ) ( types . BlobInfo , <- chan diffIDResult , error ) {
2017-03-13 16:33:17 +00:00
var getDiffIDRecorder func ( compression . DecompressorFunc ) io . Writer // = nil
2016-11-22 19:32:10 +00:00
var diffIDChan chan diffIDResult
err := errors . New ( "Internal error: unexpected panic in copyLayer" ) // For pipeWriter.CloseWithError below
if diffIDIsNeeded {
diffIDChan = make ( chan diffIDResult , 1 ) // Buffered, so that sending a value after this or our caller has failed and exited does not block.
pipeReader , pipeWriter := io . Pipe ( )
defer func ( ) { // Note that this is not the same as {defer pipeWriter.CloseWithError(err)}; we need err to be evaluated lazily.
pipeWriter . CloseWithError ( err ) // CloseWithError(nil) is equivalent to Close()
} ( )
2017-03-13 16:33:17 +00:00
getDiffIDRecorder = func ( decompressor compression . DecompressorFunc ) io . Writer {
2016-11-22 19:32:10 +00:00
// If this fails, e.g. because we have exited and due to pipeWriter.CloseWithError() above further
// reading from the pipe has failed, we don’ t really care.
// We only read from diffIDChan if the rest of the flow has succeeded, and when we do read from it,
// the return value includes an error indication, which we do check.
//
// If this gets never called, pipeReader will not be used anywhere, but pipeWriter will only be
// closed above, so we are happy enough with both pipeReader and pipeWriter to just get collected by GC.
go diffIDComputationGoroutine ( diffIDChan , pipeReader , decompressor ) // Closes pipeReader
return pipeWriter
}
}
blobInfo , err := ic . copyBlobFromStream ( srcStream , srcInfo , getDiffIDRecorder , ic . canModifyManifest ) // Sets err to nil on success
return blobInfo , diffIDChan , err
// We need the defer … pipeWriter.CloseWithError() to happen HERE so that the caller can block on reading from diffIDChan
}
// diffIDComputationGoroutine reads all input from layerStream, uncompresses using decompressor if necessary, and sends its digest, and status, if any, to dest.
2017-03-13 16:33:17 +00:00
func diffIDComputationGoroutine ( dest chan <- diffIDResult , layerStream io . ReadCloser , decompressor compression . DecompressorFunc ) {
2016-11-22 19:32:10 +00:00
result := diffIDResult {
digest : "" ,
err : errors . New ( "Internal error: unexpected panic in diffIDComputationGoroutine" ) ,
}
defer func ( ) { dest <- result } ( )
defer layerStream . Close ( ) // We do not care to bother the other end of the pipe with other failures; we send them to dest instead.
result . digest , result . err = computeDiffID ( layerStream , decompressor )
}
// computeDiffID reads all input from layerStream, uncompresses it using decompressor if necessary, and returns its digest.
2017-03-13 16:33:17 +00:00
func computeDiffID ( stream io . Reader , decompressor compression . DecompressorFunc ) ( digest . Digest , error ) {
2016-11-22 19:32:10 +00:00
if decompressor != nil {
s , err := decompressor ( stream )
if err != nil {
return "" , err
}
stream = s
}
return digest . Canonical . FromReader ( stream )
}
// copyBlobFromStream copies a blob with srcInfo (with known Digest and possibly known Size) from srcStream to dest,
// perhaps sending a copy to an io.Writer if getOriginalLayerCopyWriter != nil,
// perhaps compressing it if canCompress,
// and returns a complete blobInfo of the copied blob.
func ( ic * imageCopier ) copyBlobFromStream ( srcStream io . Reader , srcInfo types . BlobInfo ,
2017-03-13 16:33:17 +00:00
getOriginalLayerCopyWriter func ( decompressor compression . DecompressorFunc ) io . Writer ,
2016-11-22 19:32:10 +00:00
canCompress bool ) ( types . BlobInfo , error ) {
// The copying happens through a pipeline of connected io.Readers.
// === Input: srcStream
// === Process input through digestingReader to validate against the expected digest.
// Be paranoid; in case PutBlob somehow managed to ignore an error from digestingReader,
// use a separate validation failure indicator.
// Note that we don't use a stronger "validationSucceeded" indicator, because
// dest.PutBlob may detect that the layer already exists, in which case we don't
// read stream to the end, and validation does not happen.
digestingReader , err := newDigestingReader ( srcStream , srcInfo . Digest )
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , errors . Wrapf ( err , "Error preparing to verify blob %s" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
var destStream io . Reader = digestingReader
// === Detect compression of the input stream.
2017-03-13 16:33:17 +00:00
// This requires us to “peek ahead” into the stream to read the initial part, which requires us to chain through another io.Reader returned by DetectCompression.
decompressor , destStream , err := compression . DetectCompression ( destStream ) // We could skip this in some cases, but let's keep the code path uniform
2016-11-22 19:32:10 +00:00
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , errors . Wrapf ( err , "Error reading blob %s" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
isCompressed := decompressor != nil
// === Report progress using a pb.Reader.
bar := pb . New ( int ( srcInfo . Size ) ) . SetUnits ( pb . U_BYTES )
bar . Output = ic . reportWriter
bar . SetMaxWidth ( 80 )
bar . ShowTimeLeft = false
bar . ShowPercent = false
bar . Start ( )
destStream = bar . NewProxyReader ( destStream )
defer fmt . Fprint ( ic . reportWriter , "\n" )
// === Send a copy of the original, uncompressed, stream, to a separate path if necessary.
var originalLayerReader io . Reader // DO NOT USE this other than to drain the input if no other consumer in the pipeline has done so.
if getOriginalLayerCopyWriter != nil {
destStream = io . TeeReader ( destStream , getOriginalLayerCopyWriter ( decompressor ) )
originalLayerReader = destStream
}
// === Compress the layer if it is uncompressed and compression is desired
var inputInfo types . BlobInfo
if ! canCompress || isCompressed || ! ic . dest . ShouldCompressLayers ( ) {
logrus . Debugf ( "Using original blob without modification" )
inputInfo = srcInfo
} else {
logrus . Debugf ( "Compressing blob on the fly" )
pipeReader , pipeWriter := io . Pipe ( )
defer pipeReader . Close ( )
// If this fails while writing data, it will do pipeWriter.CloseWithError(); if it fails otherwise,
// e.g. because we have exited and due to pipeReader.Close() above further writing to the pipe has failed,
// we don’ t care.
go compressGoroutine ( pipeWriter , destStream ) // Closes pipeWriter
destStream = pipeReader
inputInfo . Digest = ""
inputInfo . Size = - 1
}
2017-03-13 16:33:17 +00:00
// === Report progress using the ic.progress channel, if required.
if ic . progress != nil && ic . progressInterval > 0 {
destStream = & progressReader {
source : destStream ,
channel : ic . progress ,
interval : ic . progressInterval ,
artifact : srcInfo ,
lastTime : time . Now ( ) ,
}
}
2016-11-22 19:32:10 +00:00
// === Finally, send the layer stream to dest.
uploadedInfo , err := ic . dest . PutBlob ( destStream , inputInfo )
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , errors . Wrap ( err , "Error writing blob" )
2016-11-22 19:32:10 +00:00
}
// This is fairly horrible: the writer from getOriginalLayerCopyWriter wants to consumer
// all of the input (to compute DiffIDs), even if dest.PutBlob does not need it.
// So, read everything from originalLayerReader, which will cause the rest to be
// sent there if we are not already at EOF.
if getOriginalLayerCopyWriter != nil {
logrus . Debugf ( "Consuming rest of the original blob to satisfy getOriginalLayerCopyWriter" )
_ , err := io . Copy ( ioutil . Discard , originalLayerReader )
if err != nil {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , errors . Wrapf ( err , "Error reading input blob %s" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
}
if digestingReader . validationFailed { // Coverage: This should never happen.
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , errors . Errorf ( "Internal error writing blob %s, digest verification failed but was ignored" , srcInfo . Digest )
2016-11-22 19:32:10 +00:00
}
if inputInfo . Digest != "" && uploadedInfo . Digest != inputInfo . Digest {
2016-10-17 13:53:40 +00:00
return types . BlobInfo { } , errors . Errorf ( "Internal error writing blob %s, blob with digest %s saved with digest %s" , srcInfo . Digest , inputInfo . Digest , uploadedInfo . Digest )
2016-11-22 19:32:10 +00:00
}
return uploadedInfo , nil
}
// compressGoroutine reads all input from src and writes its compressed equivalent to dest.
func compressGoroutine ( dest * io . PipeWriter , src io . Reader ) {
err := errors . New ( "Internal error: unexpected panic in compressGoroutine" )
defer func ( ) { // Note that this is not the same as {defer dest.CloseWithError(err)}; we need err to be evaluated lazily.
dest . CloseWithError ( err ) // CloseWithError(nil) is equivalent to Close()
} ( )
zipper := gzip . NewWriter ( dest )
defer zipper . Close ( )
_ , err = io . Copy ( zipper , src ) // Sets err to nil, i.e. causes dest.Close()
}