Платформа ЦРНП "Мирокод" для разработки проектов
https://git.mirocod.ru
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
366 lines
11 KiB
366 lines
11 KiB
/* |
|
* MinIO Go Library for Amazon S3 Compatible Cloud Storage |
|
* Copyright 2015-2017 MinIO, Inc. |
|
* |
|
* Licensed under the Apache License, Version 2.0 (the "License"); |
|
* you may not use this file except in compliance with the License. |
|
* You may obtain a copy of the License at |
|
* |
|
* http://www.apache.org/licenses/LICENSE-2.0 |
|
* |
|
* Unless required by applicable law or agreed to in writing, software |
|
* distributed under the License is distributed on an "AS IS" BASIS, |
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
* See the License for the specific language governing permissions and |
|
* limitations under the License. |
|
*/ |
|
|
|
package minio |
|
|
|
import ( |
|
"bytes" |
|
"context" |
|
"encoding/base64" |
|
"errors" |
|
"fmt" |
|
"io" |
|
"net/http" |
|
"sort" |
|
"time" |
|
|
|
"github.com/minio/minio-go/v7/pkg/encrypt" |
|
"github.com/minio/minio-go/v7/pkg/s3utils" |
|
"golang.org/x/net/http/httpguts" |
|
) |
|
|
|
// ReplicationStatus represents replication status of object |
|
type ReplicationStatus string |
|
|
|
const ( |
|
// ReplicationStatusPending indicates replication is pending |
|
ReplicationStatusPending ReplicationStatus = "PENDING" |
|
// ReplicationStatusComplete indicates replication completed ok |
|
ReplicationStatusComplete ReplicationStatus = "COMPLETE" |
|
// ReplicationStatusFailed indicates replication failed |
|
ReplicationStatusFailed ReplicationStatus = "FAILED" |
|
// ReplicationStatusReplica indicates object is a replica of a source |
|
ReplicationStatusReplica ReplicationStatus = "REPLICA" |
|
) |
|
|
|
// Empty returns true if no replication status set. |
|
func (r ReplicationStatus) Empty() bool { |
|
return r == "" |
|
} |
|
|
|
// AdvancedPutOptions for internal use - to be utilized by replication, ILM transition |
|
// implementation on MinIO server |
|
type AdvancedPutOptions struct { |
|
SourceVersionID string |
|
SourceETag string |
|
ReplicationStatus ReplicationStatus |
|
SourceMTime time.Time |
|
} |
|
|
|
// PutObjectOptions represents options specified by user for PutObject call |
|
type PutObjectOptions struct { |
|
UserMetadata map[string]string |
|
UserTags map[string]string |
|
Progress io.Reader |
|
ContentType string |
|
ContentEncoding string |
|
ContentDisposition string |
|
ContentLanguage string |
|
CacheControl string |
|
Mode RetentionMode |
|
RetainUntilDate time.Time |
|
ServerSideEncryption encrypt.ServerSide |
|
NumThreads uint |
|
StorageClass string |
|
WebsiteRedirectLocation string |
|
PartSize uint64 |
|
LegalHold LegalHoldStatus |
|
SendContentMd5 bool |
|
DisableMultipart bool |
|
Internal AdvancedPutOptions |
|
} |
|
|
|
// getNumThreads - gets the number of threads to be used in the multipart |
|
// put object operation |
|
func (opts PutObjectOptions) getNumThreads() (numThreads int) { |
|
if opts.NumThreads > 0 { |
|
numThreads = int(opts.NumThreads) |
|
} else { |
|
numThreads = totalWorkers |
|
} |
|
return |
|
} |
|
|
|
// Header - constructs the headers from metadata entered by user in |
|
// PutObjectOptions struct |
|
func (opts PutObjectOptions) Header() (header http.Header) { |
|
header = make(http.Header) |
|
|
|
contentType := opts.ContentType |
|
if contentType == "" { |
|
contentType = "application/octet-stream" |
|
} |
|
header.Set("Content-Type", contentType) |
|
|
|
if opts.ContentEncoding != "" { |
|
header.Set("Content-Encoding", opts.ContentEncoding) |
|
} |
|
if opts.ContentDisposition != "" { |
|
header.Set("Content-Disposition", opts.ContentDisposition) |
|
} |
|
if opts.ContentLanguage != "" { |
|
header.Set("Content-Language", opts.ContentLanguage) |
|
} |
|
if opts.CacheControl != "" { |
|
header.Set("Cache-Control", opts.CacheControl) |
|
} |
|
|
|
if opts.Mode != "" { |
|
header.Set(amzLockMode, opts.Mode.String()) |
|
} |
|
|
|
if !opts.RetainUntilDate.IsZero() { |
|
header.Set("X-Amz-Object-Lock-Retain-Until-Date", opts.RetainUntilDate.Format(time.RFC3339)) |
|
} |
|
|
|
if opts.LegalHold != "" { |
|
header.Set(amzLegalHoldHeader, opts.LegalHold.String()) |
|
} |
|
|
|
if opts.ServerSideEncryption != nil { |
|
opts.ServerSideEncryption.Marshal(header) |
|
} |
|
|
|
if opts.StorageClass != "" { |
|
header.Set(amzStorageClass, opts.StorageClass) |
|
} |
|
|
|
if opts.WebsiteRedirectLocation != "" { |
|
header.Set(amzWebsiteRedirectLocation, opts.WebsiteRedirectLocation) |
|
} |
|
|
|
if !opts.Internal.ReplicationStatus.Empty() { |
|
header.Set(amzBucketReplicationStatus, string(opts.Internal.ReplicationStatus)) |
|
} |
|
if !opts.Internal.SourceMTime.IsZero() { |
|
header.Set(minIOBucketSourceMTime, opts.Internal.SourceMTime.Format(time.RFC3339)) |
|
} |
|
if opts.Internal.SourceETag != "" { |
|
header.Set(minIOBucketSourceETag, opts.Internal.SourceETag) |
|
} |
|
if len(opts.UserTags) != 0 { |
|
header.Set(amzTaggingHeader, s3utils.TagEncode(opts.UserTags)) |
|
} |
|
|
|
for k, v := range opts.UserMetadata { |
|
if isAmzHeader(k) || isStandardHeader(k) || isStorageClassHeader(k) { |
|
header.Set(k, v) |
|
} else { |
|
header.Set("x-amz-meta-"+k, v) |
|
} |
|
} |
|
return |
|
} |
|
|
|
// validate() checks if the UserMetadata map has standard headers or and raises an error if so. |
|
func (opts PutObjectOptions) validate() (err error) { |
|
for k, v := range opts.UserMetadata { |
|
if !httpguts.ValidHeaderFieldName(k) || isStandardHeader(k) || isSSEHeader(k) || isStorageClassHeader(k) { |
|
return errInvalidArgument(k + " unsupported user defined metadata name") |
|
} |
|
if !httpguts.ValidHeaderFieldValue(v) { |
|
return errInvalidArgument(v + " unsupported user defined metadata value") |
|
} |
|
} |
|
if opts.Mode != "" && !opts.Mode.IsValid() { |
|
return errInvalidArgument(opts.Mode.String() + " unsupported retention mode") |
|
} |
|
if opts.LegalHold != "" && !opts.LegalHold.IsValid() { |
|
return errInvalidArgument(opts.LegalHold.String() + " unsupported legal-hold status") |
|
} |
|
return nil |
|
} |
|
|
|
// completedParts is a collection of parts sortable by their part numbers. |
|
// used for sorting the uploaded parts before completing the multipart request. |
|
type completedParts []CompletePart |
|
|
|
func (a completedParts) Len() int { return len(a) } |
|
func (a completedParts) Swap(i, j int) { a[i], a[j] = a[j], a[i] } |
|
func (a completedParts) Less(i, j int) bool { return a[i].PartNumber < a[j].PartNumber } |
|
|
|
// PutObject creates an object in a bucket. |
|
// |
|
// You must have WRITE permissions on a bucket to create an object. |
|
// |
|
// - For size smaller than 128MiB PutObject automatically does a |
|
// single atomic Put operation. |
|
// - For size larger than 128MiB PutObject automatically does a |
|
// multipart Put operation. |
|
// - For size input as -1 PutObject does a multipart Put operation |
|
// until input stream reaches EOF. Maximum object size that can |
|
// be uploaded through this operation will be 5TiB. |
|
func (c Client) PutObject(ctx context.Context, bucketName, objectName string, reader io.Reader, objectSize int64, |
|
opts PutObjectOptions) (info UploadInfo, err error) { |
|
if objectSize < 0 && opts.DisableMultipart { |
|
return UploadInfo{}, errors.New("object size must be provided with disable multipart upload") |
|
} |
|
|
|
err = opts.validate() |
|
if err != nil { |
|
return UploadInfo{}, err |
|
} |
|
|
|
return c.putObjectCommon(ctx, bucketName, objectName, reader, objectSize, opts) |
|
} |
|
|
|
func (c Client) putObjectCommon(ctx context.Context, bucketName, objectName string, reader io.Reader, size int64, opts PutObjectOptions) (info UploadInfo, err error) { |
|
// Check for largest object size allowed. |
|
if size > int64(maxMultipartPutObjectSize) { |
|
return UploadInfo{}, errEntityTooLarge(size, maxMultipartPutObjectSize, bucketName, objectName) |
|
} |
|
|
|
// NOTE: Streaming signature is not supported by GCS. |
|
if s3utils.IsGoogleEndpoint(*c.endpointURL) { |
|
return c.putObject(ctx, bucketName, objectName, reader, size, opts) |
|
} |
|
|
|
partSize := opts.PartSize |
|
if opts.PartSize == 0 { |
|
partSize = minPartSize |
|
} |
|
|
|
if c.overrideSignerType.IsV2() { |
|
if size >= 0 && size < int64(partSize) || opts.DisableMultipart { |
|
return c.putObject(ctx, bucketName, objectName, reader, size, opts) |
|
} |
|
return c.putObjectMultipart(ctx, bucketName, objectName, reader, size, opts) |
|
} |
|
|
|
if size < 0 { |
|
return c.putObjectMultipartStreamNoLength(ctx, bucketName, objectName, reader, opts) |
|
} |
|
|
|
if size < int64(partSize) || opts.DisableMultipart { |
|
return c.putObject(ctx, bucketName, objectName, reader, size, opts) |
|
} |
|
|
|
return c.putObjectMultipartStream(ctx, bucketName, objectName, reader, size, opts) |
|
} |
|
|
|
func (c Client) putObjectMultipartStreamNoLength(ctx context.Context, bucketName, objectName string, reader io.Reader, opts PutObjectOptions) (info UploadInfo, err error) { |
|
// Input validation. |
|
if err = s3utils.CheckValidBucketName(bucketName); err != nil { |
|
return UploadInfo{}, err |
|
} |
|
if err = s3utils.CheckValidObjectName(objectName); err != nil { |
|
return UploadInfo{}, err |
|
} |
|
|
|
// Total data read and written to server. should be equal to |
|
// 'size' at the end of the call. |
|
var totalUploadedSize int64 |
|
|
|
// Complete multipart upload. |
|
var complMultipartUpload completeMultipartUpload |
|
|
|
// Calculate the optimal parts info for a given size. |
|
totalPartsCount, partSize, _, err := optimalPartInfo(-1, opts.PartSize) |
|
if err != nil { |
|
return UploadInfo{}, err |
|
} |
|
// Initiate a new multipart upload. |
|
uploadID, err := c.newUploadID(ctx, bucketName, objectName, opts) |
|
if err != nil { |
|
return UploadInfo{}, err |
|
} |
|
|
|
defer func() { |
|
if err != nil { |
|
c.abortMultipartUpload(ctx, bucketName, objectName, uploadID) |
|
} |
|
}() |
|
|
|
// Part number always starts with '1'. |
|
partNumber := 1 |
|
|
|
// Initialize parts uploaded map. |
|
partsInfo := make(map[int]ObjectPart) |
|
|
|
// Create a buffer. |
|
buf := make([]byte, partSize) |
|
|
|
for partNumber <= totalPartsCount { |
|
length, rerr := readFull(reader, buf) |
|
if rerr == io.EOF && partNumber > 1 { |
|
break |
|
} |
|
|
|
if rerr != nil && rerr != io.ErrUnexpectedEOF && rerr != io.EOF { |
|
return UploadInfo{}, rerr |
|
} |
|
|
|
var md5Base64 string |
|
if opts.SendContentMd5 { |
|
// Calculate md5sum. |
|
hash := c.md5Hasher() |
|
hash.Write(buf[:length]) |
|
md5Base64 = base64.StdEncoding.EncodeToString(hash.Sum(nil)) |
|
hash.Close() |
|
} |
|
|
|
// Update progress reader appropriately to the latest offset |
|
// as we read from the source. |
|
rd := newHook(bytes.NewReader(buf[:length]), opts.Progress) |
|
|
|
// Proceed to upload the part. |
|
objPart, uerr := c.uploadPart(ctx, bucketName, objectName, uploadID, rd, partNumber, |
|
md5Base64, "", int64(length), opts.ServerSideEncryption) |
|
if uerr != nil { |
|
return UploadInfo{}, uerr |
|
} |
|
|
|
// Save successfully uploaded part metadata. |
|
partsInfo[partNumber] = objPart |
|
|
|
// Save successfully uploaded size. |
|
totalUploadedSize += int64(length) |
|
|
|
// Increment part number. |
|
partNumber++ |
|
|
|
// For unknown size, Read EOF we break away. |
|
// We do not have to upload till totalPartsCount. |
|
if rerr == io.EOF { |
|
break |
|
} |
|
} |
|
|
|
// Loop over total uploaded parts to save them in |
|
// Parts array before completing the multipart request. |
|
for i := 1; i < partNumber; i++ { |
|
part, ok := partsInfo[i] |
|
if !ok { |
|
return UploadInfo{}, errInvalidArgument(fmt.Sprintf("Missing part number %d", i)) |
|
} |
|
complMultipartUpload.Parts = append(complMultipartUpload.Parts, CompletePart{ |
|
ETag: part.ETag, |
|
PartNumber: part.PartNumber, |
|
}) |
|
} |
|
|
|
// Sort all completed parts. |
|
sort.Sort(completedParts(complMultipartUpload.Parts)) |
|
|
|
uploadInfo, err := c.completeMultipartUpload(ctx, bucketName, objectName, uploadID, complMultipartUpload) |
|
if err != nil { |
|
return UploadInfo{}, err |
|
} |
|
|
|
uploadInfo.Size = totalUploadedSize |
|
return uploadInfo, nil |
|
}
|
|
|