You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
298 lines
10 KiB
298 lines
10 KiB
/*
|
|
* Minio Cloud Storage, (C) 2016 Minio, Inc.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
package main
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"io"
|
|
"path"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
)
|
|
|
|
// MultipartPartInfo Info of each part kept in the multipart metadata file after
|
|
// CompleteMultipartUpload() is called.
|
|
type MultipartPartInfo struct {
|
|
PartNumber int
|
|
ETag string
|
|
Size int64
|
|
}
|
|
|
|
// MultipartObjectInfo - contents of the multipart metadata file after
|
|
// CompleteMultipartUpload() is called.
|
|
type MultipartObjectInfo struct {
|
|
Parts []MultipartPartInfo
|
|
ModTime time.Time
|
|
Size int64
|
|
MD5Sum string
|
|
}
|
|
|
|
type byMultipartFiles []string
|
|
|
|
func (files byMultipartFiles) Len() int { return len(files) }
|
|
func (files byMultipartFiles) Less(i, j int) bool {
|
|
first := strings.TrimSuffix(files[i], multipartSuffix)
|
|
second := strings.TrimSuffix(files[j], multipartSuffix)
|
|
return first < second
|
|
}
|
|
func (files byMultipartFiles) Swap(i, j int) { files[i], files[j] = files[j], files[i] }
|
|
|
|
// GetPartNumberOffset - given an offset for the whole object, return the part and offset in that part.
|
|
func (m MultipartObjectInfo) GetPartNumberOffset(offset int64) (partIndex int, partOffset int64, err error) {
|
|
partOffset = offset
|
|
for i, part := range m.Parts {
|
|
partIndex = i
|
|
if partOffset < part.Size {
|
|
return
|
|
}
|
|
partOffset -= part.Size
|
|
}
|
|
// Offset beyond the size of the object
|
|
err = errUnexpected
|
|
return
|
|
}
|
|
|
|
func partNumToPartFileName(partNum int) string {
|
|
return fmt.Sprintf("%.5d%s", partNum, multipartSuffix)
|
|
}
|
|
|
|
// ListMultipartUploads - list multipart uploads.
|
|
func (xl xlObjects) ListMultipartUploads(bucket, prefix, keyMarker, uploadIDMarker, delimiter string, maxUploads int) (ListMultipartsInfo, error) {
|
|
return listMultipartUploadsCommon(xl, bucket, prefix, keyMarker, uploadIDMarker, delimiter, maxUploads)
|
|
}
|
|
|
|
// NewMultipartUpload - initialize a new multipart upload, returns a unique id.
|
|
func (xl xlObjects) NewMultipartUpload(bucket, object string) (string, error) {
|
|
return newMultipartUploadCommon(xl.storage, bucket, object)
|
|
}
|
|
|
|
// PutObjectPart - writes the multipart upload chunks.
|
|
func (xl xlObjects) PutObjectPart(bucket, object, uploadID string, partID int, size int64, data io.Reader, md5Hex string) (string, error) {
|
|
return putObjectPartCommon(xl.storage, bucket, object, uploadID, partID, size, data, md5Hex)
|
|
}
|
|
|
|
// ListObjectParts - list object parts.
|
|
func (xl xlObjects) ListObjectParts(bucket, object, uploadID string, partNumberMarker, maxParts int) (ListPartsInfo, error) {
|
|
return listObjectPartsCommon(xl.storage, bucket, object, uploadID, partNumberMarker, maxParts)
|
|
}
|
|
|
|
// This function does the following check, suppose
|
|
// object is "a/b/c/d", stat makes sure that objects ""a/b/c""
|
|
// "a/b" and "a" do not exist.
|
|
func (xl xlObjects) parentDirIsObject(bucket, parent string) error {
|
|
var stat func(string) error
|
|
stat = func(p string) error {
|
|
if p == "." {
|
|
return nil
|
|
}
|
|
_, err := xl.getObjectInfo(bucket, p)
|
|
if err == nil {
|
|
// If there is already a file at prefix "p" return error.
|
|
return errFileAccessDenied
|
|
}
|
|
if err == errFileNotFound {
|
|
// Check if there is a file as one of the parent paths.
|
|
return stat(path.Dir(p))
|
|
}
|
|
return err
|
|
}
|
|
return stat(parent)
|
|
}
|
|
|
|
func (xl xlObjects) CompleteMultipartUpload(bucket string, object string, uploadID string, parts []completePart) (string, error) {
|
|
// Verify if bucket is valid.
|
|
if !IsValidBucketName(bucket) {
|
|
return "", BucketNameInvalid{Bucket: bucket}
|
|
}
|
|
// Verify whether the bucket exists.
|
|
if !isBucketExist(xl.storage, bucket) {
|
|
return "", BucketNotFound{Bucket: bucket}
|
|
}
|
|
if !IsValidObjectName(object) {
|
|
return "", ObjectNameInvalid{
|
|
Bucket: bucket,
|
|
Object: object,
|
|
}
|
|
}
|
|
if !isUploadIDExists(xl.storage, bucket, object, uploadID) {
|
|
return "", InvalidUploadID{UploadID: uploadID}
|
|
}
|
|
// Hold lock so that
|
|
// 1) no one aborts this multipart upload
|
|
// 2) no one does a parallel complete-multipart-upload on this multipart upload
|
|
nsMutex.Lock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object, uploadID))
|
|
defer nsMutex.Unlock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object, uploadID))
|
|
|
|
// Calculate s3 compatible md5sum for complete multipart.
|
|
s3MD5, err := completeMultipartMD5(parts...)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
var metadata = MultipartObjectInfo{}
|
|
var errs = make([]error, len(parts))
|
|
|
|
// Waitgroup to wait for go-routines.
|
|
var wg = &sync.WaitGroup{}
|
|
|
|
// Loop through all parts, validate them and then commit to disk.
|
|
for i, part := range parts {
|
|
// Construct part suffix.
|
|
partSuffix := fmt.Sprintf("%.5d.%s", part.PartNumber, part.ETag)
|
|
multipartPartFile := path.Join(mpartMetaPrefix, bucket, object, uploadID, partSuffix)
|
|
var fi FileInfo
|
|
fi, err = xl.storage.StatFile(minioMetaBucket, multipartPartFile)
|
|
if err != nil {
|
|
if err == errFileNotFound {
|
|
return "", InvalidPart{}
|
|
}
|
|
return "", err
|
|
}
|
|
// All parts except the last part has to be atleast 5MB.
|
|
if (i < len(parts)-1) && !isMinAllowedPartSize(fi.Size) {
|
|
return "", PartTooSmall{}
|
|
}
|
|
// Update metadata parts.
|
|
metadata.Parts = append(metadata.Parts, MultipartPartInfo{
|
|
PartNumber: part.PartNumber,
|
|
ETag: part.ETag,
|
|
Size: fi.Size,
|
|
})
|
|
metadata.Size += fi.Size
|
|
}
|
|
|
|
// check if an object is present as one of the parent dir.
|
|
if err = xl.parentDirIsObject(bucket, path.Dir(object)); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Save successfully calculated md5sum.
|
|
metadata.MD5Sum = s3MD5
|
|
// Save modTime as well as the current time.
|
|
metadata.ModTime = time.Now().UTC()
|
|
|
|
// Create temporary multipart meta file to write and then rename.
|
|
multipartMetaSuffix := fmt.Sprintf("%s.%s", uploadID, multipartMetaFile)
|
|
tempMultipartMetaFile := path.Join(tmpMetaPrefix, bucket, object, multipartMetaSuffix)
|
|
w, err := xl.storage.CreateFile(minioMetaBucket, tempMultipartMetaFile)
|
|
if err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
encoder := json.NewEncoder(w)
|
|
err = encoder.Encode(&metadata)
|
|
if err != nil {
|
|
if err = safeCloseAndRemove(w); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
// Close the writer.
|
|
if err = w.Close(); err != nil {
|
|
if err = safeCloseAndRemove(w); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Attempt a Rename of multipart meta file to final namespace.
|
|
multipartObjFile := path.Join(mpartMetaPrefix, bucket, object, uploadID, multipartMetaFile)
|
|
err = xl.storage.RenameFile(minioMetaBucket, tempMultipartMetaFile, minioMetaBucket, multipartObjFile)
|
|
if err != nil {
|
|
if derr := xl.storage.DeleteFile(minioMetaBucket, tempMultipartMetaFile); derr != nil {
|
|
return "", toObjectErr(err, minioMetaBucket, tempMultipartMetaFile)
|
|
}
|
|
return "", toObjectErr(err, bucket, multipartObjFile)
|
|
}
|
|
|
|
// Loop through and atomically rename the parts to their actual location.
|
|
for index, part := range parts {
|
|
wg.Add(1)
|
|
go func(index int, part completePart) {
|
|
defer wg.Done()
|
|
partSuffix := fmt.Sprintf("%.5d.%s", part.PartNumber, part.ETag)
|
|
src := path.Join(mpartMetaPrefix, bucket, object, uploadID, partSuffix)
|
|
dst := path.Join(mpartMetaPrefix, bucket, object, uploadID, partNumToPartFileName(part.PartNumber))
|
|
errs[index] = xl.storage.RenameFile(minioMetaBucket, src, minioMetaBucket, dst)
|
|
errorIf(errs[index], "Unable to rename file %s to %s.", src, dst)
|
|
}(index, part)
|
|
}
|
|
|
|
// Wait for all the renames to finish.
|
|
wg.Wait()
|
|
|
|
// Loop through errs list and return first error.
|
|
for _, err := range errs {
|
|
if err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
}
|
|
|
|
// Delete the incomplete file place holder.
|
|
uploadIDIncompletePath := path.Join(mpartMetaPrefix, bucket, object, uploadID, incompleteFile)
|
|
err = xl.storage.DeleteFile(minioMetaBucket, uploadIDIncompletePath)
|
|
if err != nil {
|
|
return "", toObjectErr(err, minioMetaBucket, uploadIDIncompletePath)
|
|
}
|
|
|
|
// Hold write lock on the destination before rename
|
|
nsMutex.Lock(bucket, object)
|
|
defer nsMutex.Unlock(bucket, object)
|
|
|
|
// Delete if an object already exists.
|
|
// FIXME: rename it to tmp file and delete only after
|
|
// the newly uploaded file is renamed from tmp location to
|
|
// the original location.
|
|
err = xl.deleteObject(bucket, object)
|
|
if err != nil && err != errFileNotFound {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
uploadIDPath := path.Join(mpartMetaPrefix, bucket, object, uploadID)
|
|
if err = xl.storage.RenameFile(minioMetaBucket, uploadIDPath, bucket, object); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Hold the lock so that two parallel complete-multipart-uploads do no
|
|
// leave a stale uploads.json behind.
|
|
nsMutex.Lock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object))
|
|
defer nsMutex.Unlock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object))
|
|
|
|
// Validate if there are other incomplete upload-id's present for
|
|
// the object, if yes do not attempt to delete 'uploads.json'.
|
|
var entries []string
|
|
if entries, err = xl.storage.ListDir(minioMetaBucket, path.Join(mpartMetaPrefix, bucket, object)); err == nil {
|
|
if len(entries) > 1 {
|
|
return s3MD5, nil
|
|
}
|
|
}
|
|
|
|
uploadsJSONPath := path.Join(mpartMetaPrefix, bucket, object, uploadsJSONFile)
|
|
err = xl.storage.DeleteFile(minioMetaBucket, uploadsJSONPath)
|
|
if err != nil {
|
|
return "", toObjectErr(err, minioMetaBucket, uploadsJSONPath)
|
|
}
|
|
|
|
// Return md5sum.
|
|
return s3MD5, nil
|
|
}
|
|
|
|
// AbortMultipartUpload - aborts a multipart upload.
|
|
func (xl xlObjects) AbortMultipartUpload(bucket, object, uploadID string) error {
|
|
return abortMultipartUploadCommon(xl.storage, bucket, object, uploadID)
|
|
}
|
|
|