/* * Minio Cloud Storage, (C) 2016 Minio, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package main import ( "encoding/json" "fmt" "io" "path" "strings" "sync" "time" ) // MultipartPartInfo Info of each part kept in the multipart metadata file after // CompleteMultipartUpload() is called. type MultipartPartInfo struct { PartNumber int ETag string Size int64 } // MultipartObjectInfo - contents of the multipart metadata file after // CompleteMultipartUpload() is called. type MultipartObjectInfo struct { Parts []MultipartPartInfo ModTime time.Time Size int64 MD5Sum string } type byMultipartFiles []string func (files byMultipartFiles) Len() int { return len(files) } func (files byMultipartFiles) Less(i, j int) bool { first := strings.TrimSuffix(files[i], multipartSuffix) second := strings.TrimSuffix(files[j], multipartSuffix) return first < second } func (files byMultipartFiles) Swap(i, j int) { files[i], files[j] = files[j], files[i] } // GetPartNumberOffset - given an offset for the whole object, return the part and offset in that part. func (m MultipartObjectInfo) GetPartNumberOffset(offset int64) (partIndex int, partOffset int64, err error) { partOffset = offset for i, part := range m.Parts { partIndex = i if partOffset < part.Size { return } partOffset -= part.Size } // Offset beyond the size of the object err = errUnexpected return } func partNumToPartFileName(partNum int) string { return fmt.Sprintf("%.5d%s", partNum, multipartSuffix) } // ListMultipartUploads - list multipart uploads. func (xl xlObjects) ListMultipartUploads(bucket, prefix, keyMarker, uploadIDMarker, delimiter string, maxUploads int) (ListMultipartsInfo, error) { return listMultipartUploadsCommon(xl, bucket, prefix, keyMarker, uploadIDMarker, delimiter, maxUploads) } // NewMultipartUpload - initialize a new multipart upload, returns a unique id. func (xl xlObjects) NewMultipartUpload(bucket, object string) (string, error) { return newMultipartUploadCommon(xl.storage, bucket, object) } // PutObjectPart - writes the multipart upload chunks. func (xl xlObjects) PutObjectPart(bucket, object, uploadID string, partID int, size int64, data io.Reader, md5Hex string) (string, error) { return putObjectPartCommon(xl.storage, bucket, object, uploadID, partID, size, data, md5Hex) } // ListObjectParts - list object parts. func (xl xlObjects) ListObjectParts(bucket, object, uploadID string, partNumberMarker, maxParts int) (ListPartsInfo, error) { return listObjectPartsCommon(xl.storage, bucket, object, uploadID, partNumberMarker, maxParts) } // This function does the following check, suppose // object is "a/b/c/d", stat makes sure that objects ""a/b/c"" // "a/b" and "a" do not exist. func (xl xlObjects) parentDirIsObject(bucket, parent string) error { var stat func(string) error stat = func(p string) error { if p == "." { return nil } _, err := xl.getObjectInfo(bucket, p) if err == nil { // If there is already a file at prefix "p" return error. return errFileAccessDenied } if err == errFileNotFound { // Check if there is a file as one of the parent paths. return stat(path.Dir(p)) } return err } return stat(parent) } func (xl xlObjects) CompleteMultipartUpload(bucket string, object string, uploadID string, parts []completePart) (string, error) { // Verify if bucket is valid. if !IsValidBucketName(bucket) { return "", BucketNameInvalid{Bucket: bucket} } // Verify whether the bucket exists. if !isBucketExist(xl.storage, bucket) { return "", BucketNotFound{Bucket: bucket} } if !IsValidObjectName(object) { return "", ObjectNameInvalid{ Bucket: bucket, Object: object, } } if !isUploadIDExists(xl.storage, bucket, object, uploadID) { return "", InvalidUploadID{UploadID: uploadID} } // Hold lock so that // 1) no one aborts this multipart upload // 2) no one does a parallel complete-multipart-upload on this multipart upload nsMutex.Lock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object, uploadID)) defer nsMutex.Unlock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object, uploadID)) // Calculate s3 compatible md5sum for complete multipart. s3MD5, err := completeMultipartMD5(parts...) if err != nil { return "", err } var metadata = MultipartObjectInfo{} var errs = make([]error, len(parts)) // Waitgroup to wait for go-routines. var wg = &sync.WaitGroup{} // Loop through all parts, validate them and then commit to disk. for i, part := range parts { // Construct part suffix. partSuffix := fmt.Sprintf("%.5d.%s", part.PartNumber, part.ETag) multipartPartFile := path.Join(mpartMetaPrefix, bucket, object, uploadID, partSuffix) var fi FileInfo fi, err = xl.storage.StatFile(minioMetaBucket, multipartPartFile) if err != nil { if err == errFileNotFound { return "", InvalidPart{} } return "", err } // All parts except the last part has to be atleast 5MB. if (i < len(parts)-1) && !isMinAllowedPartSize(fi.Size) { return "", PartTooSmall{} } // Update metadata parts. metadata.Parts = append(metadata.Parts, MultipartPartInfo{ PartNumber: part.PartNumber, ETag: part.ETag, Size: fi.Size, }) metadata.Size += fi.Size } // check if an object is present as one of the parent dir. if err = xl.parentDirIsObject(bucket, path.Dir(object)); err != nil { return "", toObjectErr(err, bucket, object) } // Save successfully calculated md5sum. metadata.MD5Sum = s3MD5 // Save modTime as well as the current time. metadata.ModTime = time.Now().UTC() // Create temporary multipart meta file to write and then rename. multipartMetaSuffix := fmt.Sprintf("%s.%s", uploadID, multipartMetaFile) tempMultipartMetaFile := path.Join(tmpMetaPrefix, bucket, object, multipartMetaSuffix) w, err := xl.storage.CreateFile(minioMetaBucket, tempMultipartMetaFile) if err != nil { return "", toObjectErr(err, bucket, object) } encoder := json.NewEncoder(w) err = encoder.Encode(&metadata) if err != nil { if err = safeCloseAndRemove(w); err != nil { return "", toObjectErr(err, bucket, object) } return "", toObjectErr(err, bucket, object) } // Close the writer. if err = w.Close(); err != nil { if err = safeCloseAndRemove(w); err != nil { return "", toObjectErr(err, bucket, object) } return "", toObjectErr(err, bucket, object) } // Attempt a Rename of multipart meta file to final namespace. multipartObjFile := path.Join(mpartMetaPrefix, bucket, object, uploadID, multipartMetaFile) err = xl.storage.RenameFile(minioMetaBucket, tempMultipartMetaFile, minioMetaBucket, multipartObjFile) if err != nil { if derr := xl.storage.DeleteFile(minioMetaBucket, tempMultipartMetaFile); derr != nil { return "", toObjectErr(err, minioMetaBucket, tempMultipartMetaFile) } return "", toObjectErr(err, bucket, multipartObjFile) } // Loop through and atomically rename the parts to their actual location. for index, part := range parts { wg.Add(1) go func(index int, part completePart) { defer wg.Done() partSuffix := fmt.Sprintf("%.5d.%s", part.PartNumber, part.ETag) src := path.Join(mpartMetaPrefix, bucket, object, uploadID, partSuffix) dst := path.Join(mpartMetaPrefix, bucket, object, uploadID, partNumToPartFileName(part.PartNumber)) errs[index] = xl.storage.RenameFile(minioMetaBucket, src, minioMetaBucket, dst) errorIf(errs[index], "Unable to rename file %s to %s.", src, dst) }(index, part) } // Wait for all the renames to finish. wg.Wait() // Loop through errs list and return first error. for _, err := range errs { if err != nil { return "", toObjectErr(err, bucket, object) } } // Delete the incomplete file place holder. uploadIDIncompletePath := path.Join(mpartMetaPrefix, bucket, object, uploadID, incompleteFile) err = xl.storage.DeleteFile(minioMetaBucket, uploadIDIncompletePath) if err != nil { return "", toObjectErr(err, minioMetaBucket, uploadIDIncompletePath) } // Hold write lock on the destination before rename nsMutex.Lock(bucket, object) defer nsMutex.Unlock(bucket, object) // Delete if an object already exists. // FIXME: rename it to tmp file and delete only after // the newly uploaded file is renamed from tmp location to // the original location. err = xl.deleteObject(bucket, object) if err != nil && err != errFileNotFound { return "", toObjectErr(err, bucket, object) } uploadIDPath := path.Join(mpartMetaPrefix, bucket, object, uploadID) if err = xl.storage.RenameFile(minioMetaBucket, uploadIDPath, bucket, object); err != nil { return "", toObjectErr(err, bucket, object) } // Hold the lock so that two parallel complete-multipart-uploads do no // leave a stale uploads.json behind. nsMutex.Lock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object)) defer nsMutex.Unlock(minioMetaBucket, pathJoin(mpartMetaPrefix, bucket, object)) // Validate if there are other incomplete upload-id's present for // the object, if yes do not attempt to delete 'uploads.json'. var entries []string if entries, err = xl.storage.ListDir(minioMetaBucket, path.Join(mpartMetaPrefix, bucket, object)); err == nil { if len(entries) > 1 { return s3MD5, nil } } uploadsJSONPath := path.Join(mpartMetaPrefix, bucket, object, uploadsJSONFile) err = xl.storage.DeleteFile(minioMetaBucket, uploadsJSONPath) if err != nil { return "", toObjectErr(err, minioMetaBucket, uploadsJSONPath) } // Return md5sum. return s3MD5, nil } // AbortMultipartUpload - aborts a multipart upload. func (xl xlObjects) AbortMultipartUpload(bucket, object, uploadID string) error { return abortMultipartUploadCommon(xl.storage, bucket, object, uploadID) }