You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
581 lines
17 KiB
581 lines
17 KiB
/*
|
|
* Minio Cloud Storage, (C) 2016 Minio, Inc.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
package main
|
|
|
|
import (
|
|
"crypto/md5"
|
|
"encoding/hex"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"path"
|
|
"path/filepath"
|
|
"strings"
|
|
"sync"
|
|
|
|
"github.com/minio/minio/pkg/mimedb"
|
|
)
|
|
|
|
const (
|
|
multipartSuffix = ".minio.multipart"
|
|
multipartMetaFile = "00000" + multipartSuffix
|
|
formatConfigFile = "format.json"
|
|
)
|
|
|
|
// xlObjects - Implements fs object layer.
|
|
type xlObjects struct {
|
|
storage StorageAPI
|
|
listObjectMap map[listParams][]*treeWalker
|
|
listObjectMapMutex *sync.Mutex
|
|
}
|
|
|
|
// errMaxDisks - returned for reached maximum of disks.
|
|
var errMaxDisks = errors.New("Number of disks are higher than supported maximum count '16'")
|
|
|
|
// errMinDisks - returned for minimum number of disks.
|
|
var errMinDisks = errors.New("Number of disks are smaller than supported minimum count '8'")
|
|
|
|
// errNumDisks - returned for odd number of disks.
|
|
var errNumDisks = errors.New("Number of disks should be multiples of '2'")
|
|
|
|
const (
|
|
// Maximum erasure blocks.
|
|
maxErasureBlocks = 16
|
|
// Minimum erasure blocks.
|
|
minErasureBlocks = 8
|
|
)
|
|
|
|
func checkSufficientDisks(disks []string) error {
|
|
// Verify total number of disks.
|
|
totalDisks := len(disks)
|
|
if totalDisks > maxErasureBlocks {
|
|
return errMaxDisks
|
|
}
|
|
if totalDisks < minErasureBlocks {
|
|
return errMinDisks
|
|
}
|
|
|
|
// isEven function to verify if a given number if even.
|
|
isEven := func(number int) bool {
|
|
return number%2 == 0
|
|
}
|
|
|
|
// Verify if we have even number of disks.
|
|
// only combination of 8, 10, 12, 14, 16 are supported.
|
|
if !isEven(totalDisks) {
|
|
return errNumDisks
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Depending on the disk type network or local, initialize storage layer.
|
|
func newStorageLayer(disk string) (storage StorageAPI, err error) {
|
|
if !strings.ContainsRune(disk, ':') || filepath.VolumeName(disk) != "" {
|
|
// Initialize filesystem storage API.
|
|
return newPosix(disk)
|
|
}
|
|
// Initialize rpc client storage API.
|
|
return newRPCClient(disk)
|
|
}
|
|
|
|
// Initialize all storage disks to bootstrap.
|
|
func bootstrapDisks(disks []string) ([]StorageAPI, error) {
|
|
storageDisks := make([]StorageAPI, len(disks))
|
|
for index, disk := range disks {
|
|
var err error
|
|
// Intentionally ignore disk not found errors while
|
|
// initializing POSIX, so that we have successfully
|
|
// initialized posix Storage. Subsequent calls to XL/Erasure
|
|
// will manage any errors related to disks.
|
|
storageDisks[index], err = newStorageLayer(disk)
|
|
if err != nil && err != errDiskNotFound {
|
|
return nil, err
|
|
}
|
|
}
|
|
return storageDisks, nil
|
|
}
|
|
|
|
// newXLObjects - initialize new xl object layer.
|
|
func newXLObjects(disks []string) (ObjectLayer, error) {
|
|
if err := checkSufficientDisks(disks); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
storageDisks, err := bootstrapDisks(disks)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Initialize object layer - like creating minioMetaBucket, cleaning up tmp files etc.
|
|
initObjectLayer(storageDisks...)
|
|
|
|
// Load saved XL format.json and validate.
|
|
newDisks, err := loadFormatXL(storageDisks)
|
|
if err != nil {
|
|
switch err {
|
|
case errUnformattedDisk:
|
|
// Save new XL format.
|
|
errSave := initFormatXL(storageDisks)
|
|
if errSave != nil {
|
|
return nil, errSave
|
|
}
|
|
newDisks = storageDisks
|
|
default:
|
|
// errCorruptedDisk - error.
|
|
return nil, fmt.Errorf("Unable to recognize backend format, %s", err)
|
|
}
|
|
}
|
|
|
|
// FIXME: healFormatXL(newDisks)
|
|
|
|
storage, err := newXL(newDisks)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Return successfully initialized object layer.
|
|
return xlObjects{
|
|
storage: storage,
|
|
listObjectMap: make(map[listParams][]*treeWalker),
|
|
listObjectMapMutex: &sync.Mutex{},
|
|
}, nil
|
|
}
|
|
|
|
/// Bucket operations
|
|
|
|
// MakeBucket - make a bucket.
|
|
func (xl xlObjects) MakeBucket(bucket string) error {
|
|
nsMutex.Lock(bucket, "")
|
|
defer nsMutex.Unlock(bucket, "")
|
|
return makeBucket(xl.storage, bucket)
|
|
}
|
|
|
|
// GetBucketInfo - get bucket info.
|
|
func (xl xlObjects) GetBucketInfo(bucket string) (BucketInfo, error) {
|
|
nsMutex.RLock(bucket, "")
|
|
defer nsMutex.RUnlock(bucket, "")
|
|
return getBucketInfo(xl.storage, bucket)
|
|
}
|
|
|
|
// ListBuckets - list buckets.
|
|
func (xl xlObjects) ListBuckets() ([]BucketInfo, error) {
|
|
return listBuckets(xl.storage)
|
|
}
|
|
|
|
// DeleteBucket - delete a bucket.
|
|
func (xl xlObjects) DeleteBucket(bucket string) error {
|
|
nsMutex.Lock(bucket, "")
|
|
nsMutex.Unlock(bucket, "")
|
|
return deleteBucket(xl.storage, bucket)
|
|
}
|
|
|
|
/// Object Operations
|
|
|
|
// GetObject - get an object.
|
|
func (xl xlObjects) GetObject(bucket, object string, startOffset int64) (io.ReadCloser, error) {
|
|
// Verify if bucket is valid.
|
|
if !IsValidBucketName(bucket) {
|
|
return nil, BucketNameInvalid{Bucket: bucket}
|
|
}
|
|
// Verify if object is valid.
|
|
if !IsValidObjectName(object) {
|
|
return nil, ObjectNameInvalid{Bucket: bucket, Object: object}
|
|
}
|
|
nsMutex.RLock(bucket, object)
|
|
defer nsMutex.RUnlock(bucket, object)
|
|
if !isMultipartObject(xl.storage, bucket, object) {
|
|
_, err := xl.storage.StatFile(bucket, object)
|
|
if err == nil {
|
|
var reader io.ReadCloser
|
|
reader, err = xl.storage.ReadFile(bucket, object, startOffset)
|
|
if err != nil {
|
|
return nil, toObjectErr(err, bucket, object)
|
|
}
|
|
return reader, nil
|
|
}
|
|
return nil, toObjectErr(err, bucket, object)
|
|
}
|
|
fileReader, fileWriter := io.Pipe()
|
|
info, err := getMultipartObjectInfo(xl.storage, bucket, object)
|
|
if err != nil {
|
|
return nil, toObjectErr(err, bucket, object)
|
|
}
|
|
partIndex, offset, err := info.GetPartNumberOffset(startOffset)
|
|
if err != nil {
|
|
return nil, toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Hold a read lock once more which can be released after the following go-routine ends.
|
|
// We hold RLock once more because the current function would return before the go routine below
|
|
// executes and hence releasing the read lock (because of defer'ed nsMutex.RUnlock() call).
|
|
nsMutex.RLock(bucket, object)
|
|
go func() {
|
|
defer nsMutex.RUnlock(bucket, object)
|
|
for ; partIndex < len(info.Parts); partIndex++ {
|
|
part := info.Parts[partIndex]
|
|
r, err := xl.storage.ReadFile(bucket, pathJoin(object, partNumToPartFileName(part.PartNumber)), offset)
|
|
if err != nil {
|
|
fileWriter.CloseWithError(err)
|
|
return
|
|
}
|
|
// Reset offset to 0 as it would be non-0 only for the first loop if startOffset is non-0.
|
|
offset = 0
|
|
if _, err = io.Copy(fileWriter, r); err != nil {
|
|
switch reader := r.(type) {
|
|
case *io.PipeReader:
|
|
reader.CloseWithError(err)
|
|
case io.ReadCloser:
|
|
reader.Close()
|
|
}
|
|
fileWriter.CloseWithError(err)
|
|
return
|
|
}
|
|
// Close the readerCloser that reads multiparts of an object from the xl storage layer.
|
|
// Not closing leaks underlying file descriptors.
|
|
r.Close()
|
|
}
|
|
fileWriter.Close()
|
|
}()
|
|
return fileReader, nil
|
|
}
|
|
|
|
// Return the partsInfo of a special multipart object.
|
|
func getMultipartObjectInfo(storage StorageAPI, bucket, object string) (info MultipartObjectInfo, err error) {
|
|
offset := int64(0)
|
|
r, err := storage.ReadFile(bucket, pathJoin(object, multipartMetaFile), offset)
|
|
if err != nil {
|
|
return MultipartObjectInfo{}, err
|
|
}
|
|
decoder := json.NewDecoder(r)
|
|
err = decoder.Decode(&info)
|
|
if err != nil {
|
|
return MultipartObjectInfo{}, err
|
|
}
|
|
return info, nil
|
|
}
|
|
|
|
// Return ObjectInfo.
|
|
func (xl xlObjects) getObjectInfo(bucket, object string) (objInfo ObjectInfo, err error) {
|
|
objInfo.Bucket = bucket
|
|
objInfo.Name = object
|
|
// First see if the object was a simple-PUT upload.
|
|
fi, err := xl.storage.StatFile(bucket, object)
|
|
if err != nil {
|
|
if err != errFileNotFound {
|
|
return ObjectInfo{}, err
|
|
}
|
|
var info MultipartObjectInfo
|
|
// Check if the object was multipart upload.
|
|
info, err = getMultipartObjectInfo(xl.storage, bucket, object)
|
|
if err != nil {
|
|
return ObjectInfo{}, err
|
|
}
|
|
objInfo.Size = info.Size
|
|
objInfo.ModTime = info.ModTime
|
|
objInfo.MD5Sum = info.MD5Sum
|
|
objInfo.ContentType = info.ContentType
|
|
objInfo.ContentEncoding = info.ContentEncoding
|
|
} else {
|
|
metadata := make(map[string]string)
|
|
offset := int64(0) // To read entire content
|
|
r, err := xl.storage.ReadFile(bucket, pathJoin(object, "meta.json"), offset)
|
|
if err != nil {
|
|
return ObjectInfo{}, toObjectErr(err, bucket, object)
|
|
}
|
|
decoder := json.NewDecoder(r)
|
|
if err = decoder.Decode(&metadata); err != nil {
|
|
return ObjectInfo{}, toObjectErr(err, bucket, object)
|
|
}
|
|
contentType := metadata["content-type"]
|
|
if len(contentType) == 0 {
|
|
contentType = "application/octet-stream"
|
|
if objectExt := filepath.Ext(object); objectExt != "" {
|
|
content, ok := mimedb.DB[strings.ToLower(strings.TrimPrefix(objectExt, "."))]
|
|
if ok {
|
|
contentType = content.ContentType
|
|
}
|
|
}
|
|
}
|
|
objInfo.Size = fi.Size
|
|
objInfo.IsDir = fi.Mode.IsDir()
|
|
objInfo.ModTime = fi.ModTime
|
|
objInfo.MD5Sum = metadata["md5Sum"]
|
|
objInfo.ContentType = contentType
|
|
objInfo.ContentEncoding = metadata["content-encoding"]
|
|
}
|
|
return objInfo, nil
|
|
}
|
|
|
|
// GetObjectInfo - get object info.
|
|
func (xl xlObjects) GetObjectInfo(bucket, object string) (ObjectInfo, error) {
|
|
// Verify if bucket is valid.
|
|
if !IsValidBucketName(bucket) {
|
|
return ObjectInfo{}, BucketNameInvalid{Bucket: bucket}
|
|
}
|
|
// Verify if object is valid.
|
|
if !IsValidObjectName(object) {
|
|
return ObjectInfo{}, ObjectNameInvalid{Bucket: bucket, Object: object}
|
|
}
|
|
nsMutex.RLock(bucket, object)
|
|
defer nsMutex.RUnlock(bucket, object)
|
|
info, err := xl.getObjectInfo(bucket, object)
|
|
if err != nil {
|
|
return ObjectInfo{}, toObjectErr(err, bucket, object)
|
|
}
|
|
return info, nil
|
|
}
|
|
|
|
// PutObject - create an object.
|
|
func (xl xlObjects) PutObject(bucket string, object string, size int64, data io.Reader, metadata map[string]string) (string, error) {
|
|
// Verify if bucket is valid.
|
|
if !IsValidBucketName(bucket) {
|
|
return "", BucketNameInvalid{Bucket: bucket}
|
|
}
|
|
// Verify bucket exists.
|
|
if !isBucketExist(xl.storage, bucket) {
|
|
return "", BucketNotFound{Bucket: bucket}
|
|
}
|
|
if !IsValidObjectName(object) {
|
|
return "", ObjectNameInvalid{
|
|
Bucket: bucket,
|
|
Object: object,
|
|
}
|
|
}
|
|
// No metadata is set, allocate a new one.
|
|
if metadata == nil {
|
|
metadata = make(map[string]string)
|
|
}
|
|
nsMutex.Lock(bucket, object)
|
|
defer nsMutex.Unlock(bucket, object)
|
|
|
|
tempObj := path.Join(tmpMetaPrefix, bucket, object)
|
|
fileWriter, err := xl.storage.CreateFile(minioMetaBucket, tempObj)
|
|
if err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Initialize md5 writer.
|
|
md5Writer := md5.New()
|
|
|
|
// Instantiate a new multi writer.
|
|
multiWriter := io.MultiWriter(md5Writer, fileWriter)
|
|
|
|
// Instantiate checksum hashers and create a multiwriter.
|
|
if size > 0 {
|
|
if _, err = io.CopyN(multiWriter, data, size); err != nil {
|
|
if clErr := safeCloseAndRemove(fileWriter); clErr != nil {
|
|
return "", toObjectErr(clErr, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
} else {
|
|
if _, err = io.Copy(multiWriter, data); err != nil {
|
|
if clErr := safeCloseAndRemove(fileWriter); clErr != nil {
|
|
return "", toObjectErr(clErr, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
}
|
|
|
|
newMD5Hex := hex.EncodeToString(md5Writer.Sum(nil))
|
|
// Update the md5sum if not set with the newly calculated one.
|
|
if len(metadata["md5Sum"]) == 0 {
|
|
metadata["md5Sum"] = newMD5Hex
|
|
}
|
|
|
|
// md5Hex representation.
|
|
md5Hex := metadata["md5Sum"]
|
|
if md5Hex != "" {
|
|
if newMD5Hex != md5Hex {
|
|
if err = safeCloseAndRemove(fileWriter); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
return "", BadDigest{md5Hex, newMD5Hex}
|
|
}
|
|
}
|
|
|
|
err = fileWriter.Close()
|
|
if err != nil {
|
|
if clErr := safeCloseAndRemove(fileWriter); clErr != nil {
|
|
return "", toObjectErr(clErr, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Check if an object is present as one of the parent dir.
|
|
if err = xl.parentDirIsObject(bucket, path.Dir(object)); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Delete if an object already exists.
|
|
// FIXME: rename it to tmp file and delete only after
|
|
// the newly uploaded file is renamed from tmp location to
|
|
// the original location.
|
|
// Verify if the object is a multipart object.
|
|
if isMultipartObject(xl.storage, bucket, object) {
|
|
err = xl.deleteMultipartObject(bucket, object)
|
|
if err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
return newMD5Hex, nil
|
|
}
|
|
err = xl.deleteObject(bucket, object)
|
|
if err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
err = xl.storage.RenameFile(minioMetaBucket, tempObj, bucket, object)
|
|
if err != nil {
|
|
if dErr := xl.storage.DeleteFile(minioMetaBucket, tempObj); dErr != nil {
|
|
return "", toObjectErr(dErr, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
tempMetaJSONFile := path.Join(tmpMetaPrefix, bucket, object, "meta.json")
|
|
metaWriter, err := xl.storage.CreateFile(minioMetaBucket, tempMetaJSONFile)
|
|
if err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
encoder := json.NewEncoder(metaWriter)
|
|
err = encoder.Encode(&metadata)
|
|
if err != nil {
|
|
if clErr := safeCloseAndRemove(metaWriter); clErr != nil {
|
|
return "", toObjectErr(clErr, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
if err = metaWriter.Close(); err != nil {
|
|
if err = safeCloseAndRemove(metaWriter); err != nil {
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
metaJSONFile := path.Join(object, "meta.json")
|
|
err = xl.storage.RenameFile(minioMetaBucket, tempMetaJSONFile, bucket, metaJSONFile)
|
|
if err != nil {
|
|
if derr := xl.storage.DeleteFile(minioMetaBucket, tempMetaJSONFile); derr != nil {
|
|
return "", toObjectErr(derr, bucket, object)
|
|
}
|
|
return "", toObjectErr(err, bucket, object)
|
|
}
|
|
|
|
// Return md5sum, successfully wrote object.
|
|
return newMD5Hex, nil
|
|
}
|
|
|
|
// isMultipartObject - verifies if an object is special multipart file.
|
|
func isMultipartObject(storage StorageAPI, bucket, object string) bool {
|
|
_, err := storage.StatFile(bucket, pathJoin(object, multipartMetaFile))
|
|
if err != nil {
|
|
if err == errFileNotFound {
|
|
return false
|
|
}
|
|
errorIf(err, "Failed to stat file "+bucket+pathJoin(object, multipartMetaFile))
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// deleteMultipartObject - deletes only multipart object.
|
|
func (xl xlObjects) deleteMultipartObject(bucket, object string) error {
|
|
// Get parts info.
|
|
info, err := getMultipartObjectInfo(xl.storage, bucket, object)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
// Range through all files and delete it.
|
|
var wg = &sync.WaitGroup{}
|
|
var errs = make([]error, len(info.Parts))
|
|
for index, part := range info.Parts {
|
|
wg.Add(1)
|
|
// Start deleting parts in routine.
|
|
go func(index int, part MultipartPartInfo) {
|
|
defer wg.Done()
|
|
partFileName := partNumToPartFileName(part.PartNumber)
|
|
errs[index] = xl.storage.DeleteFile(bucket, pathJoin(object, partFileName))
|
|
}(index, part)
|
|
}
|
|
// Wait for all the deletes to finish.
|
|
wg.Wait()
|
|
// Loop through and validate if any errors, if we are unable to remove any part return
|
|
// "unexpected" error as returning any other error might be misleading. For ex.
|
|
// if DeleteFile() had returned errFileNotFound and we return it, then client would see
|
|
// ObjectNotFound which is misleading.
|
|
for _, err := range errs {
|
|
if err != nil {
|
|
return errUnexpected
|
|
}
|
|
}
|
|
err = xl.storage.DeleteFile(bucket, pathJoin(object, multipartMetaFile))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// deleteObject - deletes a regular object.
|
|
func (xl xlObjects) deleteObject(bucket, object string) error {
|
|
metaJSONFile := path.Join(object, "meta.json")
|
|
// Ignore if meta.json file doesn't exist.
|
|
if err := xl.storage.DeleteFile(bucket, metaJSONFile); err != nil {
|
|
if err != errFileNotFound {
|
|
return err
|
|
}
|
|
}
|
|
if err := xl.storage.DeleteFile(bucket, object); err != nil {
|
|
if err != errFileNotFound {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DeleteObject - delete the object.
|
|
func (xl xlObjects) DeleteObject(bucket, object string) error {
|
|
// Verify if bucket is valid.
|
|
if !IsValidBucketName(bucket) {
|
|
return BucketNameInvalid{Bucket: bucket}
|
|
}
|
|
if !IsValidObjectName(object) {
|
|
return ObjectNameInvalid{Bucket: bucket, Object: object}
|
|
}
|
|
nsMutex.Lock(bucket, object)
|
|
defer nsMutex.Unlock(bucket, object)
|
|
// Verify if the object is a multipart object.
|
|
if isMultipartObject(xl.storage, bucket, object) {
|
|
err := xl.deleteMultipartObject(bucket, object)
|
|
if err != nil {
|
|
return toObjectErr(err, bucket, object)
|
|
}
|
|
return nil
|
|
}
|
|
err := xl.deleteObject(bucket, object)
|
|
if err != nil {
|
|
return toObjectErr(err, bucket, object)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// ListObjects - list all objects at prefix, delimited by '/'.
|
|
func (xl xlObjects) ListObjects(bucket, prefix, marker, delimiter string, maxKeys int) (ListObjectsInfo, error) {
|
|
return listObjectsCommon(xl, bucket, prefix, marker, delimiter, maxKeys)
|
|
}
|
|
|