/* * Minio Cloud Storage, (C) 2016 Minio, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package main import ( "io" "path/filepath" "strings" "sync" "github.com/Sirupsen/logrus" "github.com/minio/minio/pkg/mimedb" ) // fsObjects - Implements fs object layer. type fsObjects struct { storage StorageAPI listObjectMap map[listParams][]*treeWalker listObjectMapMutex *sync.Mutex } // FIXME: constructor should return a pointer. // newFSObjects - initialize new fs object layer. func newFSObjects(exportPath string) (ObjectLayer, error) { var storage StorageAPI var err error if !strings.ContainsRune(exportPath, ':') || filepath.VolumeName(exportPath) != "" { // Initialize filesystem storage API. storage, err = newPosix(exportPath) if err != nil { return nil, err } } else { // Initialize rpc client storage API. storage, err = newRPCClient(exportPath) if err != nil { return nil, err } } // Initialize object layer - like creating minioMetaBucket, // cleaning up tmp files etc. initObjectLayer(storage) // Return successfully initialized object layer. return fsObjects{ storage: storage, listObjectMap: make(map[listParams][]*treeWalker), listObjectMapMutex: &sync.Mutex{}, }, nil } /// Bucket operations // MakeBucket - make a bucket. func (fs fsObjects) MakeBucket(bucket string) error { return makeBucket(fs.storage, bucket) } // GetBucketInfo - get bucket info. func (fs fsObjects) GetBucketInfo(bucket string) (BucketInfo, error) { return getBucketInfo(fs.storage, bucket) } // ListBuckets - list buckets. func (fs fsObjects) ListBuckets() ([]BucketInfo, error) { return listBuckets(fs.storage) } // DeleteBucket - delete a bucket. func (fs fsObjects) DeleteBucket(bucket string) error { return deleteBucket(fs.storage, bucket) } /// Object Operations // GetObject - get an object. func (fs fsObjects) GetObject(bucket, object string, startOffset int64) (io.ReadCloser, error) { // Verify if bucket is valid. if !IsValidBucketName(bucket) { return nil, (BucketNameInvalid{Bucket: bucket}) } // Verify if object is valid. if !IsValidObjectName(object) { return nil, (ObjectNameInvalid{Bucket: bucket, Object: object}) } fileReader, err := fs.storage.ReadFile(bucket, object, startOffset) if err != nil { return nil, toObjectErr(err, bucket, object) } return fileReader, nil } // GetObjectInfo - get object info. func (fs fsObjects) GetObjectInfo(bucket, object string) (ObjectInfo, error) { // Verify if bucket is valid. if !IsValidBucketName(bucket) { return ObjectInfo{}, (BucketNameInvalid{Bucket: bucket}) } // Verify if object is valid. if !IsValidObjectName(object) { return ObjectInfo{}, (ObjectNameInvalid{Bucket: bucket, Object: object}) } fi, err := fs.storage.StatFile(bucket, object) if err != nil { return ObjectInfo{}, toObjectErr(err, bucket, object) } contentType := "application/octet-stream" if objectExt := filepath.Ext(object); objectExt != "" { content, ok := mimedb.DB[strings.ToLower(strings.TrimPrefix(objectExt, "."))] if ok { contentType = content.ContentType } } return ObjectInfo{ Bucket: bucket, Name: object, ModTime: fi.ModTime, Size: fi.Size, IsDir: fi.Mode.IsDir(), ContentType: contentType, MD5Sum: "", // Read from metadata. }, nil } // PutObject - create an object. func (fs fsObjects) PutObject(bucket string, object string, size int64, data io.Reader, metadata map[string]string) (string, error) { return putObjectCommon(fs.storage, bucket, object, size, data, metadata) } func (fs fsObjects) DeleteObject(bucket, object string) error { // Verify if bucket is valid. if !IsValidBucketName(bucket) { return BucketNameInvalid{Bucket: bucket} } if !IsValidObjectName(object) { return ObjectNameInvalid{Bucket: bucket, Object: object} } if err := fs.storage.DeleteFile(bucket, object); err != nil { return toObjectErr(err, bucket, object) } return nil } func (fs fsObjects) ListObjects(bucket, prefix, marker, delimiter string, maxKeys int) (ListObjectsInfo, error) { // Verify if bucket is valid. if !IsValidBucketName(bucket) { return ListObjectsInfo{}, BucketNameInvalid{Bucket: bucket} } // Verify whether the bucket exists. if isExist, err := isBucketExist(fs.storage, bucket); err != nil { return ListObjectsInfo{}, err } else if !isExist { return ListObjectsInfo{}, BucketNotFound{Bucket: bucket} } if !IsValidObjectPrefix(prefix) { return ListObjectsInfo{}, ObjectNameInvalid{Bucket: bucket, Object: prefix} } // Verify if delimiter is anything other than '/', which we do not support. if delimiter != "" && delimiter != slashSeparator { return ListObjectsInfo{}, UnsupportedDelimiter{ Delimiter: delimiter, } } // Verify if marker has prefix. if marker != "" { if !strings.HasPrefix(marker, prefix) { return ListObjectsInfo{}, InvalidMarkerPrefixCombination{ Marker: marker, Prefix: prefix, } } } if maxKeys == 0 { return ListObjectsInfo{}, nil } // Over flowing count - reset to maxObjectList. if maxKeys < 0 || maxKeys > maxObjectList { maxKeys = maxObjectList } // Default is recursive, if delimiter is set then list non recursive. recursive := true if delimiter == slashSeparator { recursive = false } walker := lookupTreeWalk(fs, listParams{bucket, recursive, marker, prefix}) if walker == nil { walker = startTreeWalk(fs, bucket, prefix, marker, recursive) } var fileInfos []FileInfo var eof bool var nextMarker string log.Debugf("Reading from the tree walk channel has begun.") for i := 0; i < maxKeys; { walkResult, ok := <-walker.ch if !ok { // Closed channel. eof = true break } // For any walk error return right away. if walkResult.err != nil { log.WithFields(logrus.Fields{ "bucket": bucket, "prefix": prefix, "marker": marker, "recursive": recursive, }).Debugf("Walk resulted in an error %s", walkResult.err) // File not found is a valid case. if walkResult.err == errFileNotFound { return ListObjectsInfo{}, nil } return ListObjectsInfo{}, toObjectErr(walkResult.err, bucket, prefix) } fileInfo := walkResult.fileInfo nextMarker = fileInfo.Name fileInfos = append(fileInfos, fileInfo) if walkResult.end { eof = true break } i++ } if len(fileInfos) == 0 { eof = true } params := listParams{bucket, recursive, nextMarker, prefix} log.WithFields(logrus.Fields{ "bucket": params.bucket, "recursive": params.recursive, "marker": params.marker, "prefix": params.prefix, }).Debugf("Save the tree walk into map for subsequent requests.") if !eof { saveTreeWalk(fs, params, walker) } result := ListObjectsInfo{IsTruncated: !eof} for _, fileInfo := range fileInfos { // With delimiter set we fill in NextMarker and Prefixes. if delimiter == slashSeparator { result.NextMarker = fileInfo.Name if fileInfo.Mode.IsDir() { result.Prefixes = append(result.Prefixes, fileInfo.Name) continue } } result.Objects = append(result.Objects, ObjectInfo{ Name: fileInfo.Name, ModTime: fileInfo.ModTime, Size: fileInfo.Size, IsDir: false, }) } return result, nil }