Merge pull request #307 from fkautz/pr_out_better_handling_of_metadata_in_encoded

master
Frederick F. Kautz IV 10 years ago
commit ed5051a795
  1. 77
      pkg/storage/encoded/encoded.go

@ -18,6 +18,7 @@ package encoded
import ( import (
"bytes" "bytes"
"encoding/hex"
"errors" "errors"
"github.com/minio-io/minio/pkg/donutbox" "github.com/minio-io/minio/pkg/donutbox"
"github.com/minio-io/minio/pkg/encoding/erasure" "github.com/minio-io/minio/pkg/encoding/erasure"
@ -25,6 +26,7 @@ import (
"github.com/minio-io/minio/pkg/utils/split" "github.com/minio-io/minio/pkg/utils/split"
"io" "io"
"strconv" "strconv"
"time"
) )
// StorageDriver creates a new single disk storage driver using donut without encoding. // StorageDriver creates a new single disk storage driver using donut without encoding.
@ -69,13 +71,24 @@ func (diskStorage StorageDriver) GetBucketPolicy(bucket string) (storage.BucketP
// GetObject retrieves an object and writes it to a writer // GetObject retrieves an object and writes it to a writer
func (diskStorage StorageDriver) GetObject(target io.Writer, bucket, key string) (int64, error) { func (diskStorage StorageDriver) GetObject(target io.Writer, bucket, key string) (int64, error) {
metadata, err := diskStorage.donutBox.GetObjectMetadata(bucket, key, 0) metadata, err := diskStorage.donutBox.GetObjectMetadata(bucket, key, 0)
chunkCount, err := strconv.Atoi(metadata["chunkCount"]) k, err := strconv.Atoi(metadata["erasureK"])
columnCount, err := strconv.Atoi(metadata["columnCount"]) if err != nil {
return 0, errors.New("Cannot parse erasureK")
}
m, err := strconv.Atoi(metadata["erasureM"])
if err != nil {
return 0, errors.New("Cannot parse erasureM")
}
columnCount := k + m
blockSize, err := strconv.Atoi(metadata["blockSize"]) blockSize, err := strconv.Atoi(metadata["blockSize"])
length, err := strconv.Atoi(metadata["length"])
if err != nil { if err != nil {
return 0, err return 0, errors.New("Cannot parse blockSize")
} }
size, err := strconv.Atoi(metadata["size"])
if err != nil {
return 0, errors.New("Cannot parse length")
}
chunkCount := size/blockSize + 1
var readers []io.Reader var readers []io.Reader
for column := 0; column < columnCount; column++ { for column := 0; column < columnCount; column++ {
reader, err := diskStorage.donutBox.GetObjectReader(bucket, key, uint(column)) reader, err := diskStorage.donutBox.GetObjectReader(bucket, key, uint(column))
@ -85,8 +98,8 @@ func (diskStorage StorageDriver) GetObject(target io.Writer, bucket, key string)
readers = append(readers, reader) readers = append(readers, reader)
} }
totalWritten := int64(length) totalWritten := int64(size)
totalRemaining := int64(length) totalRemaining := int64(size)
if err != err { if err != err {
return 0, err return 0, err
} }
@ -127,8 +140,22 @@ func (diskStorage StorageDriver) GetPartialObject(w io.Writer, bucket, object st
} }
// GetObjectMetadata retrieves an object's metadata // GetObjectMetadata retrieves an object's metadata
func (diskStorage StorageDriver) GetObjectMetadata(bucket, key string, prefix string) (metadata storage.ObjectMetadata, err error) { func (diskStorage StorageDriver) GetObjectMetadata(bucket, key string, prefix string) (storage.ObjectMetadata, error) {
return metadata, errors.New("Not Implemented") metadata, err := diskStorage.donutBox.GetObjectMetadata(bucket, key, 0)
if err != nil {
return storage.ObjectMetadata{}, err
}
created, err := time.Parse(metadata["created"], time.RFC3339Nano)
size, err := strconv.ParseInt(metadata["size"], 10, 64)
objectMetadata := storage.ObjectMetadata{
Bucket: bucket,
Key: key,
ContentType: metadata["contentType"],
Created: created,
ETag: hex.EncodeToString([]byte(metadata["md5"])),
Size: size,
}
return objectMetadata, nil
} }
// ListObjects lists objects // ListObjects lists objects
@ -177,12 +204,18 @@ func (diskStorage StorageDriver) CreateObject(bucketKey string, objectKey string
} }
// close connections // close connections
metadata := make(map[string]string) metadataObj := storage.ObjectMetadata{
metadata["length"] = strconv.FormatUint(totalLength, 10) Bucket: bucketKey,
metadata["blockSize"] = strconv.FormatUint(uint64(blockSize), 10) Key: objectKey,
metadata["chunkCount"] = strconv.FormatUint(uint64(chunkCount), 10)
metadata["columnCount"] = strconv.FormatUint(uint64(16), 10) ContentType: contentType,
// metadata["md5"] := md5sum Created: time.Now(),
ETag: "md5",
Size: int64(totalLength),
}
metadata := createMetadata(metadataObj, blockSize, 8, 8, "Cauchy")
for column := uint(0); column < 16; column++ { for column := uint(0); column < 16; column++ {
writers[column].SetMetadata(metadata) writers[column].SetMetadata(metadata)
} }
@ -208,3 +241,19 @@ func closeAllWritersWithError(writers []*donutbox.NewObject, err error) {
} }
} }
} }
func createMetadata(metadataObject storage.ObjectMetadata, blockSize int, k, m uint8, technique string) map[string]string {
metadata := make(map[string]string)
metadata["bucket"] = metadataObject.Bucket
metadata["key"] = metadataObject.Key
metadata["contentType"] = metadataObject.ContentType
metadata["created"] = metadataObject.Created.Format(time.RFC3339Nano)
metadata["md5"] = metadataObject.ETag
metadata["size"] = strconv.FormatInt(metadataObject.Size, 10)
metadata["blockSize"] = strconv.FormatUint(uint64(blockSize), 10)
metadata["erasureK"] = strconv.FormatUint(uint64(k), 10)
metadata["erasureM"] = strconv.FormatUint(uint64(m), 10)
metadata["erasureTechnique"] = technique
return metadata
}

Loading…
Cancel
Save