io.EOF is okay since io.ReadFull will not have read any bytes at all. Also making error channel receive only for go routine.master
parent
4ac23d747c
commit
e082f26e10
@ -1,2 +0,0 @@ |
|||||||
TESTPREFIX.* |
|
||||||
NEWFILE |
|
@ -1,179 +0,0 @@ |
|||||||
/* |
|
||||||
* Minio Cloud Storage, (C) 2014 Minio, Inc. |
|
||||||
* |
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License"); |
|
||||||
* you may not use this file except in compliance with the License. |
|
||||||
* You may obtain a copy of the License at |
|
||||||
* |
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
* |
|
||||||
* Unless required by applicable law or agreed to in writing, software |
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS, |
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
||||||
* See the License for the specific language governing permissions and |
|
||||||
* limitations under the License. |
|
||||||
*/ |
|
||||||
|
|
||||||
package split |
|
||||||
|
|
||||||
import ( |
|
||||||
"bufio" |
|
||||||
"bytes" |
|
||||||
"errors" |
|
||||||
"io" |
|
||||||
"io/ioutil" |
|
||||||
"os" |
|
||||||
"strconv" |
|
||||||
"strings" |
|
||||||
|
|
||||||
"github.com/minio/minio/pkg/iodine" |
|
||||||
) |
|
||||||
|
|
||||||
// Message - message structure for results from the Stream goroutine
|
|
||||||
type Message struct { |
|
||||||
Data []byte |
|
||||||
Err error |
|
||||||
} |
|
||||||
|
|
||||||
// Stream reads from io.Reader, splits the data into chunks, and sends
|
|
||||||
// each chunk to the channel. This method runs until an EOF or error occurs. If
|
|
||||||
// an error occurs, the method sends the error over the channel and returns.
|
|
||||||
// Before returning, the channel is always closed.
|
|
||||||
//
|
|
||||||
// The user should run this as a gorountine and retrieve the data over the
|
|
||||||
// channel.
|
|
||||||
//
|
|
||||||
// channel := make(chan Message)
|
|
||||||
// go Stream(reader, chunkSize, channel)
|
|
||||||
// for chunk := range channel {
|
|
||||||
// log.Println(chunk.Data)
|
|
||||||
// }
|
|
||||||
func Stream(reader io.Reader, chunkSize uint64) <-chan Message { |
|
||||||
ch := make(chan Message) |
|
||||||
go splitStreamGoRoutine(reader, chunkSize, ch) |
|
||||||
return ch |
|
||||||
} |
|
||||||
|
|
||||||
func splitStreamGoRoutine(reader io.Reader, chunkSize uint64, ch chan<- Message) { |
|
||||||
defer close(ch) |
|
||||||
|
|
||||||
// we read until EOF or another error
|
|
||||||
var readError error |
|
||||||
|
|
||||||
// run this until an EOF or error occurs
|
|
||||||
for readError == nil { |
|
||||||
// keep track of how much data has been read
|
|
||||||
var totalRead uint64 |
|
||||||
// Create a buffer to write the current chunk into
|
|
||||||
var bytesBuffer bytes.Buffer |
|
||||||
bytesWriter := bufio.NewWriter(&bytesBuffer) |
|
||||||
// read a full chunk
|
|
||||||
for totalRead < chunkSize && readError == nil { |
|
||||||
var currentRead int |
|
||||||
// if we didn't read a full chunk, we should attempt to read again.
|
|
||||||
// We create a byte array representing how much space is left
|
|
||||||
// unwritten in the given chunk
|
|
||||||
chunk := make([]byte, chunkSize-totalRead) |
|
||||||
currentRead, readError = reader.Read(chunk) |
|
||||||
// keep track of how much we have read in total
|
|
||||||
totalRead = totalRead + uint64(currentRead) |
|
||||||
// prune the array to only what has been read, write to chunk buffer
|
|
||||||
chunk = chunk[0:currentRead] |
|
||||||
bytesWriter.Write(chunk) |
|
||||||
} |
|
||||||
// flush stream to underlying byte buffer
|
|
||||||
bytesWriter.Flush() |
|
||||||
// if we have data available, send it over the channel
|
|
||||||
if bytesBuffer.Len() != 0 { |
|
||||||
ch <- Message{bytesBuffer.Bytes(), nil} |
|
||||||
} |
|
||||||
} |
|
||||||
// if we have an error other than an EOF, send it over the channel
|
|
||||||
if readError != io.EOF { |
|
||||||
ch <- Message{nil, readError} |
|
||||||
} |
|
||||||
} |
|
||||||
|
|
||||||
// JoinFiles reads from a given directory, joins data in chunks with prefix and sends
|
|
||||||
// an io.Reader.
|
|
||||||
//
|
|
||||||
// var err error
|
|
||||||
// for err == nil {
|
|
||||||
// buf := make([]byte, 1024*1024)
|
|
||||||
// reader := JoinFiles("mydirectory", "mypreferred-prefix")
|
|
||||||
// _, err = reader.Read(buf)
|
|
||||||
// fmt.Println(buf)
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
func JoinFiles(dirname string, inputPrefix string) io.Reader { |
|
||||||
reader, writer := io.Pipe() |
|
||||||
fileInfos, readError := ioutil.ReadDir(dirname) |
|
||||||
if readError != nil { |
|
||||||
writer.CloseWithError(readError) |
|
||||||
return nil |
|
||||||
} |
|
||||||
|
|
||||||
var newfileInfos []os.FileInfo |
|
||||||
for _, fi := range fileInfos { |
|
||||||
if strings.Contains(fi.Name(), inputPrefix) == true { |
|
||||||
newfileInfos = append(newfileInfos, fi) |
|
||||||
} |
|
||||||
} |
|
||||||
|
|
||||||
if len(newfileInfos) == 0 { |
|
||||||
nofilesError := iodine.New(errors.New("no files found for given prefix "+inputPrefix), nil) |
|
||||||
writer.CloseWithError(nofilesError) |
|
||||||
return nil |
|
||||||
} |
|
||||||
|
|
||||||
go joinFilesInGoRoutine(newfileInfos, writer) |
|
||||||
return reader |
|
||||||
} |
|
||||||
|
|
||||||
func joinFilesInGoRoutine(fileInfos []os.FileInfo, writer *io.PipeWriter) { |
|
||||||
for _, fileInfo := range fileInfos { |
|
||||||
file, err := os.Open(fileInfo.Name()) |
|
||||||
defer file.Close() |
|
||||||
for err != nil { |
|
||||||
writer.CloseWithError(err) |
|
||||||
return |
|
||||||
} |
|
||||||
_, err = io.Copy(writer, file) |
|
||||||
if err != nil { |
|
||||||
writer.CloseWithError(err) |
|
||||||
return |
|
||||||
} |
|
||||||
} |
|
||||||
writer.Close() |
|
||||||
} |
|
||||||
|
|
||||||
// FileWithPrefix - Takes a file and splits it into chunks with size chunkSize. The output
|
|
||||||
// filename is given with outputPrefix.
|
|
||||||
func FileWithPrefix(filename string, chunkSize uint64, outputPrefix string) error { |
|
||||||
// open file
|
|
||||||
file, err := os.Open(filename) |
|
||||||
defer file.Close() |
|
||||||
if err != nil { |
|
||||||
return err |
|
||||||
} |
|
||||||
|
|
||||||
if outputPrefix == "" { |
|
||||||
return errors.New("Invalid argument outputPrefix cannot be empty string") |
|
||||||
} |
|
||||||
|
|
||||||
// used to write each chunk out as a separate file. {{outputPrefix}}.{{i}}
|
|
||||||
i := 0 |
|
||||||
|
|
||||||
// write each chunk out to a separate file
|
|
||||||
for chunk := range Stream(file, chunkSize) { |
|
||||||
if chunk.Err != nil { |
|
||||||
return chunk.Err |
|
||||||
} |
|
||||||
err := ioutil.WriteFile(outputPrefix+"."+strconv.Itoa(i), chunk.Data, 0600) |
|
||||||
if err != nil { |
|
||||||
return err |
|
||||||
} |
|
||||||
i = i + 1 |
|
||||||
} |
|
||||||
return nil |
|
||||||
} |
|
@ -1,72 +0,0 @@ |
|||||||
/* |
|
||||||
* Minio Cloud Storage, (C) 2014 Minio, Inc. |
|
||||||
* |
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License"); |
|
||||||
* you may not use this file except in compliance with the License. |
|
||||||
* You may obtain a copy of the License at |
|
||||||
* |
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
* |
|
||||||
* Unless required by applicable law or agreed to in writing, software |
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS, |
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
||||||
* See the License for the specific language governing permissions and |
|
||||||
* limitations under the License. |
|
||||||
*/ |
|
||||||
|
|
||||||
package split_test |
|
||||||
|
|
||||||
import ( |
|
||||||
"bufio" |
|
||||||
"bytes" |
|
||||||
"io" |
|
||||||
"os" |
|
||||||
"strconv" |
|
||||||
"testing" |
|
||||||
|
|
||||||
"github.com/minio/minio/pkg/donut/split" |
|
||||||
. "gopkg.in/check.v1" |
|
||||||
) |
|
||||||
|
|
||||||
type MySuite struct{} |
|
||||||
|
|
||||||
var _ = Suite(&MySuite{}) |
|
||||||
|
|
||||||
func Test(t *testing.T) { TestingT(t) } |
|
||||||
|
|
||||||
func (s *MySuite) TestSplitStream(c *C) { |
|
||||||
var bytesBuffer bytes.Buffer |
|
||||||
bytesWriter := bufio.NewWriter(&bytesBuffer) |
|
||||||
for i := 0; i < 100; i++ { |
|
||||||
bytesWriter.Write([]byte(strconv.Itoa(i))) |
|
||||||
} |
|
||||||
bytesWriter.Flush() |
|
||||||
reader := bytes.NewReader(bytesBuffer.Bytes()) |
|
||||||
ch := split.Stream(reader, 25) |
|
||||||
var resultsBuffer bytes.Buffer |
|
||||||
resultsWriter := bufio.NewWriter(&resultsBuffer) |
|
||||||
for chunk := range ch { |
|
||||||
resultsWriter.Write(chunk.Data) |
|
||||||
} |
|
||||||
resultsWriter.Flush() |
|
||||||
c.Assert(bytes.Compare(bytesBuffer.Bytes(), resultsBuffer.Bytes()), Equals, 0) |
|
||||||
} |
|
||||||
|
|
||||||
func (s *MySuite) TestFileSplitJoin(c *C) { |
|
||||||
err := split.FileWithPrefix("testdata/TESTFILE", 1024, "TESTPREFIX") |
|
||||||
c.Assert(err, IsNil) |
|
||||||
err = split.FileWithPrefix("testdata/TESTFILE", 1024, "") |
|
||||||
c.Assert(err, Not(IsNil)) |
|
||||||
|
|
||||||
devnull, err := os.OpenFile(os.DevNull, 2, os.ModeAppend) |
|
||||||
defer devnull.Close() |
|
||||||
|
|
||||||
var reader io.Reader |
|
||||||
reader = split.JoinFiles(".", "ERROR") |
|
||||||
c.Assert(reader, IsNil) |
|
||||||
|
|
||||||
reader = split.JoinFiles(".", "TESTPREFIX") |
|
||||||
c.Assert(reader, Not(IsNil)) |
|
||||||
_, err = io.Copy(devnull, reader) |
|
||||||
c.Assert(err, IsNil) |
|
||||||
} |
|
File diff suppressed because it is too large
Load Diff
Loading…
Reference in new issue