You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
298 lines
9.6 KiB
298 lines
9.6 KiB
/*
|
|
* Minio Cloud Storage, (C) 2015, 2016 Minio, Inc.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
package cmd
|
|
|
|
import (
|
|
"path"
|
|
"strconv"
|
|
"sync"
|
|
"time"
|
|
|
|
. "gopkg.in/check.v1"
|
|
)
|
|
|
|
// API suite container common to both FS and XL.
|
|
type TestRPCControllerSuite struct {
|
|
serverType string
|
|
testServer TestServer
|
|
endPoint string
|
|
accessKey string
|
|
secretKey string
|
|
}
|
|
|
|
// Init and run test on XL backend.
|
|
var _ = Suite(&TestRPCControllerSuite{serverType: "XL"})
|
|
|
|
// Setting up the test suite.
|
|
// Starting the Test server with temporary FS backend.
|
|
func (s *TestRPCControllerSuite) SetUpSuite(c *C) {
|
|
s.testServer = StartTestRPCServer(c, s.serverType)
|
|
s.endPoint = s.testServer.Server.Listener.Addr().String()
|
|
s.accessKey = s.testServer.AccessKey
|
|
s.secretKey = s.testServer.SecretKey
|
|
}
|
|
|
|
// Called implicitly by "gopkg.in/check.v1" after all tests are run.
|
|
func (s *TestRPCControllerSuite) TearDownSuite(c *C) {
|
|
s.testServer.Stop()
|
|
}
|
|
|
|
// Tests to validate the correctness of lock instrumentation control RPC end point.
|
|
func (s *TestRPCControllerSuite) TestRPCControlLock(c *C) {
|
|
// enabling lock instrumentation.
|
|
globalDebugLock = true
|
|
// initializing the locks.
|
|
initNSLock(false)
|
|
// set debug lock info to `nil` so that the next tests have to initialize them again.
|
|
defer func() {
|
|
globalDebugLock = false
|
|
nsMutex.debugLockMap = nil
|
|
}()
|
|
|
|
expectedResult := []lockStateCase{
|
|
// Test case - 1.
|
|
// Case where 10 read locks are held.
|
|
// Entry for any of the 10 reads locks has to be found.
|
|
// Since they held in a loop, Lock origin for first 10 read locks (opsID 0-9) should be the same.
|
|
{
|
|
|
|
volume: "my-bucket",
|
|
path: "my-object",
|
|
opsID: "0",
|
|
readLock: true,
|
|
lockOrigin: "[lock held] in github.com/minio/minio/cmd.TestLockStats[/Users/hackintoshrao/mycode/go/src/github.com/minio/minio/cmd/namespace-lock_test.go:298]",
|
|
// expected metrics.
|
|
expectedErr: nil,
|
|
expectedLockStatus: "Running",
|
|
|
|
expectedGlobalLockCount: 10,
|
|
expectedRunningLockCount: 10,
|
|
expectedBlockedLockCount: 0,
|
|
|
|
expectedVolPathLockCount: 10,
|
|
expectedVolPathRunningCount: 10,
|
|
expectedVolPathBlockCount: 0,
|
|
},
|
|
// Test case 2.
|
|
// Testing the existance of entry for the last read lock (read lock with opsID "9").
|
|
{
|
|
|
|
volume: "my-bucket",
|
|
path: "my-object",
|
|
opsID: "9",
|
|
readLock: true,
|
|
lockOrigin: "[lock held] in github.com/minio/minio/cmd.TestLockStats[/Users/hackintoshrao/mycode/go/src/github.com/minio/minio/cmd/namespace-lock_test.go:298]",
|
|
// expected metrics.
|
|
expectedErr: nil,
|
|
expectedLockStatus: "Running",
|
|
|
|
expectedGlobalLockCount: 10,
|
|
expectedRunningLockCount: 10,
|
|
expectedBlockedLockCount: 0,
|
|
|
|
expectedVolPathLockCount: 10,
|
|
expectedVolPathRunningCount: 10,
|
|
expectedVolPathBlockCount: 0,
|
|
},
|
|
|
|
// Test case 3.
|
|
// Hold a write lock, and it should block since 10 read locks
|
|
// on <"my-bucket", "my-object"> are still held.
|
|
{
|
|
|
|
volume: "my-bucket",
|
|
path: "my-object",
|
|
opsID: "10",
|
|
readLock: false,
|
|
lockOrigin: "[lock held] in github.com/minio/minio/cmd.TestLockStats[/Users/hackintoshrao/mycode/go/src/github.com/minio/minio/cmd/namespace-lock_test.go:298]",
|
|
// expected metrics.
|
|
expectedErr: nil,
|
|
expectedLockStatus: "Blocked",
|
|
|
|
expectedGlobalLockCount: 11,
|
|
expectedRunningLockCount: 10,
|
|
expectedBlockedLockCount: 1,
|
|
|
|
expectedVolPathLockCount: 11,
|
|
expectedVolPathRunningCount: 10,
|
|
expectedVolPathBlockCount: 1,
|
|
},
|
|
|
|
// Test case 4.
|
|
// Expected result when all the read locks are released and the blocked write lock acquires the lock.
|
|
{
|
|
|
|
volume: "my-bucket",
|
|
path: "my-object",
|
|
opsID: "10",
|
|
readLock: false,
|
|
lockOrigin: "[lock held] in github.com/minio/minio/cmd.TestLockStats[/Users/hackintoshrao/mycode/go/src/github.com/minio/minio/cmd/namespace-lock_test.go:298]",
|
|
// expected metrics.
|
|
expectedErr: nil,
|
|
expectedLockStatus: "Running",
|
|
|
|
expectedGlobalLockCount: 1,
|
|
expectedRunningLockCount: 1,
|
|
expectedBlockedLockCount: 0,
|
|
|
|
expectedVolPathLockCount: 1,
|
|
expectedVolPathRunningCount: 1,
|
|
expectedVolPathBlockCount: 0,
|
|
},
|
|
// Test case - 5.
|
|
// At the end after locks are released, its verified whether the counters are set to 0.
|
|
{
|
|
|
|
volume: "my-bucket",
|
|
path: "my-object",
|
|
// expected metrics.
|
|
expectedErr: nil,
|
|
expectedLockStatus: "Blocked",
|
|
|
|
expectedGlobalLockCount: 0,
|
|
expectedRunningLockCount: 0,
|
|
expectedBlockedLockCount: 0,
|
|
},
|
|
}
|
|
|
|
// used to make sure that the tests don't end till locks held in other go routines are released.
|
|
var wg sync.WaitGroup
|
|
|
|
// Hold 5 read locks. We should find the info about these in the RPC response.
|
|
|
|
// hold 10 read locks.
|
|
// Then call the RPC control end point for obtaining lock instrumentation info.
|
|
|
|
for i := 0; i < 10; i++ {
|
|
nsMutex.RLock("my-bucket", "my-object", strconv.Itoa(i))
|
|
}
|
|
|
|
authCfg := &authConfig{
|
|
accessKey: s.accessKey,
|
|
secretKey: s.secretKey,
|
|
address: s.endPoint,
|
|
path: path.Join(reservedBucket, controlPath),
|
|
loginMethod: "Controller.LoginHandler",
|
|
}
|
|
|
|
client := newAuthClient(authCfg)
|
|
|
|
defer client.Close()
|
|
|
|
args := &GenericArgs{}
|
|
reply := &SystemLockState{}
|
|
// Call the lock instrumentation RPC end point.
|
|
err := client.Call("Controller.LockInfo", args, reply)
|
|
if err != nil {
|
|
c.Errorf("Add: expected no error but got string %q", err.Error())
|
|
}
|
|
// expected lock info.
|
|
expectedLockStats := expectedResult[0]
|
|
// verify the actual lock info with the expected one.
|
|
// verify the existance entry for first read lock (read lock with opsID "0").
|
|
verifyRPCLockInfoResponse(expectedLockStats, *reply, c, 1)
|
|
expectedLockStats = expectedResult[1]
|
|
// verify the actual lock info with the expected one.
|
|
// verify the existance entry for last read lock (read lock with opsID "9").
|
|
verifyRPCLockInfoResponse(expectedLockStats, *reply, c, 2)
|
|
|
|
// now hold a write lock in a different go routine and it should block since 10 read locks are
|
|
// still held.
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
// blocks till all read locks are released.
|
|
nsMutex.Lock("my-bucket", "my-object", strconv.Itoa(10))
|
|
// Once the above attempt to lock is unblocked/acquired, we verify the stats and release the lock.
|
|
expectedWLockStats := expectedResult[3]
|
|
// Since the write lock acquired here, the number of blocked locks should reduce by 1 and
|
|
// count of running locks should increase by 1.
|
|
|
|
// Call the RPC control handle to fetch the lock instrumentation info.
|
|
reply = &SystemLockState{}
|
|
// Call the lock instrumentation RPC end point.
|
|
err = client.Call("Controller.LockInfo", args, reply)
|
|
if err != nil {
|
|
c.Errorf("Add: expected no error but got string %q", err.Error())
|
|
}
|
|
verifyRPCLockInfoResponse(expectedWLockStats, *reply, c, 4)
|
|
|
|
// release the write lock.
|
|
nsMutex.Unlock("my-bucket", "my-object", strconv.Itoa(10))
|
|
|
|
}()
|
|
// waiting for a second so that the attempt to acquire the write lock in
|
|
// the above go routines gets blocked.
|
|
time.Sleep(1 * time.Second)
|
|
// The write lock should have got blocked by now,
|
|
// check whether the entry for one blocked lock exists.
|
|
expectedLockStats = expectedResult[2]
|
|
|
|
// Call the RPC control handle to fetch the lock instrumentation info.
|
|
reply = &SystemLockState{}
|
|
// Call the lock instrumentation RPC end point.
|
|
err = client.Call("Controller.LockInfo", args, reply)
|
|
if err != nil {
|
|
c.Errorf("Add: expected no error but got string %q", err.Error())
|
|
}
|
|
verifyRPCLockInfoResponse(expectedLockStats, *reply, c, 3)
|
|
// Release all the read locks held.
|
|
// the blocked write lock in the above go routines should get unblocked.
|
|
for i := 0; i < 10; i++ {
|
|
nsMutex.RUnlock("my-bucket", "my-object", strconv.Itoa(i))
|
|
}
|
|
wg.Wait()
|
|
// Since all the locks are released. There shouldnt be any entry in the lock info.
|
|
// and all the counters should be set to 0.
|
|
reply = &SystemLockState{}
|
|
// Call the lock instrumentation RPC end point.
|
|
err = client.Call("Controller.LockInfo", args, reply)
|
|
if err != nil {
|
|
c.Errorf("Add: expected no error but got string %q", err.Error())
|
|
}
|
|
|
|
if reply.TotalAcquiredLocks != 0 && reply.TotalLocks != 0 && reply.TotalBlockedLocks != 0 {
|
|
c.Fatalf("The counters are not reset properly after all locks are released")
|
|
}
|
|
if len(reply.LocksInfoPerObject) != 0 {
|
|
c.Fatalf("Since all locks are released there shouldn't have been any lock info entry, but found %d", len(reply.LocksInfoPerObject))
|
|
}
|
|
}
|
|
|
|
// TestControllerHandlerHealDiskMetadata - Registers and call the `HealDiskMetadataHandler`,
|
|
// asserts to validate the success.
|
|
func (s *TestRPCControllerSuite) TestControllerHandlerHealDiskMetadata(c *C) {
|
|
// The suite has already started the test RPC server, just send RPC calls.
|
|
authCfg := &authConfig{
|
|
accessKey: s.accessKey,
|
|
secretKey: s.secretKey,
|
|
address: s.endPoint,
|
|
path: path.Join(reservedBucket, controlPath),
|
|
loginMethod: "Controller.LoginHandler",
|
|
}
|
|
|
|
client := newAuthClient(authCfg)
|
|
defer client.Close()
|
|
|
|
args := &GenericArgs{}
|
|
reply := &GenericReply{}
|
|
err := client.Call("Controller.HealDiskMetadataHandler", args, reply)
|
|
|
|
if err != nil {
|
|
c.Errorf("Heal Meta Disk Handler test failed with <ERROR> %s", err.Error())
|
|
}
|
|
}
|
|
|