|
|
|
/*
|
|
|
|
* MinIO Cloud Storage, (C) 2016-2020 MinIO, Inc.
|
|
|
|
*
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
* You may obtain a copy of the License at
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
* limitations under the License.
|
|
|
|
*/
|
|
|
|
|
|
|
|
package cmd
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"sort"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/minio/minio/cmd/logger"
|
|
|
|
"github.com/minio/minio/pkg/bpool"
|
|
|
|
"github.com/minio/minio/pkg/color"
|
|
|
|
"github.com/minio/minio/pkg/dsync"
|
|
|
|
"github.com/minio/minio/pkg/madmin"
|
|
|
|
"github.com/minio/minio/pkg/sync/errgroup"
|
|
|
|
)
|
|
|
|
|
|
|
|
// OfflineDisk represents an unavailable disk.
|
|
|
|
var OfflineDisk StorageAPI // zero value is nil
|
|
|
|
|
|
|
|
// partialOperation is a successful upload/delete of an object
|
|
|
|
// but not written in all disks (having quorum)
|
|
|
|
type partialOperation struct {
|
|
|
|
bucket string
|
|
|
|
object string
|
|
|
|
versionID string
|
|
|
|
failedSet int
|
|
|
|
}
|
|
|
|
|
|
|
|
// erasureObjects - Implements ER object layer.
|
|
|
|
type erasureObjects struct {
|
|
|
|
GatewayUnsupported
|
|
|
|
|
|
|
|
setDriveCount int
|
|
|
|
defaultParityCount int
|
|
|
|
|
|
|
|
setNumber int
|
|
|
|
|
|
|
|
// getDisks returns list of storageAPIs.
|
|
|
|
getDisks func() []StorageAPI
|
|
|
|
|
|
|
|
// getLockers returns list of remote and local lockers.
|
|
|
|
getLockers func() ([]dsync.NetLocker, string)
|
|
|
|
|
|
|
|
// getEndpoints returns list of endpoint strings belonging this set.
|
|
|
|
// some may be local and some remote.
|
|
|
|
getEndpoints func() []string
|
|
|
|
|
|
|
|
// Locker mutex map.
|
|
|
|
nsMutex *nsLockMap
|
|
|
|
|
|
|
|
// Byte pools used for temporary i/o buffers.
|
|
|
|
bp *bpool.BytePoolCap
|
|
|
|
|
|
|
|
mrfOpCh chan partialOperation
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewNSLock - initialize a new namespace RWLocker instance.
|
|
|
|
func (er erasureObjects) NewNSLock(bucket string, objects ...string) RWLocker {
|
|
|
|
return er.nsMutex.NewNSLock(er.getLockers, bucket, objects...)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Shutdown function for object storage interface.
|
|
|
|
func (er erasureObjects) Shutdown(ctx context.Context) error {
|
|
|
|
// Add any object layer shutdown activities here.
|
|
|
|
closeStorageDisks(er.getDisks())
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// byDiskTotal is a collection satisfying sort.Interface.
|
|
|
|
type byDiskTotal []madmin.Disk
|
|
|
|
|
|
|
|
func (d byDiskTotal) Len() int { return len(d) }
|
|
|
|
func (d byDiskTotal) Swap(i, j int) { d[i], d[j] = d[j], d[i] }
|
|
|
|
func (d byDiskTotal) Less(i, j int) bool {
|
|
|
|
return d[i].TotalSpace < d[j].TotalSpace
|
|
|
|
}
|
|
|
|
|
|
|
|
func diskErrToDriveState(err error) (state string) {
|
|
|
|
state = madmin.DriveStateUnknown
|
|
|
|
switch {
|
|
|
|
case errors.Is(err, errDiskNotFound):
|
|
|
|
state = madmin.DriveStateOffline
|
|
|
|
case errors.Is(err, errCorruptedFormat):
|
|
|
|
state = madmin.DriveStateCorrupt
|
|
|
|
case errors.Is(err, errUnformattedDisk):
|
|
|
|
state = madmin.DriveStateUnformatted
|
|
|
|
case errors.Is(err, errDiskAccessDenied):
|
|
|
|
state = madmin.DriveStatePermission
|
|
|
|
case errors.Is(err, errFaultyDisk):
|
|
|
|
state = madmin.DriveStateFaulty
|
|
|
|
case err == nil:
|
|
|
|
state = madmin.DriveStateOk
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func getOnlineOfflineDisksStats(disksInfo []madmin.Disk) (onlineDisks, offlineDisks madmin.BackendDisks) {
|
|
|
|
onlineDisks = make(madmin.BackendDisks)
|
|
|
|
offlineDisks = make(madmin.BackendDisks)
|
|
|
|
|
|
|
|
for _, disk := range disksInfo {
|
|
|
|
ep := disk.Endpoint
|
|
|
|
if _, ok := offlineDisks[ep]; !ok {
|
|
|
|
offlineDisks[ep] = 0
|
|
|
|
}
|
|
|
|
if _, ok := onlineDisks[ep]; !ok {
|
|
|
|
onlineDisks[ep] = 0
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for the routines.
|
|
|
|
for _, disk := range disksInfo {
|
|
|
|
ep := disk.Endpoint
|
|
|
|
state := disk.State
|
|
|
|
if state != madmin.DriveStateOk && state != madmin.DriveStateUnformatted {
|
|
|
|
offlineDisks[ep]++
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
onlineDisks[ep]++
|
|
|
|
}
|
|
|
|
|
|
|
|
rootDiskCount := 0
|
|
|
|
for _, di := range disksInfo {
|
|
|
|
if di.RootDisk {
|
|
|
|
rootDiskCount++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Count offline disks as well to ensure consistent
|
|
|
|
// reportability of offline drives on local setups.
|
|
|
|
if len(disksInfo) == (rootDiskCount + offlineDisks.Sum()) {
|
|
|
|
// Success.
|
|
|
|
return onlineDisks, offlineDisks
|
|
|
|
}
|
|
|
|
|
|
|
|
// Root disk should be considered offline
|
|
|
|
for i := range disksInfo {
|
|
|
|
ep := disksInfo[i].Endpoint
|
|
|
|
if disksInfo[i].RootDisk {
|
|
|
|
offlineDisks[ep]++
|
|
|
|
onlineDisks[ep]--
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return onlineDisks, offlineDisks
|
|
|
|
}
|
|
|
|
|
|
|
|
// getDisksInfo - fetch disks info across all other storage API.
|
|
|
|
func getDisksInfo(disks []StorageAPI, endpoints []string) (disksInfo []madmin.Disk, errs []error) {
|
|
|
|
disksInfo = make([]madmin.Disk, len(disks))
|
|
|
|
|
|
|
|
g := errgroup.WithNErrs(len(disks))
|
|
|
|
for index := range disks {
|
|
|
|
index := index
|
|
|
|
g.Go(func() error {
|
|
|
|
if disks[index] == OfflineDisk {
|
|
|
|
logger.LogIf(GlobalContext, fmt.Errorf("%s: %s", errDiskNotFound, endpoints[index]))
|
|
|
|
disksInfo[index] = madmin.Disk{
|
|
|
|
State: diskErrToDriveState(errDiskNotFound),
|
|
|
|
Endpoint: endpoints[index],
|
|
|
|
}
|
|
|
|
// Storage disk is empty, perhaps ignored disk or not available.
|
|
|
|
return errDiskNotFound
|
|
|
|
}
|
|
|
|
info, err := disks[index].DiskInfo(context.TODO())
|
|
|
|
di := madmin.Disk{
|
|
|
|
Endpoint: endpoints[index],
|
|
|
|
DrivePath: info.MountPath,
|
|
|
|
TotalSpace: info.Total,
|
|
|
|
UsedSpace: info.Used,
|
|
|
|
AvailableSpace: info.Free,
|
|
|
|
UUID: info.ID,
|
|
|
|
RootDisk: info.RootDisk,
|
|
|
|
Healing: info.Healing,
|
|
|
|
State: diskErrToDriveState(err),
|
|
|
|
}
|
|
|
|
if info.Total > 0 {
|
|
|
|
di.Utilization = float64(info.Used / info.Total * 100)
|
|
|
|
}
|
|
|
|
disksInfo[index] = di
|
|
|
|
return err
|
|
|
|
}, index)
|
|
|
|
}
|
|
|
|
|
|
|
|
return disksInfo, g.Wait()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get an aggregated storage info across all disks.
|
|
|
|
func getStorageInfo(disks []StorageAPI, endpoints []string) (StorageInfo, []error) {
|
|
|
|
disksInfo, errs := getDisksInfo(disks, endpoints)
|
|
|
|
|
|
|
|
// Sort so that the first element is the smallest.
|
|
|
|
sort.Sort(byDiskTotal(disksInfo))
|
|
|
|
|
|
|
|
storageInfo := StorageInfo{
|
|
|
|
Disks: disksInfo,
|
|
|
|
}
|
|
|
|
|
|
|
|
storageInfo.Backend.Type = BackendErasure
|
|
|
|
return storageInfo, errs
|
|
|
|
}
|
|
|
|
|
|
|
|
// StorageInfo - returns underlying storage statistics.
|
|
|
|
func (er erasureObjects) StorageInfo(ctx context.Context) (StorageInfo, []error) {
|
|
|
|
disks := er.getDisks()
|
|
|
|
endpoints := er.getEndpoints()
|
|
|
|
return getStorageInfo(disks, endpoints)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (er erasureObjects) getOnlineDisksWithHealing() (newDisks []StorageAPI, healing bool) {
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
disks := er.getDisks()
|
|
|
|
infos := make([]DiskInfo, len(disks))
|
|
|
|
for _, i := range hashOrder(UTCNow().String(), len(disks)) {
|
|
|
|
i := i
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
disk := disks[i-1]
|
|
|
|
|
|
|
|
if disk == nil {
|
|
|
|
infos[i-1].Error = "nil disk"
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
di, err := disk.DiskInfo(context.Background())
|
|
|
|
if err != nil {
|
|
|
|
// - Do not consume disks which are not reachable
|
|
|
|
// unformatted or simply not accessible for some reason.
|
|
|
|
//
|
|
|
|
//
|
|
|
|
// - Future: skip busy disks
|
|
|
|
infos[i-1].Error = err.Error()
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
infos[i-1] = di
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
for i, info := range infos {
|
|
|
|
// Check if one of the drives in the set is being healed.
|
|
|
|
// this information is used by scanner to skip healing
|
|
|
|
// this erasure set while it calculates the usage.
|
|
|
|
if info.Healing || info.Error != "" {
|
|
|
|
healing = true
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
newDisks = append(newDisks, disks[i])
|
|
|
|
}
|
|
|
|
|
|
|
|
return newDisks, healing
|
|
|
|
}
|
|
|
|
|
|
|
|
// CrawlAndGetDataUsage will start crawling buckets and send updated totals as they are traversed.
|
|
|
|
// Updates are sent on a regular basis and the caller *must* consume them.
|
|
|
|
func (er erasureObjects) crawlAndGetDataUsage(ctx context.Context, buckets []BucketInfo, bf *bloomFilter, updates chan<- dataUsageCache) error {
|
|
|
|
if len(buckets) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Collect disks we can use.
|
|
|
|
disks, healing := er.getOnlineDisksWithHealing()
|
|
|
|
if len(disks) == 0 {
|
|
|
|
logger.Info(color.Green("data-crawl:") + " all disks are offline or being healed, skipping crawl")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Collect disks for healing.
|
|
|
|
allDisks := er.getDisks()
|
|
|
|
allDiskIDs := make([]string, 0, len(allDisks))
|
|
|
|
for _, disk := range allDisks {
|
|
|
|
if disk == OfflineDisk {
|
|
|
|
// its possible that disk is OfflineDisk
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
id, _ := disk.GetDiskID()
|
|
|
|
if id == "" {
|
|
|
|
// its possible that disk is unformatted
|
|
|
|
// or just went offline
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
allDiskIDs = append(allDiskIDs, id)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load bucket totals
|
|
|
|
oldCache := dataUsageCache{}
|
|
|
|
if err := oldCache.load(ctx, er, dataUsageCacheName); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// New cache..
|
|
|
|
cache := dataUsageCache{
|
|
|
|
Info: dataUsageCacheInfo{
|
|
|
|
Name: dataUsageRoot,
|
|
|
|
NextCycle: oldCache.Info.NextCycle,
|
|
|
|
},
|
|
|
|
Cache: make(map[string]dataUsageEntry, len(oldCache.Cache)),
|
|
|
|
}
|
|
|
|
bloom := bf.bytes()
|
|
|
|
|
|
|
|
// Put all buckets into channel.
|
|
|
|
bucketCh := make(chan BucketInfo, len(buckets))
|
|
|
|
// Add new buckets first
|
|
|
|
for _, b := range buckets {
|
|
|
|
if oldCache.find(b.Name) == nil {
|
|
|
|
bucketCh <- b
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add existing buckets.
|
|
|
|
for _, b := range buckets {
|
|
|
|
e := oldCache.find(b.Name)
|
|
|
|
if e != nil {
|
|
|
|
cache.replace(b.Name, dataUsageRoot, *e)
|
|
|
|
bucketCh <- b
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
close(bucketCh)
|
|
|
|
bucketResults := make(chan dataUsageEntryInfo, len(disks))
|
|
|
|
|
|
|
|
// Start async collector/saver.
|
|
|
|
// This goroutine owns the cache.
|
|
|
|
var saverWg sync.WaitGroup
|
|
|
|
saverWg.Add(1)
|
|
|
|
go func() {
|
|
|
|
const updateTime = 30 * time.Second
|
|
|
|
t := time.NewTicker(updateTime)
|
|
|
|
defer t.Stop()
|
|
|
|
defer saverWg.Done()
|
|
|
|
var lastSave time.Time
|
|
|
|
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
// Return without saving.
|
|
|
|
return
|
|
|
|
case <-t.C:
|
|
|
|
if cache.Info.LastUpdate.Equal(lastSave) {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, dataUsageCacheName))
|
|
|
|
updates <- cache.clone()
|
|
|
|
lastSave = cache.Info.LastUpdate
|
|
|
|
case v, ok := <-bucketResults:
|
|
|
|
if !ok {
|
|
|
|
// Save final state...
|
|
|
|
cache.Info.NextCycle++
|
|
|
|
cache.Info.LastUpdate = time.Now()
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, dataUsageCacheName))
|
|
|
|
updates <- cache
|
|
|
|
return
|
|
|
|
}
|
|
|
|
cache.replace(v.Name, v.Parent, v.Entry)
|
|
|
|
cache.Info.LastUpdate = time.Now()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Start one scanner per disk
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(len(disks))
|
|
|
|
for i := range disks {
|
|
|
|
go func(i int) {
|
|
|
|
defer wg.Done()
|
|
|
|
disk := disks[i]
|
|
|
|
|
|
|
|
for bucket := range bucketCh {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load cache for bucket
|
|
|
|
cacheName := pathJoin(bucket.Name, dataUsageCacheName)
|
|
|
|
cache := dataUsageCache{}
|
|
|
|
logger.LogIf(ctx, cache.load(ctx, er, cacheName))
|
|
|
|
if cache.Info.Name == "" {
|
|
|
|
cache.Info.Name = bucket.Name
|
|
|
|
}
|
|
|
|
cache.Info.BloomFilter = bloom
|
|
|
|
cache.Info.SkipHealing = healing
|
|
|
|
cache.Disks = allDiskIDs
|
|
|
|
if cache.Info.Name != bucket.Name {
|
|
|
|
logger.LogIf(ctx, fmt.Errorf("cache name mismatch: %s != %s", cache.Info.Name, bucket.Name))
|
|
|
|
cache.Info = dataUsageCacheInfo{
|
|
|
|
Name: bucket.Name,
|
|
|
|
LastUpdate: time.Time{},
|
|
|
|
NextCycle: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Calc usage
|
|
|
|
before := cache.Info.LastUpdate
|
|
|
|
var err error
|
|
|
|
cache, err = disk.CrawlAndGetDataUsage(ctx, cache)
|
|
|
|
cache.Info.BloomFilter = nil
|
|
|
|
if err != nil {
|
|
|
|
logger.LogIf(ctx, err)
|
|
|
|
if cache.Info.LastUpdate.After(before) {
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, cacheName))
|
|
|
|
}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
var root dataUsageEntry
|
|
|
|
if r := cache.root(); r != nil {
|
|
|
|
root = cache.flatten(*r)
|
|
|
|
}
|
|
|
|
bucketResults <- dataUsageEntryInfo{
|
|
|
|
Name: cache.Info.Name,
|
|
|
|
Parent: dataUsageRoot,
|
|
|
|
Entry: root,
|
|
|
|
}
|
|
|
|
// Save cache
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, cacheName))
|
|
|
|
}
|
|
|
|
}(i)
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
close(bucketResults)
|
|
|
|
saverWg.Wait()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|