config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
/*
|
|
|
|
* MinIO Cloud Storage, (C) 2016-2019 MinIO, Inc.
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
*
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
* You may obtain a copy of the License at
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
* limitations under the License.
|
|
|
|
*/
|
|
|
|
|
|
|
|
package cmd
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"encoding/json"
|
|
|
|
"fmt"
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
"os"
|
|
|
|
"path"
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
"path/filepath"
|
|
|
|
"strings"
|
|
|
|
|
|
|
|
"github.com/minio/minio/cmd/config"
|
|
|
|
"github.com/minio/minio/cmd/config/cache"
|
|
|
|
"github.com/minio/minio/cmd/config/compress"
|
|
|
|
xldap "github.com/minio/minio/cmd/config/identity/ldap"
|
|
|
|
"github.com/minio/minio/cmd/config/identity/openid"
|
|
|
|
"github.com/minio/minio/cmd/config/notify"
|
|
|
|
"github.com/minio/minio/cmd/config/policy/opa"
|
|
|
|
"github.com/minio/minio/cmd/config/storageclass"
|
|
|
|
"github.com/minio/minio/cmd/crypto"
|
|
|
|
"github.com/minio/minio/cmd/logger"
|
|
|
|
"github.com/minio/minio/pkg/auth"
|
|
|
|
"github.com/minio/minio/pkg/event"
|
|
|
|
"github.com/minio/minio/pkg/event/target"
|
|
|
|
xnet "github.com/minio/minio/pkg/net"
|
|
|
|
"github.com/minio/minio/pkg/quick"
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
)
|
|
|
|
|
|
|
|
// DO NOT EDIT following message template, please open a github issue to discuss instead.
|
|
|
|
var configMigrateMSGTemplate = "Configuration file %s migrated from version '%s' to '%s' successfully."
|
|
|
|
|
|
|
|
// Save config file to corresponding backend
|
|
|
|
func Save(configFile string, data interface{}) error {
|
|
|
|
return quick.SaveConfig(data, configFile, globalEtcdClient)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load config from backend
|
|
|
|
func Load(configFile string, data interface{}) (quick.Config, error) {
|
|
|
|
return quick.LoadConfig(configFile, globalEtcdClient, data)
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetVersion gets config version from backend
|
|
|
|
func GetVersion(configFile string) (string, error) {
|
|
|
|
return quick.GetVersion(configFile, globalEtcdClient)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Migrates all config versions from "1" to "28".
|
|
|
|
func migrateConfig() error {
|
|
|
|
// Purge all configs with version '1',
|
|
|
|
// this is a special case since version '1' used
|
|
|
|
// to be a filename 'fsUsers.json' not 'config.json'.
|
|
|
|
if err := purgeV1(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load only config version information.
|
|
|
|
version, err := GetVersion(getConfigFile())
|
|
|
|
if err != nil {
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Conditional to migrate only relevant config versions.
|
|
|
|
// Upon success migration continues to the next version in sequence.
|
|
|
|
switch version {
|
|
|
|
case "2":
|
|
|
|
// Migrate version '2' to '3'.
|
|
|
|
if err = migrateV2ToV3(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "3":
|
|
|
|
// Migrate version '3' to '4'.
|
|
|
|
if err = migrateV3ToV4(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "4":
|
|
|
|
// Migrate version '4' to '5'.
|
|
|
|
if err = migrateV4ToV5(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "5":
|
|
|
|
// Migrate version '5' to '6.
|
|
|
|
if err = migrateV5ToV6(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "6":
|
|
|
|
// Migrate version '6' to '7'.
|
|
|
|
if err = migrateV6ToV7(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "7":
|
|
|
|
// Migrate version '7' to '8'.
|
|
|
|
if err = migrateV7ToV8(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "8":
|
|
|
|
// Migrate version '8' to '9'.
|
|
|
|
if err = migrateV8ToV9(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "9":
|
|
|
|
// Migrate version '9' to '10'.
|
|
|
|
if err = migrateV9ToV10(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "10":
|
|
|
|
// Migrate version '10' to '11'.
|
|
|
|
if err = migrateV10ToV11(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "11":
|
|
|
|
// Migrate version '11' to '12'.
|
|
|
|
if err = migrateV11ToV12(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "12":
|
|
|
|
// Migrate version '12' to '13'.
|
|
|
|
if err = migrateV12ToV13(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "13":
|
|
|
|
// Migrate version '13' to '14'.
|
|
|
|
if err = migrateV13ToV14(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "14":
|
|
|
|
// Migrate version '14' to '15'.
|
|
|
|
if err = migrateV14ToV15(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "15":
|
|
|
|
// Migrate version '15' to '16'.
|
|
|
|
if err = migrateV15ToV16(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "16":
|
|
|
|
// Migrate version '16' to '17'.
|
|
|
|
if err = migrateV16ToV17(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "17":
|
|
|
|
// Migrate version '17' to '18'.
|
|
|
|
if err = migrateV17ToV18(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "18":
|
|
|
|
// Migrate version '18' to '19'.
|
|
|
|
if err = migrateV18ToV19(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "19":
|
|
|
|
if err = migrateV19ToV20(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "20":
|
|
|
|
if err = migrateV20ToV21(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "21":
|
|
|
|
if err = migrateV21ToV22(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "22":
|
|
|
|
if err = migrateV22ToV23(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "23":
|
|
|
|
if err = migrateV23ToV24(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "24":
|
|
|
|
if err = migrateV24ToV25(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "25":
|
|
|
|
if err = migrateV25ToV26(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "26":
|
|
|
|
if err = migrateV26ToV27(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "27":
|
|
|
|
if err = migrateV27ToV28(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
fallthrough
|
|
|
|
case "33":
|
|
|
|
// No migration needed. this always points to current version.
|
|
|
|
err = nil
|
|
|
|
}
|
|
|
|
return err
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
}
|
|
|
|
|
|
|
|
// Version '1' is not supported anymore and deprecated, safe to delete.
|
|
|
|
func purgeV1() error {
|
|
|
|
configFile := filepath.Join(globalConfigDir.Get(), "fsUsers.json")
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
|
|
|
|
cv1 := &configV1{}
|
|
|
|
_, err := Load(configFile, cv1)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘1’. %v", err)
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
}
|
|
|
|
if cv1.Version != "1" {
|
|
|
|
return fmt.Errorf("unrecognized config version ‘%s’", cv1.Version)
|
|
|
|
}
|
|
|
|
|
|
|
|
os.RemoveAll(configFile)
|
|
|
|
logger.Info("Removed unsupported config version ‘1’.")
|
|
|
|
return nil
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
}
|
|
|
|
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
// Version '2' to '3' config migration adds new fields and re-orders
|
|
|
|
// previous fields. Simplifies config for future additions.
|
|
|
|
func migrateV2ToV3() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv2 := &configV2{}
|
|
|
|
_, err := Load(configFile, cv2)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘2’. %v", err)
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
}
|
|
|
|
if cv2.Version != "2" {
|
|
|
|
return nil
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
}
|
|
|
|
|
|
|
|
cred, err := auth.CreateCredentials(cv2.Credentials.AccessKey, cv2.Credentials.SecretKey)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("Invalid credential in V2 configuration file. %v", err)
|
|
|
|
}
|
|
|
|
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
srvConfig := &configV3{}
|
|
|
|
srvConfig.Version = "3"
|
|
|
|
srvConfig.Addr = ":9000"
|
|
|
|
srvConfig.Credential = cred
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
srvConfig.Region = cv2.Credentials.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature V4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = consoleLoggerV1{
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
Enable: true,
|
|
|
|
Level: "fatal",
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
}
|
|
|
|
flogger := fileLoggerV1{}
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
flogger.Level = "error"
|
|
|
|
if cv2.FileLogger.Filename != "" {
|
|
|
|
flogger.Enable = true
|
|
|
|
flogger.Filename = cv2.FileLogger.Filename
|
|
|
|
}
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
srvConfig.Logger.File = flogger
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
|
|
|
|
slogger := syslogLoggerV3{}
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
slogger.Level = "debug"
|
|
|
|
if cv2.SyslogLogger.Addr != "" {
|
|
|
|
slogger.Enable = true
|
|
|
|
slogger.Addr = cv2.SyslogLogger.Addr
|
|
|
|
}
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
srvConfig.Logger.Syslog = slogger
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv2.Version, srvConfig.Version, err)
|
|
|
|
}
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv2.Version, srvConfig.Version)
|
|
|
|
return nil
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
}
|
|
|
|
|
|
|
|
// Version '3' to '4' migrates config, removes previous fields related
|
|
|
|
// to backend types and server address. This change further simplifies
|
|
|
|
// the config for future additions.
|
|
|
|
func migrateV3ToV4() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv3 := &configV3{}
|
|
|
|
_, err := Load(configFile, cv3)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘3’. %v", err)
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
}
|
|
|
|
if cv3.Version != "3" {
|
|
|
|
return nil
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
}
|
|
|
|
|
|
|
|
// Save only the new fields, ignore the rest.
|
|
|
|
srvConfig := &configV4{}
|
|
|
|
srvConfig.Version = "4"
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
srvConfig.Credential = cv3.Credential
|
|
|
|
srvConfig.Region = cv3.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
config: Migrate to the new version. Remove backend details.
Migrate to new config format v4.
```
{
"version": "4",
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"console": {
"enable": true,
"level": "fatal"
},
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
}
}
}
```
This patch also updates [minio cli spec](./minio.md)
9 years ago
|
|
|
srvConfig.Logger.Console = cv3.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv3.Logger.File
|
|
|
|
srvConfig.Logger.Syslog = cv3.Logger.Syslog
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv3.Version, srvConfig.Version, err)
|
|
|
|
}
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv3.Version, srvConfig.Version)
|
|
|
|
return nil
|
config/main: Re-write config files - add to new config v3
- New config format.
```
{
"version": "3",
"address": ":9000",
"backend": {
"type": "fs",
"disk": "/path"
},
"credential": {
"accessKey": "WLGDGYAQYIGI833EV05A",
"secretKey": "BYvgJM101sHngl2uzjXS/OBF/aMxAN06JrJ3qJlF"
},
"region": "us-east-1",
"logger": {
"file": {
"enable": false,
"fileName": "",
"level": "error"
},
"syslog": {
"enable": false,
"address": "",
"level": "debug"
},
"console": {
"enable": true,
"level": "fatal"
}
}
}
```
New command lines in lieu of supporting XL.
Minio initialize filesystem backend.
~~~
$ minio init fs <path>
~~~
Minio initialize XL backend.
~~~
$ minio init xl <url1>...<url16>
~~~
For 'fs' backend it starts the server.
~~~
$ minio server
~~~
For 'xl' backend it waits for servers to join.
~~~
$ minio server
... [PROGRESS BAR] of servers connecting
~~~
Now on other servers execute 'join' and they connect.
~~~
....
minio join <url1> -- from <url2> && minio server
minio join <url1> -- from <url3> && minio server
...
...
minio join <url1> -- from <url16> && minio server
~~~
9 years ago
|
|
|
}
|
|
|
|
|
|
|
|
// Version '4' to '5' migrates config, removes previous fields related
|
|
|
|
// to backend types and server address. This change further simplifies
|
|
|
|
// the config for future additions.
|
|
|
|
func migrateV4ToV5() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv4 := &configV4{}
|
|
|
|
_, err := Load(configFile, cv4)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘4’. %v", err)
|
|
|
|
}
|
|
|
|
if cv4.Version != "4" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Save only the new fields, ignore the rest.
|
|
|
|
srvConfig := &configV5{}
|
|
|
|
srvConfig.Version = "5"
|
|
|
|
srvConfig.Credential = cv4.Credential
|
|
|
|
srvConfig.Region = cv4.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv4.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv4.Logger.File
|
|
|
|
srvConfig.Logger.Syslog = cv4.Logger.Syslog
|
|
|
|
srvConfig.Logger.AMQP.Enable = false
|
|
|
|
srvConfig.Logger.ElasticSearch.Enable = false
|
|
|
|
srvConfig.Logger.Redis.Enable = false
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv4.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv4.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '5' to '6' migrates config, removes previous fields related
|
|
|
|
// to backend types and server address. This change further simplifies
|
|
|
|
// the config for future additions.
|
|
|
|
func migrateV5ToV6() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv5 := &configV5{}
|
|
|
|
_, err := Load(configFile, cv5)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘5’. %v", err)
|
|
|
|
}
|
|
|
|
if cv5.Version != "5" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Save only the new fields, ignore the rest.
|
|
|
|
srvConfig := &configV6{}
|
|
|
|
srvConfig.Version = "6"
|
|
|
|
srvConfig.Credential = cv5.Credential
|
|
|
|
srvConfig.Region = cv5.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv5.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv5.Logger.File
|
|
|
|
srvConfig.Logger.Syslog = cv5.Logger.Syslog
|
|
|
|
|
|
|
|
if cv5.Logger.AMQP.URL != "" {
|
|
|
|
var url *xnet.URL
|
|
|
|
if url, err = xnet.ParseURL(cv5.Logger.AMQP.URL); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
srvConfig.Notify.AMQP = map[string]target.AMQPArgs{
|
|
|
|
"1": {
|
|
|
|
Enable: cv5.Logger.AMQP.Enable,
|
|
|
|
URL: *url,
|
|
|
|
Exchange: cv5.Logger.AMQP.Exchange,
|
|
|
|
RoutingKey: cv5.Logger.AMQP.RoutingKey,
|
|
|
|
Mandatory: cv5.Logger.AMQP.Mandatory,
|
|
|
|
Immediate: cv5.Logger.AMQP.Immediate,
|
|
|
|
Durable: cv5.Logger.AMQP.Durable,
|
|
|
|
Internal: cv5.Logger.AMQP.Internal,
|
|
|
|
NoWait: cv5.Logger.AMQP.NoWait,
|
|
|
|
AutoDeleted: cv5.Logger.AMQP.AutoDeleted,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if cv5.Logger.ElasticSearch.URL != "" {
|
|
|
|
var url *xnet.URL
|
|
|
|
url, err = xnet.ParseHTTPURL(cv5.Logger.ElasticSearch.URL)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
srvConfig.Notify.ElasticSearch = map[string]target.ElasticsearchArgs{
|
|
|
|
"1": {
|
|
|
|
Enable: cv5.Logger.ElasticSearch.Enable,
|
|
|
|
URL: *url,
|
|
|
|
Index: cv5.Logger.ElasticSearch.Index,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if cv5.Logger.Redis.Addr != "" {
|
|
|
|
var addr *xnet.Host
|
|
|
|
if addr, err = xnet.ParseHost(cv5.Logger.Redis.Addr); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
srvConfig.Notify.Redis = map[string]target.RedisArgs{
|
|
|
|
"1": {
|
|
|
|
Enable: cv5.Logger.Redis.Enable,
|
|
|
|
Addr: *addr,
|
|
|
|
Password: cv5.Logger.Redis.Password,
|
|
|
|
Key: cv5.Logger.Redis.Key,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv5.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv5.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '6' to '7' migrates config, removes previous fields related
|
|
|
|
// to backend types and server address. This change further simplifies
|
|
|
|
// the config for future additions.
|
|
|
|
func migrateV6ToV7() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv6 := &configV6{}
|
|
|
|
_, err := Load(configFile, cv6)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘6’. %v", err)
|
|
|
|
}
|
|
|
|
if cv6.Version != "6" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Save only the new fields, ignore the rest.
|
|
|
|
srvConfig := &serverConfigV7{}
|
|
|
|
srvConfig.Version = "7"
|
|
|
|
srvConfig.Credential = cv6.Credential
|
|
|
|
srvConfig.Region = cv6.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv6.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv6.Logger.File
|
|
|
|
srvConfig.Logger.Syslog = cv6.Logger.Syslog
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.ElasticSearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
if len(cv6.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv6.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv6.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.ElasticSearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.ElasticSearch = cv6.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv6.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv6.Notify.Redis
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv6.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv6.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '7' to '8' migrates config, removes previous fields related
|
|
|
|
// to backend types and server address. This change further simplifies
|
|
|
|
// the config for future additions.
|
|
|
|
func migrateV7ToV8() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv7 := &serverConfigV7{}
|
|
|
|
_, err := Load(configFile, cv7)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘7’. %v", err)
|
|
|
|
}
|
|
|
|
if cv7.Version != "7" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Save only the new fields, ignore the rest.
|
|
|
|
srvConfig := &serverConfigV8{}
|
|
|
|
srvConfig.Version = "8"
|
|
|
|
srvConfig.Credential = cv7.Credential
|
|
|
|
srvConfig.Region = cv7.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv7.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv7.Logger.File
|
|
|
|
srvConfig.Logger.Syslog = cv7.Logger.Syslog
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.NATS = make(map[string]natsNotifyV1)
|
|
|
|
srvConfig.Notify.ElasticSearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
if len(cv7.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv7.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv7.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS["1"] = natsNotifyV1{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv7.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv7.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.ElasticSearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.ElasticSearch = cv7.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv7.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv7.Notify.Redis
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv7.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv7.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '8' to '9' migration. Adds postgresql notifier
|
|
|
|
// configuration, but it's otherwise the same as V8.
|
|
|
|
func migrateV8ToV9() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv8 := &serverConfigV8{}
|
|
|
|
_, err := Load(configFile, cv8)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘8’. %v", err)
|
|
|
|
}
|
|
|
|
if cv8.Version != "8" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V8 into V9 config struct
|
|
|
|
srvConfig := &serverConfigV9{}
|
|
|
|
srvConfig.Version = "9"
|
|
|
|
srvConfig.Credential = cv8.Credential
|
|
|
|
srvConfig.Region = cv8.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv8.Logger.Console
|
|
|
|
srvConfig.Logger.Console.Level = "error"
|
|
|
|
srvConfig.Logger.File = cv8.Logger.File
|
|
|
|
srvConfig.Logger.Syslog = cv8.Logger.Syslog
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv8.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv8.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv8.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]natsNotifyV1)
|
|
|
|
srvConfig.Notify.NATS["1"] = natsNotifyV1{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv8.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv8.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.ElasticSearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.ElasticSearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.ElasticSearch = cv8.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv8.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv8.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv8.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv8.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv8.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv8.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '9' to '10' migration. Remove syslog config
|
|
|
|
// but it's otherwise the same as V9.
|
|
|
|
func migrateV9ToV10() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv9 := &serverConfigV9{}
|
|
|
|
_, err := Load(configFile, cv9)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘9’. %v", err)
|
|
|
|
}
|
|
|
|
if cv9.Version != "9" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V9 into V10 config struct
|
|
|
|
srvConfig := &serverConfigV10{}
|
|
|
|
srvConfig.Version = "10"
|
|
|
|
srvConfig.Credential = cv9.Credential
|
|
|
|
srvConfig.Region = cv9.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv9.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv9.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv9.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv9.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv9.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]natsNotifyV1)
|
|
|
|
srvConfig.Notify.NATS["1"] = natsNotifyV1{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv9.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv9.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.ElasticSearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.ElasticSearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.ElasticSearch = cv9.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv9.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv9.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv9.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv9.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv9.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv9.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '10' to '11' migration. Add support for Kafka
|
|
|
|
// notifications.
|
|
|
|
func migrateV10ToV11() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv10 := &serverConfigV10{}
|
|
|
|
_, err := Load(configFile, cv10)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘10’. %v", err)
|
|
|
|
}
|
|
|
|
if cv10.Version != "10" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V10 into V11 config struct
|
|
|
|
srvConfig := &serverConfigV11{}
|
|
|
|
srvConfig.Version = "11"
|
|
|
|
srvConfig.Credential = cv10.Credential
|
|
|
|
srvConfig.Region = cv10.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv10.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv10.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv10.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv10.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv10.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]natsNotifyV1)
|
|
|
|
srvConfig.Notify.NATS["1"] = natsNotifyV1{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv10.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv10.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.ElasticSearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.ElasticSearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.ElasticSearch = cv10.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv10.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv10.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv10.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv10.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
// V10 will not have a Kafka config. So we initialize one here.
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv10.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv10.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '11' to '12' migration. Add support for NATS streaming
|
|
|
|
// notifications.
|
|
|
|
func migrateV11ToV12() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv11 := &serverConfigV11{}
|
|
|
|
_, err := Load(configFile, cv11)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘11’. %v", err)
|
|
|
|
}
|
|
|
|
if cv11.Version != "11" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V11 into V12 config struct
|
|
|
|
srvConfig := &serverConfigV12{}
|
|
|
|
srvConfig.Version = "12"
|
|
|
|
srvConfig.Credential = cv11.Credential
|
|
|
|
srvConfig.Region = cv11.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv11.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv11.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv11.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv11.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv11.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.ElasticSearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.ElasticSearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.ElasticSearch = cv11.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv11.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv11.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv11.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv11.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv11.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv11.Notify.Kafka
|
|
|
|
}
|
|
|
|
|
|
|
|
// V12 will have an updated config of nats. So we create a new one or we
|
|
|
|
// update the old one if found.
|
|
|
|
if len(cv11.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
for k, v := range cv11.Notify.NATS {
|
|
|
|
if v.Address == "" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
var addr *xnet.Host
|
|
|
|
addr, err = xnet.ParseHost(v.Address)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
n := target.NATSArgs{}
|
|
|
|
n.Enable = v.Enable
|
|
|
|
n.Address = *addr
|
|
|
|
n.Subject = v.Subject
|
|
|
|
n.Username = v.Username
|
|
|
|
n.Password = v.Password
|
|
|
|
n.Token = v.Token
|
|
|
|
n.Secure = v.Secure
|
|
|
|
n.PingInterval = v.PingInterval
|
|
|
|
srvConfig.Notify.NATS[k] = n
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv11.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv11.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '12' to '13' migration. Add support for custom webhook endpoint.
|
|
|
|
func migrateV12ToV13() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv12 := &serverConfigV12{}
|
|
|
|
_, err := Load(configFile, cv12)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘12’. %v", err)
|
|
|
|
}
|
|
|
|
if cv12.Version != "12" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V12 into V13 config struct
|
|
|
|
srvConfig := &serverConfigV13{
|
|
|
|
Logger: &loggerV7{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "13"
|
|
|
|
srvConfig.Credential = cv12.Credential
|
|
|
|
srvConfig.Region = cv12.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv12.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv12.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv12.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv12.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv12.Notify.ElasticSearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv12.Notify.ElasticSearch
|
|
|
|
}
|
|
|
|
if len(cv12.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv12.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv12.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv12.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv12.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv12.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv12.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv12.Notify.NATS
|
|
|
|
}
|
|
|
|
|
|
|
|
// V12 will not have a webhook config. So we initialize one here.
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv12.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv12.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '13' to '14' migration. Add support for browser param.
|
|
|
|
func migrateV13ToV14() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv13 := &serverConfigV13{}
|
|
|
|
_, err := Load(configFile, cv13)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘13’. %v", err)
|
|
|
|
}
|
|
|
|
if cv13.Version != "13" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V13 into V14 config struct
|
|
|
|
srvConfig := &serverConfigV14{
|
|
|
|
Logger: &loggerV7{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "14"
|
|
|
|
srvConfig.Credential = cv13.Credential
|
|
|
|
srvConfig.Region = cv13.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv13.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv13.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv13.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv13.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv13.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv13.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv13.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv13.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv13.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv13.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv13.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv13.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv13.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv13.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv13.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv13.Notify.Webhook
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the new browser parameter to true by default
|
|
|
|
srvConfig.Browser = true
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv13.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv13.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '14' to '15' migration. Add support for MySQL notifications.
|
|
|
|
func migrateV14ToV15() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv14 := &serverConfigV14{}
|
|
|
|
_, err := Load(configFile, cv14)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘14’. %v", err)
|
|
|
|
}
|
|
|
|
if cv14.Version != "14" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V14 into V15 config struct
|
|
|
|
srvConfig := &serverConfigV15{
|
|
|
|
Logger: &loggerV7{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "15"
|
|
|
|
srvConfig.Credential = cv14.Credential
|
|
|
|
srvConfig.Region = cv14.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
srvConfig.Logger.Console = cv14.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv14.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv14.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv14.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv14.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv14.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv14.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv14.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv14.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv14.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv14.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv14.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv14.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv14.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv14.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv14.Notify.Webhook
|
|
|
|
}
|
|
|
|
|
|
|
|
// V14 will not have mysql support, so we add that here.
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv14.Browser
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv14.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv14.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '15' to '16' migration. Remove log level in loggers
|
|
|
|
// and rename 'fileName' filed in File logger to 'filename'
|
|
|
|
func migrateV15ToV16() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv15 := &serverConfigV15{}
|
|
|
|
_, err := Load(configFile, cv15)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘15’. %v", err)
|
|
|
|
}
|
|
|
|
if cv15.Version != "15" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V15 into V16 config struct
|
|
|
|
srvConfig := &serverConfigV16{
|
|
|
|
Logger: &loggers{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "16"
|
|
|
|
srvConfig.Credential = cv15.Credential
|
|
|
|
srvConfig.Region = cv15.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv15.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv15.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv15.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv15.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv15.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv15.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv15.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv15.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv15.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv15.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv15.Browser
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv15.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv15.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '16' to '17' migration. Adds "format" configuration
|
|
|
|
// parameter for database targets.
|
|
|
|
func migrateV16ToV17() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv16 := &serverConfigV16{}
|
|
|
|
_, err := Load(configFile, cv16)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘16’. %v", err)
|
|
|
|
}
|
|
|
|
if cv16.Version != "16" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V16 into V17 config struct
|
|
|
|
srvConfig := &serverConfigV17{
|
|
|
|
Logger: &loggers{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "17"
|
|
|
|
srvConfig.Credential = cv16.Credential
|
|
|
|
srvConfig.Region = cv16.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
srvConfig.Logger.Console = cv16.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv16.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv16.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv16.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{}
|
|
|
|
} else {
|
|
|
|
// IMPORTANT NOTE: Future migrations should remove
|
|
|
|
// this as existing configuration will already contain
|
|
|
|
// a value for the "format" parameter.
|
|
|
|
srvConfig.Notify.Elasticsearch = cv16.Notify.Elasticsearch
|
|
|
|
for k, v := range srvConfig.Notify.Elasticsearch {
|
|
|
|
v.Format = event.NamespaceFormat
|
|
|
|
srvConfig.Notify.Elasticsearch[k] = v
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{}
|
|
|
|
} else {
|
|
|
|
// IMPORTANT NOTE: Future migrations should remove
|
|
|
|
// this as existing configuration will already contain
|
|
|
|
// a value for the "format" parameter.
|
|
|
|
srvConfig.Notify.Redis = cv16.Notify.Redis
|
|
|
|
for k, v := range srvConfig.Notify.Redis {
|
|
|
|
v.Format = event.NamespaceFormat
|
|
|
|
srvConfig.Notify.Redis[k] = v
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{}
|
|
|
|
} else {
|
|
|
|
// IMPORTANT NOTE: Future migrations should remove
|
|
|
|
// this as existing configuration will already contain
|
|
|
|
// a value for the "format" parameter.
|
|
|
|
srvConfig.Notify.PostgreSQL = cv16.Notify.PostgreSQL
|
|
|
|
for k, v := range srvConfig.Notify.PostgreSQL {
|
|
|
|
v.Format = event.NamespaceFormat
|
|
|
|
srvConfig.Notify.PostgreSQL[k] = v
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv16.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv16.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv16.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv16.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{}
|
|
|
|
} else {
|
|
|
|
// IMPORTANT NOTE: Future migrations should remove
|
|
|
|
// this as existing configuration will already contain
|
|
|
|
// a value for the "format" parameter.
|
|
|
|
srvConfig.Notify.MySQL = cv16.Notify.MySQL
|
|
|
|
for k, v := range srvConfig.Notify.MySQL {
|
|
|
|
v.Format = event.NamespaceFormat
|
|
|
|
srvConfig.Notify.MySQL[k] = v
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv16.Browser
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv16.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv16.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Version '17' to '18' migration. Adds "deliveryMode" configuration
|
|
|
|
// parameter for AMQP notification target
|
|
|
|
func migrateV17ToV18() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv17 := &serverConfigV17{}
|
|
|
|
_, err := Load(configFile, cv17)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘17’. %v", err)
|
|
|
|
}
|
|
|
|
if cv17.Version != "17" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V17 into V18 config struct
|
|
|
|
srvConfig := &serverConfigV17{
|
|
|
|
Logger: &loggers{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "18"
|
|
|
|
srvConfig.Credential = cv17.Credential
|
|
|
|
srvConfig.Region = cv17.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
srvConfig.Logger.Console = cv17.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv17.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv17.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
// New deliveryMode parameter is added for AMQP,
|
|
|
|
// default value is already 0, so nothing to
|
|
|
|
// explicitly migrate here.
|
|
|
|
srvConfig.Notify.AMQP = cv17.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv17.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv17.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv17.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv17.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv17.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv17.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv17.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv17.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv17.Browser
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv17.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv17.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV18ToV19() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv18 := &serverConfigV18{}
|
|
|
|
_, err := Load(configFile, cv18)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘18’. %v", err)
|
|
|
|
}
|
|
|
|
if cv18.Version != "18" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V18 into V19 config struct
|
|
|
|
srvConfig := &serverConfigV18{
|
|
|
|
Logger: &loggers{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "19"
|
|
|
|
srvConfig.Credential = cv18.Credential
|
|
|
|
srvConfig.Region = cv18.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
srvConfig.Logger.Console = cv18.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv18.Logger.File
|
|
|
|
|
|
|
|
// check and set notifiers config
|
|
|
|
if len(cv18.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
// New deliveryMode parameter is added for AMQP,
|
|
|
|
// default value is already 0, so nothing to
|
|
|
|
// explicitly migrate here.
|
|
|
|
srvConfig.Notify.AMQP = cv18.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv18.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv18.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv18.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv18.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv18.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv18.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv18.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv18.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
// V18 will not have mqtt support, so we add that here.
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv18.Browser
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv18.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv18.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV19ToV20() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv19 := &serverConfigV19{}
|
|
|
|
_, err := Load(configFile, cv19)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘18’. %v", err)
|
|
|
|
}
|
|
|
|
if cv19.Version != "19" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V19 into V20 config struct
|
|
|
|
srvConfig := &serverConfigV20{
|
|
|
|
Logger: &loggers{},
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "20"
|
|
|
|
srvConfig.Credential = cv19.Credential
|
|
|
|
srvConfig.Region = cv19.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
srvConfig.Logger.Console = cv19.Logger.Console
|
|
|
|
srvConfig.Logger.File = cv19.Logger.File
|
|
|
|
|
|
|
|
if len(cv19.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv19.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv19.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv19.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv19.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv19.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv19.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv19.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv19.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv19.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv19.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv19.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv19.Browser
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv19.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv19.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV20ToV21() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv20 := &serverConfigV20{}
|
|
|
|
_, err := Load(configFile, cv20)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘20’. %v", err)
|
|
|
|
}
|
|
|
|
if cv20.Version != "20" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V20 into V21 config struct
|
|
|
|
srvConfig := &serverConfigV21{
|
|
|
|
Notify: ¬ifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "21"
|
|
|
|
srvConfig.Credential = cv20.Credential
|
|
|
|
srvConfig.Region = cv20.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv20.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv20.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv20.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv20.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv20.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv20.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv20.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv20.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv20.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv20.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv20.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv20.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv20.Browser
|
|
|
|
|
|
|
|
// Load domain config from existing config in the file.
|
|
|
|
srvConfig.Domain = cv20.Domain
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv20.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv20.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV21ToV22() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv21 := &serverConfigV21{}
|
|
|
|
_, err := Load(configFile, cv21)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘21’. %v", err)
|
|
|
|
}
|
|
|
|
if cv21.Version != "21" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V21 into V22 config struct
|
|
|
|
srvConfig := &serverConfigV22{
|
|
|
|
Notify: notifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "22"
|
|
|
|
srvConfig.Credential = cv21.Credential
|
|
|
|
srvConfig.Region = cv21.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv21.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv21.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv21.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv21.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv21.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv21.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv21.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv21.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv21.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv21.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv21.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv21.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv21.Browser
|
|
|
|
|
|
|
|
// Load domain config from existing config in the file.
|
|
|
|
srvConfig.Domain = cv21.Domain
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv21.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv21.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV22ToV23() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv22 := &serverConfigV22{}
|
|
|
|
_, err := Load(configFile, cv22)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘22’. %v", err)
|
|
|
|
}
|
|
|
|
if cv22.Version != "22" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V22 into V23 config struct
|
|
|
|
srvConfig := &serverConfigV23{
|
|
|
|
Notify: notifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "23"
|
|
|
|
srvConfig.Credential = cv22.Credential
|
|
|
|
srvConfig.Region = cv22.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv22.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv22.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv22.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv22.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv22.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv22.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv22.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv22.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv22.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv22.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv22.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv22.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv22.Browser
|
|
|
|
|
|
|
|
// Load domain config from existing config in the file.
|
|
|
|
srvConfig.Domain = cv22.Domain
|
|
|
|
|
|
|
|
// Load storage class config from existing storage class config in the file
|
|
|
|
srvConfig.StorageClass.RRS = cv22.StorageClass.RRS
|
|
|
|
srvConfig.StorageClass.Standard = cv22.StorageClass.Standard
|
|
|
|
|
|
|
|
// Init cache config.For future migration, Cache config needs to be copied over from previous version.
|
|
|
|
srvConfig.Cache.Drives = []string{}
|
|
|
|
srvConfig.Cache.Exclude = []string{}
|
|
|
|
srvConfig.Cache.Expiry = 90
|
|
|
|
|
|
|
|
if err = Save(configFile, srvConfig); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv22.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv22.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV23ToV24() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv23 := &serverConfigV23{}
|
|
|
|
_, err := quick.LoadConfig(configFile, globalEtcdClient, cv23)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘23’. %v", err)
|
|
|
|
}
|
|
|
|
if cv23.Version != "23" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V23 into V24 config struct
|
|
|
|
srvConfig := &serverConfigV24{
|
|
|
|
Notify: notifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "24"
|
|
|
|
srvConfig.Credential = cv23.Credential
|
|
|
|
srvConfig.Region = cv23.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv23.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv23.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv23.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv23.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv23.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv23.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv23.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv23.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv23.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv23.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv23.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv23.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv23.Browser
|
|
|
|
|
|
|
|
// Load domain config from existing config in the file.
|
|
|
|
srvConfig.Domain = cv23.Domain
|
|
|
|
|
|
|
|
// Load storage class config from existing storage class config in the file.
|
|
|
|
srvConfig.StorageClass.RRS = cv23.StorageClass.RRS
|
|
|
|
srvConfig.StorageClass.Standard = cv23.StorageClass.Standard
|
|
|
|
|
|
|
|
// Load cache config from existing cache config in the file.
|
|
|
|
srvConfig.Cache.Drives = cv23.Cache.Drives
|
|
|
|
srvConfig.Cache.Exclude = cv23.Cache.Exclude
|
|
|
|
srvConfig.Cache.Expiry = cv23.Cache.Expiry
|
|
|
|
|
|
|
|
if err = quick.SaveConfig(srvConfig, configFile, globalEtcdClient); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv23.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv23.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV24ToV25() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv24 := &serverConfigV24{}
|
|
|
|
_, err := quick.LoadConfig(configFile, globalEtcdClient, cv24)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘24’. %v", err)
|
|
|
|
}
|
|
|
|
if cv24.Version != "24" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V24 into V25 config struct
|
|
|
|
srvConfig := &serverConfigV25{
|
|
|
|
Notify: notifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "25"
|
|
|
|
srvConfig.Credential = cv24.Credential
|
|
|
|
srvConfig.Region = cv24.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv24.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv24.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv24.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv24.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv24.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv24.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv24.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv24.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv24.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv24.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv24.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv24.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv24.Browser
|
|
|
|
|
|
|
|
// New field should be turned-off by default.
|
|
|
|
srvConfig.Worm = false // cv25.Worm should be used here
|
|
|
|
// for the next migration from v25 to v26 to persist
|
|
|
|
// local config value.
|
|
|
|
|
|
|
|
// Load domain config from existing config in the file.
|
|
|
|
srvConfig.Domain = cv24.Domain
|
|
|
|
|
|
|
|
// Load storage class config from existing storage class config in the file.
|
|
|
|
srvConfig.StorageClass.RRS = cv24.StorageClass.RRS
|
|
|
|
srvConfig.StorageClass.Standard = cv24.StorageClass.Standard
|
|
|
|
|
|
|
|
// Load cache config from existing cache config in the file.
|
|
|
|
srvConfig.Cache.Drives = cv24.Cache.Drives
|
|
|
|
srvConfig.Cache.Exclude = cv24.Cache.Exclude
|
|
|
|
srvConfig.Cache.Expiry = cv24.Cache.Expiry
|
|
|
|
|
|
|
|
if err = quick.SaveConfig(srvConfig, configFile, globalEtcdClient); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv24.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv24.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV25ToV26() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
cv25 := &serverConfigV25{}
|
|
|
|
_, err := quick.LoadConfig(configFile, globalEtcdClient, cv25)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config version ‘25’. %v", err)
|
|
|
|
}
|
|
|
|
if cv25.Version != "25" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Copy over fields from V25 into V26 config struct
|
|
|
|
srvConfig := &serverConfigV26{
|
|
|
|
Notify: notifierV3{},
|
|
|
|
}
|
|
|
|
srvConfig.Version = "26"
|
|
|
|
srvConfig.Credential = cv25.Credential
|
|
|
|
srvConfig.Region = cv25.Region
|
|
|
|
if srvConfig.Region == "" {
|
|
|
|
// Region needs to be set for AWS Signature Version 4.
|
|
|
|
srvConfig.Region = globalMinioDefaultRegion
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv25.Notify.AMQP) == 0 {
|
|
|
|
srvConfig.Notify.AMQP = make(map[string]target.AMQPArgs)
|
|
|
|
srvConfig.Notify.AMQP["1"] = target.AMQPArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.AMQP = cv25.Notify.AMQP
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.Elasticsearch) == 0 {
|
|
|
|
srvConfig.Notify.Elasticsearch = make(map[string]target.ElasticsearchArgs)
|
|
|
|
srvConfig.Notify.Elasticsearch["1"] = target.ElasticsearchArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Elasticsearch = cv25.Notify.Elasticsearch
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.Redis) == 0 {
|
|
|
|
srvConfig.Notify.Redis = make(map[string]target.RedisArgs)
|
|
|
|
srvConfig.Notify.Redis["1"] = target.RedisArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Redis = cv25.Notify.Redis
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.PostgreSQL) == 0 {
|
|
|
|
srvConfig.Notify.PostgreSQL = make(map[string]target.PostgreSQLArgs)
|
|
|
|
srvConfig.Notify.PostgreSQL["1"] = target.PostgreSQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.PostgreSQL = cv25.Notify.PostgreSQL
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.Kafka) == 0 {
|
|
|
|
srvConfig.Notify.Kafka = make(map[string]target.KafkaArgs)
|
|
|
|
srvConfig.Notify.Kafka["1"] = target.KafkaArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Kafka = cv25.Notify.Kafka
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.NATS) == 0 {
|
|
|
|
srvConfig.Notify.NATS = make(map[string]target.NATSArgs)
|
|
|
|
srvConfig.Notify.NATS["1"] = target.NATSArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.NATS = cv25.Notify.NATS
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.Webhook) == 0 {
|
|
|
|
srvConfig.Notify.Webhook = make(map[string]target.WebhookArgs)
|
|
|
|
srvConfig.Notify.Webhook["1"] = target.WebhookArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.Webhook = cv25.Notify.Webhook
|
|
|
|
}
|
|
|
|
if len(cv25.Notify.MySQL) == 0 {
|
|
|
|
srvConfig.Notify.MySQL = make(map[string]target.MySQLArgs)
|
|
|
|
srvConfig.Notify.MySQL["1"] = target.MySQLArgs{
|
|
|
|
Format: event.NamespaceFormat,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MySQL = cv25.Notify.MySQL
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(cv25.Notify.MQTT) == 0 {
|
|
|
|
srvConfig.Notify.MQTT = make(map[string]target.MQTTArgs)
|
|
|
|
srvConfig.Notify.MQTT["1"] = target.MQTTArgs{}
|
|
|
|
} else {
|
|
|
|
srvConfig.Notify.MQTT = cv25.Notify.MQTT
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load browser config from existing config in the file.
|
|
|
|
srvConfig.Browser = cv25.Browser
|
|
|
|
|
|
|
|
// Load worm config from existing config in the file.
|
|
|
|
srvConfig.Worm = cv25.Worm
|
|
|
|
|
|
|
|
// Load domain config from existing config in the file.
|
|
|
|
srvConfig.Domain = cv25.Domain
|
|
|
|
|
|
|
|
// Load storage class config from existing storage class config in the file.
|
|
|
|
srvConfig.StorageClass.RRS = cv25.StorageClass.RRS
|
|
|
|
srvConfig.StorageClass.Standard = cv25.StorageClass.Standard
|
|
|
|
|
|
|
|
// Load cache config from existing cache config in the file.
|
|
|
|
srvConfig.Cache.Drives = cv25.Cache.Drives
|
|
|
|
srvConfig.Cache.Exclude = cv25.Cache.Exclude
|
|
|
|
srvConfig.Cache.Expiry = cv25.Cache.Expiry
|
|
|
|
|
|
|
|
// Add predefined value to new server config.
|
|
|
|
srvConfig.Cache.MaxUse = 80
|
|
|
|
|
|
|
|
if err = quick.SaveConfig(srvConfig, configFile, globalEtcdClient); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘%s’ to ‘%s’. %v", cv25.Version, srvConfig.Version, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, cv25.Version, srvConfig.Version)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV26ToV27() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
// config V27 is backward compatible with V26, load the old
|
|
|
|
// config file in serverConfigV27 struct and put some examples
|
|
|
|
// in the new `logger` field
|
|
|
|
srvConfig := &serverConfigV27{}
|
|
|
|
_, err := quick.LoadConfig(configFile, globalEtcdClient, srvConfig)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if srvConfig.Version != "26" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
srvConfig.Version = "27"
|
|
|
|
// Enable console logging by default to avoid breaking users
|
|
|
|
// current deployments
|
|
|
|
srvConfig.Logger.Console.Enabled = true
|
|
|
|
srvConfig.Logger.HTTP = make(map[string]logger.HTTP)
|
|
|
|
srvConfig.Logger.HTTP["1"] = logger.HTTP{}
|
|
|
|
|
|
|
|
if err = quick.SaveConfig(srvConfig, configFile, globalEtcdClient); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘26’ to ‘27’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "26", "27")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV27ToV28() error {
|
|
|
|
configFile := getConfigFile()
|
|
|
|
|
|
|
|
// config V28 is backward compatible with V27, load the old
|
|
|
|
// config file in serverConfigV28 struct and initialize KMSConfig
|
|
|
|
|
|
|
|
srvConfig := &serverConfigV28{}
|
|
|
|
_, err := quick.LoadConfig(configFile, globalEtcdClient, srvConfig)
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if srvConfig.Version != "27" {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
srvConfig.Version = "28"
|
|
|
|
srvConfig.KMS = crypto.KMSConfig{}
|
|
|
|
if err = quick.SaveConfig(srvConfig, configFile, globalEtcdClient); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘27’ to ‘28’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "27", "28")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Migrates ${HOME}/.minio/config.json to '<export_path>/.minio.sys/config/config.json'
|
|
|
|
// if etcd is configured then migrates /config/config.json to '<export_path>/.minio.sys/config/config.json'
|
|
|
|
func migrateConfigToMinioSys(objAPI ObjectLayer) (err error) {
|
|
|
|
// Construct path to config.json for the given bucket.
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
// Verify if config was already available in .minio.sys in which case, nothing more to be done.
|
|
|
|
if err = checkConfig(context.Background(), objAPI, configFile); err != errConfigNotFound {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
defer func() {
|
|
|
|
if err == nil {
|
|
|
|
if globalEtcdClient != nil {
|
|
|
|
deleteKeyEtcd(context.Background(), globalEtcdClient, configFile)
|
|
|
|
} else {
|
|
|
|
// Rename config.json to config.json.deprecated only upon
|
|
|
|
// success of this function.
|
|
|
|
os.Rename(getConfigFile(), getConfigFile()+".deprecated")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
transactionConfigFile := configFile + ".transaction"
|
|
|
|
|
|
|
|
// As object layer's GetObject() and PutObject() take respective lock on minioMetaBucket
|
|
|
|
// and configFile, take a transaction lock to avoid data race between readConfig()
|
|
|
|
// and saveConfig().
|
|
|
|
objLock := globalNSMutex.NewNSLock(context.Background(), minioMetaBucket, transactionConfigFile)
|
|
|
|
if err = objLock.GetLock(globalOperationTimeout); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
defer objLock.Unlock()
|
|
|
|
|
|
|
|
// Verify if backend already has the file (after holding lock)
|
|
|
|
if err = checkConfig(context.Background(), objAPI, configFile); err != errConfigNotFound {
|
|
|
|
return err
|
|
|
|
} // if errConfigNotFound proceed to migrate..
|
|
|
|
|
|
|
|
var configFiles = []string{
|
|
|
|
getConfigFile(),
|
|
|
|
getConfigFile() + ".deprecated",
|
|
|
|
configFile,
|
|
|
|
}
|
|
|
|
var config = &serverConfigV27{}
|
|
|
|
for _, cfgFile := range configFiles {
|
|
|
|
if _, err = Load(cfgFile, config); err != nil {
|
|
|
|
if !os.IsNotExist(err) && !os.IsPermission(err) {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if os.IsPermission(err) {
|
|
|
|
logger.Info("Older config found but not readable %s, proceeding to initialize new config anyways", err)
|
|
|
|
}
|
|
|
|
if os.IsNotExist(err) || os.IsPermission(err) {
|
|
|
|
// Initialize the server config, if no config exists.
|
|
|
|
return newSrvConfig(objAPI)
|
|
|
|
}
|
|
|
|
return saveServerConfig(context.Background(), objAPI, config, nil)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Migrates '.minio.sys/config.json' to v33.
|
|
|
|
func migrateMinioSysConfig(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
// Check if the config version is latest, if not migrate.
|
|
|
|
ok, _, err := checkConfigVersion(objAPI, configFile, "33")
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Construct path to config.json for the given bucket.
|
|
|
|
transactionConfigFile := configFile + ".transaction"
|
|
|
|
|
|
|
|
// As object layer's GetObject() and PutObject() take respective lock on minioMetaBucket
|
|
|
|
// and configFile, take a transaction lock to avoid data race between readConfig()
|
|
|
|
// and saveConfig().
|
|
|
|
objLock := globalNSMutex.NewNSLock(context.Background(), minioMetaBucket, transactionConfigFile)
|
|
|
|
if err := objLock.GetLock(globalOperationTimeout); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
defer objLock.Unlock()
|
|
|
|
|
|
|
|
if err := migrateV27ToV28MinioSys(objAPI); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := migrateV28ToV29MinioSys(objAPI); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := migrateV29ToV30MinioSys(objAPI); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := migrateV30ToV31MinioSys(objAPI); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := migrateV31ToV32MinioSys(objAPI); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
return migrateV32ToV33MinioSys(objAPI)
|
|
|
|
}
|
|
|
|
|
|
|
|
func checkConfigVersion(objAPI ObjectLayer, configFile string, version string) (bool, []byte, error) {
|
|
|
|
data, err := readConfig(context.Background(), objAPI, configFile)
|
|
|
|
if err != nil {
|
|
|
|
return false, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
var versionConfig struct {
|
|
|
|
Version string `json:"version"`
|
|
|
|
}
|
|
|
|
|
|
|
|
vcfg := &versionConfig
|
|
|
|
if err = json.Unmarshal(data, vcfg); err != nil {
|
|
|
|
return false, nil, err
|
|
|
|
}
|
|
|
|
return vcfg.Version == version, data, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV27ToV28MinioSys(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "27")
|
|
|
|
if err == errConfigNotFound {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV28{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg.Version = "28"
|
|
|
|
cfg.KMS = crypto.KMSConfig{}
|
|
|
|
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, cfg, nil); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘27’ to ‘28’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "27", "28")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV28ToV29MinioSys(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "28")
|
|
|
|
if err == errConfigNotFound {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV29{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg.Version = "29"
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, cfg, nil); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘28’ to ‘29’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "28", "29")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV29ToV30MinioSys(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "29")
|
|
|
|
if err == errConfigNotFound {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV30{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg.Version = "30"
|
|
|
|
// Init compression config.For future migration, Compression config needs to be copied over from previous version.
|
|
|
|
cfg.Compression.Enabled = false
|
|
|
|
cfg.Compression.Extensions = strings.Split(compress.DefaultExtensions, config.ValueSeparator)
|
|
|
|
cfg.Compression.MimeTypes = strings.Split(compress.DefaultMimeTypes, config.ValueSeparator)
|
|
|
|
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, cfg, nil); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘29’ to ‘30’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "29", "30")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV30ToV31MinioSys(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "30")
|
|
|
|
if err == errConfigNotFound {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV31{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg.Version = "31"
|
|
|
|
cfg.OpenID = openid.Config{}
|
|
|
|
cfg.OpenID.JWKS.URL = &xnet.URL{}
|
|
|
|
|
|
|
|
cfg.Policy.OPA = opa.Args{
|
|
|
|
URL: &xnet.URL{},
|
|
|
|
AuthToken: "",
|
|
|
|
}
|
|
|
|
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, cfg, nil); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘30’ to ‘31’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "30", "31")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV31ToV32MinioSys(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "31")
|
|
|
|
if err == errConfigNotFound {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV32{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg.Version = "32"
|
|
|
|
cfg.Notify.NSQ = make(map[string]target.NSQArgs)
|
|
|
|
cfg.Notify.NSQ["1"] = target.NSQArgs{}
|
|
|
|
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, cfg, nil); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from ‘31’ to ‘32’. %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "31", "32")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateV32ToV33MinioSys(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "32")
|
|
|
|
if err == errConfigNotFound {
|
|
|
|
return nil
|
|
|
|
} else if err != nil {
|
|
|
|
return fmt.Errorf("Unable to load config file. %v", err)
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV33{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg.Version = "33"
|
|
|
|
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, cfg, nil); err != nil {
|
|
|
|
return fmt.Errorf("Failed to migrate config from 32 to 33 . %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info(configMigrateMSGTemplate, configFile, "32", "33")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func migrateMinioSysConfigToKV(objAPI ObjectLayer) error {
|
|
|
|
configFile := path.Join(minioConfigPrefix, minioConfigFile)
|
|
|
|
|
|
|
|
// Check if the config version is latest, if not migrate.
|
|
|
|
ok, data, err := checkConfigVersion(objAPI, configFile, "33")
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
cfg := &serverConfigV33{}
|
|
|
|
if err = json.Unmarshal(data, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
newCfg := newServerConfig()
|
|
|
|
|
|
|
|
config.SetCredentials(newCfg, cfg.Credential)
|
|
|
|
config.SetRegion(newCfg, cfg.Region)
|
|
|
|
config.SetWorm(newCfg, bool(cfg.Worm))
|
|
|
|
|
|
|
|
storageclass.SetStorageClass(newCfg, cfg.StorageClass)
|
|
|
|
|
|
|
|
for k, loggerArgs := range cfg.Logger.HTTP {
|
|
|
|
logger.SetLoggerHTTP(newCfg, k, loggerArgs)
|
|
|
|
}
|
|
|
|
for k, auditArgs := range cfg.Logger.Audit {
|
|
|
|
logger.SetLoggerHTTPAudit(newCfg, k, auditArgs)
|
|
|
|
}
|
|
|
|
|
|
|
|
crypto.SetKMSConfig(newCfg, cfg.KMS)
|
|
|
|
xldap.SetIdentityLDAP(newCfg, cfg.LDAPServerConfig)
|
|
|
|
openid.SetIdentityOpenID(newCfg, cfg.OpenID)
|
|
|
|
opa.SetPolicyOPAConfig(newCfg, cfg.Policy.OPA)
|
|
|
|
cache.SetCacheConfig(newCfg, cfg.Cache)
|
|
|
|
compress.SetCompressionConfig(newCfg, cfg.Compression)
|
|
|
|
|
|
|
|
for k, args := range cfg.Notify.AMQP {
|
|
|
|
notify.SetNotifyAMQP(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.Elasticsearch {
|
|
|
|
notify.SetNotifyES(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.Kafka {
|
|
|
|
notify.SetNotifyKafka(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.MQTT {
|
|
|
|
notify.SetNotifyMQTT(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.MySQL {
|
|
|
|
notify.SetNotifyMySQL(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.NATS {
|
|
|
|
notify.SetNotifyNATS(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.NSQ {
|
|
|
|
notify.SetNotifyNSQ(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.PostgreSQL {
|
|
|
|
notify.SetNotifyPostgres(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.Redis {
|
|
|
|
notify.SetNotifyRedis(newCfg, k, args)
|
|
|
|
}
|
|
|
|
for k, args := range cfg.Notify.Webhook {
|
|
|
|
notify.SetNotifyWebhook(newCfg, k, args)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Construct path to config.json for the given bucket.
|
|
|
|
transactionConfigFile := configFile + ".transaction"
|
|
|
|
|
|
|
|
// As object layer's GetObject() and PutObject() take respective lock on minioMetaBucket
|
|
|
|
// and configFile, take a transaction lock to avoid data race between readConfig()
|
|
|
|
// and saveConfig().
|
|
|
|
objLock := globalNSMutex.NewNSLock(context.Background(), minioMetaBucket, transactionConfigFile)
|
|
|
|
if err = objLock.GetLock(globalOperationTimeout); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
defer objLock.Unlock()
|
|
|
|
|
|
|
|
if err = saveServerConfig(context.Background(), objAPI, newCfg, cfg); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
logger.Info("Configuration file %s migrated from version '%s' to new KV format successfully.",
|
|
|
|
configFile, "33")
|
|
|
|
return nil
|
|
|
|
}
|