// Copyright (c) 2015-2024 MinIO, Inc. // // This file is part of MinIO Object Storage stack // // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU Affero General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU Affero General Public License for more details. // // You should have received a copy of the GNU Affero General Public License // along with this program. If not, see . package cmd import ( "bytes" "context" "encoding/hex" "errors" "fmt" "io" "log" "math/rand" "net" "os" "os/signal" "path/filepath" "runtime" "strings" "syscall" "time" "github.com/coreos/go-systemd/v22/daemon" "github.com/dustin/go-humanize" "github.com/minio/cli" "github.com/minio/madmin-go/v3" "github.com/minio/minio-go/v7" "github.com/minio/minio-go/v7/pkg/credentials" "github.com/minio/minio-go/v7/pkg/set" "github.com/minio/minio/internal/auth" "github.com/minio/minio/internal/bucket/bandwidth" "github.com/minio/minio/internal/color" "github.com/minio/minio/internal/config" "github.com/minio/minio/internal/handlers" "github.com/minio/minio/internal/hash/sha256" xhttp "github.com/minio/minio/internal/http" xioutil "github.com/minio/minio/internal/ioutil" "github.com/minio/minio/internal/logger" "github.com/minio/pkg/v3/certs" "github.com/minio/pkg/v3/env" "golang.org/x/exp/slices" "gopkg.in/yaml.v2" ) // ServerFlags - server command specific flags var ServerFlags = []cli.Flag{ cli.StringFlag{ Name: "config", Usage: "specify server configuration via YAML configuration", EnvVar: "MINIO_CONFIG", }, cli.StringFlag{ Name: "address", Value: ":" + GlobalMinioDefaultPort, Usage: "bind to a specific ADDRESS:PORT, ADDRESS can be an IP or hostname", EnvVar: "MINIO_ADDRESS", }, cli.IntFlag{ Name: "listeners", // Deprecated Oct 2022 Value: 1, Usage: "bind N number of listeners per ADDRESS:PORT", EnvVar: "MINIO_LISTENERS", Hidden: true, }, cli.StringFlag{ Name: "console-address", Usage: "bind to a specific ADDRESS:PORT for embedded Console UI, ADDRESS can be an IP or hostname", EnvVar: "MINIO_CONSOLE_ADDRESS", }, cli.DurationFlag{ Name: "shutdown-timeout", Value: xhttp.DefaultShutdownTimeout, Usage: "shutdown timeout to gracefully shutdown server", EnvVar: "MINIO_SHUTDOWN_TIMEOUT", Hidden: true, }, cli.DurationFlag{ Name: "idle-timeout", Value: xhttp.DefaultIdleTimeout, Usage: "idle timeout is the maximum amount of time to wait for the next request when keep-alive are enabled", EnvVar: "MINIO_IDLE_TIMEOUT", Hidden: true, }, cli.DurationFlag{ Name: "read-header-timeout", Value: xhttp.DefaultReadHeaderTimeout, Usage: "read header timeout is the amount of time allowed to read request headers", EnvVar: "MINIO_READ_HEADER_TIMEOUT", Hidden: true, }, cli.DurationFlag{ Name: "conn-user-timeout", Usage: "custom TCP_USER_TIMEOUT for socket buffers", Hidden: true, Value: 10 * time.Minute, EnvVar: "MINIO_CONN_USER_TIMEOUT", }, cli.StringFlag{ Name: "interface", Usage: "bind to right VRF device for MinIO services", Hidden: true, EnvVar: "MINIO_INTERFACE", }, cli.DurationFlag{ Name: "dns-cache-ttl", Usage: "custom DNS cache TTL", Hidden: true, Value: func() time.Duration { if orchestrated { return 30 * time.Second } return 10 * time.Minute }(), EnvVar: "MINIO_DNS_CACHE_TTL", }, cli.IntFlag{ Name: "max-idle-conns-per-host", Usage: "set a custom max idle connections per host value", Hidden: true, Value: 2048, EnvVar: "MINIO_MAX_IDLE_CONNS_PER_HOST", }, cli.StringSliceFlag{ Name: "ftp", Usage: "enable and configure an FTP(Secure) server", }, cli.StringSliceFlag{ Name: "sftp", Usage: "enable and configure an SFTP server", }, cli.StringFlag{ Name: "crossdomain-xml", Usage: "provide a custom crossdomain-xml configuration to report at http://endpoint/crossdomain.xml", Hidden: true, EnvVar: "MINIO_CROSSDOMAIN_XML", }, cli.StringFlag{ Name: "memlimit", Usage: "set global memory limit per server via GOMEMLIMIT", Hidden: true, EnvVar: "MINIO_MEMLIMIT", }, cli.IntFlag{ Name: "send-buf-size", Value: 4 * humanize.MiByte, EnvVar: "MINIO_SEND_BUF_SIZE", Hidden: true, }, cli.IntFlag{ Name: "recv-buf-size", Value: 4 * humanize.MiByte, EnvVar: "MINIO_RECV_BUF_SIZE", Hidden: true, }, cli.StringFlag{ Name: "log-dir", Usage: "specify the directory to save the server log", EnvVar: "MINIO_LOG_DIR", Hidden: true, }, cli.IntFlag{ Name: "log-size", Usage: "specify the maximum server log file size in bytes before its rotated", Value: 10 * humanize.MiByte, EnvVar: "MINIO_LOG_SIZE", Hidden: true, }, cli.BoolFlag{ Name: "log-compress", Usage: "specify if we want the rotated logs to be gzip compressed or not", EnvVar: "MINIO_LOG_COMPRESS", Hidden: true, }, cli.StringFlag{ Name: "log-prefix", Usage: "specify the log prefix name for the server log", EnvVar: "MINIO_LOG_PREFIX", Hidden: true, }, } var serverCmd = cli.Command{ Name: "server", Usage: "start object storage server", Flags: append(ServerFlags, GlobalFlags...), Action: serverMain, CustomHelpTemplate: `NAME: {{.HelpName}} - {{.Usage}} USAGE: {{.HelpName}} {{if .VisibleFlags}}[FLAGS] {{end}}DIR1 [DIR2..] {{.HelpName}} {{if .VisibleFlags}}[FLAGS] {{end}}DIR{1...64} {{.HelpName}} {{if .VisibleFlags}}[FLAGS] {{end}}DIR{1...64} DIR{65...128} DIR: DIR points to a directory on a filesystem. When you want to combine multiple drives into a single large system, pass one directory per filesystem separated by space. You may also use a '...' convention to abbreviate the directory arguments. Remote directories in a distributed setup are encoded as HTTP(s) URIs. {{if .VisibleFlags}} FLAGS: {{range .VisibleFlags}}{{.}} {{end}}{{end}} EXAMPLES: 1. Start MinIO server on "/home/shared" directory. {{.Prompt}} {{.HelpName}} /home/shared 2. Start single node server with 64 local drives "/mnt/data1" to "/mnt/data64". {{.Prompt}} {{.HelpName}} /mnt/data{1...64} 3. Start distributed MinIO server on an 32 node setup with 32 drives each, run following command on all the nodes {{.Prompt}} {{.HelpName}} http://node{1...32}.example.com/mnt/export{1...32} 4. Start distributed MinIO server in an expanded setup, run the following command on all the nodes {{.Prompt}} {{.HelpName}} http://node{1...16}.example.com/mnt/export{1...32} \ http://node{17...64}.example.com/mnt/export{1...64} 5. Start distributed MinIO server, with FTP and SFTP servers on all interfaces via port 8021, 8022 respectively {{.Prompt}} {{.HelpName}} http://node{1...4}.example.com/mnt/export{1...4} \ --ftp="address=:8021" --ftp="passive-port-range=30000-40000" \ --sftp="address=:8022" --sftp="ssh-private-key=${HOME}/.ssh/id_rsa" `, } func serverCmdArgs(ctx *cli.Context) []string { v, _, _, err := env.LookupEnv(config.EnvArgs) if err != nil { logger.FatalIf(err, "Unable to validate passed arguments in %s:%s", config.EnvArgs, os.Getenv(config.EnvArgs)) } if v == "" { v, _, _, err = env.LookupEnv(config.EnvVolumes) if err != nil { logger.FatalIf(err, "Unable to validate passed arguments in %s:%s", config.EnvVolumes, os.Getenv(config.EnvVolumes)) } } if v == "" { // Fall back to older environment value MINIO_ENDPOINTS v, _, _, err = env.LookupEnv(config.EnvEndpoints) if err != nil { logger.FatalIf(err, "Unable to validate passed arguments in %s:%s", config.EnvEndpoints, os.Getenv(config.EnvEndpoints)) } } if v == "" { if !ctx.Args().Present() || ctx.Args().First() == "help" { cli.ShowCommandHelpAndExit(ctx, ctx.Command.Name, 1) } return ctx.Args() } return strings.Fields(v) } func configCommonToSrvCtx(cf config.ServerConfigCommon, ctxt *serverCtxt) { ctxt.RootUser = cf.RootUser ctxt.RootPwd = cf.RootPwd if cf.Addr != "" { ctxt.Addr = cf.Addr } if cf.ConsoleAddr != "" { ctxt.ConsoleAddr = cf.ConsoleAddr } if cf.CertsDir != "" { ctxt.CertsDir = cf.CertsDir ctxt.certsDirSet = true } if cf.Options.FTP.Address != "" { ctxt.FTP = append(ctxt.FTP, fmt.Sprintf("address=%s", cf.Options.FTP.Address)) } if cf.Options.FTP.PassivePortRange != "" { ctxt.FTP = append(ctxt.FTP, fmt.Sprintf("passive-port-range=%s", cf.Options.FTP.PassivePortRange)) } if cf.Options.SFTP.Address != "" { ctxt.SFTP = append(ctxt.SFTP, fmt.Sprintf("address=%s", cf.Options.SFTP.Address)) } if cf.Options.SFTP.SSHPrivateKey != "" { ctxt.SFTP = append(ctxt.SFTP, fmt.Sprintf("ssh-private-key=%s", cf.Options.SFTP.SSHPrivateKey)) } } func mergeServerCtxtFromConfigFile(configFile string, ctxt *serverCtxt) error { rd, err := xioutil.ReadFile(configFile) if err != nil { return err } cfReader := bytes.NewReader(rd) cv := config.ServerConfigVersion{} if err = yaml.Unmarshal(rd, &cv); err != nil { return err } switch cv.Version { case "v1", "v2": default: return fmt.Errorf("unexpected version: %s", cv.Version) } cfCommon := config.ServerConfigCommon{} if err = yaml.Unmarshal(rd, &cfCommon); err != nil { return err } configCommonToSrvCtx(cfCommon, ctxt) v, err := env.GetInt(EnvErasureSetDriveCount, 0) if err != nil { return err } setDriveCount := uint64(v) var pools []poolArgs switch cv.Version { case "v1": cfV1 := config.ServerConfigV1{} if err = yaml.Unmarshal(rd, &cfV1); err != nil { return err } pools = make([]poolArgs, 0, len(cfV1.Pools)) for _, list := range cfV1.Pools { pools = append(pools, poolArgs{ args: list, setDriveCount: setDriveCount, }) } case "v2": cf := config.ServerConfig{} cfReader.Seek(0, io.SeekStart) if err = yaml.Unmarshal(rd, &cf); err != nil { return err } pools = make([]poolArgs, 0, len(cf.Pools)) for _, list := range cf.Pools { driveCount := list.SetDriveCount if setDriveCount > 0 { driveCount = setDriveCount } pools = append(pools, poolArgs{ args: list.Args, setDriveCount: driveCount, }) } } ctxt.Layout, err = buildDisksLayoutFromConfFile(pools) return err } func serverHandleCmdArgs(ctxt serverCtxt) { handleCommonArgs(ctxt) logger.FatalIf(CheckLocalServerAddr(globalMinioAddr), "Unable to validate passed arguments") var err error var setupType SetupType // Check and load TLS certificates. globalPublicCerts, globalTLSCerts, globalIsTLS, err = getTLSConfig() logger.FatalIf(err, "Unable to load the TLS configuration") // Check and load Root CAs. globalRootCAs, err = certs.GetRootCAs(globalCertsCADir.Get()) logger.FatalIf(err, "Failed to read root CAs (%v)", err) // Add the global public crts as part of global root CAs for _, publicCrt := range globalPublicCerts { globalRootCAs.AddCert(publicCrt) } // Register root CAs for remote ENVs env.RegisterGlobalCAs(globalRootCAs) globalEndpoints, setupType, err = createServerEndpoints(globalMinioAddr, ctxt.Layout.pools, ctxt.Layout.legacy) logger.FatalIf(err, "Invalid command line arguments") globalNodes = globalEndpoints.GetNodes() globalIsErasure = (setupType == ErasureSetupType) globalIsDistErasure = (setupType == DistErasureSetupType) if globalIsDistErasure { globalIsErasure = true } globalIsErasureSD = (setupType == ErasureSDSetupType) if globalDynamicAPIPort && globalIsDistErasure { logger.FatalIf(errInvalidArgument, "Invalid --address=\"%s\", port '0' is not allowed in a distributed erasure coded setup", ctxt.Addr) } globalLocalNodeName = GetLocalPeer(globalEndpoints, globalMinioHost, globalMinioPort) nodeNameSum := sha256.Sum256([]byte(globalLocalNodeName)) globalLocalNodeNameHex = hex.EncodeToString(nodeNameSum[:]) // Initialize, see which NIC the service is running on, and save it as global value setGlobalInternodeInterface(ctxt.Interface) globalTCPOptions = xhttp.TCPOptions{ UserTimeout: int(ctxt.UserTimeout.Milliseconds()), // FIXME: Bring this back when we have valid way to handle deadlines // DriveOPTimeout: globalDriveConfig.GetOPTimeout, Interface: ctxt.Interface, SendBufSize: ctxt.SendBufSize, RecvBufSize: ctxt.RecvBufSize, } // allow transport to be HTTP/1.1 for proxying. globalProxyEndpoints = GetProxyEndpoints(globalEndpoints) globalInternodeTransport = NewInternodeHTTPTransport(ctxt.MaxIdleConnsPerHost)() globalRemoteTargetTransport = NewRemoteTargetHTTPTransport(false)() globalForwarder = handlers.NewForwarder(&handlers.Forwarder{ PassHost: true, RoundTripper: globalRemoteTargetTransport, Logger: func(err error) { if err != nil && !errors.Is(err, context.Canceled) { proxyLogIf(GlobalContext, err) } }, }) // On macOS, if a process already listens on LOCALIPADDR:PORT, net.Listen() falls back // to IPv6 address ie minio will start listening on IPv6 address whereas another // (non-)minio process is listening on IPv4 of given port. // To avoid this error situation we check for port availability. logger.FatalIf(xhttp.CheckPortAvailability(globalMinioHost, globalMinioPort, globalTCPOptions), "Unable to start the server") } func initAllSubsystems(ctx context.Context) { // Initialize notification peer targets globalNotificationSys = NewNotificationSys(globalEndpoints) // Create new notification system globalEventNotifier = NewEventNotifier(GlobalContext) // Create new bucket metadata system. if globalBucketMetadataSys == nil { globalBucketMetadataSys = NewBucketMetadataSys() } else { // Reinitialize safely when testing. globalBucketMetadataSys.Reset() } // Create the bucket bandwidth monitor globalBucketMonitor = bandwidth.NewMonitor(ctx, uint64(totalNodeCount())) // Create a new config system. globalConfigSys = NewConfigSys() // Create new IAM system. globalIAMSys = NewIAMSys() // Create new policy system. globalPolicySys = NewPolicySys() // Create new lifecycle system. globalLifecycleSys = NewLifecycleSys() // Create new bucket encryption subsystem globalBucketSSEConfigSys = NewBucketSSEConfigSys() // Create new bucket object lock subsystem globalBucketObjectLockSys = NewBucketObjectLockSys() // Create new bucket quota subsystem globalBucketQuotaSys = NewBucketQuotaSys() // Create new bucket versioning subsystem if globalBucketVersioningSys == nil { globalBucketVersioningSys = NewBucketVersioningSys() } // Create new bucket replication subsystem globalBucketTargetSys = NewBucketTargetSys(GlobalContext) // Create new ILM tier configuration subsystem globalTierConfigMgr = NewTierConfigMgr() globalTransitionState = newTransitionState(GlobalContext) globalSiteResyncMetrics = newSiteResyncMetrics(GlobalContext) } func configRetriableErrors(err error) bool { if err == nil { return false } notInitialized := strings.Contains(err.Error(), "Server not initialized, please try again") || errors.Is(err, errServerNotInitialized) // Initializing sub-systems needs a retry mechanism for // the following reasons: // - Read quorum is lost just after the initialization // of the object layer. // - Write quorum not met when upgrading configuration // version is needed, migration is needed etc. rquorum := InsufficientReadQuorum{} wquorum := InsufficientWriteQuorum{} // One of these retriable errors shall be retried. return errors.Is(err, errDiskNotFound) || errors.Is(err, errConfigNotFound) || errors.Is(err, context.DeadlineExceeded) || errors.Is(err, errErasureWriteQuorum) || errors.Is(err, errErasureReadQuorum) || errors.Is(err, io.ErrUnexpectedEOF) || errors.As(err, &rquorum) || errors.As(err, &wquorum) || isErrObjectNotFound(err) || isErrBucketNotFound(err) || errors.Is(err, os.ErrDeadlineExceeded) || notInitialized } func bootstrapTraceMsg(msg string) { info := madmin.TraceInfo{ TraceType: madmin.TraceBootstrap, Time: UTCNow(), NodeName: globalLocalNodeName, FuncName: "BOOTSTRAP", Message: fmt.Sprintf("%s %s", getSource(2), msg), } globalBootstrapTracer.Record(info) if serverDebugLog { fmt.Println(time.Now().Round(time.Millisecond).Format(time.RFC3339), " bootstrap: ", msg) } noSubs := globalTrace.NumSubscribers(madmin.TraceBootstrap) == 0 if noSubs { return } globalTrace.Publish(info) } func bootstrapTrace(msg string, worker func()) { if serverDebugLog { fmt.Println(time.Now().Round(time.Millisecond).Format(time.RFC3339), " bootstrap: ", msg) } now := time.Now() worker() dur := time.Since(now) info := madmin.TraceInfo{ TraceType: madmin.TraceBootstrap, Time: UTCNow(), NodeName: globalLocalNodeName, FuncName: "BOOTSTRAP", Message: fmt.Sprintf("%s %s (duration: %s)", getSource(2), msg, dur), } globalBootstrapTracer.Record(info) if globalTrace.NumSubscribers(madmin.TraceBootstrap) == 0 { return } globalTrace.Publish(info) } func initServerConfig(ctx context.Context, newObject ObjectLayer) error { t1 := time.Now() r := rand.New(rand.NewSource(time.Now().UnixNano())) for { select { case <-ctx.Done(): // Retry was canceled successfully. return fmt.Errorf("Initializing sub-systems stopped gracefully %w", ctx.Err()) default: } // These messages only meant primarily for distributed setup, so only log during distributed setup. if globalIsDistErasure { logger.Info("Waiting for all MinIO sub-systems to be initialize...") } // Upon success migrating the config, initialize all sub-systems // if all sub-systems initialized successfully return right away err := initConfigSubsystem(ctx, newObject) if err == nil { // All successful return. if globalIsDistErasure { // These messages only meant primarily for distributed setup, so only log during distributed setup. logger.Info("All MinIO sub-systems initialized successfully in %s", time.Since(t1)) } return nil } if configRetriableErrors(err) { logger.Info("Waiting for all MinIO sub-systems to be initialized.. possible cause (%v)", err) time.Sleep(time.Duration(r.Float64() * float64(5*time.Second))) continue } // Any other unhandled return right here. return fmt.Errorf("Unable to initialize sub-systems: %w", err) } } func initConfigSubsystem(ctx context.Context, newObject ObjectLayer) error { // %w is used by all error returns here to make sure // we wrap the underlying error, make sure when you // are modifying this code that you do so, if and when // you want to add extra context to your error. This // ensures top level retry works accordingly. // Initialize config system. if err := globalConfigSys.Init(newObject); err != nil { if configRetriableErrors(err) { return fmt.Errorf("Unable to initialize config system: %w", err) } // Any other config errors we simply print a message and proceed forward. configLogIf(ctx, fmt.Errorf("Unable to initialize config, some features may be missing: %w", err)) } return nil } func setGlobalInternodeInterface(interfaceName string) { globalInternodeInterfaceOnce.Do(func() { if interfaceName != "" { globalInternodeInterface = interfaceName return } ip := "127.0.0.1" host, _ := mustSplitHostPort(globalLocalNodeName) if host != "" { if net.ParseIP(host) != nil { ip = host } else { ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second) defer cancel() haddrs, err := globalDNSCache.LookupHost(ctx, host) if err == nil { ip = haddrs[0] } } } ifs, _ := net.Interfaces() for _, interf := range ifs { addrs, err := interf.Addrs() if err == nil { for _, addr := range addrs { if strings.SplitN(addr.String(), "/", 2)[0] == ip { globalInternodeInterface = interf.Name } } } } }) } // Return the list of address that MinIO server needs to listen on: // - Returning 127.0.0.1 is necessary so Console will be able to send // requests to the local S3 API. // - The returned List needs to be deduplicated as well. func getServerListenAddrs() []string { // Use a string set to avoid duplication addrs := set.NewStringSet() // Listen on local interface to receive requests from Console for _, ip := range mustGetLocalIPs() { if ip != nil && ip.IsLoopback() { addrs.Add(net.JoinHostPort(ip.String(), globalMinioPort)) } } host, _ := mustSplitHostPort(globalMinioAddr) if host != "" { ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second) defer cancel() haddrs, err := globalDNSCache.LookupHost(ctx, host) if err == nil { for _, addr := range haddrs { addrs.Add(net.JoinHostPort(addr, globalMinioPort)) } } else { // Unable to lookup host in 2-secs, let it fail later anyways. addrs.Add(globalMinioAddr) } } else { addrs.Add(globalMinioAddr) } return addrs.ToSlice() } var globalLoggerOutput io.WriteCloser func initializeLogRotate(ctx *cli.Context) (io.WriteCloser, error) { lgDir := ctx.String("log-dir") if lgDir == "" { return os.Stderr, nil } lgDirAbs, err := filepath.Abs(lgDir) if err != nil { return nil, err } lgSize := ctx.Int("log-size") var fileNameFunc func() string if ctx.IsSet("log-prefix") { fileNameFunc = func() string { return fmt.Sprintf("%s-%s.log", ctx.String("log-prefix"), fmt.Sprintf("%X", time.Now().UTC().UnixNano())) } } output, err := logger.NewDir(logger.Options{ Directory: lgDirAbs, MaximumFileSize: int64(lgSize), Compress: ctx.Bool("log-compress"), FileNameFunc: fileNameFunc, }) if err != nil { return nil, err } logger.EnableJSON() return output, nil } // serverMain handler called for 'minio server' command. func serverMain(ctx *cli.Context) { var warnings []string signal.Notify(globalOSSignalCh, os.Interrupt, syscall.SIGTERM, syscall.SIGQUIT) go handleSignals() setDefaultProfilerRates() // Initialize globalConsoleSys system bootstrapTrace("newConsoleLogger", func() { output, err := initializeLogRotate(ctx) if err == nil { logger.Output = output globalConsoleSys = NewConsoleLogger(GlobalContext, output) globalLoggerOutput = output } else { logger.Output = os.Stderr globalConsoleSys = NewConsoleLogger(GlobalContext, os.Stderr) } logger.AddSystemTarget(GlobalContext, globalConsoleSys) // Set node name, only set for distributed setup. globalConsoleSys.SetNodeName(globalLocalNodeName) if err != nil { // We can only log here since we need globalConsoleSys initialized logger.Fatal(err, "invalid --logrorate-dir option") } }) // Always load ENV variables from files first. loadEnvVarsFromFiles() // Handle all server command args and build the disks layout bootstrapTrace("serverHandleCmdArgs", func() { err := buildServerCtxt(ctx, &globalServerCtxt) logger.FatalIf(err, "Unable to prepare the list of endpoints") serverHandleCmdArgs(globalServerCtxt) }) // DNS cache subsystem to reduce outgoing DNS requests runDNSCache(ctx) // Handle all server environment vars. serverHandleEnvVars() // Load the root credentials from the shell environment or from // the config file if not defined, set the default one. loadRootCredentials() // Perform any self-tests bootstrapTrace("selftests", func() { bitrotSelfTest() erasureSelfTest() compressSelfTest() }) // Initialize KMS configuration bootstrapTrace("handleKMSConfig", handleKMSConfig) // Initialize all help bootstrapTrace("initHelp", initHelp) // Initialize all sub-systems bootstrapTrace("initAllSubsystems", func() { initAllSubsystems(GlobalContext) }) // Is distributed setup, error out if no certificates are found for HTTPS endpoints. if globalIsDistErasure { if globalEndpoints.HTTPS() && !globalIsTLS { logger.Fatal(config.ErrNoCertsAndHTTPSEndpoints(nil), "Unable to start the server") } if !globalEndpoints.HTTPS() && globalIsTLS { logger.Fatal(config.ErrCertsAndHTTPEndpoints(nil), "Unable to start the server") } } var getCert certs.GetCertificateFunc if globalTLSCerts != nil { getCert = globalTLSCerts.GetCertificate } // Check for updates in non-blocking manner. go func() { if !globalServerCtxt.Quiet && !globalInplaceUpdateDisabled { // Check for new updates from dl.min.io. bootstrapTrace("checkUpdate", func() { checkUpdate(getMinioMode()) }) } }() // Set system resources to maximum. bootstrapTrace("setMaxResources", func() { _ = setMaxResources(globalServerCtxt) }) // Verify kernel release and version. if oldLinux() { warnings = append(warnings, color.YellowBold("- Detected Linux kernel version older than 4.0.0 release, there are some known potential performance problems with this kernel version. MinIO recommends a minimum of 4.x.x linux kernel version for best performance")) } maxProcs := runtime.GOMAXPROCS(0) cpuProcs := runtime.NumCPU() if maxProcs < cpuProcs { warnings = append(warnings, color.YellowBold("- Detected GOMAXPROCS(%d) < NumCPU(%d), please make sure to provide all PROCS to MinIO for optimal performance", maxProcs, cpuProcs)) } // Initialize grid bootstrapTrace("initGrid", func() { logger.FatalIf(initGlobalGrid(GlobalContext, globalEndpoints), "Unable to configure server grid RPC services") }) // Configure server. bootstrapTrace("configureServer", func() { handler, err := configureServerHandler(globalEndpoints) if err != nil { logger.Fatal(config.ErrUnexpectedError(err), "Unable to configure one of server's RPC services") } // Allow grid to start after registering all services. xioutil.SafeClose(globalGridStart) httpServer := xhttp.NewServer(getServerListenAddrs()). UseHandler(setCriticalErrorHandler(corsHandler(handler))). UseTLSConfig(newTLSConfig(getCert)). UseShutdownTimeout(globalServerCtxt.ShutdownTimeout). UseIdleTimeout(globalServerCtxt.IdleTimeout). UseReadHeaderTimeout(globalServerCtxt.ReadHeaderTimeout). UseBaseContext(GlobalContext). UseCustomLogger(log.New(io.Discard, "", 0)). // Turn-off random logging by Go stdlib UseTCPOptions(globalTCPOptions) httpServer.TCPOptions.Trace = bootstrapTraceMsg go func() { serveFn, err := httpServer.Init(GlobalContext, func(listenAddr string, err error) { bootLogIf(GlobalContext, fmt.Errorf("Unable to listen on `%s`: %v", listenAddr, err)) }) if err != nil { globalHTTPServerErrorCh <- err return } globalHTTPServerErrorCh <- serveFn() }() setHTTPServer(httpServer) }) if globalIsDistErasure { bootstrapTrace("verifying system configuration", func() { // Additionally in distributed setup, validate the setup and configuration. if err := verifyServerSystemConfig(GlobalContext, globalEndpoints, globalGrid.Load()); err != nil { logger.Fatal(err, "Unable to start the server") } }) } if !globalDisableFreezeOnBoot { // Freeze the services until the bucket notification subsystem gets initialized. bootstrapTrace("freezeServices", freezeServices) } var newObject ObjectLayer bootstrapTrace("newObjectLayer", func() { var err error newObject, err = newObjectLayer(GlobalContext, globalEndpoints) if err != nil { logFatalErrs(err, Endpoint{}, true) } }) for _, n := range globalNodes { nodeName := n.Host if n.IsLocal { nodeName = globalLocalNodeName } nodeNameSum := sha256.Sum256([]byte(nodeName + globalDeploymentID())) globalNodeNamesHex[hex.EncodeToString(nodeNameSum[:])] = struct{}{} } var err error bootstrapTrace("initServerConfig", func() { if err = initServerConfig(GlobalContext, newObject); err != nil { var cerr config.Err // For any config error, we don't need to drop into safe-mode // instead its a user error and should be fixed by user. if errors.As(err, &cerr) { logger.FatalIf(err, "Unable to initialize the server") } // If context was canceled if errors.Is(err, context.Canceled) { logger.FatalIf(err, "Server startup canceled upon user request") } bootLogIf(GlobalContext, err) } if !globalServerCtxt.StrictS3Compat { warnings = append(warnings, color.YellowBold("- Strict AWS S3 compatible incoming PUT, POST content payload validation is turned off, caution is advised do not use in production")) } }) if globalActiveCred.Equal(auth.DefaultCredentials) { msg := fmt.Sprintf("- Detected default credentials '%s', we recommend that you change these values with 'MINIO_ROOT_USER' and 'MINIO_ROOT_PASSWORD' environment variables", globalActiveCred) warnings = append(warnings, color.YellowBold(msg)) } // Initialize users credentials and policies in background right after config has initialized. go func() { bootstrapTrace("globalIAMSys.Init", func() { globalIAMSys.Init(GlobalContext, newObject, globalEtcdClient, globalRefreshIAMInterval) }) // Initialize Console UI if globalBrowserEnabled { bootstrapTrace("initConsoleServer", func() { srv, err := initConsoleServer() if err != nil { logger.FatalIf(err, "Unable to initialize console service") } setConsoleSrv(srv) go func() { logger.FatalIf(newConsoleServerFn().Serve(), "Unable to initialize console server") }() }) } // if we see FTP args, start FTP if possible if len(globalServerCtxt.FTP) > 0 { bootstrapTrace("go startFTPServer", func() { go startFTPServer(globalServerCtxt.FTP) }) } // If we see SFTP args, start SFTP if possible if len(globalServerCtxt.SFTP) > 0 { bootstrapTrace("go startSFTPServer", func() { go startSFTPServer(globalServerCtxt.SFTP) }) } }() go func() { r := rand.New(rand.NewSource(time.Now().UnixNano())) if !globalDisableFreezeOnBoot { defer bootstrapTrace("unfreezeServices", unfreezeServices) t := time.AfterFunc(5*time.Minute, func() { warnings = append(warnings, color.YellowBold("- Initializing the config subsystem is taking longer than 5 minutes. Please set '_MINIO_DISABLE_API_FREEZE_ON_BOOT=true' to not freeze the APIs")) }) defer t.Stop() } // Initialize data scanner. bootstrapTrace("initDataScanner", func() { if v := env.Get("_MINIO_SCANNER", config.EnableOn); v == config.EnableOn { initDataScanner(GlobalContext, newObject) } }) // Initialize background replication bootstrapTrace("initBackgroundReplication", func() { initBackgroundReplication(GlobalContext, newObject) }) // Initialize background ILM worker poool bootstrapTrace("initBackgroundExpiry", func() { initBackgroundExpiry(GlobalContext, newObject) }) bootstrapTrace("globalTransitionState.Init", func() { globalTransitionState.Init(newObject) }) // Initialize batch job pool. bootstrapTrace("newBatchJobPool", func() { globalBatchJobPool = newBatchJobPool(GlobalContext, newObject, 100) }) // Initialize the license update job bootstrapTrace("initLicenseUpdateJob", func() { initLicenseUpdateJob(GlobalContext, newObject) }) go func() { // Initialize transition tier configuration manager bootstrapTrace("globalTierConfigMgr.Init", func() { if err := globalTierConfigMgr.Init(GlobalContext, newObject); err != nil { bootLogIf(GlobalContext, err) } }) }() // Initialize bucket notification system. bootstrapTrace("initBucketTargets", func() { bootLogIf(GlobalContext, globalEventNotifier.InitBucketTargets(GlobalContext, newObject)) }) var buckets []BucketInfo // List buckets to initialize bucket metadata sub-sys. bootstrapTrace("listBuckets", func() { for { buckets, err = newObject.ListBuckets(GlobalContext, BucketOptions{}) if err != nil { if configRetriableErrors(err) { logger.Info("Waiting for list buckets to succeed to initialize buckets.. possible cause (%v)", err) time.Sleep(time.Duration(r.Float64() * float64(time.Second))) continue } bootLogIf(GlobalContext, fmt.Errorf("Unable to list buckets to initialize bucket metadata sub-system: %w", err)) } break } }) // Initialize bucket metadata sub-system. bootstrapTrace("globalBucketMetadataSys.Init", func() { globalBucketMetadataSys.Init(GlobalContext, buckets, newObject) }) // initialize replication resync state. bootstrapTrace("initResync", func() { globalReplicationPool.initResync(GlobalContext, buckets, newObject) }) // Initialize site replication manager after bucket metadata bootstrapTrace("globalSiteReplicationSys.Init", func() { globalSiteReplicationSys.Init(GlobalContext, newObject) }) // Initialize quota manager. bootstrapTrace("globalBucketQuotaSys.Init", func() { globalBucketQuotaSys.Init(newObject) }) // Populate existing buckets to the etcd backend if globalDNSConfig != nil { // Background this operation. bootstrapTrace("go initFederatorBackend", func() { go initFederatorBackend(buckets, newObject) }) } // Prints the formatted startup message, if err is not nil then it prints additional information as well. printStartupMessage(getAPIEndpoints(), err) // Print a warning at the end of the startup banner so it is more noticeable if newObject.BackendInfo().StandardSCParity == 0 { warnings = append(warnings, color.YellowBold("- The standard parity is set to 0. This can lead to data loss.")) } objAPI := newObjectLayerFn() if objAPI != nil { printStorageInfo(objAPI.StorageInfo(GlobalContext, true)) } if len(warnings) > 0 { logger.Info(color.Yellow("STARTUP WARNINGS:")) for _, warn := range warnings { logger.Info(warn) } } }() region := globalSite.Region() if region == "" { region = "us-east-1" } bootstrapTrace("globalMinioClient", func() { globalMinioClient, err = minio.New(globalLocalNodeName, &minio.Options{ Creds: credentials.NewStaticV4(globalActiveCred.AccessKey, globalActiveCred.SecretKey, ""), Secure: globalIsTLS, Transport: globalRemoteTargetTransport, Region: region, }) logger.FatalIf(err, "Unable to initialize MinIO client") }) go bootstrapTrace("startResourceMetricsCollection", func() { startResourceMetricsCollection() }) // Add User-Agent to differentiate the requests. globalMinioClient.SetAppInfo("minio-perf-test", ReleaseTag) if serverDebugLog { fmt.Println("== DEBUG Mode enabled ==") fmt.Println("Currently set environment settings:") ks := []string{ config.EnvAccessKey, config.EnvSecretKey, config.EnvRootUser, config.EnvRootPassword, } for _, v := range os.Environ() { // Do not print sensitive creds in debug. if slices.Contains(ks, strings.Split(v, "=")[0]) { continue } fmt.Println(v) } fmt.Println("======") } daemon.SdNotify(false, daemon.SdNotifyReady) <-globalOSSignalCh } // Initialize object layer with the supplied disks, objectLayer is nil upon any error. func newObjectLayer(ctx context.Context, endpointServerPools EndpointServerPools) (newObject ObjectLayer, err error) { return newErasureServerPools(ctx, endpointServerPools) }