package worker import ( "context" "crypto/rand" "fmt" "sync" "sync/atomic" "github.com/hashicorp/boundary/internal/cmd/config" "github.com/hashicorp/go-hclog" "github.com/hashicorp/vault/sdk/helper/base62" "github.com/hashicorp/vault/sdk/helper/mlock" ua "go.uber.org/atomic" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" ) type Worker struct { conf *Config logger hclog.Logger baseContext context.Context baseCancel context.CancelFunc started ua.Bool controllerConn *atomic.Value lastStatusSuccess *atomic.Value listeningAddress string controllerResolver *atomic.Value controllerResolverCleanup *atomic.Value jobInfoMap *sync.Map cancellationMap *sync.Map } func New(conf *Config) (*Worker, error) { w := &Worker{ conf: conf, logger: conf.Logger.Named("worker"), controllerConn: new(atomic.Value), lastStatusSuccess: new(atomic.Value), controllerResolver: new(atomic.Value), controllerResolverCleanup: new(atomic.Value), jobInfoMap: new(sync.Map), cancellationMap: new(sync.Map), } w.lastStatusSuccess.Store((*LastStatusInformation)(nil)) w.started.Store(false) w.controllerResolver.Store((*manual.Resolver)(nil)) w.controllerResolverCleanup.Store(func() {}) if conf.SecureRandomReader == nil { conf.SecureRandomReader = rand.Reader } var err error if conf.RawConfig.Worker == nil { conf.RawConfig.Worker = new(config.Worker) } if conf.RawConfig.Worker.Name == "" { if conf.RawConfig.Worker.Name, err = base62.Random(10); err != nil { return nil, fmt.Errorf("error auto-generating worker name: %w", err) } } if !conf.RawConfig.DisableMlock { // Ensure our memory usage is locked into physical RAM if err := mlock.LockMemory(); err != nil { return nil, fmt.Errorf( "Failed to lock memory: %v\n\n"+ "This usually means that the mlock syscall is not available.\n"+ "Boundary uses mlock to prevent memory from being swapped to\n"+ "disk. This requires root privileges as well as a machine\n"+ "that supports mlock. Please enable mlock on your system or\n"+ "disable Boundary from using it. To disable Boundary from using it,\n"+ "set the `disable_mlock` configuration option in your configuration\n"+ "file.", err) } } return w, nil } func (w *Worker) Start() error { if w.started.Load() { w.logger.Info("already started, skipping") return nil } w.baseContext, w.baseCancel = context.WithCancel(context.Background()) controllerResolver, controllerResolverCleanup := manual.GenerateAndRegisterManualResolver() w.controllerResolver.Store(controllerResolver) w.controllerResolverCleanup.Store(controllerResolverCleanup) if err := w.startListeners(); err != nil { return fmt.Errorf("error starting worker listeners: %w", err) } if err := w.startControllerConnections(); err != nil { return fmt.Errorf("error making controller connections: %w", err) } w.startStatusTicking(w.baseContext) w.started.Store(true) return nil } // Shutdown shuts down the workers. skipListeners can be used to not stop // listeners, useful for tests if we want to stop and start a worker. In order // to create new listeners we'd have to migrate listener setup logic here -- // doable, but work for later. func (w *Worker) Shutdown(skipListeners bool) error { if !w.started.Load() { w.logger.Info("already shut down, skipping") return nil } w.Resolver().UpdateState(resolver.State{Addresses: []resolver.Address{}}) w.controllerResolverCleanup.Load().(func())() w.baseCancel() if !skipListeners { if err := w.stopListeners(); err != nil { return fmt.Errorf("error stopping worker listeners: %w", err) } } w.listeningAddress = "" w.started.Store(false) return nil } func (w *Worker) Resolver() *manual.Resolver { raw := w.controllerResolver.Load() if raw == nil { panic("nil resolver") } return raw.(*manual.Resolver) }