summaryrefslogtreecommitdiff
path: root/pkg/adapter/containers_remote.go
diff options
context:
space:
mode:
Diffstat (limited to 'pkg/adapter/containers_remote.go')
-rw-r--r--pkg/adapter/containers_remote.go1139
1 files changed, 0 insertions, 1139 deletions
diff --git a/pkg/adapter/containers_remote.go b/pkg/adapter/containers_remote.go
deleted file mode 100644
index fc8b524d6..000000000
--- a/pkg/adapter/containers_remote.go
+++ /dev/null
@@ -1,1139 +0,0 @@
-// +build remoteclient
-
-package adapter
-
-import (
- "bufio"
- "context"
- "encoding/json"
- "fmt"
- "io"
- "os"
- "strconv"
- "syscall"
- "time"
-
- "github.com/containers/libpod/cmd/podman/cliconfig"
- "github.com/containers/libpod/cmd/podman/shared"
- "github.com/containers/libpod/libpod"
- "github.com/containers/libpod/libpod/define"
- "github.com/containers/libpod/libpod/logs"
- envLib "github.com/containers/libpod/pkg/env"
- iopodman "github.com/containers/libpod/pkg/varlink"
- "github.com/containers/libpod/pkg/varlinkapi/virtwriter"
- "github.com/cri-o/ocicni/pkg/ocicni"
- "github.com/docker/docker/pkg/term"
- "github.com/opencontainers/runtime-spec/specs-go"
- "github.com/pkg/errors"
- "github.com/sirupsen/logrus"
- "github.com/varlink/go/varlink"
- "golang.org/x/crypto/ssh/terminal"
- "k8s.io/client-go/tools/remotecommand"
-)
-
-// Inspect returns an inspect struct from varlink
-func (c *Container) Inspect(size bool) (*define.InspectContainerData, error) {
- reply, err := iopodman.ContainerInspectData().Call(c.Runtime.Conn, c.ID(), size)
- if err != nil {
- return nil, err
- }
- data := define.InspectContainerData{}
- if err := json.Unmarshal([]byte(reply), &data); err != nil {
- return nil, err
- }
- return &data, err
-}
-
-// ID returns the ID of the container
-func (c *Container) ID() string {
- return c.config.ID
-}
-
-// Restart a single container
-func (c *Container) Restart(timeout int64) error {
- _, err := iopodman.RestartContainer().Call(c.Runtime.Conn, c.ID(), timeout)
- return err
-}
-
-// Pause a container
-func (c *Container) Pause() error {
- _, err := iopodman.PauseContainer().Call(c.Runtime.Conn, c.ID())
- return err
-}
-
-// Unpause a container
-func (c *Container) Unpause() error {
- _, err := iopodman.UnpauseContainer().Call(c.Runtime.Conn, c.ID())
- return err
-}
-
-func (c *Container) PortMappings() ([]ocicni.PortMapping, error) {
- // First check if the container belongs to a network namespace (like a pod)
- // Taken from libpod portmappings()
- if len(c.config.NetNsCtr) > 0 {
- netNsCtr, err := c.Runtime.LookupContainer(c.config.NetNsCtr)
- if err != nil {
- return nil, errors.Wrapf(err, "unable to lookup network namespace for container %s", c.ID())
- }
- return netNsCtr.PortMappings()
- }
- return c.config.PortMappings, nil
-}
-
-// Config returns a container config
-func (r *LocalRuntime) Config(name string) *libpod.ContainerConfig {
- // TODO the Spec being returned is not populated. Matt and I could not figure out why. Will defer
- // further looking into it for after devconf.
- // The libpod function for this has no errors so we are kind of in a tough
- // spot here. Logging the errors for now.
- reply, err := iopodman.ContainerConfig().Call(r.Conn, name)
- if err != nil {
- logrus.Error("call to container.config failed")
- }
- data := libpod.ContainerConfig{}
- if err := json.Unmarshal([]byte(reply), &data); err != nil {
- logrus.Error("failed to unmarshal container inspect data")
- }
- return &data
-
-}
-
-// ContainerState returns the "state" of the container.
-func (r *LocalRuntime) ContainerState(name string) (*libpod.ContainerState, error) { // no-lint
- reply, err := iopodman.ContainerStateData().Call(r.Conn, name)
- if err != nil {
- return nil, err
- }
- data := libpod.ContainerState{}
- if err := json.Unmarshal([]byte(reply), &data); err != nil {
- return nil, err
- }
- return &data, err
-
-}
-
-// Spec obtains the container spec.
-func (r *LocalRuntime) Spec(name string) (*specs.Spec, error) {
- reply, err := iopodman.Spec().Call(r.Conn, name)
- if err != nil {
- return nil, err
- }
- data := specs.Spec{}
- if err := json.Unmarshal([]byte(reply), &data); err != nil {
- return nil, err
- }
- return &data, nil
-}
-
-// LookupContainers is a wrapper for LookupContainer
-func (r *LocalRuntime) LookupContainers(idsOrNames []string) ([]*Container, error) {
- var containers []*Container
- for _, name := range idsOrNames {
- ctr, err := r.LookupContainer(name)
- if err != nil {
- return nil, err
- }
- containers = append(containers, ctr)
- }
- return containers, nil
-}
-
-// LookupContainer gets basic information about container over a varlink
-// connection and then translates it to a *Container
-func (r *LocalRuntime) LookupContainer(idOrName string) (*Container, error) {
- state, err := r.ContainerState(idOrName)
- if err != nil {
- return nil, err
- }
- config := r.Config(idOrName)
- return &Container{
- remoteContainer{
- r,
- config,
- state,
- },
- }, nil
-}
-
-// GetAllContainers returns all containers in a slice
-func (r *LocalRuntime) GetAllContainers() ([]*Container, error) {
- var containers []*Container
- ctrs, err := iopodman.GetContainersByContext().Call(r.Conn, true, false, []string{})
- if err != nil {
- return nil, err
- }
- for _, ctr := range ctrs {
- container, err := r.LookupContainer(ctr)
- if err != nil {
- return nil, err
- }
- containers = append(containers, container)
- }
- return containers, nil
-}
-
-func (r *LocalRuntime) LookupContainersWithStatus(filters []string) ([]*Container, error) {
- var containers []*Container
- ctrs, err := iopodman.GetContainersByStatus().Call(r.Conn, filters)
- if err != nil {
- return nil, err
- }
- // This is not performance savvy; if this turns out to be a problematic series of lookups, we need to
- // create a new endpoint to speed things up
- for _, ctr := range ctrs {
- container, err := r.LookupContainer(ctr.Id)
- if err != nil {
- return nil, err
- }
- containers = append(containers, container)
- }
- return containers, nil
-}
-
-func (r *LocalRuntime) GetLatestContainer() (*Container, error) {
- reply, err := iopodman.GetContainersByContext().Call(r.Conn, false, true, nil)
- if err != nil {
- return nil, err
- }
- if len(reply) > 0 {
- return r.LookupContainer(reply[0])
- }
- return nil, errors.New("no containers exist")
-}
-
-// GetArtifact returns a container's artifacts
-func (c *Container) GetArtifact(name string) ([]byte, error) {
- var data []byte
- reply, err := iopodman.ContainerArtifacts().Call(c.Runtime.Conn, c.ID(), name)
- if err != nil {
- return nil, err
- }
- if err := json.Unmarshal([]byte(reply), &data); err != nil {
- return nil, err
- }
- return data, err
-}
-
-// Config returns a container's Config ... same as ctr.Config()
-func (c *Container) Config() *libpod.ContainerConfig {
- if c.config != nil {
- return c.config
- }
- return c.Runtime.Config(c.ID())
-}
-
-// Name returns the name of the container
-func (c *Container) Name() string {
- return c.config.Name
-}
-
-// StopContainers stops requested containers using varlink.
-// Returns the list of stopped container ids, map of failed to stop container ids + errors, or any non-container error
-func (r *LocalRuntime) StopContainers(ctx context.Context, cli *cliconfig.StopValues) ([]string, map[string]error, error) {
- var (
- ok = []string{}
- failures = map[string]error{}
- )
-
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, cli.All, cli.Latest, cli.InputArgs)
- if err != nil {
- return ok, failures, TranslateError(err)
- }
-
- for _, id := range ids {
- if _, err := iopodman.StopContainer().Call(r.Conn, id, int64(cli.Timeout)); err != nil {
- transError := TranslateError(err)
- if errors.Cause(transError) == define.ErrCtrStopped {
- ok = append(ok, id)
- continue
- }
- if errors.Cause(transError) == define.ErrCtrStateInvalid && cli.All {
- ok = append(ok, id)
- continue
- }
- failures[id] = err
- } else {
- // We should be using ID here because in varlink, only successful returns
- // include the string id
- ok = append(ok, id)
- }
- }
- return ok, failures, nil
-}
-
-// InitContainers initializes container(s) based on Varlink.
-// It returns a list of successful ID(s), a map of failed container ID to error,
-// or an error if a more general error occurred.
-func (r *LocalRuntime) InitContainers(ctx context.Context, cli *cliconfig.InitValues) ([]string, map[string]error, error) {
- var (
- ok = []string{}
- failures = map[string]error{}
- )
-
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, cli.All, cli.Latest, cli.InputArgs)
- if err != nil {
- return nil, nil, err
- }
-
- for _, id := range ids {
- initialized, err := iopodman.InitContainer().Call(r.Conn, id)
- if err != nil {
- if cli.All {
- switch err.(type) {
- case *iopodman.InvalidState:
- ok = append(ok, initialized)
- default:
- failures[id] = err
- }
- } else {
- failures[id] = err
- }
- } else {
- ok = append(ok, initialized)
- }
- }
- return ok, failures, nil
-}
-
-// KillContainers sends signal to container(s) based on varlink.
-// Returns list of successful id(s), map of failed id(s) + error, or error not from container
-func (r *LocalRuntime) KillContainers(ctx context.Context, cli *cliconfig.KillValues, signal syscall.Signal) ([]string, map[string]error, error) {
- var (
- ok = []string{}
- failures = map[string]error{}
- )
-
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, cli.All, cli.Latest, cli.InputArgs)
- if err != nil {
- return ok, failures, err
- }
-
- for _, id := range ids {
- killed, err := iopodman.KillContainer().Call(r.Conn, id, int64(signal))
- if err != nil {
- failures[id] = err
- } else {
- ok = append(ok, killed)
- }
- }
- return ok, failures, nil
-}
-
-// RemoveContainer removes container(s) based on varlink inputs.
-func (r *LocalRuntime) RemoveContainers(ctx context.Context, cli *cliconfig.RmValues) ([]string, map[string]error, error) {
- var (
- ok = []string{}
- failures = map[string]error{}
- )
-
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, cli.All, cli.Latest, cli.InputArgs)
- if err != nil {
- // Failed to get containers. If force is specified, get the containers ID
- // and evict them
- if !cli.Force {
- return nil, nil, TranslateError(err)
- }
-
- for _, ctr := range cli.InputArgs {
- logrus.Debugf("Evicting container %q", ctr)
- id, err := iopodman.EvictContainer().Call(r.Conn, ctr, cli.Volumes)
- if err != nil {
- failures[ctr] = errors.Wrapf(err, "Failed to evict container: %q", id)
- continue
- }
- ok = append(ok, string(id))
- }
- return ok, failures, nil
- }
-
- for _, id := range ids {
- _, err := iopodman.RemoveContainer().Call(r.Conn, id, cli.Force, cli.Volumes)
- if err != nil {
- failures[id] = err
- } else {
- ok = append(ok, id)
- }
- }
- return ok, failures, nil
-}
-
-// UmountRootFilesystems umounts container(s) root filesystems based on varlink inputs
-func (r *LocalRuntime) UmountRootFilesystems(ctx context.Context, cli *cliconfig.UmountValues) ([]string, map[string]error, error) {
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, cli.All, cli.Latest, cli.InputArgs)
- if err != nil {
- return nil, nil, err
- }
-
- var (
- ok = []string{}
- failures = map[string]error{}
- )
-
- for _, id := range ids {
- err := iopodman.UnmountContainer().Call(r.Conn, id, cli.Force)
- if err != nil {
- failures[id] = err
- } else {
- ok = append(ok, id)
- }
- }
- return ok, failures, nil
-}
-
-// WaitOnContainers waits for all given container(s) to stop.
-// interval is currently ignored.
-func (r *LocalRuntime) WaitOnContainers(ctx context.Context, cli *cliconfig.WaitValues, interval time.Duration) ([]string, map[string]error, error) {
- var (
- ok = []string{}
- failures = map[string]error{}
- )
-
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, false, cli.Latest, cli.InputArgs)
- if err != nil {
- return ok, failures, err
- }
-
- for _, id := range ids {
- stopped, err := iopodman.WaitContainer().Call(r.Conn, id, int64(interval))
- if err != nil {
- failures[id] = err
- } else {
- ok = append(ok, strconv.FormatInt(stopped, 10))
- }
- }
- return ok, failures, nil
-}
-
-// BatchContainerOp is wrapper func to mimic shared's function with a similar name meant for libpod
-func BatchContainerOp(ctr *Container, opts shared.PsOptions) (shared.BatchContainerStruct, error) {
- // TODO If pod ps ever shows container's sizes, re-enable this code; otherwise it isn't needed
- // and would be a perf hit
- // data, err := ctr.Inspect(true)
- // if err != nil {
- // return shared.BatchContainerStruct{}, err
- // }
- //
- // size := new(shared.ContainerSize)
- // size.RootFsSize = data.SizeRootFs
- // size.RwSize = data.SizeRw
-
- bcs := shared.BatchContainerStruct{
- ConConfig: ctr.config,
- ConState: ctr.state.State,
- ExitCode: ctr.state.ExitCode,
- Pid: ctr.state.PID,
- StartedTime: ctr.state.StartedTime,
- ExitedTime: ctr.state.FinishedTime,
- // Size: size,
- }
- return bcs, nil
-}
-
-// Log one or more containers over a varlink connection
-func (r *LocalRuntime) Log(c *cliconfig.LogsValues, options *logs.LogOptions) error {
- // GetContainersLogs
- reply, err := iopodman.GetContainersLogs().Send(r.Conn, uint64(varlink.More), c.InputArgs, c.Follow, c.Latest, options.Since.Format(time.RFC3339Nano), int64(c.Tail), c.Timestamps)
- if err != nil {
- return errors.Wrapf(err, "failed to get container logs")
- }
- if len(c.InputArgs) > 1 {
- options.Multi = true
- }
- for {
- log, flags, err := reply()
- if err != nil {
- return err
- }
- if log.Time == "" && log.Msg == "" {
- // We got a blank log line which can signal end of stream
- break
- }
- lTime, err := time.Parse(time.RFC3339Nano, log.Time)
- if err != nil {
- return errors.Wrapf(err, "unable to parse time of log %s", log.Time)
- }
- logLine := logs.LogLine{
- Device: log.Device,
- ParseLogType: log.ParseLogType,
- Time: lTime,
- Msg: log.Msg,
- CID: log.Cid,
- }
- fmt.Println(logLine.String(options))
- if flags&varlink.Continues == 0 {
- break
- }
- }
- return nil
-}
-
-// CreateContainer creates a container from the cli over varlink
-func (r *LocalRuntime) CreateContainer(ctx context.Context, c *cliconfig.CreateValues) (string, error) {
- results := shared.NewIntermediateLayer(&c.PodmanCommand, true)
- return iopodman.CreateContainer().Call(r.Conn, results.MakeVarlink())
-}
-
-// Run creates a container overvarlink and then starts it
-func (r *LocalRuntime) Run(ctx context.Context, c *cliconfig.RunValues, exitCode int) (int, error) {
- // TODO the exit codes for run need to be figured out for remote connections
- results := shared.NewIntermediateLayer(&c.PodmanCommand, true)
- cid, err := iopodman.CreateContainer().Call(r.Conn, results.MakeVarlink())
- if err != nil {
- return exitCode, err
- }
- if c.Bool("detach") {
- if _, err := iopodman.StartContainer().Call(r.Conn, cid); err != nil {
- return exitCode, err
- }
- fmt.Println(cid)
- return 0, nil
- }
- inputStream := os.Stdin
- // If -i is not set, clear stdin
- if !c.Bool("interactive") {
- inputStream = nil
- }
- exitChan, errChan, err := r.attach(ctx, inputStream, os.Stdout, cid, true, c.String("detach-keys"))
- if err != nil {
- return exitCode, err
- }
- exitCode = <-exitChan
- finalError := <-errChan
- return exitCode, finalError
-}
-
-func ReadExitFile(runtimeTmp, ctrID string) (int, error) {
- return 0, define.ErrNotImplemented
-}
-
-// Ps lists containers based on criteria from user
-func (r *LocalRuntime) Ps(c *cliconfig.PsValues, opts shared.PsOptions) ([]shared.PsContainerOutput, error) {
- var psContainers []shared.PsContainerOutput
- last := int64(c.Last)
- PsOpts := iopodman.PsOpts{
- All: c.All,
- Filters: &c.Filter,
- Last: &last,
- Latest: &c.Latest,
- NoTrunc: &c.NoTrunct,
- Pod: &c.Pod,
- Quiet: &c.Quiet,
- Size: &c.Size,
- Sort: &c.Sort,
- Sync: &c.Sync,
- }
- containers, err := iopodman.Ps().Call(r.Conn, PsOpts)
- if err != nil {
- return nil, err
- }
- for _, ctr := range containers {
- createdAt, err := time.Parse(time.RFC3339Nano, ctr.CreatedAt)
- if err != nil {
- return nil, err
- }
- exitedAt, err := time.Parse(time.RFC3339Nano, ctr.ExitedAt)
- if err != nil {
- return nil, err
- }
- startedAt, err := time.Parse(time.RFC3339Nano, ctr.StartedAt)
- if err != nil {
- return nil, err
- }
- containerSize := shared.ContainerSize{
- RootFsSize: ctr.RootFsSize,
- RwSize: ctr.RwSize,
- }
- state, err := define.StringToContainerStatus(ctr.State)
- if err != nil {
- return nil, err
- }
- psc := shared.PsContainerOutput{
- ID: ctr.Id,
- Image: ctr.Image,
- Command: ctr.Command,
- Created: ctr.Created,
- Ports: ctr.Ports,
- Names: ctr.Names,
- IsInfra: ctr.IsInfra,
- Status: ctr.Status,
- State: state,
- Pid: int(ctr.PidNum),
- Size: &containerSize,
- Pod: ctr.Pod,
- CreatedAt: createdAt,
- ExitedAt: exitedAt,
- StartedAt: startedAt,
- Labels: ctr.Labels,
- PID: ctr.NsPid,
- Cgroup: ctr.Cgroup,
- IPC: ctr.Ipc,
- MNT: ctr.Mnt,
- NET: ctr.Net,
- PIDNS: ctr.PidNs,
- User: ctr.User,
- UTS: ctr.Uts,
- Mounts: ctr.Mounts,
- }
- psContainers = append(psContainers, psc)
- }
- return psContainers, nil
-}
-
-// Attach to a remote terminal
-func (r *LocalRuntime) Attach(ctx context.Context, c *cliconfig.AttachValues) error {
- ctr, err := r.LookupContainer(c.InputArgs[0])
- if err != nil {
- return nil
- }
- if ctr.state.State != define.ContainerStateRunning {
- return errors.New("you can only attach to running containers")
- }
- inputStream := os.Stdin
- if c.NoStdin {
- inputStream, err = os.Open(os.DevNull)
- if err != nil {
- return err
- }
- }
- _, errChan, err := r.attach(ctx, inputStream, os.Stdout, c.InputArgs[0], false, c.DetachKeys)
- if err != nil {
- return err
- }
- return <-errChan
-}
-
-// Checkpoint one or more containers
-func (r *LocalRuntime) Checkpoint(c *cliconfig.CheckpointValues) error {
- if c.Export != "" {
- return errors.New("the remote client does not support exporting checkpoints")
- }
- if c.IgnoreRootfs {
- return errors.New("the remote client does not support --ignore-rootfs")
- }
-
- var lastError error
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, c.All, c.Latest, c.InputArgs)
- if err != nil {
- return err
- }
- if c.All {
- // We don't have a great way to get all the running containers, so need to get all and then
- // check status on them bc checkpoint considers checkpointing a stopped container an error
- var runningIds []string
- for _, id := range ids {
- ctr, err := r.LookupContainer(id)
- if err != nil {
- return err
- }
- if ctr.state.State == define.ContainerStateRunning {
- runningIds = append(runningIds, id)
- }
- }
- ids = runningIds
- }
-
- for _, id := range ids {
- if _, err := iopodman.ContainerCheckpoint().Call(r.Conn, id, c.Keep, c.Keep, c.TcpEstablished); err != nil {
- if lastError != nil {
- fmt.Fprintln(os.Stderr, lastError)
- }
- lastError = errors.Wrapf(err, "failed to checkpoint container %v", id)
- } else {
- fmt.Println(id)
- }
- }
- return lastError
-}
-
-// Restore one or more containers
-func (r *LocalRuntime) Restore(ctx context.Context, c *cliconfig.RestoreValues) error {
- if c.Import != "" {
- return errors.New("the remote client does not support importing checkpoints")
- }
- if c.IgnoreRootfs {
- return errors.New("the remote client does not support --ignore-rootfs")
- }
-
- var lastError error
- ids, err := iopodman.GetContainersByContext().Call(r.Conn, c.All, c.Latest, c.InputArgs)
- if err != nil {
- return err
- }
- if c.All {
- // We don't have a great way to get all the exited containers, so need to get all and then
- // check status on them bc checkpoint considers restoring a running container an error
- var exitedIDs []string
- for _, id := range ids {
- ctr, err := r.LookupContainer(id)
- if err != nil {
- return err
- }
- if ctr.state.State != define.ContainerStateRunning {
- exitedIDs = append(exitedIDs, id)
- }
- }
- ids = exitedIDs
- }
-
- for _, id := range ids {
- if _, err := iopodman.ContainerRestore().Call(r.Conn, id, c.Keep, c.TcpEstablished); err != nil {
- if lastError != nil {
- fmt.Fprintln(os.Stderr, lastError)
- }
- lastError = errors.Wrapf(err, "failed to restore container %v", id)
- } else {
- fmt.Println(id)
- }
- }
- return lastError
-}
-
-// Start starts an already created container
-func (r *LocalRuntime) Start(ctx context.Context, c *cliconfig.StartValues, sigProxy bool) (int, error) {
- var (
- finalErr error
- exitCode = define.ExecErrorCodeGeneric
- )
- // TODO Figure out how to deal with exit codes
- inputStream := os.Stdin
- if !c.Interactive {
- inputStream = nil
- }
-
- containerIDs, err := iopodman.GetContainersByContext().Call(r.Conn, false, c.Latest, c.InputArgs)
- if err != nil {
- return exitCode, err
- }
- if len(containerIDs) < 1 {
- return exitCode, errors.New("failed to find containers to start")
- }
- // start.go makes sure that if attach, there can be only one ctr
- if c.Attach {
- exitChan, errChan, err := r.attach(ctx, inputStream, os.Stdout, containerIDs[0], true, c.DetachKeys)
- if err != nil {
- return exitCode, nil
- }
- exitCode := <-exitChan
- err = <-errChan
- return exitCode, err
- }
-
- // TODO the notion of starting a pod container and its deps still needs to be worked through
- // Everything else is detached
- for _, cid := range containerIDs {
- reply, err := iopodman.StartContainer().Call(r.Conn, cid)
- if err != nil {
- if finalErr != nil {
- fmt.Println(err)
- }
- finalErr = err
- } else {
- fmt.Println(reply)
- }
- }
- return exitCode, finalErr
-}
-
-func (r *LocalRuntime) attach(ctx context.Context, stdin, stdout *os.File, cid string, start bool, detachKeys string) (chan int, chan error, error) {
- var (
- oldTermState *term.State
- )
- spec, err := r.Spec(cid)
- if err != nil {
- return nil, nil, err
- }
- resize := make(chan remotecommand.TerminalSize, 5)
- haveTerminal := terminal.IsTerminal(int(os.Stdin.Fd()))
-
- // Check if we are attached to a terminal. If we are, generate resize
- // events, and set the terminal to raw mode
- if haveTerminal && spec.Process.Terminal {
- cancel, oldTermState, err := handleTerminalAttach(ctx, resize)
- if err != nil {
- return nil, nil, err
- }
- defer cancel()
- defer restoreTerminal(oldTermState)
-
- logrus.SetFormatter(&RawTtyFormatter{})
- term.SetRawTerminal(os.Stdin.Fd())
- }
-
- reply, err := iopodman.Attach().Send(r.Conn, varlink.Upgrade, cid, detachKeys, start)
- if err != nil {
- restoreTerminal(oldTermState)
- return nil, nil, err
- }
-
- // See if the server accepts the upgraded connection or returns an error
- _, err = reply()
-
- if err != nil {
- restoreTerminal(oldTermState)
- return nil, nil, err
- }
-
- ecChan := make(chan int, 1)
- errChan := configureVarlinkAttachStdio(r.Conn.Reader, r.Conn.Writer, stdin, stdout, oldTermState, resize, ecChan)
- return ecChan, errChan, nil
-}
-
-// PauseContainers pauses container(s) based on CLI inputs.
-func (r *LocalRuntime) PauseContainers(ctx context.Context, cli *cliconfig.PauseValues) ([]string, map[string]error, error) {
- var (
- ok []string
- failures = map[string]error{}
- ctrs []*Container
- err error
- )
-
- if cli.All {
- filters := []string{define.ContainerStateRunning.String()}
- ctrs, err = r.LookupContainersWithStatus(filters)
- } else {
- ctrs, err = r.LookupContainers(cli.InputArgs)
- }
- if err != nil {
- return ok, failures, err
- }
-
- for _, c := range ctrs {
- c := c
- err := c.Pause()
- if err != nil {
- failures[c.ID()] = err
- } else {
- ok = append(ok, c.ID())
- }
- }
- return ok, failures, nil
-}
-
-// UnpauseContainers unpauses containers based on input
-func (r *LocalRuntime) UnpauseContainers(ctx context.Context, cli *cliconfig.UnpauseValues) ([]string, map[string]error, error) {
- var (
- ok = []string{}
- failures = map[string]error{}
- ctrs []*Container
- err error
- )
-
- maxWorkers := shared.DefaultPoolSize("unpause")
- if cli.GlobalIsSet("max-workers") {
- maxWorkers = cli.GlobalFlags.MaxWorks
- }
- logrus.Debugf("Setting maximum rm workers to %d", maxWorkers)
-
- if cli.All {
- filters := []string{define.ContainerStatePaused.String()}
- ctrs, err = r.LookupContainersWithStatus(filters)
- } else {
- ctrs, err = r.LookupContainers(cli.InputArgs)
- }
- if err != nil {
- return ok, failures, err
- }
- for _, c := range ctrs {
- c := c
- err := c.Unpause()
- if err != nil {
- failures[c.ID()] = err
- } else {
- ok = append(ok, c.ID())
- }
- }
- return ok, failures, nil
-}
-
-// Restart restarts a container over varlink
-func (r *LocalRuntime) Restart(ctx context.Context, c *cliconfig.RestartValues) ([]string, map[string]error, error) {
- var (
- containers []*Container
- restartContainers []*Container
- err error
- ok = []string{}
- failures = map[string]error{}
- )
- useTimeout := c.Flag("timeout").Changed || c.Flag("time").Changed
- inputTimeout := c.Timeout
-
- if c.Latest {
- lastCtr, err := r.GetLatestContainer()
- if err != nil {
- return nil, nil, errors.Wrapf(err, "unable to get latest container")
- }
- restartContainers = append(restartContainers, lastCtr)
- } else if c.Running {
- containers, err = r.LookupContainersWithStatus([]string{define.ContainerStateRunning.String()})
- if err != nil {
- return nil, nil, err
- }
- restartContainers = append(restartContainers, containers...)
- } else if c.All {
- containers, err = r.GetAllContainers()
- if err != nil {
- return nil, nil, err
- }
- restartContainers = append(restartContainers, containers...)
- } else {
- for _, id := range c.InputArgs {
- ctr, err := r.LookupContainer(id)
- if err != nil {
- return nil, nil, err
- }
- restartContainers = append(restartContainers, ctr)
- }
- }
-
- for _, c := range restartContainers {
- c := c
- timeout := c.config.StopTimeout
- if useTimeout {
- timeout = inputTimeout
- }
- err := c.Restart(int64(timeout))
- if err != nil {
- failures[c.ID()] = err
- } else {
- ok = append(ok, c.ID())
- }
- }
- return ok, failures, nil
-}
-
-// Top display the running processes of a container
-func (r *LocalRuntime) Top(cli *cliconfig.TopValues) ([]string, error) {
- var (
- ctr *Container
- err error
- descriptors []string
- )
- if cli.Latest {
- ctr, err = r.GetLatestContainer()
- descriptors = cli.InputArgs
- } else {
- ctr, err = r.LookupContainer(cli.InputArgs[0])
- descriptors = cli.InputArgs[1:]
- }
- if err != nil {
- return nil, err
- }
- return iopodman.Top().Call(r.Conn, ctr.ID(), descriptors)
-}
-
-// Prune removes stopped containers
-func (r *LocalRuntime) Prune(ctx context.Context, maxWorkers int, filter []string) ([]string, map[string]error, error) {
-
- var (
- ok = []string{}
- failures = map[string]error{}
- ctrs []*Container
- err error
- )
- logrus.Debugf("Setting maximum rm workers to %d", maxWorkers)
-
- filters := []string{define.ContainerStateExited.String()}
- ctrs, err = r.LookupContainersWithStatus(filters)
- if err != nil {
- return ok, failures, err
- }
- for _, c := range ctrs {
- c := c
- _, err := iopodman.RemoveContainer().Call(r.Conn, c.ID(), false, false)
- if err != nil {
- failures[c.ID()] = err
- } else {
- ok = append(ok, c.ID())
- }
- }
- return ok, failures, nil
-}
-
-// Cleanup any leftovers bits of stopped containers
-func (r *LocalRuntime) CleanupContainers(ctx context.Context, cli *cliconfig.CleanupValues) ([]string, map[string]error, error) {
- return nil, nil, errors.New("container cleanup not supported for remote clients")
-}
-
-// Port displays port information about existing containers
-func (r *LocalRuntime) Port(c *cliconfig.PortValues) ([]*Container, error) {
- var (
- containers []*Container
- err error
- )
- // This one is a bit odd because when all is used, we only use running containers.
- if !c.All {
- containers, err = r.GetContainersByContext(false, c.Latest, c.InputArgs)
- } else {
- // we need to only use running containers if all
- filters := []string{define.ContainerStateRunning.String()}
- containers, err = r.LookupContainersWithStatus(filters)
- }
- if err != nil {
- return nil, err
- }
- return containers, nil
-}
-
-// GenerateSystemd creates a systemd until for a container
-func (r *LocalRuntime) GenerateSystemd(c *cliconfig.GenerateSystemdValues) (string, error) {
- return "", errors.New("systemd generation not supported for remote clients")
-}
-
-// GetNamespaces returns namespace information about a container for PS
-func (r *LocalRuntime) GetNamespaces(container shared.PsContainerOutput) *shared.Namespace {
- ns := shared.Namespace{
- PID: container.PID,
- Cgroup: container.Cgroup,
- IPC: container.IPC,
- MNT: container.MNT,
- NET: container.NET,
- PIDNS: container.PIDNS,
- User: container.User,
- UTS: container.UTS,
- }
- return &ns
-}
-
-// Commit creates a local image from a container
-func (r *LocalRuntime) Commit(ctx context.Context, c *cliconfig.CommitValues, container, imageName string) (string, error) {
- var iid string
- reply, err := iopodman.Commit().Send(r.Conn, varlink.More, container, imageName, c.Change, c.Author, c.Message, c.Pause, c.Format)
- if err != nil {
- return "", err
- }
- for {
- responses, flags, err := reply()
- if err != nil {
- return "", err
- }
- for _, line := range responses.Logs {
- fmt.Fprintln(os.Stderr, line)
- }
- iid = responses.Id
- if flags&varlink.Continues == 0 {
- break
- }
- }
- return iid, nil
-}
-
-// ExecContainer executes a command in the container
-func (r *LocalRuntime) ExecContainer(ctx context.Context, cli *cliconfig.ExecValues) (int, error) {
- var (
- oldTermState *term.State
- ec = define.ExecErrorCodeGeneric
- )
- // default invalid command exit code
- // Validate given environment variables
- cliEnv, err := envLib.ParseSlice(cli.Env)
- if err != nil {
- return 0, errors.Wrap(err, "error parsing environment variables")
- }
- envs := envLib.Slice(cliEnv)
-
- resize := make(chan remotecommand.TerminalSize, 5)
- haveTerminal := terminal.IsTerminal(int(os.Stdin.Fd()))
-
- // Check if we are attached to a terminal. If we are, generate resize
- // events, and set the terminal to raw mode
- if haveTerminal && cli.Tty {
- cancel, oldTermState, err := handleTerminalAttach(ctx, resize)
- if err != nil {
- return ec, err
- }
- defer cancel()
- defer restoreTerminal(oldTermState)
-
- logrus.SetFormatter(&RawTtyFormatter{})
- term.SetRawTerminal(os.Stdin.Fd())
- }
-
- opts := iopodman.ExecOpts{
- Name: cli.InputArgs[0],
- Tty: cli.Tty,
- Privileged: cli.Privileged,
- Cmd: cli.InputArgs[1:],
- User: &cli.User,
- Workdir: &cli.Workdir,
- Env: &envs,
- DetachKeys: &cli.DetachKeys,
- }
-
- inputStream := os.Stdin
- if !cli.Interactive {
- inputStream = nil
- }
-
- reply, err := iopodman.ExecContainer().Send(r.Conn, varlink.Upgrade, opts)
- if err != nil {
- return ec, errors.Wrapf(err, "Exec failed to contact service for %s", cli.InputArgs)
- }
-
- _, err = reply()
- if err != nil {
- return ec, errors.Wrapf(err, "Exec operation failed for %s", cli.InputArgs)
- }
- ecChan := make(chan int, 1)
- errChan := configureVarlinkAttachStdio(r.Conn.Reader, r.Conn.Writer, inputStream, os.Stdout, oldTermState, resize, ecChan)
-
- ec = <-ecChan
- err = <-errChan
-
- return ec, err
-}
-
-func configureVarlinkAttachStdio(reader *bufio.Reader, writer *bufio.Writer, stdin *os.File, stdout *os.File, oldTermState *term.State, resize chan remotecommand.TerminalSize, ecChan chan int) chan error {
- errChan := make(chan error, 1)
- // These are the special writers that encode input from the client.
- varlinkStdinWriter := virtwriter.NewVirtWriteCloser(writer, virtwriter.ToStdin)
- varlinkResizeWriter := virtwriter.NewVirtWriteCloser(writer, virtwriter.TerminalResize)
- varlinkHangupWriter := virtwriter.NewVirtWriteCloser(writer, virtwriter.HangUpFromClient)
-
- go func() {
- // Read from the wire and direct to stdout or stderr
- err := virtwriter.Reader(reader, stdout, os.Stderr, nil, nil, ecChan)
- defer restoreTerminal(oldTermState)
- sendGenericError(ecChan)
- errChan <- err
- }()
-
- go func() {
- for termResize := range resize {
- b, err := json.Marshal(termResize)
- if err != nil {
- defer restoreTerminal(oldTermState)
- sendGenericError(ecChan)
- errChan <- err
- }
- _, err = varlinkResizeWriter.Write(b)
- if err != nil {
- defer restoreTerminal(oldTermState)
- sendGenericError(ecChan)
- errChan <- err
- }
- }
- }()
- if stdin != nil {
- // Takes stdinput and sends it over the wire after being encoded
- go func() {
- if _, err := io.Copy(varlinkStdinWriter, stdin); err != nil {
- defer restoreTerminal(oldTermState)
- sendGenericError(ecChan)
- errChan <- err
- }
- _, err := varlinkHangupWriter.Write([]byte("EOF"))
- if err != nil {
- logrus.Errorf("unable to notify server to hangup: %q", err)
- }
- err = varlinkStdinWriter.Close()
- errChan <- err
- }()
- }
- return errChan
-}
-
-func sendGenericError(ecChan chan int) {
- if ecChan != nil {
- ecChan <- define.ExecErrorCodeGeneric
- }
-}