summaryrefslogtreecommitdiff
path: root/libpod/runtime_pod_linux.go
diff options
context:
space:
mode:
Diffstat (limited to 'libpod/runtime_pod_linux.go')
-rw-r--r--libpod/runtime_pod_linux.go249
1 files changed, 249 insertions, 0 deletions
diff --git a/libpod/runtime_pod_linux.go b/libpod/runtime_pod_linux.go
new file mode 100644
index 000000000..0edac9dbe
--- /dev/null
+++ b/libpod/runtime_pod_linux.go
@@ -0,0 +1,249 @@
+// +build linux
+
+package libpod
+
+import (
+ "context"
+ "path"
+ "path/filepath"
+ "strings"
+
+ "github.com/containerd/cgroups"
+ "github.com/pkg/errors"
+ "github.com/sirupsen/logrus"
+)
+
+// NewPod makes a new, empty pod
+func (r *Runtime) NewPod(options ...PodCreateOption) (*Pod, error) {
+ r.lock.Lock()
+ defer r.lock.Unlock()
+
+ if !r.valid {
+ return nil, ErrRuntimeStopped
+ }
+
+ pod, err := newPod(r.lockDir, r)
+ if err != nil {
+ return nil, errors.Wrapf(err, "error creating pod")
+ }
+
+ for _, option := range options {
+ if err := option(pod); err != nil {
+ return nil, errors.Wrapf(err, "error running pod create option")
+ }
+ }
+
+ if pod.config.Name == "" {
+ name, err := r.generateName()
+ if err != nil {
+ return nil, err
+ }
+ pod.config.Name = name
+ }
+
+ pod.valid = true
+
+ // Check CGroup parent sanity, and set it if it was not set
+ switch r.config.CgroupManager {
+ case CgroupfsCgroupsManager:
+ if pod.config.CgroupParent == "" {
+ pod.config.CgroupParent = CgroupfsDefaultCgroupParent
+ } else if strings.HasSuffix(path.Base(pod.config.CgroupParent), ".slice") {
+ return nil, errors.Wrapf(ErrInvalidArg, "systemd slice received as cgroup parent when using cgroupfs")
+ }
+ // Creating CGroup path is currently a NOOP until proper systemd
+ // cgroup management is merged
+ case SystemdCgroupsManager:
+ if pod.config.CgroupParent == "" {
+ pod.config.CgroupParent = SystemdDefaultCgroupParent
+ } else if len(pod.config.CgroupParent) < 6 || !strings.HasSuffix(path.Base(pod.config.CgroupParent), ".slice") {
+ return nil, errors.Wrapf(ErrInvalidArg, "did not receive systemd slice as cgroup parent when using systemd to manage cgroups")
+ }
+ // If we are set to use pod cgroups, set the cgroup parent that
+ // all containers in the pod will share
+ // No need to create it with cgroupfs - the first container to
+ // launch should do it for us
+ if pod.config.UsePodCgroup {
+ pod.state.CgroupPath = filepath.Join(pod.config.CgroupParent, pod.ID())
+ }
+ default:
+ return nil, errors.Wrapf(ErrInvalidArg, "unsupported CGroup manager: %s - cannot validate cgroup parent", r.config.CgroupManager)
+ }
+
+ if err := r.state.AddPod(pod); err != nil {
+ return nil, errors.Wrapf(err, "error adding pod to state")
+ }
+
+ return nil, ErrNotImplemented
+}
+
+// RemovePod removes a pod
+// If removeCtrs is specified, containers will be removed
+// Otherwise, a pod that is not empty will return an error and not be removed
+// If force is specified with removeCtrs, all containers will be stopped before
+// being removed
+// Otherwise, the pod will not be removed if any containers are running
+func (r *Runtime) RemovePod(ctx context.Context, p *Pod, removeCtrs, force bool) error {
+ r.lock.Lock()
+ defer r.lock.Unlock()
+
+ if !r.valid {
+ return ErrRuntimeStopped
+ }
+
+ p.lock.Lock()
+ defer p.lock.Unlock()
+
+ if !p.valid {
+ return ErrPodRemoved
+ }
+
+ ctrs, err := r.state.PodContainers(p)
+ if err != nil {
+ return err
+ }
+
+ numCtrs := len(ctrs)
+
+ if !removeCtrs && numCtrs > 0 {
+ return errors.Wrapf(ErrCtrExists, "pod %s contains containers and cannot be removed", p.ID())
+ }
+
+ // Go through and lock all containers so we can operate on them all at once
+ dependencies := make(map[string][]string)
+ for _, ctr := range ctrs {
+ ctr.lock.Lock()
+ defer ctr.lock.Unlock()
+
+ // Sync all containers
+ if err := ctr.syncContainer(); err != nil {
+ return err
+ }
+
+ // Check if the container is in a good state to be removed
+ if ctr.state.State == ContainerStatePaused {
+ return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains paused container %s, cannot remove", p.ID(), ctr.ID())
+ }
+
+ if ctr.state.State == ContainerStateUnknown {
+ return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains container %s with invalid state", p.ID(), ctr.ID())
+ }
+
+ // If the container is running and force is not set we can't do anything
+ if ctr.state.State == ContainerStateRunning && !force {
+ return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains container %s which is running", p.ID(), ctr.ID())
+ }
+
+ // If the container has active exec sessions and force is not set we can't do anything
+ if len(ctr.state.ExecSessions) != 0 && !force {
+ return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains container %s which has active exec sessions", p.ID(), ctr.ID())
+ }
+
+ deps, err := r.state.ContainerInUse(ctr)
+ if err != nil {
+ return err
+ }
+ dependencies[ctr.ID()] = deps
+ }
+
+ // Check if containers have dependencies
+ // If they do, and the dependencies are not in the pod, error
+ for ctr, deps := range dependencies {
+ for _, dep := range deps {
+ if _, ok := dependencies[dep]; !ok {
+ return errors.Wrapf(ErrCtrExists, "container %s depends on container %s not in pod %s", ctr, dep, p.ID())
+ }
+ }
+ }
+
+ // First loop through all containers and stop them
+ // Do not remove in this loop to ensure that we don't remove unless all
+ // containers are in a good state
+ if force {
+ for _, ctr := range ctrs {
+ // If force is set and the container is running, stop it now
+ if ctr.state.State == ContainerStateRunning {
+ if err := r.ociRuntime.stopContainer(ctr, ctr.StopTimeout()); err != nil {
+ return errors.Wrapf(err, "error stopping container %s to remove pod %s", ctr.ID(), p.ID())
+ }
+
+ // Sync again to pick up stopped state
+ if err := ctr.syncContainer(); err != nil {
+ return err
+ }
+ }
+ // If the container has active exec sessions, stop them now
+ if len(ctr.state.ExecSessions) != 0 {
+ if err := r.ociRuntime.execStopContainer(ctr, ctr.StopTimeout()); err != nil {
+ return err
+ }
+ }
+ }
+ }
+
+ // Start removing containers
+ // We can remove containers even if they have dependencies now
+ // As we have guaranteed their dependencies are in the pod
+ for _, ctr := range ctrs {
+ // Clean up network namespace, cgroups, mounts
+ if err := ctr.cleanup(); err != nil {
+ return err
+ }
+
+ // Stop container's storage
+ if err := ctr.teardownStorage(); err != nil {
+ return err
+ }
+
+ // Delete the container from runtime (only if we are not
+ // ContainerStateConfigured)
+ if ctr.state.State != ContainerStateConfigured {
+ if err := ctr.delete(ctx); err != nil {
+ return err
+ }
+ }
+ }
+
+ // Remove containers from the state
+ if err := r.state.RemovePodContainers(p); err != nil {
+ return err
+ }
+
+ // Mark containers invalid
+ for _, ctr := range ctrs {
+ ctr.valid = false
+ }
+
+ // Remove pod cgroup, if present
+ if p.state.CgroupPath != "" {
+ switch p.runtime.config.CgroupManager {
+ case SystemdCgroupsManager:
+ // NOOP for now, until proper systemd cgroup management
+ // is implemented
+ case CgroupfsCgroupsManager:
+ // Delete the cgroupfs cgroup
+ logrus.Debugf("Removing pod cgroup %s", p.state.CgroupPath)
+
+ cgroup, err := cgroups.Load(cgroups.V1, cgroups.StaticPath(p.state.CgroupPath))
+ if err != nil && err != cgroups.ErrCgroupDeleted {
+ return err
+ } else if err == nil {
+ if err := cgroup.Delete(); err != nil {
+ return err
+ }
+ }
+ default:
+ return errors.Wrapf(ErrInvalidArg, "unknown cgroups manager %s specified", p.runtime.config.CgroupManager)
+ }
+ }
+
+ // Remove pod from state
+ if err := r.state.RemovePod(p); err != nil {
+ return err
+ }
+
+ // Mark pod invalid
+ p.valid = false
+
+ return nil
+}