mirror of
https://github.com/fnproject/fn.git
synced 2022-10-28 21:29:17 +03:00
Now obsoleted driver.PrepareCookie() call handled image and container creation. In agent, going forward we will need finer grained control over the timeouts implied by the contexts. For this reason, with this change, we split PrepareCookie() into Validate/Pull/Create calls under Cookie interface.
361 lines
10 KiB
Go
361 lines
10 KiB
Go
package docker
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"net/http"
|
|
"path"
|
|
"strings"
|
|
|
|
"github.com/fnproject/fn/api/agent/drivers"
|
|
"github.com/fnproject/fn/api/common"
|
|
"github.com/fnproject/fn/api/models"
|
|
|
|
"github.com/fsouza/go-dockerclient"
|
|
"github.com/sirupsen/logrus"
|
|
"go.opencensus.io/trace"
|
|
)
|
|
|
|
// A cookie identifies a unique request to run a task.
|
|
type cookie struct {
|
|
// namespace id used from prefork pool if applicable
|
|
poolId string
|
|
// network name from docker networks if applicable
|
|
netId string
|
|
|
|
// docker container create options created by Driver.CreateCookie, required for Driver.Prepare()
|
|
opts docker.CreateContainerOptions
|
|
// task associated with this cookie
|
|
task drivers.ContainerTask
|
|
// pointer to docker driver
|
|
drv *DockerDriver
|
|
|
|
// do we need to remove container at exit?
|
|
isCreated bool
|
|
|
|
imgReg string
|
|
imgRepo string
|
|
imgTag string
|
|
imgAuthConf *docker.AuthConfiguration
|
|
}
|
|
|
|
func (c *cookie) configureLogger(log logrus.FieldLogger) {
|
|
|
|
conf := c.task.LoggerConfig()
|
|
if conf.URL == "" {
|
|
c.opts.HostConfig.LogConfig = docker.LogConfig{
|
|
Type: "none",
|
|
}
|
|
return
|
|
}
|
|
|
|
c.opts.HostConfig.LogConfig = docker.LogConfig{
|
|
Type: "syslog",
|
|
Config: map[string]string{
|
|
"syslog-address": conf.URL,
|
|
"syslog-facility": "user",
|
|
"syslog-format": "rfc5424",
|
|
},
|
|
}
|
|
|
|
tags := make([]string, 0, len(conf.Tags))
|
|
for _, pair := range conf.Tags {
|
|
tags = append(tags, fmt.Sprintf("%s=%s", pair.Name, pair.Value))
|
|
}
|
|
if len(tags) > 0 {
|
|
c.opts.HostConfig.LogConfig.Config["tag"] = strings.Join(tags, ",")
|
|
}
|
|
}
|
|
|
|
func (c *cookie) configureMem(log logrus.FieldLogger) {
|
|
if c.task.Memory() == 0 {
|
|
return
|
|
}
|
|
|
|
mem := int64(c.task.Memory())
|
|
|
|
c.opts.Config.Memory = mem
|
|
c.opts.Config.MemorySwap = mem // disables swap
|
|
c.opts.Config.KernelMemory = mem
|
|
}
|
|
|
|
func (c *cookie) configureFsSize(log logrus.FieldLogger) {
|
|
if c.task.FsSize() == 0 {
|
|
return
|
|
}
|
|
|
|
// If defined, impose file system size limit. In MB units.
|
|
if c.opts.HostConfig.StorageOpt == nil {
|
|
c.opts.HostConfig.StorageOpt = make(map[string]string)
|
|
}
|
|
|
|
opt := fmt.Sprintf("%vM", c.task.FsSize())
|
|
log.WithFields(logrus.Fields{"size": opt, "call_id": c.task.Id()}).Debug("setting storage option")
|
|
c.opts.HostConfig.StorageOpt["size"] = opt
|
|
}
|
|
|
|
func (c *cookie) configureTmpFs(log logrus.FieldLogger) {
|
|
// if RO Root is NOT enabled and TmpFsSize does not have any limit, then we do not need
|
|
// any tmpfs in the container since function can freely write whereever it wants.
|
|
if c.task.TmpFsSize() == 0 && !c.drv.conf.EnableReadOnlyRootFs {
|
|
return
|
|
}
|
|
|
|
if c.opts.HostConfig.Tmpfs == nil {
|
|
c.opts.HostConfig.Tmpfs = make(map[string]string)
|
|
}
|
|
|
|
var tmpFsOption string
|
|
if c.task.TmpFsSize() != 0 {
|
|
if c.drv.conf.MaxTmpFsInodes != 0 {
|
|
tmpFsOption = fmt.Sprintf("size=%dm,nr_inodes=%d", c.task.TmpFsSize(), c.drv.conf.MaxTmpFsInodes)
|
|
} else {
|
|
tmpFsOption = fmt.Sprintf("size=%dm", c.task.TmpFsSize())
|
|
}
|
|
}
|
|
|
|
log.WithFields(logrus.Fields{"target": "/tmp", "options": tmpFsOption, "call_id": c.task.Id()}).Debug("setting tmpfs")
|
|
c.opts.HostConfig.Tmpfs["/tmp"] = tmpFsOption
|
|
}
|
|
|
|
func (c *cookie) configureIOFS(log logrus.FieldLogger) {
|
|
path := c.task.UDSDockerPath()
|
|
if path == "" {
|
|
// TODO this should be required soon-ish
|
|
return
|
|
}
|
|
|
|
bind := fmt.Sprintf("%s:%s", path, c.task.UDSDockerDest())
|
|
c.opts.HostConfig.Binds = append(c.opts.HostConfig.Binds, bind)
|
|
}
|
|
|
|
func (c *cookie) configureVolumes(log logrus.FieldLogger) {
|
|
if len(c.task.Volumes()) == 0 {
|
|
return
|
|
}
|
|
|
|
if c.opts.Config.Volumes == nil {
|
|
c.opts.Config.Volumes = map[string]struct{}{}
|
|
}
|
|
|
|
for _, mapping := range c.task.Volumes() {
|
|
hostDir := mapping[0]
|
|
containerDir := mapping[1]
|
|
c.opts.Config.Volumes[containerDir] = struct{}{}
|
|
mapn := fmt.Sprintf("%s:%s", hostDir, containerDir)
|
|
c.opts.HostConfig.Binds = append(c.opts.HostConfig.Binds, mapn)
|
|
log.WithFields(logrus.Fields{"volumes": mapn, "call_id": c.task.Id()}).Debug("setting volumes")
|
|
}
|
|
}
|
|
|
|
func (c *cookie) configureCPU(log logrus.FieldLogger) {
|
|
// Translate milli cpus into CPUQuota & CPUPeriod (see Linux cGroups CFS cgroup v1 documentation)
|
|
// eg: task.CPUQuota() of 8000 means CPUQuota of 8 * 100000 usecs in 100000 usec period,
|
|
// which is approx 8 CPUS in CFS world.
|
|
// Also see docker run options --cpu-quota and --cpu-period
|
|
if c.task.CPUs() == 0 {
|
|
return
|
|
}
|
|
|
|
quota := int64(c.task.CPUs() * 100)
|
|
period := int64(100000)
|
|
|
|
log.WithFields(logrus.Fields{"quota": quota, "period": period, "call_id": c.task.Id()}).Debug("setting CPU")
|
|
c.opts.HostConfig.CPUQuota = quota
|
|
c.opts.HostConfig.CPUPeriod = period
|
|
}
|
|
|
|
func (c *cookie) configureWorkDir(log logrus.FieldLogger) {
|
|
wd := c.task.WorkDir()
|
|
if wd == "" {
|
|
return
|
|
}
|
|
|
|
log.WithFields(logrus.Fields{"wd": wd, "call_id": c.task.Id()}).Debug("setting work dir")
|
|
c.opts.Config.WorkingDir = wd
|
|
}
|
|
|
|
func (c *cookie) configureHostname(log logrus.FieldLogger) {
|
|
// hostname and container NetworkMode is not compatible.
|
|
if c.opts.HostConfig.NetworkMode != "" {
|
|
return
|
|
}
|
|
|
|
log.WithFields(logrus.Fields{"hostname": c.drv.hostname, "call_id": c.task.Id()}).Debug("setting hostname")
|
|
c.opts.Config.Hostname = c.drv.hostname
|
|
}
|
|
|
|
func (c *cookie) configureCmd(log logrus.FieldLogger) {
|
|
if c.task.Command() == "" {
|
|
return
|
|
}
|
|
|
|
// NOTE: this is hyper-sensitive and may not be correct like this even, but it passes old tests
|
|
cmd := strings.Fields(c.task.Command())
|
|
log.WithFields(logrus.Fields{"call_id": c.task.Id(), "cmd": cmd, "len": len(cmd)}).Debug("docker command")
|
|
c.opts.Config.Cmd = cmd
|
|
}
|
|
|
|
func (c *cookie) configureEnv(log logrus.FieldLogger) {
|
|
if len(c.task.EnvVars()) == 0 {
|
|
return
|
|
}
|
|
|
|
if c.opts.Config.Env == nil {
|
|
c.opts.Config.Env = make([]string, 0, len(c.task.EnvVars()))
|
|
}
|
|
|
|
for name, val := range c.task.EnvVars() {
|
|
c.opts.Config.Env = append(c.opts.Config.Env, name+"="+val)
|
|
}
|
|
}
|
|
|
|
// implements Cookie
|
|
func (c *cookie) Close(ctx context.Context) error {
|
|
var err error
|
|
if c.isCreated {
|
|
err = c.drv.removeContainer(ctx, c.task.Id())
|
|
}
|
|
c.drv.unpickPool(c)
|
|
c.drv.unpickNetwork(c)
|
|
return err
|
|
}
|
|
|
|
// implements Cookie
|
|
func (c *cookie) Run(ctx context.Context) (drivers.WaitResult, error) {
|
|
return c.drv.run(ctx, c.task.Id(), c.task)
|
|
}
|
|
|
|
// implements Cookie
|
|
func (c *cookie) ContainerOptions() interface{} {
|
|
return c.opts
|
|
}
|
|
|
|
// implements Cookie
|
|
func (c *cookie) Freeze(ctx context.Context) error {
|
|
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"stack": "Freeze"})
|
|
log.WithFields(logrus.Fields{"call_id": c.task.Id()}).Debug("docker pause")
|
|
|
|
err := c.drv.docker.PauseContainer(c.task.Id(), ctx)
|
|
if err != nil {
|
|
log.WithError(err).WithFields(logrus.Fields{"call_id": c.task.Id()}).Error("error pausing container")
|
|
}
|
|
return err
|
|
}
|
|
|
|
// implements Cookie
|
|
func (c *cookie) Unfreeze(ctx context.Context) error {
|
|
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"stack": "Unfreeze"})
|
|
log.WithFields(logrus.Fields{"call_id": c.task.Id()}).Debug("docker unpause")
|
|
|
|
err := c.drv.docker.UnpauseContainer(c.task.Id(), ctx)
|
|
if err != nil {
|
|
log.WithError(err).WithFields(logrus.Fields{"call_id": c.task.Id()}).Error("error unpausing container")
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (c *cookie) ValidateImage(ctx context.Context) (bool, error) {
|
|
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"stack": "ValidateImage"})
|
|
log.WithFields(logrus.Fields{"call_id": c.task.Id()}).Debug("docker auth and inspect image")
|
|
|
|
// ask for docker creds before looking for image, as the tasker may need to
|
|
// validate creds even if the image is downloaded.
|
|
config := findRegistryConfig(c.imgReg, c.drv.auths)
|
|
|
|
if task, ok := c.task.(Auther); ok {
|
|
_, span := trace.StartSpan(ctx, "docker_auth")
|
|
authConfig, err := task.DockerAuth()
|
|
span.End()
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if authConfig != nil {
|
|
config = authConfig
|
|
}
|
|
}
|
|
|
|
c.imgAuthConf = config
|
|
|
|
// see if we already have it
|
|
_, err := c.drv.docker.InspectImage(ctx, c.task.Image())
|
|
if err == docker.ErrNoSuchImage {
|
|
return true, nil
|
|
}
|
|
return false, err
|
|
}
|
|
|
|
func (c *cookie) PullImage(ctx context.Context) error {
|
|
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"stack": "PullImage"})
|
|
|
|
cfg := c.imgAuthConf
|
|
if cfg == nil {
|
|
log.Fatal("invalid usage: call ValidateImage first")
|
|
}
|
|
|
|
repo := path.Join(c.imgReg, c.imgRepo)
|
|
|
|
log = common.Logger(ctx).WithFields(logrus.Fields{"registry": cfg.ServerAddress, "username": cfg.Username, "image": c.task.Image()})
|
|
log.WithFields(logrus.Fields{"call_id": c.task.Id()}).Debug("docker pull")
|
|
|
|
err := c.drv.docker.PullImage(docker.PullImageOptions{Repository: repo, Tag: c.imgTag, Context: ctx}, *cfg)
|
|
if err != nil {
|
|
log.WithError(err).Error("Failed to pull image")
|
|
|
|
// TODO need to inspect for hub or network errors and pick; for now, assume
|
|
// 500 if not a docker error
|
|
msg := err.Error()
|
|
code := http.StatusInternalServerError
|
|
if dErr, ok := err.(*docker.Error); ok {
|
|
msg = dockerMsg(dErr)
|
|
code = dErr.Status // 401/404
|
|
}
|
|
|
|
return models.NewAPIError(code, fmt.Errorf("Failed to pull image '%s': %s", c.task.Image(), msg))
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (c *cookie) CreateContainer(ctx context.Context) error {
|
|
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"stack": "CreateContainer"})
|
|
log.WithFields(logrus.Fields{"call_id": c.task.Id()}).Debug("docker create container")
|
|
|
|
// here let's assume we have created container, logically this should be after 'CreateContainer', but we
|
|
// are not 100% sure that *any* failure to CreateContainer does not ever leave a container around especially
|
|
// going through fsouza+docker-api.
|
|
c.isCreated = true
|
|
|
|
c.opts.Context = ctx
|
|
_, err := c.drv.docker.CreateContainer(c.opts)
|
|
if err != nil {
|
|
// since we retry under the hood, if the container gets created and retry fails, we can just ignore error
|
|
if err != docker.ErrContainerAlreadyExists {
|
|
log.WithError(err).Error("Could not create container")
|
|
// NOTE: if the container fails to create we don't really want to show to user since they aren't directly configuring the container
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// removes docker err formatting: 'API Error (code) {"message":"..."}'
|
|
func dockerMsg(derr *docker.Error) string {
|
|
// derr.Message is a JSON response from docker, which has a "message" field we want to extract if possible.
|
|
// this is pretty lame, but it is what it is
|
|
var v struct {
|
|
Msg string `json:"message"`
|
|
}
|
|
|
|
err := json.Unmarshal([]byte(derr.Message), &v)
|
|
if err != nil {
|
|
// If message was not valid JSON, the raw body is still better than nothing.
|
|
return derr.Message
|
|
}
|
|
return v.Msg
|
|
}
|
|
|
|
var _ drivers.Cookie = &cookie{}
|