mirror of
https://github.com/fnproject/fn.git
synced 2022-10-28 21:29:17 +03:00
this patch gets rid of max concurrency for functions altogether, as discussed, since it will be challenging to support across functions nodes. as a result of doing so, the previous version of functions would fall over when offered 1000 functions, so there was some work needed in order to push this through. further work is necessary as docker basically falls over when trying to start enough containers at the same time, and with this patch essentially every function can scale infinitely. it seems like we could add some kind of adaptive restrictions based on task run length and configured wait time so that fast running functions will line up to run in a hot container instead of them all creating new hot containers. this patch takes a first cut at whacking out some of the insanity that was the previous concurrency model, which was problematic in that it limited concurrency significantly across all functions since every task went through the same unbuffered channel, which could create blocking issues for all functions if the channel is not picked off fast enough (it's not apparent that this was impossible in the previous implementation). in any event, each request has a goroutine already, there's no reason not to use it. not too hard to wrap a map in a lock, not sure what the benefits were (added insanity?) in effect this is marginally easier to understand and less insane (marginally). after getting rid of max c this adds a blocking mechanism for the first invocation of any function so that all other hot functions will wait on the first one to finish to avoid a herd issue (was making docker die...) -- this could be slightly improved, but works in a pinch. reduced some memory usage by having redundant maps of htfnsvr's and task.Requests (by a factor of 2!). cleaned up some of the protocol stuff, need to clean this up further. anyway, it's a first cut. have another patch that rewrites all of it but was getting into rabbit hole territory, would be happy to oblige if anybody else has problems understanding this rat's nest of channels. there is a good bit of work left to make this prod ready (regardless of removing max c). a warning that this will break the db schemas, didn't put the effort in to add migration stuff since this isn't deployed anywhere in prod... TODO need to clean out the htfnmgr bucket with LRU TODO need to clean up runner interface TODO need to unify the task running paths across protocols TODO need to move the ram checking stuff into worker for noted reasons TODO need better elasticity of hot f(x) containers
315 lines
10 KiB
Go
315 lines
10 KiB
Go
package runner
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/Sirupsen/logrus"
|
|
"github.com/go-openapi/strfmt"
|
|
uuid "github.com/satori/go.uuid"
|
|
"gitlab-odx.oracle.com/odx/functions/api/models"
|
|
"gitlab-odx.oracle.com/odx/functions/api/runner/drivers"
|
|
"gitlab-odx.oracle.com/odx/functions/api/runner/protocol"
|
|
"gitlab-odx.oracle.com/odx/functions/api/runner/task"
|
|
)
|
|
|
|
// hot functions - theory of operation
|
|
//
|
|
// A function is converted into a hot function if its `Format` is either
|
|
// a streamable format/protocol. At the very first task request a hot
|
|
// container shall be started and run it. Each hot function has an internal
|
|
// clock that actually halts the container if it goes idle long enough. In the
|
|
// absence of workload, it just stops the whole clockwork.
|
|
//
|
|
// Internally, the hot function uses a modified Config whose Stdin and Stdout
|
|
// are bound to an internal pipe. This internal pipe is fed with incoming tasks
|
|
// Stdin and feeds incoming tasks with Stdout.
|
|
//
|
|
// Each execution is the alternation of feeding hot functions stdin with tasks
|
|
// stdin, and reading the answer back from containers stdout. For all `Format`s
|
|
// we send embedded into the message metadata to help the container to know when
|
|
// to stop reading from its stdin and Functions expect the container to do the
|
|
// same. Refer to api/runner/protocol.go for details of these communications.
|
|
//
|
|
// hot functions implementation relies in two moving parts (drawn below):
|
|
// htfnmgr and htfn. Refer to their respective comments for
|
|
// details.
|
|
// │
|
|
// Incoming
|
|
// Task
|
|
// │
|
|
// ┌──────▼────────┐
|
|
// ┌┴──────────────┐│
|
|
// │ Per Function ││ non-streamable f()
|
|
// ┌───────│ Container │├──────┐───────────────┐
|
|
// │ │ Manager ├┘ │ │
|
|
// │ └───────────────┘ │ │
|
|
// │ │ │ │
|
|
// ▼ ▼ ▼ ▼
|
|
// ┌───────────┐ ┌───────────┐ ┌───────────┐ ┌───────────┐
|
|
// │ Hot │ │ Hot │ │ Hot │ │ Cold │
|
|
// │ Function │ │ Function │ │ Function │ │ Function │
|
|
// └───────────┘ └───────────┘ └───────────┘ └───────────┘
|
|
// Timeout
|
|
// Terminate
|
|
// (internal clock)
|
|
|
|
// RunTrackedTask is just a wrapper for shared logic for async/sync runners
|
|
func (rnr *Runner) RunTrackedTask(newTask *models.Task, ctx context.Context, cfg *task.Config, ds models.Datastore) (drivers.RunResult, error) {
|
|
startedAt := strfmt.DateTime(time.Now())
|
|
newTask.StartedAt = startedAt
|
|
|
|
result, err := rnr.RunTask(ctx, cfg)
|
|
|
|
completedAt := strfmt.DateTime(time.Now())
|
|
status := result.Status()
|
|
newTask.CompletedAt = completedAt
|
|
newTask.Status = status
|
|
|
|
err = ds.InsertTask(ctx, newTask)
|
|
// TODO we should just log this error not return it to user? just issue storing task status but task is run
|
|
|
|
return result, err
|
|
}
|
|
|
|
// RunTask will dispatch a task specified by cfg to a hot container, if possible,
|
|
// that already exists or will create a new container to run a task and then run it.
|
|
// TODO XXX (reed): merge this and RunTrackedTask to reduce surface area...
|
|
func (rnr *Runner) RunTask(ctx context.Context, cfg *task.Config) (drivers.RunResult, error) {
|
|
rnr.Start() // TODO layering issue ???
|
|
defer rnr.Complete()
|
|
|
|
tresp := make(chan task.Response)
|
|
treq := task.Request{Ctx: ctx, Config: cfg, Response: tresp}
|
|
tasks := rnr.hcmgr.getPipe(ctx, rnr, cfg)
|
|
if tasks == nil {
|
|
// TODO get rid of this to use herd stuff
|
|
go runTaskReq(rnr, treq)
|
|
} else {
|
|
tasks <- treq
|
|
}
|
|
resp := <-treq.Response
|
|
return resp.Result, resp.Err
|
|
}
|
|
|
|
// htfnmgr tracks all hot functions, used to funnel kittens into existing tubes
|
|
// XXX (reed): this map grows unbounded, need to add LRU but need to make
|
|
// sure that no functions are running when we evict
|
|
type htfnmgr struct {
|
|
sync.RWMutex
|
|
hc map[string]*htfnsvr
|
|
}
|
|
|
|
func (h *htfnmgr) getPipe(ctx context.Context, rnr *Runner, cfg *task.Config) chan<- task.Request {
|
|
isStream := protocol.IsStreamable(protocol.Protocol(cfg.Format))
|
|
if !isStream {
|
|
// TODO stop doing this, to prevent herds
|
|
return nil
|
|
}
|
|
|
|
h.RLock()
|
|
if h.hc == nil {
|
|
h.RUnlock()
|
|
h.Lock()
|
|
if h.hc == nil {
|
|
h.hc = make(map[string]*htfnsvr)
|
|
}
|
|
h.Unlock()
|
|
h.RLock()
|
|
}
|
|
|
|
// TODO(ccirello): re-implement this without memory allocation (fmt.Sprint)
|
|
fn := fmt.Sprint(cfg.AppName, ",", cfg.Path, cfg.Image, cfg.Timeout, cfg.Memory, cfg.Format)
|
|
svr, ok := h.hc[fn]
|
|
h.RUnlock()
|
|
if !ok {
|
|
h.Lock()
|
|
svr, ok = h.hc[fn]
|
|
if !ok {
|
|
svr = newhtfnsvr(ctx, cfg, rnr)
|
|
h.hc[fn] = svr
|
|
}
|
|
h.Unlock()
|
|
}
|
|
|
|
return svr.tasksin
|
|
}
|
|
|
|
// htfnsvr is part of htfnmgr, abstracted apart for simplicity, its only
|
|
// purpose is to test for hot functions saturation and try starting as many as
|
|
// needed. In case of absence of workload, it will stop trying to start new hot
|
|
// containers.
|
|
type htfnsvr struct {
|
|
cfg *task.Config
|
|
rnr *Runner
|
|
// TODO sharing with only a channel among hot containers will result in
|
|
// inefficient recycling of containers, we need a stack not a queue, so that
|
|
// when a lot of hot containers are up and throughput drops they don't all
|
|
// find a task every few seconds and stay up for a lot longer than we really
|
|
// need them.
|
|
tasksin chan task.Request
|
|
tasksout chan task.Request
|
|
first chan struct{}
|
|
once sync.Once // TODO this really needs to happen any time runner count goes to 0
|
|
}
|
|
|
|
func newhtfnsvr(ctx context.Context, cfg *task.Config, rnr *Runner) *htfnsvr {
|
|
svr := &htfnsvr{
|
|
cfg: cfg,
|
|
rnr: rnr,
|
|
tasksin: make(chan task.Request),
|
|
tasksout: make(chan task.Request, 1),
|
|
first: make(chan struct{}, 1),
|
|
}
|
|
svr.first <- struct{}{} // prime so that 1 thread will start the first container, others will wait
|
|
|
|
// This pipe will take all incoming tasks and just forward them to the
|
|
// started hot functions. The catch here is that it feeds a buffered
|
|
// channel from an unbuffered one. And this buffered channel is
|
|
// then used to determine the presence of running hot functions.
|
|
// If no hot function is available, tasksout will fill up to its
|
|
// capacity and pipe() will start them.
|
|
go svr.pipe(context.Background()) // XXX (reed): real context for adding consuela
|
|
return svr
|
|
}
|
|
|
|
func (svr *htfnsvr) pipe(ctx context.Context) {
|
|
for {
|
|
select {
|
|
case t := <-svr.tasksin:
|
|
svr.tasksout <- t
|
|
|
|
// TODO move checking for ram up here? then we can wait for hot functions to open up instead of always
|
|
// trying to make new ones if all hot functions are busy (and if machine is full and all functions are
|
|
// hot then most new hot functions are going to time out waiting to get available ram)
|
|
// TODO need to add some kind of metering here, we could track average run time and # of runners
|
|
select {
|
|
case _, ok := <-svr.first: // wait for >= 1 to be up to avoid herd
|
|
if ok || len(svr.tasksout) > 0 {
|
|
svr.launch(ctx)
|
|
}
|
|
case <-ctx.Done(): // TODO we should prob watch the task timeout not just the pipe...
|
|
return
|
|
}
|
|
case <-ctx.Done():
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
func (svr *htfnsvr) launch(ctx context.Context) {
|
|
hc := newhtfn(
|
|
svr.cfg,
|
|
svr.tasksout,
|
|
svr.rnr,
|
|
func() { svr.once.Do(func() { close(svr.first) }) },
|
|
)
|
|
go hc.serve(ctx)
|
|
}
|
|
|
|
// htfn is one instance of a hot container, which may or may not be running a
|
|
// task. If idle long enough, it will stop. It uses route configuration to
|
|
// determine which protocol to use.
|
|
type htfn struct {
|
|
id string
|
|
cfg *task.Config
|
|
proto protocol.ContainerIO
|
|
tasks <-chan task.Request
|
|
once func()
|
|
|
|
// Receiving side of the container.
|
|
containerIn io.Reader
|
|
containerOut io.Writer
|
|
|
|
rnr *Runner
|
|
}
|
|
|
|
func newhtfn(cfg *task.Config, tasks <-chan task.Request, rnr *Runner, once func()) *htfn {
|
|
stdinr, stdinw := io.Pipe()
|
|
stdoutr, stdoutw := io.Pipe()
|
|
|
|
return &htfn{
|
|
id: uuid.NewV5(uuid.Nil, fmt.Sprintf("%s%s%d", cfg.AppName, cfg.Path, time.Now().Unix())).String(),
|
|
cfg: cfg,
|
|
proto: protocol.New(protocol.Protocol(cfg.Format), stdinw, stdoutr),
|
|
tasks: tasks,
|
|
once: once,
|
|
|
|
containerIn: stdinr,
|
|
containerOut: stdoutw,
|
|
|
|
rnr: rnr,
|
|
}
|
|
}
|
|
|
|
func (hc *htfn) serve(ctx context.Context) {
|
|
lctx, cancel := context.WithCancel(ctx)
|
|
defer cancel()
|
|
cfg := *hc.cfg
|
|
logger := logrus.WithFields(logrus.Fields{"hot_id": hc.id, "app": cfg.AppName, "route": cfg.Path, "image": cfg.Image, "memory": cfg.Memory, "format": cfg.Format, "idle_timeout": cfg.IdleTimeout})
|
|
|
|
go func() {
|
|
for {
|
|
select {
|
|
case <-lctx.Done():
|
|
return
|
|
case <-time.After(cfg.IdleTimeout):
|
|
logger.Info("Canceling inactive hot function")
|
|
cancel()
|
|
case t := <-hc.tasks:
|
|
err := hc.proto.Dispatch(lctx, t)
|
|
status := "success"
|
|
if err != nil {
|
|
status = "error"
|
|
logrus.WithField("ctx", lctx).Info("task failed")
|
|
}
|
|
hc.once()
|
|
|
|
t.Response <- task.Response{
|
|
&runResult{StatusValue: status, error: err},
|
|
err,
|
|
}
|
|
}
|
|
}
|
|
}()
|
|
|
|
cfg.Env["FN_FORMAT"] = cfg.Format
|
|
cfg.Timeout = 0 // add a timeout to simulate ab.end. failure.
|
|
cfg.Stdin = hc.containerIn
|
|
cfg.Stdout = hc.containerOut
|
|
// NOTE: cfg.Stderr is overwritten in rnr.Run()
|
|
|
|
result, err := hc.rnr.run(lctx, &cfg)
|
|
if err != nil {
|
|
logger.WithError(err).Error("hot function failure detected")
|
|
}
|
|
logger.WithField("result", result).Info("hot function terminated")
|
|
}
|
|
|
|
// TODO make Default protocol a real thing and get rid of this in favor of Dispatch
|
|
func runTaskReq(rnr *Runner, t task.Request) {
|
|
result, err := rnr.run(t.Ctx, t.Config)
|
|
select {
|
|
case t.Response <- task.Response{result, err}:
|
|
close(t.Response)
|
|
default:
|
|
}
|
|
}
|
|
|
|
type runResult struct {
|
|
error
|
|
StatusValue string
|
|
}
|
|
|
|
func (r *runResult) Error() string {
|
|
if r.error == nil {
|
|
return ""
|
|
}
|
|
return r.error.Error()
|
|
}
|
|
|
|
func (r *runResult) Status() string { return r.StatusValue }
|