mirror of
https://github.com/fnproject/fn.git
synced 2022-10-28 21:29:17 +03:00
functions: common concurrency stream for sync and async (#314)
* functions: add bounded concurrency * functions: plug runners to sync and async interfaces * functions: update documentation about the new env var * functions: fix test flakiness * functions: the runner is self-regulated, no need to set a number of runners * functions: push the execution to the background on incoming requests * functions: ensure async tasks are always on * functions: add prioritization to tasks consumption Ensure that Sync tasks are consumed before Async tasks. Also, fixes termination races problems for free. * functions: remove stale comments * functions: improve mem availability calculation * functions: parallel run for async tasks * functions: check for memory availability before pulling async task * functions: comment about rnr.hasAvailableMemory and sync.Cond * functions: implement memory check for async runners using Cond vars * functions: code grooming - remove unnecessary goroutines - fix stale docs - reorganize import group * Revert "functions: implement memory check for async runners using Cond vars" This reverts commit 922e64032201a177c03ce6a46240925e3d35430d. * Revert "functions: comment about rnr.hasAvailableMemory and sync.Cond" This reverts commit 49ad7d52d341f12da9603b1a1df9d145871f0e0a. * functions: set a minimum memory availability for sync * functions: simplify the implementation by removing the priority queue * functions: code grooming - code deduplication - review waitgroups Waits
This commit is contained in:
committed by
Seif Lotfy سيف لطفي
parent
c1f361dd0c
commit
9d06b6e687
@@ -15,7 +15,6 @@ import (
|
||||
"github.com/Sirupsen/logrus"
|
||||
"github.com/iron-io/functions/api/models"
|
||||
"github.com/iron-io/runner/common"
|
||||
"github.com/iron-io/runner/drivers"
|
||||
)
|
||||
|
||||
func getTask(ctx context.Context, url string) (*models.Task, error) {
|
||||
@@ -85,30 +84,14 @@ func deleteTask(url string, task *models.Task) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func runTask(ctx context.Context, task *models.Task) (drivers.RunResult, error) {
|
||||
// Set up runner and process task
|
||||
cfg := getCfg(task)
|
||||
rnr, err := New(NewMetricLogger())
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return rnr.Run(ctx, cfg)
|
||||
}
|
||||
|
||||
// RunAsyncRunner pulls tasks off a queue and processes them
|
||||
func RunAsyncRunner(ctx context.Context, tasksrv string, n int) {
|
||||
func RunAsyncRunner(ctx context.Context, tasksrv string, tasks chan TaskRequest, rnr *Runner) {
|
||||
u, h := tasksrvURL(tasksrv)
|
||||
if isHostOpen(h) {
|
||||
return
|
||||
}
|
||||
|
||||
var wg sync.WaitGroup
|
||||
for i := 0; i < n; i++ {
|
||||
wg.Add(1)
|
||||
go startAsyncRunners(ctx, &wg, i, u, runTask)
|
||||
}
|
||||
|
||||
wg.Wait()
|
||||
startAsyncRunners(ctx, u, tasks, rnr)
|
||||
<-ctx.Done()
|
||||
}
|
||||
|
||||
@@ -121,16 +104,21 @@ func isHostOpen(host string) bool {
|
||||
return available
|
||||
}
|
||||
|
||||
// todo: not a big fan of this anonymous function for testing, should use an interface and make a Mock object for testing - TR
|
||||
func startAsyncRunners(ctx context.Context, wg *sync.WaitGroup, i int, url string, runTask func(ctx context.Context, task *models.Task) (drivers.RunResult, error)) {
|
||||
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"async_runner": i})
|
||||
defer wg.Done()
|
||||
func startAsyncRunners(ctx context.Context, url string, tasks chan TaskRequest, rnr *Runner) {
|
||||
var wg sync.WaitGroup
|
||||
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"runner": "async"})
|
||||
for {
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
wg.Wait()
|
||||
return
|
||||
|
||||
default:
|
||||
if !rnr.hasAsyncAvailableMemory() {
|
||||
log.Debug("memory full")
|
||||
time.Sleep(1 * time.Second)
|
||||
continue
|
||||
}
|
||||
task, err := getTask(ctx, url)
|
||||
if err != nil {
|
||||
if err, ok := err.(net.Error); ok && err.Timeout() {
|
||||
@@ -148,11 +136,16 @@ func startAsyncRunners(ctx context.Context, wg *sync.WaitGroup, i int, url strin
|
||||
|
||||
ctx, log := common.LoggerWithFields(ctx, logrus.Fields{"call_id": task.ID})
|
||||
log.Debug("Running task:", task.ID)
|
||||
// Process Task
|
||||
if _, err := runTask(ctx, task); err != nil {
|
||||
log.WithError(err).Error("Cannot run task")
|
||||
continue
|
||||
}
|
||||
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
// Process Task
|
||||
if _, err := RunTask(tasks, ctx, getCfg(task)); err != nil {
|
||||
log.WithError(err).Error("Cannot run task")
|
||||
}
|
||||
}()
|
||||
|
||||
log.Debug("Processed task")
|
||||
|
||||
// Delete task from queue
|
||||
@@ -160,8 +153,8 @@ func startAsyncRunners(ctx context.Context, wg *sync.WaitGroup, i int, url strin
|
||||
log.WithError(err).Error("Cannot delete task")
|
||||
continue
|
||||
}
|
||||
|
||||
log.Info("Task complete")
|
||||
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -10,7 +10,6 @@ import (
|
||||
"math/rand"
|
||||
"net/http"
|
||||
"net/http/httptest"
|
||||
"sync"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
@@ -18,7 +17,6 @@ import (
|
||||
"github.com/gin-gonic/gin"
|
||||
"github.com/iron-io/functions/api/models"
|
||||
"github.com/iron-io/functions/api/mqs"
|
||||
"github.com/iron-io/runner/drivers"
|
||||
)
|
||||
|
||||
func setLogBuffer() *bytes.Buffer {
|
||||
@@ -92,22 +90,6 @@ func getTestServer(mockTasks []*models.Task) *httptest.Server {
|
||||
return httptest.NewServer(r)
|
||||
}
|
||||
|
||||
var helloImage = "iron/hello"
|
||||
|
||||
func TestRunTask(t *testing.T) {
|
||||
mockTask := getMockTask()
|
||||
mockTask.Image = &helloImage
|
||||
|
||||
result, err := runTask(context.Background(), &mockTask)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
if result.Status() != "success" {
|
||||
t.Errorf("TestRunTask failed to execute runTask")
|
||||
}
|
||||
}
|
||||
|
||||
func TestGetTask(t *testing.T) {
|
||||
buf := setLogBuffer()
|
||||
mockTask := getMockTask()
|
||||
@@ -206,19 +188,35 @@ func TestTasksrvURL(t *testing.T) {
|
||||
}
|
||||
}
|
||||
|
||||
func testRunner(t *testing.T) *Runner {
|
||||
r, err := New(NewMetricLogger())
|
||||
if err != nil {
|
||||
t.Fatal("Test: failed to create new runner")
|
||||
}
|
||||
return r
|
||||
}
|
||||
|
||||
func TestAsyncRunnersGracefulShutdown(t *testing.T) {
|
||||
buf := setLogBuffer()
|
||||
mockTask := getMockTask()
|
||||
ts := getTestServer([]*models.Task{&mockTask})
|
||||
defer ts.Close()
|
||||
|
||||
ctx, _ := context.WithTimeout(context.Background(), 2*time.Second)
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(1)
|
||||
go startAsyncRunners(ctx, &wg, 0, ts.URL+"/tasks", func(ctx context.Context, task *models.Task) (drivers.RunResult, error) {
|
||||
return nil, nil
|
||||
})
|
||||
wg.Wait()
|
||||
tasks := make(chan TaskRequest)
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second)
|
||||
defer cancel()
|
||||
defer close(tasks)
|
||||
go func() {
|
||||
for t := range tasks {
|
||||
t.Response <- TaskResponse{
|
||||
Result: nil,
|
||||
Err: nil,
|
||||
}
|
||||
|
||||
}
|
||||
}()
|
||||
|
||||
startAsyncRunners(ctx, ts.URL+"/tasks", tasks, testRunner(t))
|
||||
|
||||
if err := ctx.Err(); err != context.DeadlineExceeded {
|
||||
t.Log(buf.String())
|
||||
|
||||
@@ -115,10 +115,17 @@ func (r *Runner) queueHandler() {
|
||||
}
|
||||
}
|
||||
|
||||
func (r *Runner) hasAsyncAvailableMemory() bool {
|
||||
r.usedMemMutex.RLock()
|
||||
defer r.usedMemMutex.RUnlock()
|
||||
// reserve at least half of the memory for sync
|
||||
return (r.availableMem/2)-r.usedMem > 0
|
||||
}
|
||||
|
||||
func (r *Runner) checkRequiredMem(req uint64) bool {
|
||||
r.usedMemMutex.RLock()
|
||||
defer r.usedMemMutex.RUnlock()
|
||||
return r.availableMem-r.usedMem/int64(req)*1024*1024 > 0
|
||||
return (r.availableMem-r.usedMem)/int64(req)*1024*1024 > 0
|
||||
}
|
||||
|
||||
func (r *Runner) addUsedMem(used int64) {
|
||||
@@ -136,7 +143,7 @@ func (r *Runner) checkMemAndUse(req uint64) bool {
|
||||
|
||||
used := int64(req) * 1024 * 1024
|
||||
|
||||
if r.availableMem-r.usedMem/used < 0 {
|
||||
if (r.availableMem-r.usedMem)/used < 0 {
|
||||
return false
|
||||
}
|
||||
|
||||
|
||||
52
api/runner/worker.go
Normal file
52
api/runner/worker.go
Normal file
@@ -0,0 +1,52 @@
|
||||
package runner
|
||||
|
||||
import (
|
||||
"context"
|
||||
"sync"
|
||||
|
||||
"github.com/iron-io/runner/drivers"
|
||||
)
|
||||
|
||||
type TaskRequest struct {
|
||||
Ctx context.Context
|
||||
Config *Config
|
||||
Response chan TaskResponse
|
||||
}
|
||||
|
||||
type TaskResponse struct {
|
||||
Result drivers.RunResult
|
||||
Err error
|
||||
}
|
||||
|
||||
// StartWorkers handle incoming tasks and spawns self-regulating container
|
||||
// workers.
|
||||
func StartWorkers(ctx context.Context, rnr *Runner, tasks <-chan TaskRequest) {
|
||||
var wg sync.WaitGroup
|
||||
for {
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
wg.Wait()
|
||||
return
|
||||
case task := <-tasks:
|
||||
wg.Add(1)
|
||||
go func(task TaskRequest) {
|
||||
defer wg.Done()
|
||||
result, err := rnr.Run(task.Ctx, task.Config)
|
||||
select {
|
||||
case task.Response <- TaskResponse{result, err}:
|
||||
close(task.Response)
|
||||
default:
|
||||
}
|
||||
}(task)
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
func RunTask(tasks chan TaskRequest, ctx context.Context, cfg *Config) (drivers.RunResult, error) {
|
||||
tresp := make(chan TaskResponse)
|
||||
treq := TaskRequest{Ctx: ctx, Config: cfg, Response: tresp}
|
||||
tasks <- treq
|
||||
resp := <-treq.Response
|
||||
return resp.Result, resp.Err
|
||||
}
|
||||
Reference in New Issue
Block a user