commit a561ef354139e5124fbd1ed35c3f9fb6836b574b
parent 807a8e1cf6f50b57bef1a5e9bfa6568e9fc1eb45
Author: kim <89579420+NyaaaWhatsUpDoc@users.noreply.github.com>
Date: Sat, 7 May 2022 16:36:01 +0100
[feature] Update media manager to use internal/worker package (#543)
* update media manager to use internal/worker package, update worker with better logging
Signed-off-by: kim <grufwub@gmail.com>
* fix Queue() trace log message format operators
Signed-off-by: kim <grufwub@gmail.com>
* update media manager comment to match updated worker implementation
Signed-off-by: kim <grufwub@gmail.com>
Diffstat:
3 files changed, 83 insertions(+), 129 deletions(-)
diff --git a/internal/media/manager.go b/internal/media/manager.go
@@ -20,18 +20,16 @@ package media
import (
"context"
- "errors"
"fmt"
- "runtime"
"time"
- "codeberg.org/gruf/go-runners"
"codeberg.org/gruf/go-store/kv"
"github.com/robfig/cron/v3"
"github.com/sirupsen/logrus"
"github.com/spf13/viper"
"github.com/superseriousbusiness/gotosocial/internal/config"
"github.com/superseriousbusiness/gotosocial/internal/db"
+ "github.com/superseriousbusiness/gotosocial/internal/worker"
)
// Manager provides an interface for managing media: parsing, storing, and retrieving media objects like photos, videos, and gifs.
@@ -72,14 +70,6 @@ type Manager interface {
// 'Pruning' in this context means removing the locally stored data of the attachment (both thumbnail and full size),
// and setting 'cached' to false on the associated attachment.
PruneRemote(ctx context.Context, olderThanDays int) (int, error)
- // NumWorkers returns the total number of workers available to this manager.
- NumWorkers() int
- // QueueSize returns the total capacity of the queue.
- QueueSize() int
- // JobsQueued returns the number of jobs currently in the task queue.
- JobsQueued() int
- // ActiveWorkers returns the number of workers currently performing jobs.
- ActiveWorkers() int
// Stop stops the underlying worker pool of the manager. It should be called
// when closing GoToSocial in order to cleanly finish any in-progress jobs.
// It will block until workers are finished processing.
@@ -89,47 +79,54 @@ type Manager interface {
type manager struct {
db db.DB
storage *kv.KVStore
- pool runners.WorkerPool
+ emojiWorker *worker.Worker[*ProcessingEmoji]
+ mediaWorker *worker.Worker[*ProcessingMedia]
stopCronJobs func() error
- numWorkers int
- queueSize int
}
// NewManager returns a media manager with the given db and underlying storage.
//
// A worker pool will also be initialized for the manager, to ensure that only
-// a limited number of media will be processed in parallel.
-//
-// The number of workers will be the number of CPUs available to the Go runtime,
-// divided by 2 (rounding down, but always at least 1).
-//
-// The length of the queue will be the number of workers multiplied by 10.
-//
-// So for an 8 core machine, the media manager will get 4 workers, and a queue of length 40.
-// For a 4 core machine, this will be 2 workers, and a queue length of 20.
-// For a single or 2-core machine, the media manager will get 1 worker, and a queue of length 10.
+// a limited number of media will be processed in parallel. The numbers of workers
+// is determined from the $GOMAXPROCS environment variable (usually no. CPU cores).
+// See internal/worker.New() documentation for further information.
func NewManager(database db.DB, storage *kv.KVStore) (Manager, error) {
- // configure the worker pool
- // make sure we always have at least 1 worker even on single-core machines
- numWorkers := runtime.NumCPU() / 2
- if numWorkers == 0 {
- numWorkers = 1
- }
- queueSize := numWorkers * 10
-
m := &manager{
- db: database,
- storage: storage,
- pool: runners.NewWorkerPool(numWorkers, queueSize),
- numWorkers: numWorkers,
- queueSize: queueSize,
+ db: database,
+ storage: storage,
}
- // start the worker pool
- if start := m.pool.Start(); !start {
- return nil, errors.New("could not start worker pool")
+ // Prepare the media worker pool
+ m.mediaWorker = worker.New[*ProcessingMedia](-1, 10)
+ m.mediaWorker.SetProcessor(func(ctx context.Context, media *ProcessingMedia) error {
+ if err := ctx.Err(); err != nil {
+ return err
+ }
+ if _, err := media.LoadAttachment(ctx); err != nil {
+ return fmt.Errorf("error loading media %s: %v", media.AttachmentID(), err)
+ }
+ return nil
+ })
+
+ // Prepare the emoji worker pool
+ m.emojiWorker = worker.New[*ProcessingEmoji](-1, 10)
+ m.emojiWorker.SetProcessor(func(ctx context.Context, emoji *ProcessingEmoji) error {
+ if err := ctx.Err(); err != nil {
+ return err
+ }
+ if _, err := emoji.LoadEmoji(ctx); err != nil {
+ return fmt.Errorf("error loading emoji %s: %v", emoji.EmojiID(), err)
+ }
+ return nil
+ })
+
+ // Start the worker pools
+ if err := m.mediaWorker.Start(); err != nil {
+ return nil, err
+ }
+ if err := m.emojiWorker.Start(); err != nil {
+ return nil, err
}
- logrus.Debugf("started media manager worker pool with %d workers and queue capacity of %d", numWorkers, queueSize)
// start remote cache cleanup cronjob if configured
cacheCleanupDays := viper.GetInt(config.Keys.MediaRemoteCacheDays)
@@ -188,22 +185,7 @@ func (m *manager) ProcessMedia(ctx context.Context, data DataFunc, postData Post
if err != nil {
return nil, err
}
-
- logrus.Tracef("ProcessMedia: about to enqueue media with attachmentID %s, queue length is %d", processingMedia.AttachmentID(), m.pool.Queue())
- m.pool.Enqueue(func(innerCtx context.Context) {
- select {
- case <-innerCtx.Done():
- // if the inner context is done that means the worker pool is closing, so we should just return
- return
- default:
- // start loading the media already for the caller's convenience
- if _, err := processingMedia.LoadAttachment(innerCtx); err != nil {
- logrus.Errorf("ProcessMedia: error processing media with attachmentID %s: %s", processingMedia.AttachmentID(), err)
- }
- }
- })
- logrus.Tracef("ProcessMedia: succesfully queued media with attachmentID %s, queue length is %d", processingMedia.AttachmentID(), m.pool.Queue())
-
+ m.mediaWorker.Queue(processingMedia)
return processingMedia, nil
}
@@ -212,22 +194,7 @@ func (m *manager) ProcessEmoji(ctx context.Context, data DataFunc, postData Post
if err != nil {
return nil, err
}
-
- logrus.Tracef("ProcessEmoji: about to enqueue emoji with id %s, queue length is %d", processingEmoji.EmojiID(), m.pool.Queue())
- m.pool.Enqueue(func(innerCtx context.Context) {
- select {
- case <-innerCtx.Done():
- // if the inner context is done that means the worker pool is closing, so we should just return
- return
- default:
- // start loading the emoji already for the caller's convenience
- if _, err := processingEmoji.LoadEmoji(innerCtx); err != nil {
- logrus.Errorf("ProcessEmoji: error processing emoji with id %s: %s", processingEmoji.EmojiID(), err)
- }
- }
- })
- logrus.Tracef("ProcessEmoji: succesfully queued emoji with id %s, queue length is %d", processingEmoji.EmojiID(), m.pool.Queue())
-
+ m.emojiWorker.Queue(processingEmoji)
return processingEmoji, nil
}
@@ -236,51 +203,26 @@ func (m *manager) RecacheMedia(ctx context.Context, data DataFunc, postData Post
if err != nil {
return nil, err
}
-
- logrus.Tracef("RecacheMedia: about to enqueue recache with attachmentID %s, queue length is %d", processingRecache.AttachmentID(), m.pool.Queue())
- m.pool.Enqueue(func(innerCtx context.Context) {
- select {
- case <-innerCtx.Done():
- // if the inner context is done that means the worker pool is closing, so we should just return
- return
- default:
- // start loading the media already for the caller's convenience
- if _, err := processingRecache.LoadAttachment(innerCtx); err != nil {
- logrus.Errorf("RecacheMedia: error processing recache with attachmentID %s: %s", processingRecache.AttachmentID(), err)
- }
- }
- })
- logrus.Tracef("RecacheMedia: succesfully queued recache with attachmentID %s, queue length is %d", processingRecache.AttachmentID(), m.pool.Queue())
-
+ m.mediaWorker.Queue(processingRecache)
return processingRecache, nil
}
-func (m *manager) NumWorkers() int {
- return m.numWorkers
-}
-
-func (m *manager) QueueSize() int {
- return m.queueSize
-}
-
-func (m *manager) JobsQueued() int {
- return m.pool.Queue()
-}
+func (m *manager) Stop() error {
+ // Stop media and emoji worker pools
+ mediaErr := m.mediaWorker.Stop()
+ emojiErr := m.emojiWorker.Stop()
-func (m *manager) ActiveWorkers() int {
- return m.pool.Workers()
-}
+ var cronErr error
-func (m *manager) Stop() error {
- logrus.Info("stopping media manager worker pool")
- if !m.pool.Stop() {
- return errors.New("could not stop media manager worker pool")
+ if m.stopCronJobs != nil {
+ // only set if cache prune age > 0
+ cronErr = m.stopCronJobs()
}
- if m.stopCronJobs != nil { // only defined if cron jobs are actually running
- logrus.Info("stopping media manager cache cleanup jobs")
- return m.stopCronJobs()
+ if mediaErr != nil {
+ return mediaErr
+ } else if emojiErr != nil {
+ return emojiErr
}
-
- return nil
+ return cronErr
}
diff --git a/internal/media/manager_test.go b/internal/media/manager_test.go
@@ -443,8 +443,6 @@ func (suite *ManagerTestSuite) TestSimpleJpegQueueSpamming() {
}
for _, processingMedia := range inProcess {
- fmt.Printf("\n\n\nactive workers: %d, queue length: %d\n\n\n", suite.manager.ActiveWorkers(), suite.manager.JobsQueued())
-
// fetch the attachment id from the processing media
attachmentID := processingMedia.AttachmentID()
diff --git a/internal/worker/workers.go b/internal/worker/workers.go
@@ -3,6 +3,8 @@ package worker
import (
"context"
"errors"
+ "fmt"
+ "path"
"reflect"
"runtime"
@@ -17,34 +19,44 @@ type Worker[MsgType any] struct {
prefix string // contains type prefix for logging
}
-// New returns a new Worker[MsgType] with given number of workers and queue size
-// (see runners.WorkerPool for more information on args). If args < 1 then suitable
-// defaults are determined from the runtime's GOMAXPROCS variable.
-func New[MsgType any](workers int, queue int) *Worker[MsgType] {
+// New returns a new Worker[MsgType] with given number of workers and queue ratio,
+// where the queue ratio is multiplied by no. workers to get queue size. If args < 1
+// then suitable defaults are determined from the runtime's GOMAXPROCS variable.
+func New[MsgType any](workers int, queueRatio int) *Worker[MsgType] {
+ var zero MsgType
+
if workers < 1 {
// ensure sensible workers
workers = runtime.GOMAXPROCS(0)
}
- if queue < 1 {
- // ensure sensible queue
- queue = workers * 100
+ if queueRatio < 1 {
+ // ensure sensible ratio
+ queueRatio = 100
}
+ // Calculate the short type string for the msg type
+ msgType := reflect.TypeOf(zero).String()
+ _, msgType = path.Split(msgType)
+
w := &Worker[MsgType]{
- workers: runners.NewWorkerPool(workers, queue),
+ workers: runners.NewWorkerPool(workers, workers*queueRatio),
process: nil,
- prefix: reflect.TypeOf(Worker[MsgType]{}).String(), //nolint
+ prefix: fmt.Sprintf("worker.Worker[%s]", msgType),
}
// Log new worker creation with type prefix
- logrus.Infof("%s created with workers=%d queue=%d", w.prefix, workers, queue)
+ logrus.Infof("%s created with workers=%d queue=%d",
+ w.prefix,
+ workers,
+ workers*queueRatio,
+ )
return w
}
// Start will attempt to start the underlying worker pool, or return error.
func (w *Worker[MsgType]) Start() error {
- logrus.Info(w.prefix, "starting")
+ logrus.Infof("%s starting", w.prefix)
// Check processor was set
if w.process == nil {
@@ -61,7 +73,7 @@ func (w *Worker[MsgType]) Start() error {
// Stop will attempt to stop the underlying worker pool, or return error.
func (w *Worker[MsgType]) Stop() error {
- logrus.Info(w.prefix, "stopping")
+ logrus.Infof("%s stopping", w.prefix)
// Attempt to stop pool
if !w.workers.Stop() {
@@ -74,17 +86,19 @@ func (w *Worker[MsgType]) Stop() error {
// SetProcessor will set the Worker's processor function, which is called for each queued message.
func (w *Worker[MsgType]) SetProcessor(fn func(context.Context, MsgType) error) {
if w.process != nil {
- logrus.Panic(w.prefix, "Worker.process is already set")
+ logrus.Panicf("%s Worker.process is already set", w.prefix)
}
w.process = fn
}
// Queue will queue provided message to be processed with there's a free worker.
func (w *Worker[MsgType]) Queue(msg MsgType) {
- logrus.Tracef("%s queueing message: %+v", w.prefix, msg)
+ logrus.Tracef("%s queueing message (workers=%d queue=%d): %+v",
+ w.prefix, w.workers.Workers(), w.workers.Queue(), msg,
+ )
w.workers.Enqueue(func(ctx context.Context) {
if err := w.process(ctx, msg); err != nil {
- logrus.Error(err)
+ logrus.Errorf("%s %v", w.prefix, err)
}
})
}