summaryrefslogtreecommitdiff
path: root/internal/transport/delivery/delivery.go
diff options
context:
space:
mode:
Diffstat (limited to 'internal/transport/delivery/delivery.go')
-rw-r--r--internal/transport/delivery/delivery.go331
1 files changed, 78 insertions, 253 deletions
diff --git a/internal/transport/delivery/delivery.go b/internal/transport/delivery/delivery.go
index 1e9126b2e..1e3ebb054 100644
--- a/internal/transport/delivery/delivery.go
+++ b/internal/transport/delivery/delivery.go
@@ -18,16 +18,13 @@
package delivery
import (
- "context"
- "slices"
+ "bytes"
+ "encoding/json"
+ "io"
+ "net/http"
"time"
- "codeberg.org/gruf/go-runners"
- "codeberg.org/gruf/go-structr"
- "github.com/superseriousbusiness/gotosocial/internal/gtscontext"
"github.com/superseriousbusiness/gotosocial/internal/httpclient"
- "github.com/superseriousbusiness/gotosocial/internal/queue"
- "github.com/superseriousbusiness/gotosocial/internal/util"
)
// Delivery wraps an httpclient.Request{}
@@ -36,6 +33,9 @@ import (
// be indexed (and so, dropped from queue)
// by any of these possible ID IRIs.
type Delivery struct {
+ // PubKeyID is the signing public key
+ // ID of the actor performing request.
+ PubKeyID string
// ActorID contains the ActivityPub
// actor ID IRI (if any) of the activity
@@ -61,273 +61,98 @@ type Delivery struct {
next time.Time
}
-func (dlv *Delivery) backoff() time.Duration {
- if dlv.next.IsZero() {
- return 0
- }
- return time.Until(dlv.next)
+// delivery is an internal type
+// for Delivery{} that provides
+// a json serialize / deserialize
+// able shape that minimizes data.
+type delivery struct {
+ PubKeyID string `json:"pub_key_id,omitempty"`
+ ActorID string `json:"actor_id,omitempty"`
+ ObjectID string `json:"object_id,omitempty"`
+ TargetID string `json:"target_id,omitempty"`
+ Method string `json:"method,omitempty"`
+ Header map[string][]string `json:"header,omitempty"`
+ URL string `json:"url,omitempty"`
+ Body []byte `json:"body,omitempty"`
}
-// WorkerPool wraps multiple Worker{}s in
-// a singular struct for easy multi start/stop.
-type WorkerPool struct {
+// Serialize will serialize the delivery data as data blob for storage,
+// note that this will flatten some of the data, dropping signing funcs.
+func (dlv *Delivery) Serialize() ([]byte, error) {
+ var body []byte
- // Client defines httpclient.Client{}
- // passed to each of delivery pool Worker{}s.
- Client *httpclient.Client
+ if dlv.Request.GetBody != nil {
+ // Fetch a fresh copy of request body.
+ rbody, err := dlv.Request.GetBody()
+ if err != nil {
+ return nil, err
+ }
- // Queue is the embedded queue.StructQueue{}
- // passed to each of delivery pool Worker{}s.
- Queue queue.StructQueue[*Delivery]
+ // Read request body into memory.
+ body, err = io.ReadAll(rbody)
- // internal fields.
- workers []*Worker
-}
+ // Done with body.
+ _ = rbody.Close()
-// Init will initialize the Worker{} pool
-// with given http client, request queue to pull
-// from and number of delivery workers to spawn.
-func (p *WorkerPool) Init(client *httpclient.Client) {
- p.Client = client
- p.Queue.Init(structr.QueueConfig[*Delivery]{
- Indices: []structr.IndexConfig{
- {Fields: "ActorID", Multiple: true},
- {Fields: "ObjectID", Multiple: true},
- {Fields: "TargetID", Multiple: true},
- },
- })
-}
-
-// Start will attempt to start 'n' Worker{}s.
-func (p *WorkerPool) Start(n int) {
- // Check whether workers are
- // set (is already running).
- ok := (len(p.workers) > 0)
- if ok {
- return
+ if err != nil {
+ return nil, err
+ }
}
- // Allocate new workers slice.
- p.workers = make([]*Worker, n)
- for i := range p.workers {
-
- // Allocate new Worker{}.
- p.workers[i] = new(Worker)
- p.workers[i].Client = p.Client
- p.workers[i].Queue = &p.Queue
-
- // Attempt to start worker.
- // Return bool not useful
- // here, as true = started,
- // false = already running.
- _ = p.workers[i].Start()
- }
+ // Marshal as internal JSON type.
+ return json.Marshal(delivery{
+ PubKeyID: dlv.PubKeyID,
+ ActorID: dlv.ActorID,
+ ObjectID: dlv.ObjectID,
+ TargetID: dlv.TargetID,
+ Method: dlv.Request.Method,
+ Header: dlv.Request.Header,
+ URL: dlv.Request.URL.String(),
+ Body: body,
+ })
}
-// Stop will attempt to stop contained Worker{}s.
-func (p *WorkerPool) Stop() {
- // Check whether workers are
- // set (is currently running).
- ok := (len(p.workers) == 0)
- if ok {
- return
- }
-
- // Stop all running workers.
- for i := range p.workers {
+// Deserialize will attempt to deserialize a blob of task data,
+// which will involve unflattening previously serialized data and
+// leave delivery incomplete, still requiring signing func setup.
+func (dlv *Delivery) Deserialize(data []byte) error {
+ var idlv delivery
- // return bool not useful
- // here, as true = stopped,
- // false = never running.
- _ = p.workers[i].Stop()
+ // Unmarshal as internal JSON type.
+ err := json.Unmarshal(data, &idlv)
+ if err != nil {
+ return err
}
- // Unset workers slice.
- p.workers = p.workers[:0]
-}
-
-// Worker wraps an httpclient.Client{} to feed
-// from queue.StructQueue{} for ActivityPub reqs
-// to deliver. It does so while prioritizing new
-// queued requests over backlogged retries.
-type Worker struct {
-
- // Client is the httpclient.Client{} that
- // delivery worker will use for requests.
- Client *httpclient.Client
-
- // Queue is the Delivery{} message queue
- // that delivery worker will feed from.
- Queue *queue.StructQueue[*Delivery]
+ // Copy over simplest fields.
+ dlv.PubKeyID = idlv.PubKeyID
+ dlv.ActorID = idlv.ActorID
+ dlv.ObjectID = idlv.ObjectID
+ dlv.TargetID = idlv.TargetID
- // internal fields.
- backlog []*Delivery
- service runners.Service
-}
-
-// Start will attempt to start the Worker{}.
-func (w *Worker) Start() bool {
- return w.service.GoRun(w.run)
-}
+ var body io.Reader
-// Stop will attempt to stop the Worker{}.
-func (w *Worker) Stop() bool {
- return w.service.Stop()
-}
-
-// run wraps process to restart on any panic.
-func (w *Worker) run(ctx context.Context) {
- if w.Client == nil || w.Queue == nil {
- panic("not yet initialized")
- }
- util.Must(func() { w.process(ctx) })
-}
-
-// process is the main delivery worker processing routine.
-func (w *Worker) process(ctx context.Context) bool {
- if w.Client == nil || w.Queue == nil {
- // we perform this check here just
- // to ensure the compiler knows these
- // variables aren't nil in the loop,
- // even if already checked by caller.
- panic("not yet initialized")
+ if idlv.Body != nil {
+ // Create new body reader from data.
+ body = bytes.NewReader(idlv.Body)
}
-loop:
- for {
- // Get next delivery.
- dlv, ok := w.next(ctx)
- if !ok {
- return true
- }
-
- // Check whether backoff required.
- const min = 100 * time.Millisecond
- if d := dlv.backoff(); d > min {
-
- // Start backoff sleep timer.
- backoff := time.NewTimer(d)
-
- select {
- case <-ctx.Done():
- // Main ctx
- // cancelled.
- backoff.Stop()
- return true
-
- case <-w.Queue.Wait():
- // A new message was
- // queued, re-add this
- // to backlog + retry.
- w.pushBacklog(dlv)
- backoff.Stop()
- continue loop
-
- case <-backoff.C:
- // success!
- }
- }
-
- // Attempt delivery of AP request.
- rsp, retry, err := w.Client.DoOnce(
- &dlv.Request,
- )
-
- if err == nil {
- // Ensure body closed.
- _ = rsp.Body.Close()
- continue loop
- }
-
- if !retry {
- // Drop deliveries when no
- // retry requested, or they
- // reached max (either).
- continue loop
- }
-
- // Determine next delivery attempt.
- backoff := dlv.Request.BackOff()
- dlv.next = time.Now().Add(backoff)
-
- // Push to backlog.
- w.pushBacklog(dlv)
+ // Create a new request object from unmarshaled details.
+ r, err := http.NewRequest(idlv.Method, idlv.URL, body)
+ if err != nil {
+ return err
}
-}
-
-// next gets the next available delivery, blocking until available if necessary.
-func (w *Worker) next(ctx context.Context) (*Delivery, bool) {
-loop:
- for {
- // Try pop next queued.
- dlv, ok := w.Queue.Pop()
-
- if !ok {
- // Check the backlog.
- if len(w.backlog) > 0 {
-
- // Sort by 'next' time.
- sortDeliveries(w.backlog)
-
- // Pop next delivery.
- dlv := w.popBacklog()
-
- return dlv, true
- }
- select {
- // Backlog is empty, we MUST
- // block until next enqueued.
- case <-w.Queue.Wait():
- continue loop
+ // Wrap request in httpclient type.
+ dlv.Request = httpclient.WrapRequest(r)
- // Worker was stopped.
- case <-ctx.Done():
- return nil, false
- }
- }
-
- // Replace request context for worker state canceling.
- ctx := gtscontext.WithValues(ctx, dlv.Request.Context())
- dlv.Request.Request = dlv.Request.Request.WithContext(ctx)
-
- return dlv, true
- }
+ return nil
}
-// popBacklog pops next available from the backlog.
-func (w *Worker) popBacklog() *Delivery {
- if len(w.backlog) == 0 {
- return nil
+// backoff returns a valid (>= 0) backoff duration.
+func (dlv *Delivery) backoff() time.Duration {
+ if dlv.next.IsZero() {
+ return 0
}
-
- // Pop from backlog.
- dlv := w.backlog[0]
-
- // Shift backlog down by one.
- copy(w.backlog, w.backlog[1:])
- w.backlog = w.backlog[:len(w.backlog)-1]
-
- return dlv
-}
-
-// pushBacklog pushes the given delivery to backlog.
-func (w *Worker) pushBacklog(dlv *Delivery) {
- w.backlog = append(w.backlog, dlv)
-}
-
-// sortDeliveries sorts deliveries according
-// to when is the first requiring re-attempt.
-func sortDeliveries(d []*Delivery) {
- slices.SortFunc(d, func(a, b *Delivery) int {
- const k = +1
- switch {
- case a.next.Before(b.next):
- return +k
- case b.next.Before(a.next):
- return -k
- default:
- return 0
- }
- })
+ return time.Until(dlv.next)
}