| // Package dataloader provides a generic batching and caching mechanism for loading data | |
| // efficiently. It reduces the number of database queries by batching multiple individual | |
| // requests together and caching results to avoid duplicate fetches. | |
| // | |
| // The primary use case is in GraphQL resolvers where you might have N+1 query problems. | |
| // Instead of making individual database calls for each item, the dataloader batches | |
| // requests together and executes them in a single operation. | |
| // | |
| // Basic usage example: | |
| // | |
| // // Create a loader for User entities | |
| // userLoader := dataloader.NewStoreLoader(ctx, userStore.FindMany, func(u User) string { return u.ID }) | |
| // | |
| // // Use in resolvers - these calls will be batched together | |
| // user1, err := userLoader.FetchOne(ctx, "user-id-1") | |
| // user2, err := userLoader.FetchOne(ctx, "user-id-2") | |
| // user3, err := userLoader.FetchOne(ctx, "user-id-3") | |
| // | |
| // // For database operations with a DB connection: | |
| // dbLoader := dataloader.NewStoreLoaderWithDB(ctx, db, dbStore.FindMany, func(u User) string { return u.ID }) | |
| // | |
| // For parameterized queries, use FetcherParam: | |
| // | |
| // // Create a parameterized loader | |
| // type UserParams struct { Active bool } | |
| // paramLoader := dataloader.NewStoreLoaderParam(ctx, | |
| // func(ctx context.Context, param UserParams, ids []string) ([]User, error) { | |
| // return userStore.FindManyFiltered(ctx, ids, param.Active) | |
| // }, | |
| // func(u User) string { return u.ID }, | |
| // ) | |
| // activeUser, err := paramLoader.FetchOneParam(ctx, "user-id", UserParams{Active: true}) | |
| package dataloader | |
| import ( | |
| "context" | |
| "sync" | |
| "time" | |
| "github.com/target/goalert/gadb" | |
| ) | |
| type Loader[K comparable, V any] = Fetcher[K, V] | |
| type ( | |
| FetchFunc[K comparable, V any] func(context.Context, []K) ([]V, error) | |
| IDFunc[K comparable, V any] func(V) K | |
| ) | |
| // NewStoreLoader creates a new Fetcher for loading data from a store without parameters. | |
| // It's a convenience function for the common case where you only need to batch by ID. | |
| // | |
| // The fetchMany function should return values in any order, and the Fetcher will | |
| // map them back to the correct requests using the ID field. | |
| func NewStoreLoader[V any, K comparable](ctx context.Context, fetchMany FetchFunc[K, V], idFunc IDFunc[K, V]) *Fetcher[K, V] { | |
| return &Fetcher[K, V]{ | |
| MaxBatch: 100, | |
| Delay: 5 * time.Millisecond, | |
| FetchFunc: func(ctx context.Context, ids []K) ([]V, error) { return fetchMany(ctx, ids) }, | |
| IDFunc: idFunc, | |
| } | |
| } | |
| // NewStoreLoaderWithDB creates a new Fetcher that automatically passes a database | |
| // connection to the fetch function. This is a convenience wrapper around NewStoreLoader | |
| // for database-based loading operations. | |
| func NewStoreLoaderWithDB[V any, K comparable]( | |
| ctx context.Context, | |
| db gadb.DBTX, | |
| fetchMany func(context.Context, gadb.DBTX, []K) ([]V, error), | |
| idFunc IDFunc[K, V], | |
| ) *Fetcher[K, V] { | |
| return NewStoreLoader(ctx, func(ctx context.Context, ids []K) ([]V, error) { | |
| return fetchMany(ctx, db, ids) | |
| }, idFunc) | |
| } | |
| // Fetcher provides batched loading of data with caching. It batches individual requests | |
| // together to reduce the number of calls to the underlying data source, and caches | |
| // results to avoid duplicate fetches within the same Fetcher instance. | |
| // | |
| // Type parameters: | |
| // - K: The type of the unique identifier (key) for items being fetched | |
| // - V: The type of values being fetched | |
| // | |
| // The Fetcher is safe for concurrent use. All methods can be called from multiple | |
| // goroutines simultaneously. | |
| type Fetcher[K comparable, V any] struct { | |
| // FetchFunc is called to retrieve data for a batch of IDs. | |
| // It should return values in any order - the Fetcher will map them back to requests | |
| // using the ID extracted via IDFunc. | |
| FetchFunc func(ctx context.Context, ids []K) ([]V, error) | |
| // IDFunc extracts the unique identifier from a value. This function is required | |
| // and is used to match returned values back to their corresponding requests. | |
| IDFunc func(V) K | |
| // MaxBatch sets the maximum number of IDs to include in a single batch. | |
| // When this limit is reached, a new batch is started immediately. | |
| MaxBatch int | |
| // Delay specifies how long to wait before executing a batch. This allows | |
| // multiple requests to accumulate into a single batch, improving efficiency. | |
| Delay time.Duration | |
| cache map[K]*result[K, V] | |
| currentBatch *batch[K, V] | |
| mx sync.Mutex | |
| wg sync.WaitGroup | |
| } | |
| // result holds the outcome of fetching a single item, including the value, any error, | |
| // and a channel to signal completion. | |
| type result[K comparable, V any] struct { | |
| id K | |
| value *V | |
| err error | |
| done chan struct{} | |
| } | |
| // wait blocks until the result is available or the context is cancelled. | |
| // It returns the fetched value and any error that occurred during fetching. | |
| func (r *result[K, V]) wait(ctx context.Context) (*V, error) { | |
| select { | |
| case <-ctx.Done(): | |
| return nil, ctx.Err() | |
| case <-r.done: | |
| return r.value, r.err | |
| } | |
| } | |
| // Close waits for all pending batches to complete. This should be called when | |
| // the Fetcher is no longer needed to ensure proper cleanup and prevent | |
| // goroutine leaks. | |
| func (f *Fetcher[K, V]) Close() { | |
| // Wait for all batches to complete | |
| f.wg.Wait() | |
| } | |
| // fetchAll executes a batch request and distributes the results to waiting goroutines. | |
| // It calls FetchFunc with all the IDs in the batch, then matches returned values | |
| // back to their corresponding requests using IDFunc. | |
| func (f *Fetcher[K, V]) fetchAll(ctx context.Context, batch *batch[K, V]) { | |
| defer f.wg.Done() | |
| f.mx.Lock() | |
| // Before fetching, we check if this is the current batch, if so, | |
| // we remove the pointer so that it isn't changed by another goroutine. | |
| if f.currentBatch == batch { | |
| f.currentBatch = nil // reset for next batch | |
| } | |
| f.mx.Unlock() | |
| // Fetch ouside of the lock to avoid blocking other requests. | |
| values, err := f.FetchFunc(ctx, batch.ids) | |
| f.mx.Lock() | |
| defer f.mx.Unlock() | |
| if err != nil { | |
| // In the error case, we close all results with the error since there is no value to return. | |
| for _, res := range batch.results { | |
| res.err = err | |
| close(res.done) | |
| } | |
| return | |
| } | |
| // Match returned values to their corresponding requests | |
| for _, val := range values { | |
| id := f.IDFunc(val) | |
| res, ok := f.cache[id] | |
| if !ok { | |
| // we didn't ask for this ID, ignore it | |
| continue | |
| } | |
| res.value = &val | |
| } | |
| // Close all result channels - missing values will have nil value | |
| for _, res := range batch.results { | |
| close(res.done) | |
| } | |
| } | |
| // FetchOne retrieves a single value by its ID. Multiple calls to FetchOne within | |
| // the same batch window (defined by Delay) will be batched together into a single | |
| // call to FetchFunc. | |
| // | |
| // The method returns a pointer to the value if found, or nil if not found. | |
| // An error is returned if the fetch operation fails or the context is cancelled. | |
| func (f *Fetcher[K, V]) FetchOne(ctx context.Context, id K) (*V, error) { | |
| f.mx.Lock() | |
| if f.cache == nil { | |
| f.cache = make(map[K]*result[K, V]) | |
| } | |
| if res, ok := f.cache[id]; ok { | |
| // easy path, already cached | |
| f.mx.Unlock() | |
| return res.wait(ctx) | |
| } | |
| // create the entry for the new ID | |
| if f.currentBatch == nil { | |
| newBatch := new(batch[K, V]) | |
| f.currentBatch = newBatch | |
| f.wg.Add(1) | |
| time.AfterFunc(f.Delay, func() { | |
| f.fetchAll(ctx, newBatch) | |
| }) | |
| } | |
| res := f.currentBatch.Add(ctx, id) | |
| f.cache[id] = res | |
| if len(f.currentBatch.ids) >= f.MaxBatch { | |
| f.currentBatch = nil // reset for next batch | |
| } | |
| f.mx.Unlock() | |
| return res.wait(ctx) | |
| } | |