Spaces:
Configuration error
Configuration error
| package backend | |
| import ( | |
| "context" | |
| "fmt" | |
| "github.com/mudler/LocalAI/core/config" | |
| "github.com/mudler/LocalAI/pkg/grpc/proto" | |
| model "github.com/mudler/LocalAI/pkg/model" | |
| ) | |
| func TokenMetrics( | |
| modelFile string, | |
| loader *model.ModelLoader, | |
| appConfig *config.ApplicationConfig, | |
| backendConfig config.BackendConfig) (*proto.MetricsResponse, error) { | |
| opts := ModelOptions(backendConfig, appConfig, []model.Option{ | |
| model.WithModel(modelFile), | |
| }) | |
| model, err := loader.BackendLoader(opts...) | |
| if err != nil { | |
| return nil, err | |
| } | |
| if model == nil { | |
| return nil, fmt.Errorf("could not loadmodel model") | |
| } | |
| res, err := model.GetTokenMetrics(context.Background(), &proto.MetricsRequest{}) | |
| return res, err | |
| } | |