Spaces:
Configuration error
Configuration error
| package backend | |
| import ( | |
| "github.com/mudler/LocalAI/core/config" | |
| "github.com/mudler/LocalAI/core/schema" | |
| "github.com/mudler/LocalAI/pkg/grpc" | |
| model "github.com/mudler/LocalAI/pkg/model" | |
| ) | |
| func ModelTokenize(s string, loader *model.ModelLoader, backendConfig config.BackendConfig, appConfig *config.ApplicationConfig) (schema.TokenizeResponse, error) { | |
| modelFile := backendConfig.Model | |
| var inferenceModel grpc.Backend | |
| var err error | |
| opts := ModelOptions(backendConfig, appConfig, []model.Option{ | |
| model.WithModel(modelFile), | |
| }) | |
| if backendConfig.Backend == "" { | |
| inferenceModel, err = loader.GreedyLoader(opts...) | |
| } else { | |
| opts = append(opts, model.WithBackendString(backendConfig.Backend)) | |
| inferenceModel, err = loader.BackendLoader(opts...) | |
| } | |
| if err != nil { | |
| return schema.TokenizeResponse{}, err | |
| } | |
| predictOptions := gRPCPredictOpts(backendConfig, loader.ModelPath) | |
| predictOptions.Prompt = s | |
| // tokenize the string | |
| resp, err := inferenceModel.TokenizeString(appConfig.Context, predictOptions) | |
| if err != nil { | |
| return schema.TokenizeResponse{}, err | |
| } | |
| return schema.TokenizeResponse{ | |
| Tokens: resp.Tokens, | |
| }, nil | |
| } | |