Spaces:
Running
Running
Amlan-109
feat: Initial commit of LocalAI Amlan Edition with premium branding and personalization
750bbe6
| package cli | |
| import ( | |
| "encoding/json" | |
| "errors" | |
| "fmt" | |
| "os" | |
| "path/filepath" | |
| "strings" | |
| "github.com/mholt/archiver/v3" | |
| "github.com/mudler/xlog" | |
| gguf "github.com/gpustack/gguf-parser-go" | |
| cliContext "github.com/mudler/LocalAI/core/cli/context" | |
| "github.com/mudler/LocalAI/core/config" | |
| "github.com/mudler/LocalAI/core/gallery" | |
| "github.com/mudler/LocalAI/pkg/downloader" | |
| "github.com/mudler/LocalAI/pkg/oci" | |
| "github.com/mudler/LocalAI/pkg/system" | |
| ) | |
| type UtilCMD struct { | |
| GGUFInfo GGUFInfoCMD `cmd:"" name:"gguf-info" help:"Get information about a GGUF file"` | |
| CreateOCIImage CreateOCIImageCMD `cmd:"" name:"create-oci-image" help:"Create an OCI image from a file or a directory"` | |
| HFScan HFScanCMD `cmd:"" name:"hf-scan" help:"Checks installed models for known security issues. WARNING: this is a best-effort feature and may not catch everything!"` | |
| UsecaseHeuristic UsecaseHeuristicCMD `cmd:"" name:"usecase-heuristic" help:"Checks a specific model config and prints what usecase LocalAI will offer for it."` | |
| } | |
| type GGUFInfoCMD struct { | |
| Args []string `arg:"" optional:"" name:"args" help:"Arguments to pass to the utility command"` | |
| Header bool `optional:"" default:"false" name:"header" help:"Show header information"` | |
| } | |
| type HFScanCMD struct { | |
| ModelsPath string `env:"LOCALAI_MODELS_PATH,MODELS_PATH" type:"path" default:"${basepath}/models" help:"Path containing models used for inferencing" group:"storage"` | |
| Galleries string `env:"LOCALAI_GALLERIES,GALLERIES" help:"JSON list of galleries" group:"models" default:"${galleries}"` | |
| ToScan []string `arg:""` | |
| } | |
| type UsecaseHeuristicCMD struct { | |
| ConfigName string `name:"The config file to check"` | |
| ModelsPath string `env:"LOCALAI_MODELS_PATH,MODELS_PATH" type:"path" default:"${basepath}/models" help:"Path containing models used for inferencing" group:"storage"` | |
| } | |
| type CreateOCIImageCMD struct { | |
| Input []string `arg:"" help:"Input file or directory to create an OCI image from"` | |
| Output string `default:"image.tar" help:"Output OCI image name"` | |
| ImageName string `default:"localai" help:"Image name"` | |
| Platform string `default:"linux/amd64" help:"Platform of the image"` | |
| } | |
| func (u *CreateOCIImageCMD) Run(ctx *cliContext.Context) error { | |
| xlog.Info("Creating OCI image from input") | |
| dir, err := os.MkdirTemp("", "localai") | |
| if err != nil { | |
| return err | |
| } | |
| defer os.RemoveAll(dir) | |
| err = archiver.Archive(u.Input, filepath.Join(dir, "archive.tar")) | |
| if err != nil { | |
| return err | |
| } | |
| xlog.Info("Creating OCI image", "output", u.Output, "input", u.Input) | |
| platform := strings.Split(u.Platform, "/") | |
| if len(platform) != 2 { | |
| return fmt.Errorf("invalid platform: %s", u.Platform) | |
| } | |
| return oci.CreateTar(filepath.Join(dir, "archive.tar"), u.Output, u.ImageName, platform[1], platform[0]) | |
| } | |
| func (u *GGUFInfoCMD) Run(ctx *cliContext.Context) error { | |
| if len(u.Args) == 0 { | |
| return fmt.Errorf("no GGUF file provided") | |
| } | |
| // We try to guess only if we don't have a template defined already | |
| f, err := gguf.ParseGGUFFile(u.Args[0]) | |
| if err != nil { | |
| // Only valid for gguf files | |
| xlog.Error("guessDefaultsFromFile: not a GGUF file") | |
| return err | |
| } | |
| xlog.Info("GGUF file loaded", "file", u.Args[0], "eosTokenID", f.Tokenizer().EOSTokenID, "bosTokenID", f.Tokenizer().BOSTokenID, "modelName", f.Metadata().Name, "architecture", f.Architecture().Architecture) | |
| xlog.Info("Tokenizer", "tokenizer", fmt.Sprintf("%+v", f.Tokenizer())) | |
| xlog.Info("Architecture", "architecture", fmt.Sprintf("%+v", f.Architecture())) | |
| v, exists := f.Header.MetadataKV.Get("tokenizer.chat_template") | |
| if exists { | |
| xlog.Info("chat_template", "template", v.ValueString()) | |
| } | |
| if u.Header { | |
| for _, metadata := range f.Header.MetadataKV { | |
| xlog.Info("metadata", "key", metadata.Key, "value", metadata.Value) | |
| } | |
| // log.Info().Any("header", fmt.Sprintf("%+v", f.Header)).Msg("Header") | |
| } | |
| return nil | |
| } | |
| func (hfscmd *HFScanCMD) Run(ctx *cliContext.Context) error { | |
| systemState, err := system.GetSystemState( | |
| system.WithModelPath(hfscmd.ModelsPath), | |
| ) | |
| if err != nil { | |
| return err | |
| } | |
| xlog.Info("LocalAI Security Scanner - This is BEST EFFORT functionality! Currently limited to huggingface models!") | |
| if len(hfscmd.ToScan) == 0 { | |
| xlog.Info("Checking all installed models against galleries") | |
| var galleries []config.Gallery | |
| if err := json.Unmarshal([]byte(hfscmd.Galleries), &galleries); err != nil { | |
| xlog.Error("unable to load galleries", "error", err) | |
| } | |
| err := gallery.SafetyScanGalleryModels(galleries, systemState) | |
| if err == nil { | |
| xlog.Info("No security warnings were detected for your installed models. Please note that this is a BEST EFFORT tool, and all issues may not be detected.") | |
| } else { | |
| xlog.Error("! WARNING ! A known-vulnerable model is installed!", "error", err) | |
| } | |
| return err | |
| } else { | |
| var errs error = nil | |
| for _, uri := range hfscmd.ToScan { | |
| xlog.Info("scanning specific uri", "uri", uri) | |
| scanResults, err := downloader.HuggingFaceScan(downloader.URI(uri)) | |
| if err != nil && errors.Is(err, downloader.ErrUnsafeFilesFound) { | |
| xlog.Error("! WARNING ! A known-vulnerable model is included in this repo!", "error", err, "clamAV", scanResults.ClamAVInfectedFiles, "pickles", scanResults.DangerousPickles) | |
| errs = errors.Join(errs, err) | |
| } | |
| } | |
| if errs != nil { | |
| return errs | |
| } | |
| xlog.Info("No security warnings were detected for your installed models. Please note that this is a BEST EFFORT tool, and all issues may not be detected.") | |
| return nil | |
| } | |
| } | |
| func (uhcmd *UsecaseHeuristicCMD) Run(ctx *cliContext.Context) error { | |
| if len(uhcmd.ConfigName) == 0 { | |
| xlog.Error("ConfigName is a required parameter") | |
| return fmt.Errorf("config name is a required parameter") | |
| } | |
| if len(uhcmd.ModelsPath) == 0 { | |
| xlog.Error("ModelsPath is a required parameter") | |
| return fmt.Errorf("model path is a required parameter") | |
| } | |
| bcl := config.NewModelConfigLoader(uhcmd.ModelsPath) | |
| err := bcl.ReadModelConfig(uhcmd.ConfigName) | |
| if err != nil { | |
| xlog.Error("error while loading backend", "error", err, "ConfigName", uhcmd.ConfigName) | |
| return err | |
| } | |
| bc, exists := bcl.GetModelConfig(uhcmd.ConfigName) | |
| if !exists { | |
| xlog.Error("ConfigName not found", "ConfigName", uhcmd.ConfigName) | |
| } | |
| for name, uc := range config.GetAllModelConfigUsecases() { | |
| if bc.HasUsecases(uc) { | |
| xlog.Info("Usecase", "usecase", name) | |
| } | |
| } | |
| xlog.Info("---") | |
| return nil | |
| } | |