Spaces:
Running
Running
Amlan-109
feat: Initial commit of LocalAI Amlan Edition with premium branding and personalization
750bbe6 | package localai | |
| import ( | |
| "github.com/labstack/echo/v4" | |
| "github.com/mudler/LocalAI/core/config" | |
| "github.com/mudler/LocalAI/core/schema" | |
| "github.com/mudler/LocalAI/pkg/model" | |
| ) | |
| // SystemInformations returns the system informations | |
| // @Summary Show the LocalAI instance information | |
| // @Success 200 {object} schema.SystemInformationResponse "Response" | |
| // @Router /system [get] | |
| func SystemInformations(ml *model.ModelLoader, appConfig *config.ApplicationConfig) echo.HandlerFunc { | |
| return func(c echo.Context) error { | |
| availableBackends := []string{} | |
| loadedModels := ml.ListLoadedModels() | |
| for b := range appConfig.ExternalGRPCBackends { | |
| availableBackends = append(availableBackends, b) | |
| } | |
| for b := range ml.GetAllExternalBackends(nil) { | |
| availableBackends = append(availableBackends, b) | |
| } | |
| sysmodels := []schema.SysInfoModel{} | |
| for _, m := range loadedModels { | |
| sysmodels = append(sysmodels, schema.SysInfoModel{ID: m.ID}) | |
| } | |
| return c.JSON(200, | |
| schema.SystemInformationResponse{ | |
| Backends: availableBackends, | |
| Models: sysmodels, | |
| }, | |
| ) | |
| } | |
| } | |