MuseEye / model_allowlists /1_0_6.json
NoExitPrior's picture
clean main: MuseEye HF version (fonts removed)
f5f56a5
{
"models": [
{
"name": "Gemma-3n-E2B-it",
"modelId": "google/gemma-3n-E2B-it-litert-lm",
"modelFile": "gemma-3n-E2B-it-int4.litertlm",
"description": "Preview version of [Gemma 3n E2B](https://ai.google.dev/gemma/docs/gemma-3n) ready for deployment on Android using the [MediaPipe LLM Inference API](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference). The current checkpoint supports text, vision, and audio input, with 4096 context length.",
"sizeInBytes": 3388604416,
"minDeviceMemoryInGb": 8,
"commitHash": "73b019b63436d346f68dd9c1dbfd117eb264d888",
"llmSupportImage": true,
"llmSupportAudio": true,
"defaultConfig": {
"topK": 64,
"topP": 0.95,
"temperature": 1.0,
"maxTokens": 4096,
"accelerators": "cpu,gpu"
},
"taskTypes": ["llm_chat", "llm_prompt_lab", "llm_ask_image", "llm_ask_audio"],
"bestForTaskTypes": ["llm_ask_image", "llm_ask_audio"]
},
{
"name": "Gemma-3n-E4B-it",
"modelId": "google/gemma-3n-E4B-it-litert-lm",
"modelFile": "gemma-3n-E4B-it-int4.litertlm",
"description": "Preview version of [Gemma 3n E4B](https://ai.google.dev/gemma/docs/gemma-3n) ready for deployment on Android using the [MediaPipe LLM Inference API](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference). The current checkpoint supports text, vision, and audio input, with 4096 context length.",
"sizeInBytes": 4652318720,
"minDeviceMemoryInGb": 12,
"commitHash": "3d0179a0648381585ab337e170b7517aae8e0ce4",
"llmSupportImage": true,
"llmSupportAudio": true,
"defaultConfig": {
"topK": 64,
"topP": 0.95,
"temperature": 1.0,
"maxTokens": 4096,
"accelerators": "cpu,gpu"
},
"taskTypes": ["llm_chat", "llm_prompt_lab", "llm_ask_image", "llm_ask_audio"]
},
{
"name": "Gemma3-1B-IT",
"modelId": "litert-community/Gemma3-1B-IT",
"modelFile": "gemma3-1b-it-int4.litertlm",
"description": "A variant of [google/Gemma-3-1B-IT](https://huggingface.co/google/Gemma-3-1B-IT) with 4-bit quantization ready for deployment on Android using the [MediaPipe LLM Inference API](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference)",
"sizeInBytes": 584417280,
"minDeviceMemoryInGb": 6,
"commitHash": "42d538a932e8d5b12e6b3b455f5572560bd60b2c",
"defaultConfig": {
"topK": 64,
"topP": 0.95,
"temperature": 1.0,
"maxTokens": 1024,
"accelerators": "gpu,cpu"
},
"taskTypes": ["llm_chat", "llm_prompt_lab"],
"bestForTaskTypes": ["llm_chat", "llm_prompt_lab"]
}
]
}