Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

efops
/
marziel-8b-custom

Text Generation
MLX
Safetensors
GGUF
English
llama
vllm
4-bit precision
local-ai
private
maritime
vessel-tracking
osint
conversational
compressed-tensors
Model card Files Files and versions
xet
Community
marziel-8b-custom
  • 1 contributor
History: 20 commits
efops's picture
efops
v0.5.9: semantic intent routing
9cd2a84 verified about 9 hours ago
  • .gitattributes
    1.63 kB
    Initial release v0.5.0 (Clean history) 1 day ago
  • README.md
    13.2 kB
    v0.5.9: semantic intent routing about 9 hours ago
  • chat_template.jinja
    4.61 kB
    Initial release v0.5.0 (Clean history) 1 day ago
  • config.json
    1.81 kB
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • generation_config.json
    155 Bytes
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • marziel-8b-custom.gguf
    4.92 GB
    xet
    Initial release v0.5.0 (Clean history) 1 day ago
  • model-00001-of-00002.safetensors
    4.65 GB
    xet
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • model-00002-of-00002.safetensors
    1.05 GB
    xet
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • model.safetensors.index.json
    64.6 kB
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • recipe.yaml
    170 Bytes
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • special_tokens_map.json
    296 Bytes
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago
  • tokenizer.json
    17.2 MB
    xet
    Initial release v0.5.0 (Clean history) 1 day ago
  • tokenizer_config.json
    50.5 kB
    v0.5.8: GPTQ W4A16 quantized model for vLLM CPU (~4GB) about 11 hours ago