Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

majentik
/
Nemotron-3-Nano-4B-RotorQuant-MLX-2bit

Text Generation
MLX
Safetensors
nemotron_h
rotorquant
kv-cache-quantization
nemotron
nvidia
mamba2
hybrid
quantized
2bit
conversational
custom_code
2-bit
Model card Files Files and versions
xet
Community
Nemotron-3-Nano-4B-RotorQuant-MLX-2bit
1.26 GB
Ctrl+K
Ctrl+K
  • 1 contributor
History: 3 commits
majentik's picture
majentik
Add model card
0b99ae5 verified 2 days ago
  • .gitattributes
    1.57 kB
    Add MLX quantized model weights 3 days ago
  • README.md
    4.19 kB
    Add model card 2 days ago
  • __init__.py
    0 Bytes
    Add MLX quantized model weights 3 days ago
  • chat_template.jinja
    10.5 kB
    Add MLX quantized model weights 3 days ago
  • config.json
    1.6 kB
    Add MLX quantized model weights 3 days ago
  • configuration_nemotron_h.py
    12.1 kB
    Add MLX quantized model weights 3 days ago
  • generation_config.json
    188 Bytes
    Add MLX quantized model weights 3 days ago
  • model.safetensors
    1.24 GB
    xet
    Add MLX quantized model weights 3 days ago
  • model.safetensors.index.json
    31.3 kB
    Add MLX quantized model weights 3 days ago
  • modeling_nemotron_h.py
    78.6 kB
    Add MLX quantized model weights 3 days ago
  • nano_v3_reasoning_parser.py
    798 Bytes
    Add MLX quantized model weights 3 days ago
  • tokenizer.json
    17.1 MB
    xet
    Add MLX quantized model weights 3 days ago
  • tokenizer_config.json
    372 Bytes
    Add MLX quantized model weights 3 days ago