File size: 1,364 Bytes
fdb3f63
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
tags:
- neuron
- optimized
- aws-neuron
- text-generation
base_model: HuggingFaceTB/SmolLM2-135M-Instruct
---

# Neuron-Optimized HuggingFaceTB/SmolLM2-135M-Instruct

This repository contains AWS Neuron-optimized files for [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct).

## Model Details

- **Base Model**: [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct)
- **Task**: text-generation
- **Optimization**: AWS Neuron compilation
- **Generated by**: [badaoui](https://huggingface.co/badaoui)
- **Generated using**: [Optimum Neuron Compiler Space](https://huggingface.co/spaces/optimum/neuron-export)

## Usage

This model has been optimized for AWS Neuron devices (Inferentia/Trainium). To use it:

```python
from optimum.neuron import NeuronModelForCausalLM

model = NeuronModelForCausalLM.from_pretrained("badaoui/HuggingFaceTB-SmolLM2-135M-Instruct-neuron")
```

## Performance

These files are pre-compiled for AWS Neuron devices and should provide improved inference performance compared to the original model when deployed on Inferentia or Trainium instances.

## Original Model

For the original model, training details, and more information, please visit: [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct)