Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -61,7 +61,7 @@ Below this is the role you are to play.
|
|
| 61 |
# Install flash-attn
|
| 62 |
subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
|
| 63 |
# Initialize the model pipeline
|
| 64 |
-
generator = pipeline('text-generation', model='Locutusque/Thespis-Llama-3.1-8B', torch_dtype=torch.bfloat16, token=os.getenv("TOKEN")
|
| 65 |
@spaces.GPU
|
| 66 |
def generate_text(prompt, system_prompt, temperature, top_p, top_k, repetition_penalty, max_length):
|
| 67 |
"""
|
|
|
|
| 61 |
# Install flash-attn
|
| 62 |
subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
|
| 63 |
# Initialize the model pipeline
|
| 64 |
+
generator = pipeline('text-generation', model='Locutusque/Thespis-Llama-3.1-8B', torch_dtype=torch.bfloat16, token=os.getenv("TOKEN"))
|
| 65 |
@spaces.GPU
|
| 66 |
def generate_text(prompt, system_prompt, temperature, top_p, top_k, repetition_penalty, max_length):
|
| 67 |
"""
|