Spaces:
Sleeping
Sleeping
Update AI_Agent/llm_adapters/hf_adapter.py
Browse files
AI_Agent/llm_adapters/hf_adapter.py
CHANGED
|
@@ -1,4 +1,3 @@
|
|
| 1 |
-
# AI_Agent/llm_adapters/hf_adapter.py
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 3 |
import torch
|
| 4 |
import asyncio
|
|
@@ -9,8 +8,8 @@ class HuggingFaceAdapter:
|
|
| 9 |
self.tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 10 |
self.model = AutoModelForCausalLM.from_pretrained(
|
| 11 |
model_name,
|
| 12 |
-
|
| 13 |
-
device_map="auto"
|
| 14 |
)
|
| 15 |
|
| 16 |
async def generate(self, prompt: str, max_tokens=500):
|
|
|
|
|
|
|
| 1 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 2 |
import torch
|
| 3 |
import asyncio
|
|
|
|
| 8 |
self.tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 9 |
self.model = AutoModelForCausalLM.from_pretrained(
|
| 10 |
model_name,
|
| 11 |
+
dtype=torch.float16, # updated from torch_dtype to dtype
|
| 12 |
+
device_map="auto" # requires accelerate
|
| 13 |
)
|
| 14 |
|
| 15 |
async def generate(self, prompt: str, max_tokens=500):
|