GPT4All-Model / README.md
JonusNattapong's picture
Upload trained model
3a089e5 verified
# GPT4All-Model (Hanuman)
This folder contains the files needed to load and run the custom Hanuman model.
Included files:
- `pytorch_model.bin` — model weights
- `config.json` — model configuration
- `tokenizer.json`, `tokenizer_config.json`, `special_tokens_map.json` — tokenizer files
- `modeling.py` — custom `Hanuman` model implementation
- `hanuman_loader.py` — convenience loader (optional)
Quick usage (local files in this folder):
```python
# inference_local.py
from transformers import AutoTokenizer
from modeling import Hanuman
import torch
# load tokenizer from local folder
tokenizer = AutoTokenizer.from_pretrained('.')
# load model using the provided helper
model = Hanuman.from_pretrained('.', map_location='cpu')
prompt = "สวัสดีครับ ช่วยอธิบายสั้น ๆ เกี่ยวกับประเทศไทย"
inputs = tokenizer(prompt, return_tensors='pt')
outputs = model.generate(inputs['input_ids'], max_new_tokens=50, temperature=1.2, top_k=50, top_p=0.95)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
Or load from the Hugging Face Hub (if this folder was uploaded to the hub as the repo root):
```python
# inference_from_hub.py
from transformers import AutoTokenizer
from hanuman_loader import HanumanModel
repo_id = "ZombitX64/GPT4All-Model"
# tokenizer will download from HF
tokenizer = AutoTokenizer.from_pretrained(repo_id)
# HanumanModel downloads weights and modeling.py dynamically
model_wrapper = HanumanModel.from_pretrained(repo_id, map_location='cpu')
model = model_wrapper.model
prompt = "สวัสดีครับ ช่วยสรุปประเทศไทยสั้น ๆ"
inputs = tokenizer(prompt, return_tensors='pt')
outputs = model.generate(inputs['input_ids'], max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
Notes:
- This repo uses a custom model class (`Hanuman`) — users must keep `modeling.py` or use the provided `hanuman_loader.py` that dynamically imports it.
- For CPU inference, install a CPU build of PyTorch. For GPU, install the appropriate CUDA-enabled PyTorch.