|
|
--- |
|
|
license: other |
|
|
language: |
|
|
- en |
|
|
base_model: |
|
|
- meta-llama/Llama-3.1-8B-Instruct |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- character |
|
|
--- |
|
|
# SydneyBot - LLaMA 8B Model (v1) |
|
|
|
|
|
## Model Description |
|
|
This is a fine-tuned version of the LLaMA 8B model, trained to emulate the personality of a fictional character named Sydney. The model is trained for conversational AI and supports text generation tasks. |
|
|
|
|
|
- **Architecture**: LLaMA 8B |
|
|
- **Fine-tuned On**: Custom dataset representing the personality of Sydney |
|
|
- **Size**: 8B parameters |
|
|
- **Task**: Text generation (Causal Language Modeling) |
|
|
|
|
|
## Intended Use |
|
|
- **Primary Use**: This model is intended for text generation, including role-playing chat, dialogue systems, and storytelling. |
|
|
- **How to Use**: The model can be used via the Hugging Face Inference API or integrated into custom applications using transformers. |
|
|
|
|
|
## Example Usage: |
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained("Eschatol/SydneyBot") |
|
|
tokenizer = AutoTokenizer.from_pretrained("Eschatol/SydneyBot") |
|
|
|
|
|
inputs = tokenizer("Hello, Sydney!", return_tensors="pt") |
|
|
outputs = model.generate(inputs["input_ids"], max_length=50) |
|
|
print(tokenizer.decode(outputs[0])) |