VIA-01 / README.md
Invescoz's picture
Update README.md
479811e verified
---
license: apache-2.0
language:
- en
library_name: transformers
pipeline_tag: text-generation
tags:
- Vidyut
- Rapnss
---
# *First Indian text-generation llm model with 7.24B parameters*
VIA-01 by Rapnss
VIA-01 is a state-of-the-art language model crafted by Rapnss, designed for exceptional performance in conversational tasks and code generation. Engineered for speed and versatility, VIA-01 delivers precise, high-quality responses across diverse applications, from answering complex questions to generating efficient code snippets.
Features
Dual-Purpose Excellence: Seamlessly blends natural language understanding with robust code generation, ideal for developers and conversational AI use cases.
Optimized Performance: Lightweight design ensures fast inference, even in resource-constrained environments.
Easy Integration: Compatible with Hugging Face Transformers for straightforward deployment.
Rapnss Innovation: A unique AI built to inspire creativity and productivity.
Usage
Get started with VIA-01 using the following Python code:
```python
from transformers import pipeline
import torch
# Initialize the pipeline
pipe = pipeline(
"text-generation",
model="rapnss/VIA-01",
torch_dtype=torch.float16,
device_map="auto",
max_new_tokens=15
)
# Generate a response
prompt = "Write a Python function to sort a list:"
response = pipe(prompt)[0]['generated_text']
print(response)
```
Example Output:
```
Write a Python function to sort a list:
def sort_list(arr):
return sorted(arr)
```
```
## Installation
Install required dependencies:
```bash
pip install transformers torch accelerate gradio
```
# Performance
Inference Speed: Optimized for low-latency responses, typically ~20-40 seconds on standard CPU hardware (e.g., Hugging Face free Space). For sub-10-second responses, use a GPU-enabled environment (e.g., Hugging Face Pro Space).
Model Size: ~8GB, balanced for efficiency and performance.
Capabilities: Excels in conversational queries, technical problem-solving, and code generation tasks like writing functions or debugging snippets.
# Try It Out
Interact with VIA-01 via our Hugging Face Space, featuring a Gradio interface for real-time testing.
Limitations
Response Length: Short responses (up to 15 tokens) recommended for optimal speed on free-tier hosting.
Hardware: CPU-based inference may be slower than GPU; performance varies with deployment setup.
# License
Licensed under the Apache 2.0 License, enabling flexible use and redistribution.
Contact
Created by Rapnss. For inquiries or feedback, reach out via Hugging Face or the VIA-01 Space.
Made with ❤️ by Rapnss.