| | --- |
| | license: apache-2.0 |
| | pipeline_tag: text-generation |
| | library_name: transformers |
| | tags: |
| | - mistral |
| | - chatbot |
| | --- |
| | # Roy-v1 |
| |
|
| | **Roy** is a personal AI assistant model created and fine-tuned by **Souvik Pramanick**. |
| | Designed to be helpful, conversational, and practical for everyday tasks such as learning, coding, problem solving, and general assistance. |
| |
|
| | --- |
| |
|
| | ## Creator |
| |
|
| | **Founder & Trainer:** |
| | **Souvik Pramanick** |
| | GitHub: https://github.com/Souvik18p |
| | HuggingFace: https://huggingface.co/souvik18 |
| |
|
| | Roy is an independent project built with the vision of creating a smart, friendly, and customizable AI assistant. |
| |
|
| | --- |
| |
|
| | ## What Roy Can Do |
| |
|
| | Roy is capable of: |
| |
|
| | - Natural conversation and assistance |
| | - Answering general knowledge questions |
| | - Solving math and logical problems |
| | - Helping with coding and debugging |
| | - Writing emails, stories, and content |
| | - Explaining concepts in simple language |
| | - Brainstorming ideas and learning support |
| |
|
| | --- |
| |
|
| | ## Model Details |
| |
|
| | - **Model Name:** Roy-v1 |
| | - **Parameters:** 7B |
| | - **Architecture:** LLaMA-based |
| | - **Tensor Type:** F16 |
| | - **Format:** Safetensors |
| | - **License:** Open for community usage |
| |
|
| | Base Model: **souvik18/Roy-v1** |
| |
|
| | ## Quantized Versions (Community) |
| |
|
| | Thanks to [@mradermacher](https://huggingface.co/mradermacher) for providing GGUF quants of Roy-v1: |
| |
|
| | **https://huggingface.co/mradermacher/Roy-v1-GGUF** |
| |
|
| | These versions allow Roy to run on: |
| | - CPU only systems |
| | - Low VRAM GPUs |
| | - Mobile / local apps via llama.cpp, ollama, koboldcpp |
| |
|
| |
|
| | ## Quick Usage |
| |
|
| | ### Using HuggingFace Transformers |
| |
|
| | ```python |
| | !pip install -U transformers datasets accelerate bitsandbytes peft huggingface_hub |
| | |
| | from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig |
| | import torch |
| | |
| | MODEL_ID = "souvik18/Roy-v1" |
| | |
| | # 4bit config – works best on Kaggle |
| | bnb_config = BitsAndBytesConfig( |
| | load_in_4bit=True, |
| | bnb_4bit_compute_dtype=torch.float16, |
| | bnb_4bit_use_double_quant=True, |
| | ) |
| | |
| | print(" Loading tokenizer...") |
| | tokenizer = AutoTokenizer.from_pretrained(MODEL_ID) |
| | tokenizer.pad_token = tokenizer.eos_token |
| | |
| | print(" Loading model (4bit)...") |
| | model = AutoModelForCausalLM.from_pretrained( |
| | MODEL_ID, |
| | quantization_config=bnb_config, |
| | device_map="auto" |
| | ) |
| | |
| | print("\n Roy-v1 Loaded Successfully!") |
| | |
| | while True: |
| | text = input("You: ") |
| | if text.lower() in ["exit","quit"]: |
| | break |
| | |
| | prompt = f"[INST] {text} [/INST]" |
| | |
| | inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| | |
| | with torch.no_grad(): |
| | out = model.generate( |
| | **inputs, |
| | max_new_tokens=200, |
| | temperature=0.7, |
| | top_p=0.9, |
| | do_sample=True |
| | ) |
| | |
| | |
| | |
| | |