| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | metrics: |
| | - accuracy |
| | base_model: |
| | - meta-llama/Llama-3.1-8B-Instruct |
| | pipeline_tag: reinforcement-learning |
| | --- |
| | # Hibernates-2B-R1-V1 |
| |
|
| | A highly efficient 2B parameter language model optimized for reasoning and dialogue tasks. |
| |
|
| | ## Model Overview |
| |
|
| | Hibernates-2B is a custom transformer architecture designed for advanced language understanding and generation. Built with performance and efficiency in mind, it leverages state-of-the-art techniques for natural language processing. |
| |
|
| | ### Key Features |
| | - 2B Parameters |
| | - 4096 Token Context Window |
| | - Custom Transformer Architecture |
| | - Optimized for CPU and GPU Inference |
| | - Multi-Turn Dialogue Support |
| |
|
| | ## Technical Specifications |
| |
|
| | - **Architecture**: Custom Transformer |
| | - **Parameters**: 2 Billion |
| | - **Context Length**: 4096 tokens |
| | - **Model Type**: Decoder-only |
| | - **Tokenizer**: Custom WordPiece |
| | - **Format**: SafeTensors |
| |
|
| | ## Usage Guide |
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | import torch |
| | |
| | # Load model and tokenizer |
| | model_id = "Hibernates-2B-R1-V1" |
| | tokenizer = AutoTokenizer.from_pretrained(model_id) |
| | model = AutoModelForCausalLM.from_pretrained( |
| | model_id, |
| | torch_dtype=torch.float16, |
| | device_map="auto" |
| | ) |
| | |
| | # Example conversation |
| | messages = [ |
| | {"role": "system", "content": "You are a helpful AI assistant."}, |
| | {"role": "user", "content": "How can you help me today?"} |
| | ] |
| | |
| | # Generate response |
| | input_text = tokenizer.apply_chat_template(messages, tokenize=False) |
| | inputs = tokenizer(input_text, return_tensors="pt").to(model.device) |
| | outputs = model.generate( |
| | inputs["input_ids"], |
| | max_new_tokens=512, |
| | temperature=0.7, |
| | top_p=0.95 |
| | ) |
| | response = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| | ``` |
| |
|
| | ## Performance Characteristics |
| |
|
| | ### Strengths |
| | - Efficient Resource Usage |
| | - Strong Reasoning Capabilities |
| | - Multi-Turn Dialogue |
| | - Context Awareness |
| | - Instruction Following |
| |
|
| | ### Considerations |
| | - Resource Requirements: 8GB+ GPU RAM recommended |
| | - Task Specificity: Best suited for dialogue and reasoning tasks |
| | - Language Support: Primary focus on English |
| | - Model Size: Optimized for balance of performance and efficiency |
| |
|
| | ## License and Usage |
| | - Research and commercial use permitted |
| | - Attribution appreciated but not required |
| | - No warranty provided |
| |
|
| | ## Citation |
| |
|
| | If you use this model in your research, please cite: |
| |
|
| | ```bibtex |
| | @software{hibernates2b_2024, |
| | title={Hibernates-2B: Efficient Language Model for Reasoning}, |
| | year={2024}, |
| | version={R1-V1} |
| | } |
| | ``` |
| |
|
| | ## Acknowledgments |
| | Built using PyTorch and Hugging Face Transformers. Special thanks to the open-source AI community. |
| |
|
| | ## Download Instructions |
| |
|
| | Due to file size limitations, the model files are hosted externally. Download them from: |
| |
|
| | 1. [model-00001-of-00002.safetensors](https://huggingface.co/HibernatesAI/Hibernates-2B-R1-V1/blob/main/model-00001-of-00002.safetensors) |
| | 2. [model-00002-of-00002.safetensors](https://huggingface.co/HibernatesAI/Hibernates-2B-R1-V1/blob/main/model-00002-of-00002.safetensors) |
| |
|
| | Place these files in the root directory of the project before running. |