Upload README.md with huggingface_hub
Browse files
README.md
ADDED
|
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
---
|
| 3 |
+
|
| 4 |
+
language: en
|
| 5 |
+
license: mit
|
| 6 |
+
tags:
|
| 7 |
+
- chain-of-thought
|
| 8 |
+
- structured-response
|
| 9 |
+
- causal-lm
|
| 10 |
+
- text-generation
|
| 11 |
+
datasets:
|
| 12 |
+
- diverse
|
| 13 |
+
pipeline_tag: text-generation
|
| 14 |
+
model_name: state-0
|
| 15 |
+
library_name: transformers
|
| 16 |
+
metrics:
|
| 17 |
+
- accuracy
|
| 18 |
+
- character
|
| 19 |
+
inference: true
|
| 20 |
+
|
| 21 |
+
---
|
| 22 |
+
|
| 23 |
+
[](https://hf.co/QuantFactory)
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
# QuantFactory/state-0-GGUF
|
| 27 |
+
This is quantized version of [Exthalpy/state-0](https://huggingface.co/Exthalpy/state-0) created using llama.cpp
|
| 28 |
+
|
| 29 |
+
# Original Model Card
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
# State-0: A chain-of-thoughts-based 8B alternative to GPT-o1
|
| 36 |
+
|
| 37 |
+
[](https://colab.research.google.com/drive/124hfluZIrtVeZ-gWJEz6C_6nhfFpUBhY?usp=sharing)
|
| 38 |
+
|
| 39 |
+
[](https://exthalpy.com/2024/09/18/introducing-state-0-exthalpys-advanced-chain-of-thought-ai-model-on-hugging-face/)
|
| 40 |
+
|
| 41 |
+
|
| 42 |
+
## Model Card
|
| 43 |
+
|
| 44 |
+
- **Model Name**: State-0
|
| 45 |
+
- **Version**: 1.0
|
| 46 |
+
- **Author**: Udit Akhouri
|
| 47 |
+
- **Hugging Face Model Page**: [Exthalpy/state-0](https://huggingface.co/Exthalpy/state-0/)
|
| 48 |
+
- **Architecture**: 8b core parameters with an additional 40 million parameters
|
| 49 |
+
- **Training Data**: Diverse datasets across various domains
|
| 50 |
+
- **Capabilities**: Chain-of-thought reasoning, Socratic instincts, in-depth and structured responses
|
| 51 |
+
- **Competitive Benchmark**: Capable of matching and surpassing the reasoning ability of GPT-4o1
|
| 52 |
+
- **Applications**: Educational tools, research, analytical problem-solving, and more
|
| 53 |
+
- **License**: MIT License
|
| 54 |
+
|
| 55 |
+
## Abstract
|
| 56 |
+
|
| 57 |
+
State-0 is a novel chain-of-thought language model, designed to emulate structured human-like reasoning in its responses. Inspired from the robust architecture of Llama 3.1 8b and enhanced with over 40 million additional parameters, State-0 achieves a significant leap in cognitive capabilities. It incorporates "Socratic instincts" to dissect complex queries methodically and arrive at well-rounded conclusions. Competing with the reasoning prowess of GPT-4o1, State-0 not only provides accurate answers but also elucidates the logical pathways leading to those answers, making it a powerful tool for applications requiring in-depth analysis and clarity.
|
| 58 |
+
|
| 59 |
+
## 1. Introduction
|
| 60 |
+
|
| 61 |
+
The field of natural language processing (NLP) has been significantly advanced by large language models (LLMs) capable of generating human-like text. However, most LLMs still lack the ability to break down complex queries into multiple facets, analyze them, and synthesize a comprehensive answer. State-0 addresses this limitation by employing a chain-of-thought reasoning mechanism combined with Socratic instincts. This paper introduces the architecture, training, and capabilities of State-0, demonstrating its competitive edge against models like GPT-4o1 in structured thinking and problem-solving.
|
| 62 |
+
|
| 63 |
+
## 2. Model Architecture
|
| 64 |
+
|
| 65 |
+
State-0, fundamentally inspired by Llama 3.1 8b, is augmented with over 40 million chain-of-thoughts and socratic instinct parameters. These parameters are meticulously trained to enhance the model's ability to reason, question, and deduce, drawing from vast datasets encompassing diverse fields of knowledge.
|
| 66 |
+
|
| 67 |
+
### 2.1 Enhancements Over Llama 3.1 8b
|
| 68 |
+
|
| 69 |
+
- **Additional Parameters**: State-0 incorporates 40 million additional parameters specifically fine-tuned to improve logical reasoning and analytical depth.
|
| 70 |
+
- **Chain-of-Thought Mechanism**: The model leverages a multi-step process for breaking down queries into manageable components, similar to human analytical thinking.
|
| 71 |
+
- **Socratic Instincts**: Inspired by the Socratic method, State-0 probes different aspects of a question, leading to a comprehensive and reasoned conclusion.
|
| 72 |
+
|
| 73 |
+
## 3. Model Capabilities
|
| 74 |
+
|
| 75 |
+
### 3.1 Chain-of-Thought Reasoning
|
| 76 |
+
|
| 77 |
+
State-0 excels at decomposing complex questions into simpler elements. It addresses these components individually and synthesizes the answers into a coherent response.
|
| 78 |
+
|
| 79 |
+
**Example**:
|
| 80 |
+
**Prompt**: Is milk a good source of protein?
|
| 81 |
+
|
| 82 |
+
**Response**:
|
| 83 |
+
1. **Definition**: Milk is a dairy product produced by mammals, including cows, goats, and sheep.
|
| 84 |
+
2. **Protein Content**: Milk is a rich source of protein, providing essential amino acids.
|
| 85 |
+
3. **Importance**: Protein is necessary for muscle growth, tissue repair, and immune function.
|
| 86 |
+
**Conclusion**: Milk is a valuable source of protein and contributes to overall health.
|
| 87 |
+
|
| 88 |
+
### 3.2 Competing with GPT-4o1
|
| 89 |
+
|
| 90 |
+
State-0 demonstrates competitive performance in terms of analytical depth and reasoning, often surpassing models like GPT-4o1 in its ability to provide contextually relevant and logically sound answers.
|
| 91 |
+
|
| 92 |
+
## 4. Getting Started
|
| 93 |
+
|
| 94 |
+
State-0 is available for use via the Hugging Face `transformers` library. This section outlines the installation and usage process for integrating State-0 into your projects.
|
| 95 |
+
|
| 96 |
+
### 4.1 Installation
|
| 97 |
+
|
| 98 |
+
Ensure you have the `transformers` library installed:
|
| 99 |
+
|
| 100 |
+
```bash
|
| 101 |
+
pip install transformers
|
| 102 |
+
```
|
| 103 |
+
|
| 104 |
+
### 4.2 Usage
|
| 105 |
+
|
| 106 |
+
#### High-Level Pipeline
|
| 107 |
+
|
| 108 |
+
State-0 can be easily used with the high-level pipeline API for text generation:
|
| 109 |
+
|
| 110 |
+
```python
|
| 111 |
+
from transformers import pipeline
|
| 112 |
+
|
| 113 |
+
pipe = pipeline("text-generation", model="uditakhouri/state-0")
|
| 114 |
+
response = pipe("Is milk a good source of protein?")
|
| 115 |
+
print(response)
|
| 116 |
+
```
|
| 117 |
+
|
| 118 |
+
#### Direct Model Loading
|
| 119 |
+
|
| 120 |
+
For more control, State-0 can be loaded directly using the following code:
|
| 121 |
+
|
| 122 |
+
```python
|
| 123 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 124 |
+
|
| 125 |
+
tokenizer = AutoTokenizer.from_pretrained("uditakhouri/state-0")
|
| 126 |
+
model = AutoModelForCausalLM.from_pretrained("uditakhouri/state-0")
|
| 127 |
+
|
| 128 |
+
input_text = "Is milk a good source of protein?"
|
| 129 |
+
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
| 130 |
+
|
| 131 |
+
output = model.generate(input_ids, max_length=100)
|
| 132 |
+
response = tokenizer.decode(output[0], skip_special_tokens=True)
|
| 133 |
+
print(response)
|
| 134 |
+
```
|
| 135 |
+
|
| 136 |
+
## 5. Training Details
|
| 137 |
+
|
| 138 |
+
State-0 was trained using a diverse set of datasets, fine-tuned to enhance its reasoning and conversational abilities. The training process focused on:
|
| 139 |
+
- Reinforcement Learning from Human Feedback (RLHF) for nuanced responses.
|
| 140 |
+
- Incorporating various fields of knowledge, from basic concepts to complex theories, to create a versatile reasoning engine.
|
| 141 |
+
|
| 142 |
+
## 6. Socratic Instincts
|
| 143 |
+
|
| 144 |
+
Inspired by the Socratic method, State-0 is designed to think through different scenarios and perspectives before arriving at an answer. This is achieved through:
|
| 145 |
+
- **Multi-Step Processing**: Breaking down a question into smaller parts, analyzing each component, and then synthesizing an answer.
|
| 146 |
+
- **Self-Interrogation**: The model internally queries different aspects of a topic, ensuring a balanced and well-thought-out response.
|
| 147 |
+
|
| 148 |
+
## 7. Evaluation and Results
|
| 149 |
+
|
| 150 |
+
State-0 has been rigorously tested against existing models like GPT-4o1, showing a high level of competence in chain-of-thought reasoning. It provides not only accurate answers but also the logical pathway leading to those answers, setting a new benchmark in LLM reasoning.
|
| 151 |
+
|
| 152 |
+
## 8. Conclusion
|
| 153 |
+
|
| 154 |
+
State-0 represents a significant advancement in the field of NLP by integrating chain-of-thought reasoning and Socratic instincts into its framework. With its enhanced parameters and structured analytical capabilities, State-0 is a formidable model for applications that demand a deep and reasoned understanding of complex queries.
|
| 155 |
+
|
| 156 |
+
## 9. Future Work
|
| 157 |
+
|
| 158 |
+
Future versions of State-0 aim to further enhance its reasoning capabilities by incorporating more advanced cognitive models and expanding its knowledge base.
|
| 159 |
+
|
| 160 |
+
## 10. License
|
| 161 |
+
|
| 162 |
+
State-0 is released under the MIT License.
|
| 163 |
+
|
| 164 |
+
## 11. References
|
| 165 |
+
|
| 166 |
+
For a complete list of references and further reading, please visit the model's page on [Hugging Face](https://huggingface.co/uditakhouri/state-0).
|
| 167 |
+
|
| 168 |
+
## 12. Contact
|
| 169 |
+
|
| 170 |
+
For inquiries, collaborations, or further information, please contact Udit Akhouri.
|