Update README.md
Browse files
README.md
CHANGED
|
@@ -8,7 +8,7 @@ pipeline_tag: text-generation
|
|
| 8 |
---
|
| 9 |
# Model Card for Model ID
|
| 10 |
|
| 11 |
-
The model `Precacons/ReasonGPT-
|
| 12 |
|
| 13 |
|
| 14 |
## Model Details
|
|
@@ -47,7 +47,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
|
|
| 47 |
|
| 48 |
### Limitations
|
| 49 |
|
| 50 |
-
**ReasonGPT-
|
| 51 |
|
| 52 |
1. **Calculation Accuracy**:
|
| 53 |
- Due to its small size, the model may not perform complex calculations with high accuracy. It is optimized for reasoning and explanations rather than precise numerical computations.
|
|
@@ -59,7 +59,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
|
|
| 59 |
- With a smaller parameter size, the model may have limitations in understanding and generating contextually rich and nuanced responses compared to larger models.
|
| 60 |
|
| 61 |
4. **Bias and Fairness**:
|
| 62 |
-
- Like all language models, ReasonGPT-
|
| 63 |
|
| 64 |
5. **Resource Constraints**:
|
| 65 |
- While the model is designed to be efficient, it still requires a GPU for optimal performance. Users with limited computational resources may experience slower inference times.
|
|
@@ -70,7 +70,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
|
|
| 70 |
import predacons
|
| 71 |
|
| 72 |
# Load the model and tokenizer
|
| 73 |
-
model_path = "ReasonGPT-
|
| 74 |
model = predacons.load_model(model_path = model_path)
|
| 75 |
tokenizer = predacons.load_tokenizer(model_path)
|
| 76 |
|
|
@@ -87,7 +87,7 @@ generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
| 87 |
print(generated_text)
|
| 88 |
```
|
| 89 |
|
| 90 |
-
This example demonstrates how to load the `ReasonGPT-
|
| 91 |
|
| 92 |
|
| 93 |
|
|
|
|
| 8 |
---
|
| 9 |
# Model Card for Model ID
|
| 10 |
|
| 11 |
+
The model `Precacons/ReasonGPT-2B-4bit` is a lightweight language model based on the GEMMA architecture. It is designed to provide reasoning and explanations for any given problem. Despite its powerful capabilities, it is very compact, with a size of just 2.16 GB, making it efficient for deployment and use in various applications.
|
| 12 |
|
| 13 |
|
| 14 |
## Model Details
|
|
|
|
| 47 |
|
| 48 |
### Limitations
|
| 49 |
|
| 50 |
+
**ReasonGPT-2B-4bit** is a compact model designed for efficiency, but it comes with certain limitations:
|
| 51 |
|
| 52 |
1. **Calculation Accuracy**:
|
| 53 |
- Due to its small size, the model may not perform complex calculations with high accuracy. It is optimized for reasoning and explanations rather than precise numerical computations.
|
|
|
|
| 59 |
- With a smaller parameter size, the model may have limitations in understanding and generating contextually rich and nuanced responses compared to larger models.
|
| 60 |
|
| 61 |
4. **Bias and Fairness**:
|
| 62 |
+
- Like all language models, ReasonGPT-2B-4bit may exhibit biases present in the training data. Users should be cautious of potential biases in the generated outputs.
|
| 63 |
|
| 64 |
5. **Resource Constraints**:
|
| 65 |
- While the model is designed to be efficient, it still requires a GPU for optimal performance. Users with limited computational resources may experience slower inference times.
|
|
|
|
| 70 |
import predacons
|
| 71 |
|
| 72 |
# Load the model and tokenizer
|
| 73 |
+
model_path = "ReasonGPT-2B-4bit"
|
| 74 |
model = predacons.load_model(model_path = model_path)
|
| 75 |
tokenizer = predacons.load_tokenizer(model_path)
|
| 76 |
|
|
|
|
| 87 |
print(generated_text)
|
| 88 |
```
|
| 89 |
|
| 90 |
+
This example demonstrates how to load the `ReasonGPT-2B-4bit` model and use it to generate an explanation for a given query, keeping in mind the limitations mentioned above.
|
| 91 |
|
| 92 |
|
| 93 |
|