Update README.md
Browse files
README.md
CHANGED
|
@@ -7,6 +7,7 @@ library_name: transformers
|
|
| 7 |
tags:
|
| 8 |
- gemma2
|
| 9 |
- text-generation-inference
|
|
|
|
| 10 |
---
|
| 11 |

|
| 12 |
|
|
@@ -110,4 +111,4 @@ print(tokenizer.decode(outputs[0]))
|
|
| 110 |
For optimal performance on domain-specific tasks, fine-tuning on relevant datasets is required, which demands additional computational resources and expertise.
|
| 111 |
|
| 112 |
7. **Context Length Limitation:**
|
| 113 |
-
The model’s ability to process long documents is limited by its maximum context window size. If the input exceeds this limit, truncation may lead to loss of important information.
|
|
|
|
| 7 |
tags:
|
| 8 |
- gemma2
|
| 9 |
- text-generation-inference
|
| 10 |
+
- f16
|
| 11 |
---
|
| 12 |

|
| 13 |
|
|
|
|
| 111 |
For optimal performance on domain-specific tasks, fine-tuning on relevant datasets is required, which demands additional computational resources and expertise.
|
| 112 |
|
| 113 |
7. **Context Length Limitation:**
|
| 114 |
+
The model’s ability to process long documents is limited by its maximum context window size. If the input exceeds this limit, truncation may lead to loss of important information.
|