Update README.md
Browse files
README.md
CHANGED
|
@@ -428,7 +428,7 @@ These are qualitative observations from the model author based on manual use. Th
|
|
| 428 |
## Bias, Risks, and Limitations
|
| 429 |
|
| 430 |
- **Weak safety generalization.** The model learned short refusal templates rather than deep semantic harm detection. Paraphrased or novel harmful prompts frequently bypass refusals.
|
| 431 |
-
- **
|
| 432 |
- **All base model limitations apply.** 512-token context, limited world knowledge, occasional hallucination — see the [Stentor-30M model card](https://huggingface.co/StentorLabs/Stentor-30M) for full details.
|
| 433 |
- **No RLHF.** SFT only — no preference-based alignment was applied.
|
| 434 |
- **Dataset biases.** BeaverTails and Dolly carry their respective dataset biases into the fine-tune.
|
|
|
|
| 428 |
## Bias, Risks, and Limitations
|
| 429 |
|
| 430 |
- **Weak safety generalization.** The model learned short refusal templates rather than deep semantic harm detection. Paraphrased or novel harmful prompts frequently bypass refusals.
|
| 431 |
+
- **Rare self termination.** Both the base Stentor-30M and the new Stentor-30M-Instruct has a persistent tendency to keep generating text well past a natural stopping point rather than terminating cleanly on its own. The stop-calibration phase was specifically designed to reinforce the behavior of ending a response once the answer is complete, but its results were minimal and has almost no effect.
|
| 432 |
- **All base model limitations apply.** 512-token context, limited world knowledge, occasional hallucination — see the [Stentor-30M model card](https://huggingface.co/StentorLabs/Stentor-30M) for full details.
|
| 433 |
- **No RLHF.** SFT only — no preference-based alignment was applied.
|
| 434 |
- **Dataset biases.** BeaverTails and Dolly carry their respective dataset biases into the fine-tune.
|