Update README.md
Browse files
README.md
CHANGED
|
@@ -34,7 +34,7 @@ The model is trained on a dataset provided in CSV format (with a 'data' column)
|
|
| 34 |
- **Language(s) (NLP):** English
|
| 35 |
- **License:** Apache-2.0
|
| 36 |
|
| 37 |
-
### Model Sources
|
| 38 |
|
| 39 |
- **Repository:** https://huggingface.co/aarohanverma/lstm-next-word-predictor
|
| 40 |
- **Demo:** [LSTM Next Word Predictor Demo](https://huggingface.co/spaces/aarohanverma/lstm-next-word-predictor-demo)
|
|
@@ -49,7 +49,7 @@ The model is trained on a dataset provided in CSV format (with a 'data' column)
|
|
| 49 |
|
| 50 |
This model can be directly used for next word prediction in text autocompletion.
|
| 51 |
|
| 52 |
-
### Downstream Use
|
| 53 |
|
| 54 |
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 55 |
|
|
@@ -153,7 +153,7 @@ To run the model, follow these steps:
|
|
| 153 |
|
| 154 |
- Interpretability techniques such as examining predicted token distributions can be applied to further understand model behavior.
|
| 155 |
|
| 156 |
-
## Technical Specifications
|
| 157 |
|
| 158 |
### Model Architecture and Objective
|
| 159 |
|
|
@@ -166,7 +166,16 @@ To run the model, follow these steps:
|
|
| 166 |
|
| 167 |
**BibTeX:**
|
| 168 |
|
| 169 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 170 |
|
| 171 |
## Model Card Contact
|
| 172 |
|
|
|
|
| 34 |
- **Language(s) (NLP):** English
|
| 35 |
- **License:** Apache-2.0
|
| 36 |
|
| 37 |
+
### Model Sources
|
| 38 |
|
| 39 |
- **Repository:** https://huggingface.co/aarohanverma/lstm-next-word-predictor
|
| 40 |
- **Demo:** [LSTM Next Word Predictor Demo](https://huggingface.co/spaces/aarohanverma/lstm-next-word-predictor-demo)
|
|
|
|
| 49 |
|
| 50 |
This model can be directly used for next word prediction in text autocompletion.
|
| 51 |
|
| 52 |
+
### Downstream Use
|
| 53 |
|
| 54 |
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 55 |
|
|
|
|
| 153 |
|
| 154 |
- Interpretability techniques such as examining predicted token distributions can be applied to further understand model behavior.
|
| 155 |
|
| 156 |
+
## Technical Specifications
|
| 157 |
|
| 158 |
### Model Architecture and Objective
|
| 159 |
|
|
|
|
| 166 |
|
| 167 |
**BibTeX:**
|
| 168 |
|
| 169 |
+
```bibtex
|
| 170 |
+
@misc {aarohan_verma_2025,
|
| 171 |
+
author = { {Aarohan Verma} },
|
| 172 |
+
title = { lstm-next-word-predictor },
|
| 173 |
+
year = 2025,
|
| 174 |
+
url = { https://huggingface.co/aarohanverma/lstm-next-word-predictor },
|
| 175 |
+
doi = { 10.57967/hf/4882 },
|
| 176 |
+
publisher = { Hugging Face }
|
| 177 |
+
}
|
| 178 |
+
```
|
| 179 |
|
| 180 |
## Model Card Contact
|
| 181 |
|