Pushed by DataDreamer
Browse filesUpdate datadreamer.json
- datadreamer.json +79 -0
datadreamer.json
ADDED
|
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_card": {
|
| 3 |
+
"Date & Time": "2024-07-22T14:22:53.866930",
|
| 4 |
+
"Model Card": [
|
| 5 |
+
"https://huggingface.co/google/t5-v1_1-base"
|
| 6 |
+
],
|
| 7 |
+
"License Information": [
|
| 8 |
+
"apache-2.0"
|
| 9 |
+
],
|
| 10 |
+
"Citation Information": [
|
| 11 |
+
"\n@inproceedings{Wolf_Transformers_State-of-the-Art_Natural_2020,\n author = {Wolf, Thomas and Debut, Lysandre and Sanh, Victor and Chaumond, Julien",
|
| 12 |
+
"\n@Misc{peft,\n title = {PEFT: State-of-the-art Parameter-Efficient Fine-Tuning methods},\n author = {Sourab Mangrulkar and Sylvain Gugger and Lysandre Debut and Younes"
|
| 13 |
+
]
|
| 14 |
+
},
|
| 15 |
+
"data_card": {
|
| 16 |
+
"Generate Research Paper Abstracts": {
|
| 17 |
+
"Date & Time": "2024-07-22T13:38:20.754417",
|
| 18 |
+
"Model Name": [
|
| 19 |
+
"openai-community/gpt2"
|
| 20 |
+
],
|
| 21 |
+
"Model Card": [
|
| 22 |
+
"https://huggingface.co/openai-community/gpt2"
|
| 23 |
+
],
|
| 24 |
+
"License Information": [
|
| 25 |
+
"mit"
|
| 26 |
+
],
|
| 27 |
+
"Citation Information": [
|
| 28 |
+
"\n@inproceedings{Wolf_Transformers_State-of-the-Art_Natural_2020,\n author = {Wolf, Thomas and Debut, Lysandre and Sanh, Victor and Chaumond, Julien",
|
| 29 |
+
"@article{radford2019language,\n title={Language Models are Unsupervised Multitask Learners},\n author={Radford, Alec and Wu, Jeff and Child, Rewon and Luan, David and Amodei, Dario and Sutskever, Ilya},\n year={2019}\n}"
|
| 30 |
+
]
|
| 31 |
+
},
|
| 32 |
+
"Generate LinkedIn post from Abstracts": {
|
| 33 |
+
"Date & Time": "2024-07-22T13:41:35.938694",
|
| 34 |
+
"Model Name": [
|
| 35 |
+
"openai-community/gpt2"
|
| 36 |
+
],
|
| 37 |
+
"Model Card": [
|
| 38 |
+
"https://huggingface.co/openai-community/gpt2"
|
| 39 |
+
],
|
| 40 |
+
"License Information": [
|
| 41 |
+
"mit"
|
| 42 |
+
],
|
| 43 |
+
"Citation Information": [
|
| 44 |
+
"\n@inproceedings{Wolf_Transformers_State-of-the-Art_Natural_2020,\n author = {Wolf, Thomas and Debut, Lysandre and Sanh, Victor and Chaumond, Julien",
|
| 45 |
+
"@article{radford2019language,\n title={Language Models are Unsupervised Multitask Learners},\n author={Radford, Alec and Wu, Jeff and Child, Rewon and Luan, David and Amodei, Dario and Sutskever, Ilya},\n year={2019}\n}"
|
| 46 |
+
]
|
| 47 |
+
},
|
| 48 |
+
"Generate LinkedIn post from Abstracts (train split)": {
|
| 49 |
+
"Date & Time": "2024-07-22T13:41:35.979386"
|
| 50 |
+
}
|
| 51 |
+
},
|
| 52 |
+
"__version__": "0.28.0",
|
| 53 |
+
"datetime": "2024-07-22T13:41:38.886949",
|
| 54 |
+
"type": "TrainHFFineTune",
|
| 55 |
+
"name": "Train an Abstract => LinkedIn Post Model",
|
| 56 |
+
"version": 1.0,
|
| 57 |
+
"fingerprint": "39a75dc45318b92e",
|
| 58 |
+
"req_versions": {
|
| 59 |
+
"dill": "0.3.8",
|
| 60 |
+
"sqlitedict": "2.1.0",
|
| 61 |
+
"torch": "2.3.1",
|
| 62 |
+
"numpy": "1.26.4",
|
| 63 |
+
"transformers": "4.42.4",
|
| 64 |
+
"datasets": "2.20.0",
|
| 65 |
+
"huggingface_hub": "0.24.0",
|
| 66 |
+
"accelerate": "0.32.1",
|
| 67 |
+
"peft": "0.11.1",
|
| 68 |
+
"tiktoken": "0.7.0",
|
| 69 |
+
"tokenizers": "0.19.1",
|
| 70 |
+
"openai": "1.37.0",
|
| 71 |
+
"ctransformers": "0.2.27",
|
| 72 |
+
"optimum": "1.21.2",
|
| 73 |
+
"bitsandbytes": "0.42.0",
|
| 74 |
+
"litellm": "1.31.14",
|
| 75 |
+
"trl": "0.8.1",
|
| 76 |
+
"setfit": "1.0.3"
|
| 77 |
+
},
|
| 78 |
+
"interpreter": "3.10.10 (main, Jul 17 2023, 08:26:46) [Clang 14.0.3 (clang-1403.0.22.14.1)]"
|
| 79 |
+
}
|