Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +83 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/adapter_model.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/model-00001-of-00002.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/model-00002-of-00002.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/adapter_model.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/model-00001-of-00002.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/model-00002-of-00002.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/adapter_model.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/model-00002-of-00002.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/adapter_model.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/training_args.bin +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/model-00002-of-00002.safetensors +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/tokenizer.json +3 -0
- FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/adapter_model.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/model-00004-of-00004.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/adapter_model.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/model-00004-of-00004.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/adapter_model.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/model-00004-of-00004.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/adapter_model.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/tokenizer.json +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/training_args.bin +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr/model-00004-of-00004.safetensors +3 -0
- FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr/tokenizer.json +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,86 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
FP_models/scalable/Qwen-Qwen2.5-3B-Instruct_scalable/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 51 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
FP_models/scalable/meta-llama-Llama-3.1-8B-Instruct_scalable/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 54 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 57 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 58 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 59 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 60 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
FP_models/scalable/meta-llama-Llama-3.2-1B-Instruct_scalable/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 63 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 64 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 65 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 66 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 67 |
+
FP_models/IF/meta-llama/Llama-3.2-1B-Instruct_IF/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 68 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 69 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 70 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 71 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 72 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 73 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 74 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 75 |
+
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 76 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 77 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 78 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 79 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 80 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 81 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 82 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 83 |
+
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 84 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 85 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 86 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 87 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 88 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 89 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 90 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 91 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ft-LucieFr-NoReg/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 92 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 93 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 94 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 95 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 96 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 97 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 98 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 99 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-LucieFr-Al4-OWT-TV/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 100 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 101 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 102 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 103 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 104 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 105 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 106 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 107 |
+
Grogros/dm-llama3.2-1BI-LucieFr-Al4-OWT-TV/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 108 |
+
Grogros/dmWM-Llama-3.1-8B-French/Dolly-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 109 |
+
Grogros/dmWM-Llama-3.1-8B-French/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 110 |
+
Grogros/dmWM-Llama-3.1-8B-French/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 111 |
+
Grogros/dmWM-Llama-3.1-8B-French/OpenMathInstruct-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 112 |
+
Grogros/dmWM-Llama-3.1-8B-French/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 113 |
+
Grogros/dmWM-Llama-3.1-8B-French/WildchatFr-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 114 |
+
Grogros/dmWM-Llama-3.1-8B-French/AlpacaGPT4-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 115 |
+
Grogros/dmWM-Llama-3.1-8B-French/Dolly/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 116 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ALL/AlpacaGPT4/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 117 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ALL/OpenMathInstruct/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 118 |
+
Grogros/dmWM-Qwen-Qwen2.5-3B-Instruct-ALL/WildchatFr/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a683bc584a717a6019165da46f0cc693bbadb0835ac36dbb033164f90c523625
|
| 3 |
+
size 239536272
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:960d8454a7ded595e485996ba6c065736dd7705d96397e1e138dd6023d094b79
|
| 3 |
+
size 6289
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/model-00001-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b60816c7a93f92ce6c4e2ea18a878a4c12c14cd254258c72c14f2532166ff51e
|
| 3 |
+
size 4957560304
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e5bd62389ca5a26bf2c42df90243af6450f650036fecfd38c06b590c8a11ae5
|
| 3 |
+
size 1214366696
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/AlpacaGPT4/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:264b171efed71e4acd4bbb5a39e5ed6609ad63d9d962650bbeda24c4e70e0020
|
| 3 |
+
size 6289
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9fee13e146731fc85983e5e02e0e05a0cf4e813c33e02199cb6b805b541b090
|
| 3 |
+
size 239536272
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0177c1f70dea40359640d3e554ee768e2d0eb5b0a3fe852a51fd23a7e43dd287
|
| 3 |
+
size 6289
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/model-00001-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e590e5645875eb75757312e8b27b001045409516e8a0f4cd946cf74a4c629772
|
| 3 |
+
size 4957560304
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dd97a6a3c5cded8da891323cd4e8a443a5462d72c5059b3c01c63d619c96b3dc
|
| 3 |
+
size 1214366696
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/Dolly/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e132aed90c522af88a45638b0fdb50385a8331428ffc2c280ddb00618e618fd1
|
| 3 |
+
size 6225
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2dd0e49d24de969b0d51cd433c82262da12aa7d47271c5493aa8563e373e23fa
|
| 3 |
+
size 239536272
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f29e30e2a219840081ef0c08002cd4994930216892bd66e5e01d6bc897e17a1a
|
| 3 |
+
size 6289
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:638461daba5ddb94becacbc3ed00002fff11b5dbe613d0175bf9c5d34f5aa25e
|
| 3 |
+
size 1214366696
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/OpenMathInstruct/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:68c5eddb4aaa0d454a1b2be8c89895e04f9d2b1783c4126cfd1e44cdf3dae92b
|
| 3 |
+
size 6289
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d1a3e5e128c85f5bb13368eda72c66b10a886f573f2af61b4b66092a5f941b7
|
| 3 |
+
size 239536272
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b2819807a1b29d6c344f76dba9977db787e5d3f1f89c79902723cfb1f68a24f5
|
| 3 |
+
size 6289
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:99fc4c9b8f32e43a4bf8f45bc7b6781213c968bbcb69fe7e4ebb89e187af3f89
|
| 3 |
+
size 1214366696
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
| 3 |
+
size 11421896
|
FP_models/IF/Qwen/Qwen2.5-3B-Instruct_IF/WildchatFr/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:252e1c3953ee1cfcb860261822b870232abf290ef09ea5772a9412c59a8c3a14
|
| 3 |
+
size 6289
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22013cec8b271cabbd46cdc994c1ae2c1e14fdbac472b611facff8a62ec37533
|
| 3 |
+
size 335604696
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e439fd408d1a9aff05eba13e1e47e1c6796a7f987a6586c67cc0611f89fe10a
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/model-00004-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c9aaaee14b268cab4641cb8268916d971d4916ffa8cc10a6a7fe68bbef0b1dc
|
| 3 |
+
size 1168138808
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/AlpacaGPT4/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f21cbf99d0c378d73e32383e94b1c589a54b71a0020d5d753d5a1fc7bbc175d2
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f61014b2557b84fc64920f2bc5dae84fb36b209572a05f86d26d731c85a527fd
|
| 3 |
+
size 335604696
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:051fbb0b63f5d4449001e328fc9589d63b1d8c0dc02fba0917cf9ae9af453014
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/model-00004-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d3af0f308831c36b67c8ef959346f79d5c37f09da413db1f50d4f420f8f5d3b6
|
| 3 |
+
size 1168138808
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/Dolly/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f96b531bc358ea1422497e324d926df111234b99765923a68aae0fc1ee0f614
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7090771fd02510cbec2afa824e9e51b25eeff9510591a4759b0a4da81027c0a5
|
| 3 |
+
size 335604696
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f4c772fcc4c62e1911c7d22a6f5198c0584c57aac742a9c4c014269c2d4e6e1
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/model-00004-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08caa8bd2e20e0cb5ade12d22225e8042911684d31fb02765913644d6f083ffd
|
| 3 |
+
size 1168138808
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/OpenMathInstruct/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fd7a370d5d0643fa629ff76b09fa5c60cd9d482a05e8068bbaa765dc58f6301c
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:268ebb9bc132dd4713ed2ea7d913125c00f64cd07ff5caa6a53ea6595dcf8e82
|
| 3 |
+
size 335604696
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr-lora/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5c4ff7aba26707f7e2d7fe7dd6aed0b4abf33d7c2b2b00c582b0f9b366f1f7d
|
| 3 |
+
size 6225
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr/model-00004-of-00004.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:481d48bba3816f9eba2509ab51c8dacd5800d81157b2c0de5821063b85c53c85
|
| 3 |
+
size 1168138808
|
FP_models/IF/meta-llama/Llama-3.1-8B-Instruct_IF/WildchatFr/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
|
| 3 |
+
size 17209920
|