omitakahiro commited on
Commit
c4e5ef1
·
verified ·
1 Parent(s): 10cf9e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md CHANGED
@@ -62,6 +62,43 @@ English data is sampled from [RedPajama-Data](https://github.com/togethercompute
62
  - Container: [Pytorch NGC Container](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/pytorch)
63
  - Library: [Megatron-LM](https://github.com/NVIDIA/Megatron-LM)
64
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
65
  ## License
66
 
67
  [MIT](https://opensource.org/licenses/MIT)
 
62
  - Container: [Pytorch NGC Container](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/pytorch)
63
  - Library: [Megatron-LM](https://github.com/NVIDIA/Megatron-LM)
64
 
65
+ ## Performance
66
+
67
+ **Stockmark Business Questions**
68
+
69
+ Dataset: https://huggingface.co/datasets/stockmark/business-questions
70
+
71
+ | model | accuracy |
72
+ |:---:|:---:|
73
+ |stockmark-100b-instruct| 0.90 |
74
+ |stockmark-13b-instruct| 0.80 |
75
+ |GPT-3.5-turbo[^1]| 0.42 |
76
+
77
+ [^1]: 0613
78
+
79
+ **Japanese Vicuna QA Benchmark**
80
+
81
+ We exclud categories that require calculation and coding, and use remaining 60 questions for evaluation.
82
+
83
+ GitHub: https://github.com/ku-nlp/ja-vicuna-qa-benchmark
84
+
85
+ | model | average score |
86
+ |:---:|:---:|
87
+ |stockmark-100b-instruct| 5.97 |
88
+ |tokyotech-llm/Swallow-70b-instruct-hf| 5.59 |
89
+ |GPT-3.5 (text-davinci-003)| 5.08 |
90
+
91
+ **Inference speed**
92
+
93
+ | model | time [s] for genrating 100 characters in Japanese |
94
+ |:---:|:---:|
95
+ |stockmark-100b-instruct| 1.86 |
96
+ | gpt-3.5-turbo | 2.15 |
97
+ | gpt-4-turbo | 5.48 |
98
+ |tokyotech-llm/Swallow-70b-instruct-hf| 2.22 |
99
+
100
+ For local LLMs, we measured the inference time using AWS Inferentia2.
101
+
102
  ## License
103
 
104
  [MIT](https://opensource.org/licenses/MIT)