MERA-evaluation commited on
Commit
9919bd3
·
verified ·
1 Parent(s): 373baa6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -828,7 +828,7 @@ MERA (Multimodal Evaluation for Russian-language Architectures) is a new open be
828
 
829
  *MERA benchmark brings together all industry and academic players in one place to study the capabilities of fundamental models, draw attention to AI problems, develop collaboration within the Russian Federation and in the international arena, and create an independent unified system for measuring all current models.*
830
 
831
- The benchmark covers 21 evaluation tasks comprising knowledge about the world, logic, reasoning, AI ethics, and other domains. Each task is supplied with a dataset and a human-level score on this task. NB that 4 datasets are diagnostic and not used in the overall model evaluation.
832
 
833
  ## MERA tasks & datasets
834
 
 
828
 
829
  *MERA benchmark brings together all industry and academic players in one place to study the capabilities of fundamental models, draw attention to AI problems, develop collaboration within the Russian Federation and in the international arena, and create an independent unified system for measuring all current models.*
830
 
831
+ The benchmark covers 23 evaluation tasks comprising knowledge about the world, logic, reasoning, AI ethics, and other domains. Each task is supplied with a dataset and a human-level score on this task. NB that 4 datasets are diagnostic and not used in the overall model evaluation.
832
 
833
  ## MERA tasks & datasets
834