Add Farsi(Persian) MTEB - FaMTEB

#6
This view is limited to 50 files because it contains too many changes. See the raw diff here.
Files changed (50) hide show
  1. .codecarbon.lock +0 -0
  2. .github/pull_request_template.md +15 -8
  3. .github/workflows/generate_cached_results.yml +0 -105
  4. .github/workflows/model-results-comparison.yaml +0 -89
  5. .github/workflows/stale_pr.yml +0 -20
  6. .github/workflows/test.yml +5 -17
  7. .gitignore +0 -6
  8. README.md +2 -1
  9. load_external.py +253 -0
  10. makefile +4 -3
  11. paths.json +0 -0
  12. pyproject.toml +2 -18
  13. results.py +0 -3
  14. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonCounterfactualVNClassification.json +0 -95
  15. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonPolarityVNClassification.json +0 -95
  16. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonReviewsVNClassification.json +0 -73
  17. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ArguAna-VN.json +0 -158
  18. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AskUbuntuDupQuestions-VN.json +0 -26
  19. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/BIOSSES-VN.json +0 -26
  20. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Banking77VNClassification.json +0 -73
  21. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackAndroid-VN.json +0 -54
  22. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackGis-VN.json +0 -54
  23. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackMathematica-VN.json +0 -54
  24. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackPhysics-VN.json +0 -54
  25. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackProgrammers-VN.json +0 -54
  26. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackStats-VN.json +0 -54
  27. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackTex-VN.json +0 -54
  28. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackUnix-VN.json +0 -54
  29. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWebmasters-VN.json +0 -54
  30. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWordpress-VN.json +0 -54
  31. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ClimateFEVER-VN.json +0 -54
  32. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/DBPedia-VN.json +0 -54
  33. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/EmotionVNClassification.json +0 -137
  34. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FEVER-VN.json +0 -54
  35. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FiQA2018-VN.json +0 -54
  36. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/GreenNodeTableMarkdownRetrieval.json +0 -158
  37. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/HotpotQA-VN.json +0 -54
  38. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ImdbVNClassification.json +0 -95
  39. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MSMARCO-VN.json +0 -54
  40. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPDomainVNClassification.json +0 -73
  41. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPIntentVNClassification.json +0 -73
  42. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveIntentVNClassification.json +0 -73
  43. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveScenarioVNClassification.json +0 -73
  44. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NFCorpus-VN.json +0 -54
  45. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NQ-VN.json +0 -54
  46. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Quora-VN.json +0 -54
  47. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClustering-VN.json +0 -32
  48. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClusteringP2P-VN.json +0 -32
  49. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SCIDOCS-VN.json +0 -54
  50. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SICK-R-VN.json +0 -26
.codecarbon.lock ADDED
File without changes
.github/pull_request_template.md CHANGED
@@ -1,11 +1,18 @@
1
- <!-- Description of the PR goes here -->
2
 
3
- ### Checklist
4
 
5
- - [ ] My model has a model sheet, report, or similar
6
- - [ ] My model has a reference implementation in [`mteb/models/model_implementations/`](https://github.com/embeddings-benchmark/mteb/tree/main/mteb/models/model_implementations), this can be as an API. Instruction on how to add a model can be found [here](https://embeddings-benchmark.github.io/mteb/contributing/adding_a_model/)
7
- - [ ] No, but there is an existing PR ___
8
- - [ ] The results submitted are obtained using the reference implementation
9
- - [ ] My model is available, either as a publicly accessible API or publicly on e.g., Huggingface
10
- - [ ] I *solemnly swear* that for all results submitted I have not trained on the evaluation dataset including training splits. If I have, I have disclosed it clearly.
11
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- If you are submitting a dataset or a model for the model registry, please use the corresponding checklists below. Otherwise, feel free to remove them. -->
2
 
3
+ <!-- add additional description, question etc. related to the new dataset -->
4
 
 
 
 
 
 
 
5
 
6
+ ## Checklist
7
+ <!-- Please do not delete this -->
8
+
9
+ - [ ] Run tests locally to make sure nothing is broken using `make test`.
10
+ - [ ] Run the results files checker `make pre-push`.
11
+
12
+ ### Adding a model checklist
13
+ <!--
14
+ When adding a model to the model registry
15
+ see also https://github.com/embeddings-benchmark/mteb/blob/main/docs/reproducible_workflow.md
16
+ -->
17
+
18
+ - [ ] I have added model implementation to `mteb/models/` [directory](https://github.com/embeddings-benchmark/mteb/tree/main/mteb/models). Instruction to add a model can be found [here](https://github.com/embeddings-benchmark/mteb/blob/main/docs/reproducible_workflow.md) in the following PR ____
.github/workflows/generate_cached_results.yml DELETED
@@ -1,105 +0,0 @@
1
- name: Generate Cached Results
2
-
3
- on:
4
- push:
5
- branches: [main]
6
- # Allow manual trigger for testing
7
- workflow_dispatch:
8
-
9
- jobs:
10
- generate-cache:
11
- runs-on: ubuntu-latest
12
- steps:
13
- - name: Free disk space
14
- run: |
15
- sudo rm -rf /usr/share/dotnet
16
- sudo rm -rf /opt/ghc
17
- sudo rm -rf /usr/local/share/boost
18
- docker system prune -af
19
-
20
- - name: Checkout repository
21
- uses: actions/checkout@v4
22
- with:
23
- fetch-depth: 0
24
- token: ${{ secrets.GITHUB_TOKEN }}
25
-
26
- - name: Setup Python
27
- uses: actions/setup-python@v5
28
- with:
29
- python-version: '3.10'
30
- cache: 'pip'
31
-
32
- - name: Install dependencies
33
- run: |
34
- pip install git+https://github.com/embeddings-benchmark/mteb.git
35
-
36
- - name: Generate cached results
37
- run: |
38
- python scripts/generate_cached_results.py
39
- env:
40
- PYTHONUNBUFFERED: 1
41
-
42
- - name: Configure Git
43
- run: |
44
- git config --global user.name "github-actions[bot]"
45
- git config --global user.email "github-actions[bot]@users.noreply.github.com"
46
-
47
- - name: Update cached-data branch
48
- env:
49
- GIT_LFS_SKIP_SMUDGE: "1"
50
- run: |
51
- # Check if __cached_results.json.gz was created
52
- if [ ! -f "__cached_results.json.gz" ]; then
53
- echo "❌ Cached results file not found"
54
- exit 1
55
- fi
56
-
57
- # Get file size for logging
58
- FILE_SIZE=$(stat -f%z __cached_results.json.gz 2>/dev/null || stat -c%s __cached_results.json.gz)
59
- echo "📦 Generated cache file: $(echo "scale=1; $FILE_SIZE/1024/1024" | bc -l)MB"
60
-
61
- # Move the generated file outside repository before rebuilding branch
62
- CACHE_TMP_PATH="$RUNNER_TEMP/__cached_results.json.gz"
63
- mv __cached_results.json.gz "$CACHE_TMP_PATH"
64
- README_TMP_PATH="$RUNNER_TEMP/cached-data-README.md"
65
-
66
- # Preserve README.md from cached-data branch (if it exists) without modifications
67
- if git ls-remote --exit-code --heads origin cached-data >/dev/null 2>&1; then
68
- git fetch origin cached-data:refs/remotes/origin/cached-data
69
- if git cat-file -e origin/cached-data:README.md 2>/dev/null; then
70
- git show origin/cached-data:README.md > "$README_TMP_PATH"
71
- fi
72
- fi
73
-
74
- # Rebuild cached-data as a fresh orphan branch so history always has one commit
75
- git checkout --orphan cached-data
76
- git rm -rf . 2>/dev/null || true
77
-
78
- # Restore and stage cache file
79
- mv "$CACHE_TMP_PATH" __cached_results.json.gz
80
- git add __cached_results.json.gz
81
-
82
- # Restore preserved README.md exactly as it was in cached-data
83
- if [ -f "$README_TMP_PATH" ]; then
84
- cp "$README_TMP_PATH" README.md
85
- git add README.md
86
- fi
87
-
88
- # Commit with timestamp and file size
89
- TIMESTAMP=$(date -u '+%Y-%m-%d %H:%M:%S UTC')
90
- COMMIT_MSG="Update cached results - $TIMESTAMP ($(echo "scale=1; $FILE_SIZE/1024/1024" | bc -l)MB)"
91
- git commit -m "$COMMIT_MSG"
92
-
93
- # Force push keeps remote branch as a single-commit history
94
- git push origin cached-data --force
95
- echo "✅ Successfully replaced cached-data branch with a single commit"
96
-
97
- - name: Report status
98
- if: always()
99
- run: |
100
- if [ -f "__cached_results.json.gz" ]; then
101
- FILE_SIZE=$(stat -f%z __cached_results.json.gz 2>/dev/null || stat -c%s __cached_results.json.gz)
102
- echo "✅ Workflow completed. Cache file size: $(echo "scale=1; $FILE_SIZE/1024/1024" | bc -l)MB"
103
- else
104
- echo "❌ Workflow failed - no cache file generated"
105
- fi
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.github/workflows/model-results-comparison.yaml DELETED
@@ -1,89 +0,0 @@
1
- name: Model Results Comparison
2
-
3
- on:
4
- pull_request_target:
5
- types: [opened, synchronize, edited]
6
- paths:
7
- - 'results/**/*.json'
8
- workflow_dispatch:
9
- inputs:
10
- reference_models:
11
- description: 'Space-separated list of reference models for comparison'
12
- required: true
13
- type: string
14
- default: 'intfloat/multilingual-e5-large google/gemini-embedding-001'
15
- pull_request_number:
16
- description: 'The pull request number to comment on (required if triggered manually)'
17
- required: false
18
- type: string
19
-
20
- permissions:
21
- contents: read
22
- pull-requests: write
23
-
24
- jobs:
25
- compare-results:
26
- runs-on: ubuntu-latest
27
-
28
- steps:
29
- - name: Free disk space
30
- run: |
31
- sudo rm -rf /usr/share/dotnet
32
- sudo rm -rf /opt/ghc
33
- sudo rm -rf /usr/local/share/boost
34
- docker system prune -af
35
-
36
- - name: Checkout code
37
- uses: actions/checkout@v4
38
- with:
39
- # IMPORTANT: For pull_request_target, check out the PR branch explicitly
40
- ref: ${{ github.event.pull_request.head.sha }}
41
- fetch-depth: 0
42
-
43
- - name: Fetch origin main
44
- run: git fetch origin main
45
-
46
- - name: Install uv and set the Python ${{ matrix.python-version }} version
47
- uses: astral-sh/setup-uv@v7
48
- with:
49
- python-version: ${{ matrix.python-version }}
50
-
51
- - name: Install dependencies
52
- shell: bash
53
- run: |
54
- make install-for-tests
55
-
56
- - name: Generate model comparison
57
- env:
58
- REFERENCE_MODELS: ${{ github.event.inputs.reference_models || 'intfloat/multilingual-e5-large google/gemini-embedding-001' }}
59
- run: |
60
- uv run python scripts/create_pr_results_comment.py --reference-models $REFERENCE_MODELS --output model-comparison.md
61
-
62
- - name: Upload comparison report
63
- uses: actions/upload-artifact@v4
64
- with:
65
- name: model-comparison
66
- path: model-comparison.md
67
-
68
- - name: Determine PR Number
69
- id: pr_info
70
- run: |
71
- if [ "${{ github.event_name }}" == "pull_request_target" ]; then
72
- echo "pr_number=${{ github.event.number }}" >> $GITHUB_OUTPUT
73
- elif [ "${{ github.event_name }}" == "workflow_dispatch" ] && [ -n "${{ github.event.inputs.pull_request_number }}" ]; then
74
- echo "pr_number=${{ github.event.inputs.pull_request_number }}" >> $GITHUB_OUTPUT
75
- else
76
- echo "pr_number=" >> $GITHUB_OUTPUT
77
- fi
78
-
79
- - name: Post PR comment
80
- if: steps.pr_info.outputs.pr_number != ''
81
- env:
82
- GITHUB_TOKEN: ${{ github.token }}
83
- run: |
84
- echo "Truncating model-comparison.md to fit GitHub API limit..."
85
- head -c 65300 model-comparison.md > model-comparison-truncated.md
86
- echo -e "\n\n---\n**Note:** Content truncated due to GitHub API limits. See the full report in the workflow artifacts." >> model-comparison-truncated.md
87
- mv model-comparison-truncated.md model-comparison.md
88
-
89
- gh pr comment ${{ steps.pr_info.outputs.pr_number }} --body-file model-comparison.md --create-if-none --edit-last
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.github/workflows/stale_pr.yml DELETED
@@ -1,20 +0,0 @@
1
- name: Label and Close stale PRs
2
- on:
3
- schedule:
4
- - cron: '30 1 * * *' # runs daily at 1:30 AM
5
- jobs:
6
- stale:
7
- runs-on: ubuntu-latest
8
- permissions:
9
- actions: write
10
- issues: write
11
- pull-requests: write
12
- steps:
13
- - uses: actions/stale@v9.1.0
14
- with:
15
- stale-pr-label: 'stale'
16
- stale-issue-label: 'stale'
17
- days-before-pr-stale: 14
18
- days-before-pr-close: 7
19
- stale-pr-message: 'This pull request has been automatically marked as stale due to inactivity.'
20
- close-pr-message: 'This pull request has been automatically closed due to inactivity.'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.github/workflows/test.yml CHANGED
@@ -2,6 +2,7 @@
2
  # 1) install Python dependencies
3
  # 2) run make test
4
 
 
5
  name: Test
6
  on:
7
  push:
@@ -15,24 +16,13 @@ jobs:
15
  fail-fast: false
16
  matrix:
17
  os: [ubuntu-latest] #, macos-latest, windows-latest]
18
- python-version: ["3.10"]
19
  steps:
20
- - name: Free disk space
21
- run: |
22
- sudo rm -rf /usr/share/dotnet
23
- sudo rm -rf /opt/ghc
24
- sudo rm -rf /usr/local/share/boost
25
- docker system prune -af
26
-
27
  - uses: actions/checkout@v3
28
- with:
29
- fetch-depth: 0 # to allow us to examine the git diff
30
-
31
- - name: Install uv and set the Python ${{ matrix.python-version }} version
32
- uses: astral-sh/setup-uv@v7
33
- with:
34
- python-version: ${{ matrix.python-version }}
35
 
 
 
 
36
  - name: Install dependencies
37
  shell: bash
38
  run: |
@@ -40,7 +30,5 @@ jobs:
40
 
41
  - name: Run tests
42
  shell: bash
43
- env:
44
- PR_BASE_SHA: ${{ github.event.pull_request.base.sha }}
45
  run: |
46
  make test
 
2
  # 1) install Python dependencies
3
  # 2) run make test
4
 
5
+
6
  name: Test
7
  on:
8
  push:
 
16
  fail-fast: false
17
  matrix:
18
  os: [ubuntu-latest] #, macos-latest, windows-latest]
19
+ python-version: ["3.9"]
20
  steps:
 
 
 
 
 
 
 
21
  - uses: actions/checkout@v3
 
 
 
 
 
 
 
22
 
23
+ - name: Setup Python ${{ matrix.python-version }}
24
+ uses: actions/setup-python@v4
25
+
26
  - name: Install dependencies
27
  shell: bash
28
  run: |
 
30
 
31
  - name: Run tests
32
  shell: bash
 
 
33
  run: |
34
  make test
.gitignore CHANGED
@@ -20,9 +20,3 @@ tmp.py
20
 
21
  # UV
22
  uv.lock
23
-
24
- build/
25
- results.egg-info/
26
-
27
- # codecarbon
28
- .codecarbon.lock
 
20
 
21
  # UV
22
  uv.lock
 
 
 
 
 
 
README.md CHANGED
@@ -5,7 +5,8 @@ submission_name: MTEB
5
  ---
6
 
7
  > [!NOTE]
8
- > Previously, it was possible to submit model results to MTEB by adding them to the metadata of the model card on huggingface. However, this is no longer possible as we want to ensure that we can match the results with the model implementation. If you want to add your model, please follow the [guide](https://embeddings-benchmark.github.io/mteb/contributing/adding_a_model/) on how to do so.
 
9
 
10
  This repository contains the results of the embedding benchmark evaluated using the package `mteb`.
11
 
 
5
  ---
6
 
7
  > [!NOTE]
8
+ > Previously it was possible to submit models results to MTEB by adding the results to the model metadata. This is no longer an option as we want to ensure high quality metadata.
9
+
10
 
11
  This repository contains the results of the embedding benchmark evaluated using the package `mteb`.
12
 
load_external.py ADDED
@@ -0,0 +1,253 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __future__ import annotations
2
+
3
+ import json
4
+ import logging
5
+ import math
6
+ import re
7
+ from pathlib import Path
8
+ from typing import Any
9
+
10
+ from huggingface_hub import HfApi, get_hf_file_metadata, hf_hub_download, hf_hub_url
11
+ from huggingface_hub.errors import NotASafetensorsRepoError
12
+ from huggingface_hub.hf_api import ModelInfo
13
+ from huggingface_hub.repocard import metadata_load
14
+ from mteb import ModelMeta, get_task
15
+
16
+ API = HfApi()
17
+ logger = logging.getLogger(__name__)
18
+
19
+
20
+ library_mapping = {
21
+ "sentence-transformers": "Sentence Transformers",
22
+ }
23
+
24
+
25
+ def get_model_dir(model_id: str) -> Path:
26
+ external_result_dir = Path("results") / model_id.replace("/", "__") / "external"
27
+ return external_result_dir
28
+
29
+
30
+ renamed_tasks = {
31
+ "NorwegianParliament": "NorwegianParliamentClassification",
32
+ "CMedQAv2": "CMedQAv2-reranking",
33
+ "CMedQAv1": "CMedQAv1-reranking",
34
+ "8TagsClustering": "EightTagsClustering",
35
+ "PPC": "PpcPC",
36
+ "PawsX": "PawsXParaphraseIdentification",
37
+ }
38
+
39
+
40
+ def simplify_dataset_name(name: str) -> str:
41
+ task_name = name.replace("MTEB ", "").split()[0]
42
+ return renamed_tasks.get(task_name, task_name)
43
+
44
+
45
+ def get_model_parameters_memory(model_info: ModelInfo) -> tuple[int| None, float|None]:
46
+ try:
47
+ safetensors = API.get_safetensors_metadata(model_info.id)
48
+ num_parameters = sum(safetensors.parameter_count.values())
49
+ return num_parameters, round(num_parameters * 4 / 1024 ** 3, 2)
50
+ except NotASafetensorsRepoError as e:
51
+ logger.info(f"Could not find SafeTensors metadata for {model_info.id}")
52
+
53
+ filenames = [sib.rfilename for sib in model_info.siblings]
54
+ if "pytorch_model.bin" in filenames:
55
+ url = hf_hub_url(model_info.id, filename="pytorch_model.bin")
56
+ meta = get_hf_file_metadata(url)
57
+ bytes_per_param = 4
58
+ num_params = round(meta.size / bytes_per_param)
59
+ size_gb = round(meta.size * (4 / bytes_per_param) / 1024 ** 3, 2)
60
+ return num_params, size_gb
61
+ if "pytorch_model.bin.index.json" in filenames:
62
+ index_path = hf_hub_download(model_info.id, filename="pytorch_model.bin.index.json")
63
+ size = json.load(open(index_path))
64
+ bytes_per_param = 4
65
+ if "metadata" in size and "total_size" in size["metadata"]:
66
+ return round(size["metadata"]["total_size"] / bytes_per_param), round(size["metadata"]["total_size"] / 1024 ** 3, 2)
67
+ logger.info(f"Could not find the model parameters for {model_info.id}")
68
+ return None, None
69
+
70
+
71
+ def get_dim_seq_size(model: ModelInfo) -> tuple[str | None, str | None, int, float]:
72
+ siblings = model.siblings or []
73
+ filenames = [sib.rfilename for sib in siblings]
74
+ dim, seq = None, None
75
+ for filename in filenames:
76
+ if re.match(r"\d+_Pooling/config.json", filename):
77
+ st_config_path = hf_hub_download(model.id, filename=filename)
78
+ dim = json.load(open(st_config_path)).get("word_embedding_dimension", None)
79
+ break
80
+ for filename in filenames:
81
+ if re.match(r"\d+_Dense/config.json", filename):
82
+ st_config_path = hf_hub_download(model.id, filename=filename)
83
+ dim = json.load(open(st_config_path)).get("out_features", dim)
84
+ if "config.json" in filenames:
85
+ config_path = hf_hub_download(model.id, filename="config.json")
86
+ config = json.load(open(config_path))
87
+ if not dim:
88
+ dim = config.get("hidden_dim", config.get("hidden_size", config.get("d_model", None)))
89
+ seq = config.get("n_positions", config.get("max_position_embeddings", config.get("n_ctx", config.get("seq_length", None))))
90
+
91
+ parameters, memory = get_model_parameters_memory(model)
92
+ return dim, seq, parameters, memory
93
+
94
+
95
+ def create_model_meta(model_info: ModelInfo) -> ModelMeta | None:
96
+ readme_path = hf_hub_download(model_info.id, filename="README.md", etag_timeout=30)
97
+ meta = metadata_load(readme_path)
98
+ dim, seq, parameters, memory = None, None, None, None
99
+ try:
100
+ dim, seq, parameters, memory = get_dim_seq_size(model_info)
101
+ except Exception as e:
102
+ logger.error(f"Error getting model parameters for {model_info.id}, {e}")
103
+
104
+ release_date = str(model_info.created_at.date()) if model_info.created_at else ""
105
+ library = [library_mapping[model_info.library_name]] if model_info.library_name in library_mapping else []
106
+ languages = meta.get("language", [])
107
+ if not isinstance(languages, list) and isinstance(languages, str):
108
+ languages = [languages]
109
+ # yaml transforms norwegian `no` to False
110
+ for i in range(len(languages)):
111
+ if languages[i] is False:
112
+ languages[i] = "no"
113
+
114
+ model_meta = ModelMeta(
115
+ name=model_info.id,
116
+ revision=model_info.sha,
117
+ release_date=release_date,
118
+ open_weights=True,
119
+ framework=library,
120
+ license=meta.get("license", None),
121
+ embed_dim=dim,
122
+ max_tokens=seq,
123
+ n_parameters=parameters,
124
+ languages=languages,
125
+ )
126
+ return model_meta
127
+
128
+
129
+ def parse_readme(model_info: ModelInfo) -> dict[str, dict[str, Any]] | None:
130
+ model_id = model_info.id
131
+ try:
132
+ readme_path = hf_hub_download(model_info.id, filename="README.md", etag_timeout=30)
133
+ except Exception:
134
+ logger.warning(f"ERROR: Could not fetch metadata for {model_id}, trying again")
135
+ readme_path = hf_hub_download(model_id, filename="README.md", etag_timeout=30)
136
+ meta = metadata_load(readme_path)
137
+ if "model-index" not in meta:
138
+ logger.info(f"Could not find model-index in {model_id}")
139
+ return
140
+ model_index = meta["model-index"][0]
141
+ model_name_from_readme = model_index.get("name", None)
142
+ orgs = ["Alibaba-NLP", "HIT-TMG", "McGill-NLP", "Snowflake", "facebook", "jinaai", "nomic-ai"]
143
+ is_org = any([model_id.startswith(org) for org in orgs])
144
+ # There a lot of reuploads with tunes, quantization, etc. We only want the original model
145
+ # to prevent this most of the time we can check if the model name from the readme is the same as the model id
146
+ # but some orgs have a different naming in their readme
147
+ if model_name_from_readme and not model_info.id.endswith(model_name_from_readme) and not is_org:
148
+ logger.warning(f"Model name mismatch: {model_info.id} vs {model_name_from_readme}")
149
+ return
150
+ results = model_index.get("results", [])
151
+ model_results = {}
152
+ for result in results:
153
+ dataset = result["dataset"]
154
+ dataset_type = simplify_dataset_name(dataset["name"])
155
+
156
+ if dataset_type not in model_results:
157
+ output_dict = {
158
+ "dataset_revision": dataset.get("revision", ""),
159
+ "task_name": simplify_dataset_name(dataset["name"]),
160
+ "evaluation_time": None,
161
+ "mteb_version": None,
162
+ "scores": {},
163
+ }
164
+ else:
165
+ output_dict = model_results[dataset_type]
166
+
167
+ try:
168
+ mteb_task = get_task(output_dict["task_name"])
169
+ except Exception:
170
+ logger.warning(f"Error getting task for {model_id} {output_dict['task_name']}")
171
+ continue
172
+
173
+ mteb_task_metadata = mteb_task.metadata
174
+ mteb_task_eval_languages = mteb_task_metadata.eval_langs
175
+
176
+ scores_dict = output_dict["scores"]
177
+ current_split = dataset["split"]
178
+ current_config = dataset.get("config", "")
179
+ cur_split_metrics = {
180
+ "hf_subset": current_config,
181
+ "languages": mteb_task_eval_languages if isinstance(mteb_task_eval_languages, list) else mteb_task_eval_languages.get(current_config, ["None"]),
182
+ }
183
+ for metric in result["metrics"]:
184
+ if isinstance(metric["value"], (float, int)):
185
+ cur_split_metrics[metric["type"]] = metric["value"] / 100
186
+ else:
187
+ cur_split_metrics[metric["type"]] = metric["value"]
188
+
189
+ main_score_str = "main_score"
190
+ if main_score_str not in cur_split_metrics:
191
+ # old sts and sum_eval have cos_sim_pearson, but in model_meta cosine_spearman is main_score
192
+ for old_metric, new_metric in zip(["cos_sim_pearson", "cos_sim_spearman"], ["cosine_pearson", "cosine_spearman"]):
193
+ if old_metric in cur_split_metrics:
194
+ cur_split_metrics[new_metric] = cur_split_metrics[old_metric]
195
+
196
+ if mteb_task.metadata.main_score not in cur_split_metrics:
197
+ logger.warning(f"Could not find main score for {model_id} {output_dict['task_name']}, mteb task {mteb_task.metadata.name}. Main score: {mteb_task.metadata.main_score}. Metrics: {cur_split_metrics}, result {result['metrics']}")
198
+ continue
199
+
200
+ cur_split_metrics[main_score_str] = cur_split_metrics.get(mteb_task.metadata.main_score, None)
201
+ split_metrics = scores_dict.get(current_split, [])
202
+ split_metrics.append(cur_split_metrics)
203
+ scores_dict[current_split] = split_metrics
204
+ model_results[dataset_type] = output_dict
205
+ return model_results
206
+
207
+
208
+ def get_mteb_data() -> None:
209
+ models = sorted(list(API.list_models(filter="mteb", full=True)), key=lambda x: x.id)
210
+ # models = [model for model in models if model.id == "ai-forever/ru-en-RoSBERTa"]
211
+ for i, model_info in enumerate(models, start=1):
212
+ logger.info(f"[{i}/{len(models)}] Processing {model_info.id}")
213
+ model_path = get_model_dir(model_info.id)
214
+ if (model_path / "model_meta.json").exists() and len(list(model_path.glob("*.json"))) > 1:
215
+ logger.info(f"Model meta already exists for {model_info.id}")
216
+ # continue
217
+ if model_info.id.lower().endswith("gguf"):
218
+ logger.info(f"Skipping {model_info.id} GGUF model")
219
+ continue
220
+
221
+ spam_users = ["ILKT", "fine-tuned", "mlx-community"]
222
+ is_spam = False
223
+ for spam_user in spam_users:
224
+ if model_info.id.startswith(spam_user):
225
+ logger.info(f"Skipping {model_info.id}")
226
+ is_spam = True
227
+ continue
228
+ if is_spam:
229
+ continue
230
+ model_meta = create_model_meta(model_info)
231
+ model_results = parse_readme(model_info)
232
+
233
+ if not model_meta or not model_results:
234
+ logger.warning(f"Could not get model meta or results for {model_info.id}")
235
+ continue
236
+
237
+ if not model_path.exists():
238
+ model_path.mkdir(parents=True, exist_ok=True)
239
+
240
+ model_meta_path = model_path / "model_meta.json"
241
+ with model_meta_path.open("w") as f:
242
+ json.dump(model_meta.model_dump(), f, indent=4)
243
+
244
+ for model_result in model_results:
245
+ task_name = model_results[model_result]["task_name"]
246
+ result_file = model_path / f"{task_name}.json"
247
+ with result_file.open("w") as f:
248
+ json.dump(model_results[model_result], f, indent=4)
249
+
250
+
251
+ if __name__ == "__main__":
252
+ logging.basicConfig(level=logging.INFO)
253
+ get_mteb_data()
makefile CHANGED
@@ -1,11 +1,12 @@
1
  install-for-tests:
2
  @echo "--- Installing dependencies for tests ---"
3
- uv sync --group dev --group pr-comment
 
4
 
5
  test:
6
  @echo "--- Running tests ---"
7
- uv run --no-sync pytest
8
 
9
  pre-push:
10
  @echo "--- Running pre-push commands ---"
11
- python reduce_large_json_files.py
 
1
  install-for-tests:
2
  @echo "--- Installing dependencies for tests ---"
3
+ pip install pip --upgrade
4
+ pip install .
5
 
6
  test:
7
  @echo "--- Running tests ---"
8
+ pytest
9
 
10
  pre-push:
11
  @echo "--- Running pre-push commands ---"
12
+ python reduce_large_json_files.py
paths.json CHANGED
The diff for this file is too large to render. See raw diff
 
pyproject.toml CHANGED
@@ -3,21 +3,5 @@ name = "results"
3
  version = "0.1.0"
4
  description = "The result repository for mteb"
5
  readme = "README.md"
6
- requires-python = ">=3.10"
7
- dependencies = [
8
- "mteb>=2.0",
9
- ]
10
-
11
- [dependency-groups]
12
- dev = [
13
- "pytest>=8.3.4",
14
- ]
15
- lint = [
16
- "ruff>=0.14.9",
17
- ]
18
- pr-comment = [
19
- "tabulate>=0.9.0",
20
- ]
21
-
22
- [tool.uv.sources]
23
- mteb = { git = "https://github.com/embeddings-benchmark/mteb" }
 
3
  version = "0.1.0"
4
  description = "The result repository for mteb"
5
  readme = "README.md"
6
+ requires-python = ">=3.9"
7
+ dependencies = ["mteb[dev]>=1.13.0"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results.py CHANGED
@@ -84,7 +84,6 @@ EVAL_LANGS = [
84
  "est-eng",
85
  "eus-eng",
86
  "fa",
87
- "fas-Arab",
88
  "fao-eng",
89
  "fi",
90
  "fin-eng",
@@ -241,7 +240,6 @@ VALIDATION_SPLIT = [
241
  "LEMBSummScreenFDRetrieval",
242
  "MSMARCO",
243
  "MSMARCO-PL",
244
- "MSMARCO-Fa",
245
  "MultilingualSentiment",
246
  "Ocnli",
247
  "TNews",
@@ -257,7 +255,6 @@ DEV_SPLIT = [
257
  "MMarcoRetrieval",
258
  "MSMARCO",
259
  "MSMARCO-PL",
260
- "MSMARCO-Fa",
261
  "T2Reranking",
262
  "T2Retrieval",
263
  "VideoRetrieval",
 
84
  "est-eng",
85
  "eus-eng",
86
  "fa",
 
87
  "fao-eng",
88
  "fi",
89
  "fin-eng",
 
240
  "LEMBSummScreenFDRetrieval",
241
  "MSMARCO",
242
  "MSMARCO-PL",
 
243
  "MultilingualSentiment",
244
  "Ocnli",
245
  "TNews",
 
255
  "MMarcoRetrieval",
256
  "MSMARCO",
257
  "MSMARCO-PL",
 
258
  "T2Reranking",
259
  "T2Retrieval",
260
  "VideoRetrieval",
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonCounterfactualVNClassification.json DELETED
@@ -1,95 +0,0 @@
1
- {
2
- "dataset_revision": "b48bc27d383cfca5b6a47135a52390fa5f66b253",
3
- "task_name": "AmazonCounterfactualVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.619742,
9
- "f1": 0.561867,
10
- "f1_weighted": 0.660984,
11
- "ap": 0.257592,
12
- "ap_weighted": 0.257592,
13
- "scores_per_experiment": [
14
- {
15
- "accuracy": 0.587983,
16
- "f1": 0.555485,
17
- "f1_weighted": 0.631829,
18
- "ap": 0.276246,
19
- "ap_weighted": 0.276246
20
- },
21
- {
22
- "accuracy": 0.736052,
23
- "f1": 0.655812,
24
- "f1_weighted": 0.761372,
25
- "ap": 0.318313,
26
- "ap_weighted": 0.318313
27
- },
28
- {
29
- "accuracy": 0.652361,
30
- "f1": 0.594521,
31
- "f1_weighted": 0.691796,
32
- "ap": 0.280129,
33
- "ap_weighted": 0.280129
34
- },
35
- {
36
- "accuracy": 0.51073,
37
- "f1": 0.482958,
38
- "f1_weighted": 0.559073,
39
- "ap": 0.224991,
40
- "ap_weighted": 0.224991
41
- },
42
- {
43
- "accuracy": 0.622318,
44
- "f1": 0.539112,
45
- "f1_weighted": 0.6635,
46
- "ap": 0.219724,
47
- "ap_weighted": 0.219724
48
- },
49
- {
50
- "accuracy": 0.583691,
51
- "f1": 0.53557,
52
- "f1_weighted": 0.630528,
53
- "ap": 0.241011,
54
- "ap_weighted": 0.241011
55
- },
56
- {
57
- "accuracy": 0.690987,
58
- "f1": 0.626203,
59
- "f1_weighted": 0.725049,
60
- "ap": 0.302755,
61
- "ap_weighted": 0.302755
62
- },
63
- {
64
- "accuracy": 0.594421,
65
- "f1": 0.549602,
66
- "f1_weighted": 0.639849,
67
- "ap": 0.255337,
68
- "ap_weighted": 0.255337
69
- },
70
- {
71
- "accuracy": 0.618026,
72
- "f1": 0.556143,
73
- "f1_weighted": 0.661415,
74
- "ap": 0.244741,
75
- "ap_weighted": 0.244741
76
- },
77
- {
78
- "accuracy": 0.600858,
79
- "f1": 0.523266,
80
- "f1_weighted": 0.645433,
81
- "ap": 0.212676,
82
- "ap_weighted": 0.212676
83
- }
84
- ],
85
- "main_score": 0.619742,
86
- "hf_subset": "default",
87
- "languages": [
88
- "vie-Latn"
89
- ]
90
- }
91
- ]
92
- },
93
- "evaluation_time": 1.7541632652282715,
94
- "kg_co2_emissions": null
95
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonPolarityVNClassification.json DELETED
@@ -1,95 +0,0 @@
1
- {
2
- "dataset_revision": "4e9a0d6e6bd97ab32f23c50c043d751eed2a5f8a",
3
- "task_name": "AmazonPolarityVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.887758,
9
- "f1": 0.887555,
10
- "f1_weighted": 0.88755,
11
- "ap": 0.85126,
12
- "ap_weighted": 0.85126,
13
- "scores_per_experiment": [
14
- {
15
- "accuracy": 0.887756,
16
- "f1": 0.887564,
17
- "f1_weighted": 0.88753,
18
- "ap": 0.86203,
19
- "ap_weighted": 0.86203
20
- },
21
- {
22
- "accuracy": 0.886623,
23
- "f1": 0.886583,
24
- "f1_weighted": 0.886599,
25
- "ap": 0.840929,
26
- "ap_weighted": 0.840929
27
- },
28
- {
29
- "accuracy": 0.887771,
30
- "f1": 0.887457,
31
- "f1_weighted": 0.887413,
32
- "ap": 0.86644,
33
- "ap_weighted": 0.86644
34
- },
35
- {
36
- "accuracy": 0.879607,
37
- "f1": 0.879353,
38
- "f1_weighted": 0.879312,
39
- "ap": 0.852779,
40
- "ap_weighted": 0.852779
41
- },
42
- {
43
- "accuracy": 0.894125,
44
- "f1": 0.894014,
45
- "f1_weighted": 0.893989,
46
- "ap": 0.86726,
47
- "ap_weighted": 0.86726
48
- },
49
- {
50
- "accuracy": 0.870804,
51
- "f1": 0.870803,
52
- "f1_weighted": 0.870801,
53
- "ap": 0.827334,
54
- "ap_weighted": 0.827334
55
- },
56
- {
57
- "accuracy": 0.896821,
58
- "f1": 0.896763,
59
- "f1_weighted": 0.896744,
60
- "ap": 0.86776,
61
- "ap_weighted": 0.86776
62
- },
63
- {
64
- "accuracy": 0.901696,
65
- "f1": 0.901691,
66
- "f1_weighted": 0.901696,
67
- "ap": 0.863676,
68
- "ap_weighted": 0.863676
69
- },
70
- {
71
- "accuracy": 0.90081,
72
- "f1": 0.900772,
73
- "f1_weighted": 0.900787,
74
- "ap": 0.858558,
75
- "ap_weighted": 0.858558
76
- },
77
- {
78
- "accuracy": 0.871574,
79
- "f1": 0.870546,
80
- "f1_weighted": 0.870631,
81
- "ap": 0.805833,
82
- "ap_weighted": 0.805833
83
- }
84
- ],
85
- "main_score": 0.887758,
86
- "hf_subset": "default",
87
- "languages": [
88
- "vie-Latn"
89
- ]
90
- }
91
- ]
92
- },
93
- "evaluation_time": 1072.2194910049438,
94
- "kg_co2_emissions": null
95
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonReviewsVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "27da94deb6d4f44af789a3d70750fa506b79f189",
3
- "task_name": "AmazonReviewsVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.444801,
9
- "f1": 0.426392,
10
- "f1_weighted": 0.432322,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.450058,
14
- "f1": 0.42758,
15
- "f1_weighted": 0.434702
16
- },
17
- {
18
- "accuracy": 0.470502,
19
- "f1": 0.448032,
20
- "f1_weighted": 0.455157
21
- },
22
- {
23
- "accuracy": 0.436624,
24
- "f1": 0.42657,
25
- "f1_weighted": 0.430991
26
- },
27
- {
28
- "accuracy": 0.451811,
29
- "f1": 0.435001,
30
- "f1_weighted": 0.441204
31
- },
32
- {
33
- "accuracy": 0.462325,
34
- "f1": 0.433835,
35
- "f1_weighted": 0.440433
36
- },
37
- {
38
- "accuracy": 0.439836,
39
- "f1": 0.413124,
40
- "f1_weighted": 0.419804
41
- },
42
- {
43
- "accuracy": 0.418808,
44
- "f1": 0.415917,
45
- "f1_weighted": 0.420269
46
- },
47
- {
48
- "accuracy": 0.463785,
49
- "f1": 0.457309,
50
- "f1_weighted": 0.461625
51
- },
52
- {
53
- "accuracy": 0.439252,
54
- "f1": 0.428232,
55
- "f1_weighted": 0.433021
56
- },
57
- {
58
- "accuracy": 0.415012,
59
- "f1": 0.378321,
60
- "f1_weighted": 0.386011
61
- }
62
- ],
63
- "main_score": 0.444801,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 11.328449964523315,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ArguAna-VN.json DELETED
@@ -1,158 +0,0 @@
1
- {
2
- "dataset_revision": "2a5133a05d7430e6f353497b1624a6e73148105b",
3
- "task_name": "ArguAna-VN",
4
- "mteb_version": "1.38.39",
5
- "scores": {
6
- "test": [
7
- {
8
- "ndcg_at_1": 0.00154,
9
- "ndcg_at_3": 0.25091,
10
- "ndcg_at_5": 0.3202,
11
- "ndcg_at_10": 0.37932,
12
- "ndcg_at_20": 0.40551,
13
- "ndcg_at_100": 0.42317,
14
- "ndcg_at_1000": 0.42594,
15
- "map_at_1": 0.00154,
16
- "map_at_3": 0.18906,
17
- "map_at_5": 0.22748,
18
- "map_at_10": 0.25228,
19
- "map_at_20": 0.25969,
20
- "map_at_100": 0.26227,
21
- "map_at_1000": 0.26239,
22
- "recall_at_1": 0.00154,
23
- "recall_at_3": 0.43012,
24
- "recall_at_5": 0.59846,
25
- "recall_at_10": 0.77915,
26
- "recall_at_20": 0.88108,
27
- "recall_at_100": 0.97452,
28
- "recall_at_1000": 0.99537,
29
- "precision_at_1": 0.00154,
30
- "precision_at_3": 0.14337,
31
- "precision_at_5": 0.11969,
32
- "precision_at_10": 0.07792,
33
- "precision_at_20": 0.04405,
34
- "precision_at_100": 0.00975,
35
- "precision_at_1000": 0.001,
36
- "mrr_at_1": 0.001544,
37
- "mrr_at_3": 0.18906,
38
- "mrr_at_5": 0.227477,
39
- "mrr_at_10": 0.252276,
40
- "mrr_at_20": 0.259688,
41
- "mrr_at_100": 0.262271,
42
- "mrr_at_1000": 0.262394,
43
- "nauc_ndcg_at_1_max": -0.15791,
44
- "nauc_ndcg_at_1_std": -0.309824,
45
- "nauc_ndcg_at_1_diff1": -0.573698,
46
- "nauc_ndcg_at_3_max": 0.050466,
47
- "nauc_ndcg_at_3_std": -0.047425,
48
- "nauc_ndcg_at_3_diff1": -0.053972,
49
- "nauc_ndcg_at_5_max": 0.051704,
50
- "nauc_ndcg_at_5_std": -0.034976,
51
- "nauc_ndcg_at_5_diff1": -0.067507,
52
- "nauc_ndcg_at_10_max": 0.051058,
53
- "nauc_ndcg_at_10_std": 0.002573,
54
- "nauc_ndcg_at_10_diff1": -0.115774,
55
- "nauc_ndcg_at_20_max": 0.060309,
56
- "nauc_ndcg_at_20_std": -0.002941,
57
- "nauc_ndcg_at_20_diff1": -0.112784,
58
- "nauc_ndcg_at_100_max": 0.064026,
59
- "nauc_ndcg_at_100_std": -0.001192,
60
- "nauc_ndcg_at_100_diff1": -0.102481,
61
- "nauc_ndcg_at_1000_max": 0.055098,
62
- "nauc_ndcg_at_1000_std": -0.009017,
63
- "nauc_ndcg_at_1000_diff1": -0.099237,
64
- "nauc_map_at_1_max": -0.15791,
65
- "nauc_map_at_1_std": -0.309824,
66
- "nauc_map_at_1_diff1": -0.573698,
67
- "nauc_map_at_3_max": 0.049703,
68
- "nauc_map_at_3_std": -0.048378,
69
- "nauc_map_at_3_diff1": -0.054478,
70
- "nauc_map_at_5_max": 0.049706,
71
- "nauc_map_at_5_std": -0.041634,
72
- "nauc_map_at_5_diff1": -0.062748,
73
- "nauc_map_at_10_max": 0.049333,
74
- "nauc_map_at_10_std": -0.023585,
75
- "nauc_map_at_10_diff1": -0.08462,
76
- "nauc_map_at_20_max": 0.052529,
77
- "nauc_map_at_20_std": -0.026038,
78
- "nauc_map_at_20_diff1": -0.082893,
79
- "nauc_map_at_100_max": 0.052676,
80
- "nauc_map_at_100_std": -0.025392,
81
- "nauc_map_at_100_diff1": -0.081601,
82
- "nauc_map_at_1000_max": 0.05229,
83
- "nauc_map_at_1000_std": -0.025648,
84
- "nauc_map_at_1000_diff1": -0.081523,
85
- "nauc_recall_at_1_max": -0.15791,
86
- "nauc_recall_at_1_std": -0.309824,
87
- "nauc_recall_at_1_diff1": -0.573698,
88
- "nauc_recall_at_3_max": 0.051543,
89
- "nauc_recall_at_3_std": -0.046082,
90
- "nauc_recall_at_3_diff1": -0.053542,
91
- "nauc_recall_at_5_max": 0.055409,
92
- "nauc_recall_at_5_std": -0.022324,
93
- "nauc_recall_at_5_diff1": -0.077261,
94
- "nauc_recall_at_10_max": 0.055321,
95
- "nauc_recall_at_10_std": 0.07644,
96
- "nauc_recall_at_10_diff1": -0.207529,
97
- "nauc_recall_at_20_max": 0.098502,
98
- "nauc_recall_at_20_std": 0.101036,
99
- "nauc_recall_at_20_diff1": -0.255702,
100
- "nauc_recall_at_100_max": 0.346732,
101
- "nauc_recall_at_100_std": 0.450867,
102
- "nauc_recall_at_100_diff1": -0.448661,
103
- "nauc_recall_at_1000_max": -0.18723,
104
- "nauc_recall_at_1000_std": 0.764955,
105
- "nauc_recall_at_1000_diff1": -1.251363,
106
- "nauc_precision_at_1_max": -0.15791,
107
- "nauc_precision_at_1_std": -0.309824,
108
- "nauc_precision_at_1_diff1": -0.573698,
109
- "nauc_precision_at_3_max": 0.051543,
110
- "nauc_precision_at_3_std": -0.046082,
111
- "nauc_precision_at_3_diff1": -0.053542,
112
- "nauc_precision_at_5_max": 0.055409,
113
- "nauc_precision_at_5_std": -0.022324,
114
- "nauc_precision_at_5_diff1": -0.077261,
115
- "nauc_precision_at_10_max": 0.055321,
116
- "nauc_precision_at_10_std": 0.07644,
117
- "nauc_precision_at_10_diff1": -0.207529,
118
- "nauc_precision_at_20_max": 0.098502,
119
- "nauc_precision_at_20_std": 0.101036,
120
- "nauc_precision_at_20_diff1": -0.255702,
121
- "nauc_precision_at_100_max": 0.346732,
122
- "nauc_precision_at_100_std": 0.450867,
123
- "nauc_precision_at_100_diff1": -0.448661,
124
- "nauc_precision_at_1000_max": -0.18723,
125
- "nauc_precision_at_1000_std": 0.764955,
126
- "nauc_precision_at_1000_diff1": -1.251363,
127
- "nauc_mrr_at_1_max": -0.15791,
128
- "nauc_mrr_at_1_std": -0.309824,
129
- "nauc_mrr_at_1_diff1": -0.573698,
130
- "nauc_mrr_at_3_max": 0.049703,
131
- "nauc_mrr_at_3_std": -0.048378,
132
- "nauc_mrr_at_3_diff1": -0.054478,
133
- "nauc_mrr_at_5_max": 0.049706,
134
- "nauc_mrr_at_5_std": -0.041634,
135
- "nauc_mrr_at_5_diff1": -0.062748,
136
- "nauc_mrr_at_10_max": 0.049333,
137
- "nauc_mrr_at_10_std": -0.023585,
138
- "nauc_mrr_at_10_diff1": -0.08462,
139
- "nauc_mrr_at_20_max": 0.052529,
140
- "nauc_mrr_at_20_std": -0.026038,
141
- "nauc_mrr_at_20_diff1": -0.082893,
142
- "nauc_mrr_at_100_max": 0.052676,
143
- "nauc_mrr_at_100_std": -0.025392,
144
- "nauc_mrr_at_100_diff1": -0.081601,
145
- "nauc_mrr_at_1000_max": 0.05229,
146
- "nauc_mrr_at_1000_std": -0.025648,
147
- "nauc_mrr_at_1000_diff1": -0.081523,
148
- "main_score": 0.37932,
149
- "hf_subset": "default",
150
- "languages": [
151
- "vie-Latn"
152
- ]
153
- }
154
- ]
155
- },
156
- "evaluation_time": 56.45738124847412,
157
- "kg_co2_emissions": null
158
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AskUbuntuDupQuestions-VN.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "dataset_revision": "5cfaa5c07252d30c37302bfc056f0d85884971a1",
3
- "task_name": "AskUbuntuDupQuestions-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "map": 0.618748,
9
- "mrr": 0.729745,
10
- "nAUC_map_max": 0.263989,
11
- "nAUC_map_std": 0.233967,
12
- "nAUC_map_diff1": 0.170011,
13
- "nAUC_mrr_max": 0.350929,
14
- "nAUC_mrr_std": 0.29905,
15
- "nAUC_mrr_diff1": 0.207218,
16
- "main_score": 0.618748,
17
- "hf_subset": "default",
18
- "languages": [
19
- "vie-Latn"
20
- ]
21
- }
22
- ]
23
- },
24
- "evaluation_time": 6.192819356918335,
25
- "kg_co2_emissions": null
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/BIOSSES-VN.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "dataset_revision": "1dae4a6df91c0852680cd4ab48c8c1d8a9ed49b2",
3
- "task_name": "BIOSSES-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "pearson": 0.799628,
9
- "spearman": 0.781415,
10
- "cosine_pearson": 0.799628,
11
- "cosine_spearman": 0.781415,
12
- "manhattan_pearson": 0.777069,
13
- "manhattan_spearman": 0.777819,
14
- "euclidean_pearson": 0.780537,
15
- "euclidean_spearman": 0.781415,
16
- "main_score": 0.781415,
17
- "hf_subset": "default",
18
- "languages": [
19
- "vie-Latn"
20
- ]
21
- }
22
- ]
23
- },
24
- "evaluation_time": 0.6975255012512207,
25
- "kg_co2_emissions": null
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Banking77VNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "42541b07c25a49604be129fba6d70b752be229c1",
3
- "task_name": "Banking77VNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.792094,
9
- "f1": 0.777269,
10
- "f1_weighted": 0.78413,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.784693,
14
- "f1": 0.766683,
15
- "f1_weighted": 0.774623
16
- },
17
- {
18
- "accuracy": 0.795627,
19
- "f1": 0.776761,
20
- "f1_weighted": 0.784439
21
- },
22
- {
23
- "accuracy": 0.796468,
24
- "f1": 0.781336,
25
- "f1_weighted": 0.787568
26
- },
27
- {
28
- "accuracy": 0.799832,
29
- "f1": 0.791148,
30
- "f1_weighted": 0.79406
31
- },
32
- {
33
- "accuracy": 0.77418,
34
- "f1": 0.752046,
35
- "f1_weighted": 0.764798
36
- },
37
- {
38
- "accuracy": 0.785114,
39
- "f1": 0.775827,
40
- "f1_weighted": 0.77941
41
- },
42
- {
43
- "accuracy": 0.799832,
44
- "f1": 0.785842,
45
- "f1_weighted": 0.792011
46
- },
47
- {
48
- "accuracy": 0.801514,
49
- "f1": 0.788659,
50
- "f1_weighted": 0.795154
51
- },
52
- {
53
- "accuracy": 0.80614,
54
- "f1": 0.795751,
55
- "f1_weighted": 0.800618
56
- },
57
- {
58
- "accuracy": 0.777544,
59
- "f1": 0.758633,
60
- "f1_weighted": 0.768624
61
- }
62
- ],
63
- "main_score": 0.792094,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 12.62467908859253,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackAndroid-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "4a022e7213ccc05ee970a176abd0293b3a0a2da0",
3
- "mteb_dataset_name": "CQADupstackAndroid-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackAndroid-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 44.89,
10
- "map_at_1": 0.26147,
11
- "map_at_10": 0.35997,
12
- "map_at_100": 0.3737,
13
- "map_at_1000": 0.37509,
14
- "map_at_20": 0.36694,
15
- "map_at_3": 0.32705,
16
- "map_at_5": 0.34598,
17
- "mrr_at_1": 0.32474,
18
- "mrr_at_10": 0.41319,
19
- "mrr_at_100": 0.42197,
20
- "mrr_at_1000": 0.42238,
21
- "mrr_at_20": 0.41806,
22
- "mrr_at_3": 0.38431,
23
- "mrr_at_5": 0.4008,
24
- "ndcg_at_1": 0.32474,
25
- "ndcg_at_10": 0.41928,
26
- "ndcg_at_100": 0.47686,
27
- "ndcg_at_1000": 0.49942,
28
- "ndcg_at_20": 0.43829,
29
- "ndcg_at_3": 0.36824,
30
- "ndcg_at_5": 0.39274,
31
- "precision_at_1": 0.32474,
32
- "precision_at_10": 0.08041,
33
- "precision_at_100": 0.01424,
34
- "precision_at_1000": 0.00203,
35
- "precision_at_20": 0.04828,
36
- "precision_at_3": 0.17468,
37
- "precision_at_5": 0.12921,
38
- "recall_at_1": 0.26147,
39
- "recall_at_10": 0.53732,
40
- "recall_at_100": 0.7876,
41
- "recall_at_1000": 0.93133,
42
- "recall_at_20": 0.60459,
43
- "recall_at_3": 0.38587,
44
- "recall_at_5": 0.4548,
45
- "hf_subset": "default",
46
- "main_score": 0.41928,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackGis-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "755156d548a8288efdb29b80bad302750ab33977",
3
- "mteb_dataset_name": "CQADupstackGis-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackGis-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 199.51,
10
- "map_at_1": 0.20556,
11
- "map_at_10": 0.2769,
12
- "map_at_100": 0.28656,
13
- "map_at_1000": 0.28748,
14
- "map_at_20": 0.28257,
15
- "map_at_3": 0.25622,
16
- "map_at_5": 0.26698,
17
- "mrr_at_1": 0.22308,
18
- "mrr_at_10": 0.2932,
19
- "mrr_at_100": 0.30271,
20
- "mrr_at_1000": 0.3034,
21
- "mrr_at_20": 0.29919,
22
- "mrr_at_3": 0.27256,
23
- "mrr_at_5": 0.28372,
24
- "ndcg_at_1": 0.22308,
25
- "ndcg_at_10": 0.31907,
26
- "ndcg_at_100": 0.36814,
27
- "ndcg_at_1000": 0.39334,
28
- "ndcg_at_20": 0.33973,
29
- "ndcg_at_3": 0.27709,
30
- "ndcg_at_5": 0.29583,
31
- "precision_at_1": 0.22308,
32
- "precision_at_10": 0.05015,
33
- "precision_at_100": 0.00785,
34
- "precision_at_1000": 0.00105,
35
- "precision_at_20": 0.02969,
36
- "precision_at_3": 0.11949,
37
- "precision_at_5": 0.08308,
38
- "recall_at_1": 0.20556,
39
- "recall_at_10": 0.43315,
40
- "recall_at_100": 0.66017,
41
- "recall_at_1000": 0.85128,
42
- "recall_at_20": 0.51194,
43
- "recall_at_3": 0.31881,
44
- "recall_at_5": 0.36374,
45
- "hf_subset": "default",
46
- "main_score": 0.31907,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackMathematica-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "d0cc9b60ba66faa3fb21cb9a54ef969af548b312",
3
- "mteb_dataset_name": "CQADupstackMathematica-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackMathematica-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 111.08,
10
- "map_at_1": 0.11446,
11
- "map_at_10": 0.17372,
12
- "map_at_100": 0.18428,
13
- "map_at_1000": 0.18553,
14
- "map_at_20": 0.17952,
15
- "map_at_3": 0.15416,
16
- "map_at_5": 0.16377,
17
- "mrr_at_1": 0.14499,
18
- "mrr_at_10": 0.20717,
19
- "mrr_at_100": 0.21714,
20
- "mrr_at_1000": 0.21797,
21
- "mrr_at_20": 0.21261,
22
- "mrr_at_3": 0.18799,
23
- "mrr_at_5": 0.19737,
24
- "ndcg_at_1": 0.14499,
25
- "ndcg_at_10": 0.21439,
26
- "ndcg_at_100": 0.26912,
27
- "ndcg_at_1000": 0.30206,
28
- "ndcg_at_20": 0.23465,
29
- "ndcg_at_3": 0.17684,
30
- "ndcg_at_5": 0.19131,
31
- "precision_at_1": 0.14499,
32
- "precision_at_10": 0.04009,
33
- "precision_at_100": 0.00755,
34
- "precision_at_1000": 0.00117,
35
- "precision_at_20": 0.02516,
36
- "precision_at_3": 0.08529,
37
- "precision_at_5": 0.06226,
38
- "recall_at_1": 0.11446,
39
- "recall_at_10": 0.30977,
40
- "recall_at_100": 0.55949,
41
- "recall_at_1000": 0.8006,
42
- "recall_at_20": 0.38463,
43
- "recall_at_3": 0.20025,
44
- "recall_at_5": 0.2377,
45
- "hf_subset": "default",
46
- "main_score": 0.21439,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackPhysics-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "8b6b68b59933cc72985f674f76c80a678c27d6be",
3
- "mteb_dataset_name": "CQADupstackPhysics-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackPhysics-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 190.76,
10
- "map_at_1": 0.21216,
11
- "map_at_10": 0.29794,
12
- "map_at_100": 0.31226,
13
- "map_at_1000": 0.31362,
14
- "map_at_20": 0.30679,
15
- "map_at_3": 0.26528,
16
- "map_at_5": 0.28309,
17
- "mrr_at_1": 0.27629,
18
- "mrr_at_10": 0.36405,
19
- "mrr_at_100": 0.37387,
20
- "mrr_at_1000": 0.37455,
21
- "mrr_at_20": 0.37036,
22
- "mrr_at_3": 0.34158,
23
- "mrr_at_5": 0.35354,
24
- "ndcg_at_1": 0.27629,
25
- "ndcg_at_10": 0.35518,
26
- "ndcg_at_100": 0.4152,
27
- "ndcg_at_1000": 0.44354,
28
- "ndcg_at_20": 0.38148,
29
- "ndcg_at_3": 0.30811,
30
- "ndcg_at_5": 0.32809,
31
- "precision_at_1": 0.27629,
32
- "precision_at_10": 0.07072,
33
- "precision_at_100": 0.01196,
34
- "precision_at_1000": 0.00166,
35
- "precision_at_20": 0.04381,
36
- "precision_at_3": 0.15258,
37
- "precision_at_5": 0.11134,
38
- "recall_at_1": 0.21216,
39
- "recall_at_10": 0.46074,
40
- "recall_at_100": 0.71265,
41
- "recall_at_1000": 0.9034,
42
- "recall_at_20": 0.55265,
43
- "recall_at_3": 0.32342,
44
- "recall_at_5": 0.38045,
45
- "hf_subset": "default",
46
- "main_score": 0.35518,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackProgrammers-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "1a628c4e61f71ffdb7707d6d4024d25cfe68215a",
3
- "mteb_dataset_name": "CQADupstackProgrammers-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackProgrammers-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 166.63,
10
- "map_at_1": 0.19431,
11
- "map_at_10": 0.27708,
12
- "map_at_100": 0.2901,
13
- "map_at_1000": 0.29131,
14
- "map_at_20": 0.28417,
15
- "map_at_3": 0.25104,
16
- "map_at_5": 0.26674,
17
- "mrr_at_1": 0.23328,
18
- "mrr_at_10": 0.31688,
19
- "mrr_at_100": 0.32747,
20
- "mrr_at_1000": 0.32809,
21
- "mrr_at_20": 0.32311,
22
- "mrr_at_3": 0.29201,
23
- "mrr_at_5": 0.30685,
24
- "ndcg_at_1": 0.23328,
25
- "ndcg_at_10": 0.3271,
26
- "ndcg_at_100": 0.38812,
27
- "ndcg_at_1000": 0.41372,
28
- "ndcg_at_20": 0.35064,
29
- "ndcg_at_3": 0.28162,
30
- "ndcg_at_5": 0.30513,
31
- "precision_at_1": 0.23328,
32
- "precision_at_10": 0.06101,
33
- "precision_at_100": 0.01064,
34
- "precision_at_1000": 0.00145,
35
- "precision_at_20": 0.03695,
36
- "precision_at_3": 0.13431,
37
- "precision_at_5": 0.09886,
38
- "recall_at_1": 0.19431,
39
- "recall_at_10": 0.43628,
40
- "recall_at_100": 0.70198,
41
- "recall_at_1000": 0.87329,
42
- "recall_at_20": 0.52302,
43
- "recall_at_3": 0.31336,
44
- "recall_at_5": 0.37245,
45
- "hf_subset": "default",
46
- "main_score": 0.3271,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackStats-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "6b8164f3af61f3bb7728724229ba36213fb46c25",
3
- "mteb_dataset_name": "CQADupstackStats-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackStats-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 234.79,
10
- "map_at_1": 0.16556,
11
- "map_at_10": 0.22848,
12
- "map_at_100": 0.23963,
13
- "map_at_1000": 0.24038,
14
- "map_at_20": 0.23554,
15
- "map_at_3": 0.20422,
16
- "map_at_5": 0.21873,
17
- "mrr_at_1": 0.18874,
18
- "mrr_at_10": 0.2507,
19
- "mrr_at_100": 0.26037,
20
- "mrr_at_1000": 0.26091,
21
- "mrr_at_20": 0.25655,
22
- "mrr_at_3": 0.22572,
23
- "mrr_at_5": 0.24078,
24
- "ndcg_at_1": 0.18874,
25
- "ndcg_at_10": 0.26859,
26
- "ndcg_at_100": 0.32125,
27
- "ndcg_at_1000": 0.34208,
28
- "ndcg_at_20": 0.29217,
29
- "ndcg_at_3": 0.22063,
30
- "ndcg_at_5": 0.2463,
31
- "precision_at_1": 0.18874,
32
- "precision_at_10": 0.04603,
33
- "precision_at_100": 0.00768,
34
- "precision_at_1000": 0.00102,
35
- "precision_at_20": 0.02881,
36
- "precision_at_3": 0.09492,
37
- "precision_at_5": 0.07219,
38
- "recall_at_1": 0.16556,
39
- "recall_at_10": 0.3683,
40
- "recall_at_100": 0.60969,
41
- "recall_at_1000": 0.76575,
42
- "recall_at_20": 0.45836,
43
- "recall_at_3": 0.24264,
44
- "recall_at_5": 0.30273,
45
- "hf_subset": "default",
46
- "main_score": 0.26859,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackTex-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "aec43e5ae40451526528b3fc80dd5983ec388e21",
3
- "mteb_dataset_name": "CQADupstackTex-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackTex-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 490.61,
10
- "map_at_1": 0.14818,
11
- "map_at_10": 0.20724,
12
- "map_at_100": 0.21763,
13
- "map_at_1000": 0.2188,
14
- "map_at_20": 0.21326,
15
- "map_at_3": 0.18685,
16
- "map_at_5": 0.19831,
17
- "mrr_at_1": 0.18582,
18
- "mrr_at_10": 0.24641,
19
- "mrr_at_100": 0.2559,
20
- "mrr_at_1000": 0.25671,
21
- "mrr_at_20": 0.25211,
22
- "mrr_at_3": 0.2268,
23
- "mrr_at_5": 0.23826,
24
- "ndcg_at_1": 0.18582,
25
- "ndcg_at_10": 0.24781,
26
- "ndcg_at_100": 0.29847,
27
- "ndcg_at_1000": 0.32928,
28
- "ndcg_at_20": 0.26771,
29
- "ndcg_at_3": 0.21233,
30
- "ndcg_at_5": 0.22894,
31
- "precision_at_1": 0.18582,
32
- "precision_at_10": 0.04523,
33
- "precision_at_100": 0.00827,
34
- "precision_at_1000": 0.00126,
35
- "precision_at_20": 0.02799,
36
- "precision_at_3": 0.10025,
37
- "precision_at_5": 0.07364,
38
- "recall_at_1": 0.14818,
39
- "recall_at_10": 0.33285,
40
- "recall_at_100": 0.56243,
41
- "recall_at_1000": 0.78623,
42
- "recall_at_20": 0.40797,
43
- "recall_at_3": 0.22972,
44
- "recall_at_5": 0.27396,
45
- "hf_subset": "default",
46
- "main_score": 0.24781,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackUnix-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "f8b884697871cb38901139f2435c273135f83a3f",
3
- "mteb_dataset_name": "CQADupstackUnix-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackUnix-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 248.55,
10
- "map_at_1": 0.22331,
11
- "map_at_10": 0.29694,
12
- "map_at_100": 0.30728,
13
- "map_at_1000": 0.30838,
14
- "map_at_20": 0.30234,
15
- "map_at_3": 0.27118,
16
- "map_at_5": 0.28451,
17
- "mrr_at_1": 0.26481,
18
- "mrr_at_10": 0.33574,
19
- "mrr_at_100": 0.34457,
20
- "mrr_at_1000": 0.34526,
21
- "mrr_at_20": 0.34089,
22
- "mrr_at_3": 0.31177,
23
- "mrr_at_5": 0.32465,
24
- "ndcg_at_1": 0.26481,
25
- "ndcg_at_10": 0.34515,
26
- "ndcg_at_100": 0.39566,
27
- "ndcg_at_1000": 0.42141,
28
- "ndcg_at_20": 0.36389,
29
- "ndcg_at_3": 0.29735,
30
- "ndcg_at_5": 0.31759,
31
- "precision_at_1": 0.26481,
32
- "precision_at_10": 0.05925,
33
- "precision_at_100": 0.00946,
34
- "precision_at_1000": 0.00128,
35
- "precision_at_20": 0.03458,
36
- "precision_at_3": 0.13301,
37
- "precision_at_5": 0.09432,
38
- "recall_at_1": 0.22331,
39
- "recall_at_10": 0.45476,
40
- "recall_at_100": 0.67928,
41
- "recall_at_1000": 0.86037,
42
- "recall_at_20": 0.52341,
43
- "recall_at_3": 0.32383,
44
- "recall_at_5": 0.37485,
45
- "hf_subset": "default",
46
- "main_score": 0.34515,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWebmasters-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "482d6e560d977b137e435d33379c5a8049e70e8d",
3
- "mteb_dataset_name": "CQADupstackWebmasters-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackWebmasters-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 73.56,
10
- "map_at_1": 0.18814,
11
- "map_at_10": 0.26309,
12
- "map_at_100": 0.27633,
13
- "map_at_1000": 0.2786,
14
- "map_at_20": 0.26934,
15
- "map_at_3": 0.23491,
16
- "map_at_5": 0.25194,
17
- "mrr_at_1": 0.225,
18
- "mrr_at_10": 0.30061,
19
- "mrr_at_100": 0.31017,
20
- "mrr_at_1000": 0.31098,
21
- "mrr_at_20": 0.30585,
22
- "mrr_at_3": 0.27083,
23
- "mrr_at_5": 0.29083,
24
- "ndcg_at_1": 0.225,
25
- "ndcg_at_10": 0.31666,
26
- "ndcg_at_100": 0.36984,
27
- "ndcg_at_1000": 0.40278,
28
- "ndcg_at_20": 0.33293,
29
- "ndcg_at_3": 0.26741,
30
- "ndcg_at_5": 0.29431,
31
- "precision_at_1": 0.225,
32
- "precision_at_10": 0.062,
33
- "precision_at_100": 0.01322,
34
- "precision_at_1000": 0.00234,
35
- "precision_at_20": 0.039,
36
- "precision_at_3": 0.125,
37
- "precision_at_5": 0.097,
38
- "recall_at_1": 0.18814,
39
- "recall_at_10": 0.42675,
40
- "recall_at_100": 0.66891,
41
- "recall_at_1000": 0.88531,
42
- "recall_at_20": 0.49,
43
- "recall_at_3": 0.28062,
44
- "recall_at_5": 0.35206,
45
- "hf_subset": "default",
46
- "main_score": 0.31666,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWordpress-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "2230f80e1baf42aa005731ca86577621c566fcd7",
3
- "mteb_dataset_name": "CQADupstackWordpress-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackWordpress-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 283.45,
10
- "map_at_1": 0.13534,
11
- "map_at_10": 0.20314,
12
- "map_at_100": 0.21388,
13
- "map_at_1000": 0.21491,
14
- "map_at_20": 0.20936,
15
- "map_at_3": 0.18317,
16
- "map_at_5": 0.19319,
17
- "mrr_at_1": 0.15183,
18
- "mrr_at_10": 0.22404,
19
- "mrr_at_100": 0.23378,
20
- "mrr_at_1000": 0.23451,
21
- "mrr_at_20": 0.22996,
22
- "mrr_at_3": 0.20201,
23
- "mrr_at_5": 0.21366,
24
- "ndcg_at_1": 0.15183,
25
- "ndcg_at_10": 0.24744,
26
- "ndcg_at_100": 0.29839,
27
- "ndcg_at_1000": 0.32591,
28
- "ndcg_at_20": 0.2689,
29
- "ndcg_at_3": 0.20599,
30
- "ndcg_at_5": 0.22338,
31
- "precision_at_1": 0.15183,
32
- "precision_at_10": 0.04162,
33
- "precision_at_100": 0.0073,
34
- "precision_at_1000": 0.00109,
35
- "precision_at_20": 0.02592,
36
- "precision_at_3": 0.09511,
37
- "precision_at_5": 0.06702,
38
- "recall_at_1": 0.13534,
39
- "recall_at_10": 0.36562,
40
- "recall_at_100": 0.59327,
41
- "recall_at_1000": 0.80055,
42
- "recall_at_20": 0.44546,
43
- "recall_at_3": 0.24908,
44
- "recall_at_5": 0.29228,
45
- "hf_subset": "default",
46
- "main_score": 0.24744,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ClimateFEVER-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "42328bf787e17b1ad1a88be4f5e87ea9fb668511",
3
- "mteb_dataset_name": "ClimateFEVER-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "ClimateFEVER-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 16316.19,
10
- "map_at_1": 0.04967,
11
- "map_at_10": 0.08868,
12
- "map_at_100": 0.09661,
13
- "map_at_1000": 0.0982,
14
- "map_at_20": 0.09205,
15
- "map_at_3": 0.07457,
16
- "map_at_5": 0.08228,
17
- "mrr_at_1": 0.11353,
18
- "mrr_at_10": 0.18208,
19
- "mrr_at_100": 0.19152,
20
- "mrr_at_1000": 0.1925,
21
- "mrr_at_20": 0.18658,
22
- "mrr_at_3": 0.15849,
23
- "mrr_at_5": 0.17107,
24
- "ndcg_at_1": 0.11353,
25
- "ndcg_at_10": 0.13249,
26
- "ndcg_at_100": 0.17814,
27
- "ndcg_at_1000": 0.21712,
28
- "ndcg_at_20": 0.14482,
29
- "ndcg_at_3": 0.10497,
30
- "ndcg_at_5": 0.1154,
31
- "precision_at_1": 0.11353,
32
- "precision_at_10": 0.04223,
33
- "precision_at_100": 0.00927,
34
- "precision_at_1000": 0.00164,
35
- "precision_at_20": 0.02643,
36
- "precision_at_3": 0.08053,
37
- "precision_at_5": 0.06267,
38
- "recall_at_1": 0.04967,
39
- "recall_at_10": 0.16757,
40
- "recall_at_100": 0.33663,
41
- "recall_at_1000": 0.5625,
42
- "recall_at_20": 0.20262,
43
- "recall_at_3": 0.09955,
44
- "recall_at_5": 0.12846,
45
- "hf_subset": "default",
46
- "main_score": 0.13249,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/DBPedia-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "c3e20179fbcee16217ef9461a14a54b7faca9b63",
3
- "mteb_dataset_name": "DBPedia-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "DBPedia-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 4419.42,
10
- "map_at_1": 0.07486,
11
- "map_at_10": 0.15539,
12
- "map_at_100": 0.21047,
13
- "map_at_1000": 0.22274,
14
- "map_at_20": 0.17743,
15
- "map_at_3": 0.11208,
16
- "map_at_5": 0.13146,
17
- "mrr_at_1": 0.53106,
18
- "mrr_at_10": 0.6375,
19
- "mrr_at_100": 0.64307,
20
- "mrr_at_1000": 0.64321,
21
- "mrr_at_20": 0.64143,
22
- "mrr_at_3": 0.61439,
23
- "mrr_at_5": 0.62805,
24
- "ndcg_at_1": 0.43634,
25
- "ndcg_at_10": 0.34204,
26
- "ndcg_at_100": 0.38478,
27
- "ndcg_at_1000": 0.45509,
28
- "ndcg_at_20": 0.34051,
29
- "ndcg_at_3": 0.37097,
30
- "ndcg_at_5": 0.35581,
31
- "precision_at_1": 0.53106,
32
- "precision_at_10": 0.26553,
33
- "precision_at_100": 0.08469,
34
- "precision_at_1000": 0.01695,
35
- "precision_at_20": 0.2028,
36
- "precision_at_3": 0.39648,
37
- "precision_at_5": 0.34472,
38
- "recall_at_1": 0.07486,
39
- "recall_at_10": 0.21316,
40
- "recall_at_100": 0.44686,
41
- "recall_at_1000": 0.67695,
42
- "recall_at_20": 0.27749,
43
- "recall_at_3": 0.12701,
44
- "recall_at_5": 0.16045,
45
- "hf_subset": "default",
46
- "main_score": 0.34204,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/EmotionVNClassification.json DELETED
@@ -1,137 +0,0 @@
1
- {
2
- "dataset_revision": "797a93c0dd755ebf5818fbf54d0e0a024df9216d",
3
- "task_name": "EmotionVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "validation": [
7
- {
8
- "accuracy": 0.44375,
9
- "f1": 0.390648,
10
- "f1_weighted": 0.463387,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.472656,
14
- "f1": 0.404343,
15
- "f1_weighted": 0.498363
16
- },
17
- {
18
- "accuracy": 0.43125,
19
- "f1": 0.354818,
20
- "f1_weighted": 0.446879
21
- },
22
- {
23
- "accuracy": 0.473438,
24
- "f1": 0.402721,
25
- "f1_weighted": 0.496464
26
- },
27
- {
28
- "accuracy": 0.422656,
29
- "f1": 0.384927,
30
- "f1_weighted": 0.440412
31
- },
32
- {
33
- "accuracy": 0.461719,
34
- "f1": 0.414061,
35
- "f1_weighted": 0.476512
36
- },
37
- {
38
- "accuracy": 0.405469,
39
- "f1": 0.381572,
40
- "f1_weighted": 0.421386
41
- },
42
- {
43
- "accuracy": 0.478906,
44
- "f1": 0.391987,
45
- "f1_weighted": 0.487998
46
- },
47
- {
48
- "accuracy": 0.396875,
49
- "f1": 0.378969,
50
- "f1_weighted": 0.422998
51
- },
52
- {
53
- "accuracy": 0.463281,
54
- "f1": 0.413623,
55
- "f1_weighted": 0.487997
56
- },
57
- {
58
- "accuracy": 0.43125,
59
- "f1": 0.379461,
60
- "f1_weighted": 0.454856
61
- }
62
- ],
63
- "main_score": 0.44375,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ],
70
- "test": [
71
- {
72
- "accuracy": 0.446822,
73
- "f1": 0.380728,
74
- "f1_weighted": 0.468056,
75
- "scores_per_experiment": [
76
- {
77
- "accuracy": 0.481395,
78
- "f1": 0.401385,
79
- "f1_weighted": 0.508444
80
- },
81
- {
82
- "accuracy": 0.432558,
83
- "f1": 0.352439,
84
- "f1_weighted": 0.446656
85
- },
86
- {
87
- "accuracy": 0.475969,
88
- "f1": 0.394022,
89
- "f1_weighted": 0.49485
90
- },
91
- {
92
- "accuracy": 0.413953,
93
- "f1": 0.362915,
94
- "f1_weighted": 0.439652
95
- },
96
- {
97
- "accuracy": 0.455039,
98
- "f1": 0.399925,
99
- "f1_weighted": 0.475066
100
- },
101
- {
102
- "accuracy": 0.397674,
103
- "f1": 0.362269,
104
- "f1_weighted": 0.412897
105
- },
106
- {
107
- "accuracy": 0.490698,
108
- "f1": 0.389544,
109
- "f1_weighted": 0.498402
110
- },
111
- {
112
- "accuracy": 0.402326,
113
- "f1": 0.361925,
114
- "f1_weighted": 0.434939
115
- },
116
- {
117
- "accuracy": 0.473643,
118
- "f1": 0.405323,
119
- "f1_weighted": 0.496273
120
- },
121
- {
122
- "accuracy": 0.444961,
123
- "f1": 0.377529,
124
- "f1_weighted": 0.473386
125
- }
126
- ],
127
- "main_score": 0.446822,
128
- "hf_subset": "default",
129
- "languages": [
130
- "vie-Latn"
131
- ]
132
- }
133
- ]
134
- },
135
- "evaluation_time": 16.90190815925598,
136
- "kg_co2_emissions": null
137
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FEVER-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "a543dd8b98aed3603110c01d26db05ba39b87d49",
3
- "mteb_dataset_name": "FEVER-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "FEVER-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 7374.25,
10
- "map_at_1": 0.31332,
11
- "map_at_10": 0.42736,
12
- "map_at_100": 0.43506,
13
- "map_at_1000": 0.43552,
14
- "map_at_20": 0.432,
15
- "map_at_3": 0.3997,
16
- "map_at_5": 0.41616,
17
- "mrr_at_1": 0.33676,
18
- "mrr_at_10": 0.45185,
19
- "mrr_at_100": 0.45905,
20
- "mrr_at_1000": 0.45943,
21
- "mrr_at_20": 0.45621,
22
- "mrr_at_3": 0.42454,
23
- "mrr_at_5": 0.44056,
24
- "ndcg_at_1": 0.33676,
25
- "ndcg_at_10": 0.4881,
26
- "ndcg_at_100": 0.52495,
27
- "ndcg_at_1000": 0.53726,
28
- "ndcg_at_20": 0.50418,
29
- "ndcg_at_3": 0.43293,
30
- "ndcg_at_5": 0.46137,
31
- "precision_at_1": 0.33676,
32
- "precision_at_10": 0.0715,
33
- "precision_at_100": 0.00916,
34
- "precision_at_1000": 0.00104,
35
- "precision_at_20": 0.03934,
36
- "precision_at_3": 0.1826,
37
- "precision_at_5": 0.12488,
38
- "recall_at_1": 0.31332,
39
- "recall_at_10": 0.6531,
40
- "recall_at_100": 0.8205,
41
- "recall_at_1000": 0.91405,
42
- "recall_at_20": 0.71409,
43
- "recall_at_3": 0.50359,
44
- "recall_at_5": 0.57214,
45
- "hf_subset": "default",
46
- "main_score": 0.4881,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FiQA2018-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "6c3cdf6f102151dbbbbc1d2cf999b305eba44dae",
3
- "mteb_dataset_name": "FiQA2018-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "FiQA2018-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 149.94,
10
- "map_at_1": 0.15915,
11
- "map_at_10": 0.23933,
12
- "map_at_100": 0.25351,
13
- "map_at_1000": 0.25561,
14
- "map_at_20": 0.24686,
15
- "map_at_3": 0.21014,
16
- "map_at_5": 0.22637,
17
- "mrr_at_1": 0.29024,
18
- "mrr_at_10": 0.3643,
19
- "mrr_at_100": 0.37292,
20
- "mrr_at_1000": 0.37358,
21
- "mrr_at_20": 0.36883,
22
- "mrr_at_3": 0.34301,
23
- "mrr_at_5": 0.35396,
24
- "ndcg_at_1": 0.29024,
25
- "ndcg_at_10": 0.29939,
26
- "ndcg_at_100": 0.35992,
27
- "ndcg_at_1000": 0.39932,
28
- "ndcg_at_20": 0.32076,
29
- "ndcg_at_3": 0.27282,
30
- "ndcg_at_5": 0.27997,
31
- "precision_at_1": 0.29024,
32
- "precision_at_10": 0.08153,
33
- "precision_at_100": 0.01478,
34
- "precision_at_1000": 0.00216,
35
- "precision_at_20": 0.05013,
36
- "precision_at_3": 0.17326,
37
- "precision_at_5": 0.12823,
38
- "recall_at_1": 0.15915,
39
- "recall_at_10": 0.35196,
40
- "recall_at_100": 0.57857,
41
- "recall_at_1000": 0.81862,
42
- "recall_at_20": 0.41776,
43
- "recall_at_3": 0.24858,
44
- "recall_at_5": 0.29204,
45
- "hf_subset": "default",
46
- "main_score": 0.29939,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/GreenNodeTableMarkdownRetrieval.json DELETED
@@ -1,158 +0,0 @@
1
- {
2
- "dataset_revision": "d86a4dad9fd7c70359f617d86984395ea89be1c5",
3
- "task_name": "GreenNodeTableMarkdownRetrieval",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "ndcg_at_1": 0.30846,
9
- "ndcg_at_3": 0.36243,
10
- "ndcg_at_5": 0.37957,
11
- "ndcg_at_10": 0.39719,
12
- "ndcg_at_20": 0.412,
13
- "ndcg_at_100": 0.43624,
14
- "ndcg_at_1000": 0.45896,
15
- "map_at_1": 0.30846,
16
- "map_at_3": 0.3491,
17
- "map_at_5": 0.35861,
18
- "map_at_10": 0.36591,
19
- "map_at_20": 0.36997,
20
- "map_at_100": 0.3732,
21
- "map_at_1000": 0.37392,
22
- "recall_at_1": 0.30846,
23
- "recall_at_3": 0.40102,
24
- "recall_at_5": 0.44265,
25
- "recall_at_10": 0.49702,
26
- "recall_at_20": 0.55553,
27
- "recall_at_100": 0.68819,
28
- "recall_at_1000": 0.87447,
29
- "precision_at_1": 0.30846,
30
- "precision_at_3": 0.13367,
31
- "precision_at_5": 0.08853,
32
- "precision_at_10": 0.0497,
33
- "precision_at_20": 0.02778,
34
- "precision_at_100": 0.00688,
35
- "precision_at_1000": 0.00087,
36
- "mrr_at_1": 0.308457,
37
- "mrr_at_3": 0.349096,
38
- "mrr_at_5": 0.35861,
39
- "mrr_at_10": 0.365907,
40
- "mrr_at_20": 0.369975,
41
- "mrr_at_100": 0.373196,
42
- "mrr_at_1000": 0.373924,
43
- "nauc_ndcg_at_1_max": 0.139786,
44
- "nauc_ndcg_at_1_std": -0.039386,
45
- "nauc_ndcg_at_1_diff1": 0.733187,
46
- "nauc_ndcg_at_3_max": 0.09619,
47
- "nauc_ndcg_at_3_std": -0.017705,
48
- "nauc_ndcg_at_3_diff1": 0.676034,
49
- "nauc_ndcg_at_5_max": 0.08391,
50
- "nauc_ndcg_at_5_std": -0.006374,
51
- "nauc_ndcg_at_5_diff1": 0.667409,
52
- "nauc_ndcg_at_10_max": 0.071841,
53
- "nauc_ndcg_at_10_std": 0.004193,
54
- "nauc_ndcg_at_10_diff1": 0.657513,
55
- "nauc_ndcg_at_20_max": 0.06409,
56
- "nauc_ndcg_at_20_std": 0.014012,
57
- "nauc_ndcg_at_20_diff1": 0.653131,
58
- "nauc_ndcg_at_100_max": 0.059821,
59
- "nauc_ndcg_at_100_std": 0.028492,
60
- "nauc_ndcg_at_100_diff1": 0.649499,
61
- "nauc_ndcg_at_1000_max": 0.066935,
62
- "nauc_ndcg_at_1000_std": 0.028327,
63
- "nauc_ndcg_at_1000_diff1": 0.655572,
64
- "nauc_map_at_1_max": 0.139786,
65
- "nauc_map_at_1_std": -0.039386,
66
- "nauc_map_at_1_diff1": 0.733187,
67
- "nauc_map_at_3_max": 0.106754,
68
- "nauc_map_at_3_std": -0.023135,
69
- "nauc_map_at_3_diff1": 0.689956,
70
- "nauc_map_at_5_max": 0.099963,
71
- "nauc_map_at_5_std": -0.016825,
72
- "nauc_map_at_5_diff1": 0.685233,
73
- "nauc_map_at_10_max": 0.09492,
74
- "nauc_map_at_10_std": -0.012515,
75
- "nauc_map_at_10_diff1": 0.681245,
76
- "nauc_map_at_20_max": 0.092859,
77
- "nauc_map_at_20_std": -0.00995,
78
- "nauc_map_at_20_diff1": 0.68014,
79
- "nauc_map_at_100_max": 0.092295,
80
- "nauc_map_at_100_std": -0.008175,
81
- "nauc_map_at_100_diff1": 0.67967,
82
- "nauc_map_at_1000_max": 0.092486,
83
- "nauc_map_at_1000_std": -0.008127,
84
- "nauc_map_at_1000_diff1": 0.679852,
85
- "nauc_recall_at_1_max": 0.139786,
86
- "nauc_recall_at_1_std": -0.039386,
87
- "nauc_recall_at_1_diff1": 0.733187,
88
- "nauc_recall_at_3_max": 0.066333,
89
- "nauc_recall_at_3_std": -0.002289,
90
- "nauc_recall_at_3_diff1": 0.636658,
91
- "nauc_recall_at_5_max": 0.036325,
92
- "nauc_recall_at_5_std": 0.025109,
93
- "nauc_recall_at_5_diff1": 0.615216,
94
- "nauc_recall_at_10_max": -0.001472,
95
- "nauc_recall_at_10_std": 0.058617,
96
- "nauc_recall_at_10_diff1": 0.583252,
97
- "nauc_recall_at_20_max": -0.037741,
98
- "nauc_recall_at_20_std": 0.102985,
99
- "nauc_recall_at_20_diff1": 0.560586,
100
- "nauc_recall_at_100_max": -0.089231,
101
- "nauc_recall_at_100_std": 0.224089,
102
- "nauc_recall_at_100_diff1": 0.515154,
103
- "nauc_recall_at_1000_max": -0.101471,
104
- "nauc_recall_at_1000_std": 0.430692,
105
- "nauc_recall_at_1000_diff1": 0.493278,
106
- "nauc_precision_at_1_max": 0.139786,
107
- "nauc_precision_at_1_std": -0.039386,
108
- "nauc_precision_at_1_diff1": 0.733187,
109
- "nauc_precision_at_3_max": 0.066333,
110
- "nauc_precision_at_3_std": -0.002289,
111
- "nauc_precision_at_3_diff1": 0.636658,
112
- "nauc_precision_at_5_max": 0.036325,
113
- "nauc_precision_at_5_std": 0.025109,
114
- "nauc_precision_at_5_diff1": 0.615216,
115
- "nauc_precision_at_10_max": -0.001472,
116
- "nauc_precision_at_10_std": 0.058617,
117
- "nauc_precision_at_10_diff1": 0.583252,
118
- "nauc_precision_at_20_max": -0.037741,
119
- "nauc_precision_at_20_std": 0.102985,
120
- "nauc_precision_at_20_diff1": 0.560586,
121
- "nauc_precision_at_100_max": -0.089231,
122
- "nauc_precision_at_100_std": 0.224089,
123
- "nauc_precision_at_100_diff1": 0.515154,
124
- "nauc_precision_at_1000_max": -0.101471,
125
- "nauc_precision_at_1000_std": 0.430692,
126
- "nauc_precision_at_1000_diff1": 0.493278,
127
- "nauc_mrr_at_1_max": 0.139786,
128
- "nauc_mrr_at_1_std": -0.039386,
129
- "nauc_mrr_at_1_diff1": 0.733187,
130
- "nauc_mrr_at_3_max": 0.106754,
131
- "nauc_mrr_at_3_std": -0.023135,
132
- "nauc_mrr_at_3_diff1": 0.689956,
133
- "nauc_mrr_at_5_max": 0.099963,
134
- "nauc_mrr_at_5_std": -0.016825,
135
- "nauc_mrr_at_5_diff1": 0.685233,
136
- "nauc_mrr_at_10_max": 0.09492,
137
- "nauc_mrr_at_10_std": -0.012515,
138
- "nauc_mrr_at_10_diff1": 0.681245,
139
- "nauc_mrr_at_20_max": 0.092859,
140
- "nauc_mrr_at_20_std": -0.00995,
141
- "nauc_mrr_at_20_diff1": 0.68014,
142
- "nauc_mrr_at_100_max": 0.092295,
143
- "nauc_mrr_at_100_std": -0.008175,
144
- "nauc_mrr_at_100_diff1": 0.67967,
145
- "nauc_mrr_at_1000_max": 0.092486,
146
- "nauc_mrr_at_1000_std": -0.008127,
147
- "nauc_mrr_at_1000_diff1": 0.679852,
148
- "main_score": 0.39719,
149
- "hf_subset": "default",
150
- "languages": [
151
- "vie-Latn"
152
- ]
153
- }
154
- ]
155
- },
156
- "evaluation_time": 716.343478679657,
157
- "kg_co2_emissions": null
158
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/HotpotQA-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "8a5220c7af5084f0d5d2afeb74f9c2b41b759ff0",
3
- "mteb_dataset_name": "HotpotQA-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "HotpotQA-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 13567.89,
10
- "map_at_1": 0.39226,
11
- "map_at_10": 0.6178,
12
- "map_at_100": 0.62675,
13
- "map_at_1000": 0.62738,
14
- "map_at_20": 0.6231,
15
- "map_at_3": 0.58315,
16
- "map_at_5": 0.60581,
17
- "mrr_at_1": 0.78453,
18
- "mrr_at_10": 0.84187,
19
- "mrr_at_100": 0.84371,
20
- "mrr_at_1000": 0.84377,
21
- "mrr_at_20": 0.8431,
22
- "mrr_at_3": 0.8326,
23
- "mrr_at_5": 0.83842,
24
- "ndcg_at_1": 0.78453,
25
- "ndcg_at_10": 0.70074,
26
- "ndcg_at_100": 0.73145,
27
- "ndcg_at_1000": 0.74324,
28
- "ndcg_at_20": 0.71379,
29
- "ndcg_at_3": 0.65249,
30
- "ndcg_at_5": 0.6806,
31
- "precision_at_1": 0.78453,
32
- "precision_at_10": 0.14613,
33
- "precision_at_100": 0.01701,
34
- "precision_at_1000": 0.00186,
35
- "precision_at_20": 0.07725,
36
- "precision_at_3": 0.4166,
37
- "precision_at_5": 0.27211,
38
- "recall_at_1": 0.39226,
39
- "recall_at_10": 0.73066,
40
- "recall_at_100": 0.85026,
41
- "recall_at_1000": 0.92794,
42
- "recall_at_20": 0.7725,
43
- "recall_at_3": 0.6249,
44
- "recall_at_5": 0.68027,
45
- "hf_subset": "default",
46
- "main_score": 0.70074,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ImdbVNClassification.json DELETED
@@ -1,95 +0,0 @@
1
- {
2
- "dataset_revision": "0dccb383ee26c90c99d03c8674cf40de642f099a",
3
- "task_name": "ImdbVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.830619,
9
- "f1": 0.830218,
10
- "f1_weighted": 0.830237,
11
- "ap": 0.775496,
12
- "ap_weighted": 0.775496,
13
- "scores_per_experiment": [
14
- {
15
- "accuracy": 0.845523,
16
- "f1": 0.845006,
17
- "f1_weighted": 0.844953,
18
- "ap": 0.811288,
19
- "ap_weighted": 0.811288
20
- },
21
- {
22
- "accuracy": 0.812101,
23
- "f1": 0.810477,
24
- "f1_weighted": 0.810581,
25
- "ap": 0.740588,
26
- "ap_weighted": 0.740588
27
- },
28
- {
29
- "accuracy": 0.837824,
30
- "f1": 0.83732,
31
- "f1_weighted": 0.837373,
32
- "ap": 0.77416,
33
- "ap_weighted": 0.77416
34
- },
35
- {
36
- "accuracy": 0.860559,
37
- "f1": 0.860469,
38
- "f1_weighted": 0.860448,
39
- "ap": 0.820455,
40
- "ap_weighted": 0.820455
41
- },
42
- {
43
- "accuracy": 0.839409,
44
- "f1": 0.839408,
45
- "f1_weighted": 0.839411,
46
- "ap": 0.787281,
47
- "ap_weighted": 0.787281
48
- },
49
- {
50
- "accuracy": 0.816856,
51
- "f1": 0.816719,
52
- "f1_weighted": 0.81669,
53
- "ap": 0.767843,
54
- "ap_weighted": 0.767843
55
- },
56
- {
57
- "accuracy": 0.81826,
58
- "f1": 0.817479,
59
- "f1_weighted": 0.81755,
60
- "ap": 0.751218,
61
- "ap_weighted": 0.751218
62
- },
63
- {
64
- "accuracy": 0.826729,
65
- "f1": 0.82647,
66
- "f1_weighted": 0.82651,
67
- "ap": 0.765153,
68
- "ap_weighted": 0.765153
69
- },
70
- {
71
- "accuracy": 0.816403,
72
- "f1": 0.816309,
73
- "f1_weighted": 0.816333,
74
- "ap": 0.756783,
75
- "ap_weighted": 0.756783
76
- },
77
- {
78
- "accuracy": 0.832526,
79
- "f1": 0.832525,
80
- "f1_weighted": 0.832524,
81
- "ap": 0.780192,
82
- "ap_weighted": 0.780192
83
- }
84
- ],
85
- "main_score": 0.830619,
86
- "hf_subset": "default",
87
- "languages": [
88
- "vie-Latn"
89
- ]
90
- }
91
- ]
92
- },
93
- "evaluation_time": 162.1266748905182,
94
- "kg_co2_emissions": null
95
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MSMARCO-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "85d1ad4cc9070b8d019d65f5af1631a2ab91e294",
3
- "mteb_dataset_name": "MSMARCO-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "MSMARCO-VN",
6
- "scores": {
7
- "dev": [
8
- {
9
- "evaluation_time": 22839.72,
10
- "map_at_1": 0.14118,
11
- "map_at_10": 0.24393,
12
- "map_at_100": 0.25563,
13
- "map_at_1000": 0.25629,
14
- "map_at_20": 0.25091,
15
- "map_at_3": 0.20812,
16
- "map_at_5": 0.22766,
17
- "mrr_at_1": 0.14551,
18
- "mrr_at_10": 0.24818,
19
- "mrr_at_100": 0.25954,
20
- "mrr_at_1000": 0.26016,
21
- "mrr_at_20": 0.25497,
22
- "mrr_at_3": 0.21241,
23
- "mrr_at_5": 0.23198,
24
- "ndcg_at_1": 0.14551,
25
- "ndcg_at_10": 0.30497,
26
- "ndcg_at_100": 0.36328,
27
- "ndcg_at_1000": 0.38075,
28
- "ndcg_at_20": 0.33003,
29
- "ndcg_at_3": 0.2309,
30
- "ndcg_at_5": 0.26603,
31
- "precision_at_1": 0.14551,
32
- "precision_at_10": 0.05166,
33
- "precision_at_100": 0.00808,
34
- "precision_at_1000": 0.00095,
35
- "precision_at_20": 0.03099,
36
- "precision_at_3": 0.10149,
37
- "precision_at_5": 0.07867,
38
- "recall_at_1": 0.14118,
39
- "recall_at_10": 0.49585,
40
- "recall_at_100": 0.76965,
41
- "recall_at_1000": 0.90623,
42
- "recall_at_20": 0.59377,
43
- "recall_at_3": 0.29241,
44
- "recall_at_5": 0.37722,
45
- "hf_subset": "default",
46
- "main_score": 0.30497,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPDomainVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "6e1ec8c54c018151c77472d94b1c0765230cf6ca",
3
- "task_name": "MTOPDomainVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.855346,
9
- "f1": 0.853271,
10
- "f1_weighted": 0.854398,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.841622,
14
- "f1": 0.833588,
15
- "f1_weighted": 0.841051
16
- },
17
- {
18
- "accuracy": 0.845535,
19
- "f1": 0.84299,
20
- "f1_weighted": 0.845127
21
- },
22
- {
23
- "accuracy": 0.860658,
24
- "f1": 0.858643,
25
- "f1_weighted": 0.859362
26
- },
27
- {
28
- "accuracy": 0.865097,
29
- "f1": 0.860371,
30
- "f1_weighted": 0.864139
31
- },
32
- {
33
- "accuracy": 0.85351,
34
- "f1": 0.85694,
35
- "f1_weighted": 0.852418
36
- },
37
- {
38
- "accuracy": 0.851704,
39
- "f1": 0.850781,
40
- "f1_weighted": 0.850459
41
- },
42
- {
43
- "accuracy": 0.867279,
44
- "f1": 0.863823,
45
- "f1_weighted": 0.865732
46
- },
47
- {
48
- "accuracy": 0.868558,
49
- "f1": 0.862947,
50
- "f1_weighted": 0.866173
51
- },
52
- {
53
- "accuracy": 0.828681,
54
- "f1": 0.837955,
55
- "f1_weighted": 0.829348
56
- },
57
- {
58
- "accuracy": 0.870815,
59
- "f1": 0.864666,
60
- "f1_weighted": 0.870169
61
- }
62
- ],
63
- "main_score": 0.855346,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 9.854939937591553,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPIntentVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "c4e81a5c9a813a0142d905e261e5a446cc6fbc4a",
3
- "task_name": "MTOPIntentVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.580114,
9
- "f1": 0.394653,
10
- "f1_weighted": 0.619897,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.558573,
14
- "f1": 0.394329,
15
- "f1_weighted": 0.598478
16
- },
17
- {
18
- "accuracy": 0.591528,
19
- "f1": 0.401524,
20
- "f1_weighted": 0.634855
21
- },
22
- {
23
- "accuracy": 0.580618,
24
- "f1": 0.387417,
25
- "f1_weighted": 0.616344
26
- },
27
- {
28
- "accuracy": 0.601685,
29
- "f1": 0.407789,
30
- "f1_weighted": 0.648841
31
- },
32
- {
33
- "accuracy": 0.569408,
34
- "f1": 0.394753,
35
- "f1_weighted": 0.606008
36
- },
37
- {
38
- "accuracy": 0.578963,
39
- "f1": 0.379034,
40
- "f1_weighted": 0.616886
41
- },
42
- {
43
- "accuracy": 0.58739,
44
- "f1": 0.400041,
45
- "f1_weighted": 0.622262
46
- },
47
- {
48
- "accuracy": 0.568731,
49
- "f1": 0.392647,
50
- "f1_weighted": 0.610166
51
- },
52
- {
53
- "accuracy": 0.562862,
54
- "f1": 0.393473,
55
- "f1_weighted": 0.60385
56
- },
57
- {
58
- "accuracy": 0.601384,
59
- "f1": 0.395527,
60
- "f1_weighted": 0.64128
61
- }
62
- ],
63
- "main_score": 0.580114,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 16.95852541923523,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveIntentVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "35c7ced69f958dbbaa24f792db4a9250e461866d",
3
- "task_name": "MassiveIntentVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.67741,
9
- "f1": 0.643766,
10
- "f1_weighted": 0.665189,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.679933,
14
- "f1": 0.670853,
15
- "f1_weighted": 0.66368
16
- },
17
- {
18
- "accuracy": 0.705717,
19
- "f1": 0.65224,
20
- "f1_weighted": 0.698288
21
- },
22
- {
23
- "accuracy": 0.675448,
24
- "f1": 0.627179,
25
- "f1_weighted": 0.662605
26
- },
27
- {
28
- "accuracy": 0.707399,
29
- "f1": 0.675465,
30
- "f1_weighted": 0.699657
31
- },
32
- {
33
- "accuracy": 0.659193,
34
- "f1": 0.624411,
35
- "f1_weighted": 0.641452
36
- },
37
- {
38
- "accuracy": 0.661996,
39
- "f1": 0.611837,
40
- "f1_weighted": 0.654907
41
- },
42
- {
43
- "accuracy": 0.679372,
44
- "f1": 0.646968,
45
- "f1_weighted": 0.669935
46
- },
47
- {
48
- "accuracy": 0.665359,
49
- "f1": 0.646588,
50
- "f1_weighted": 0.653284
51
- },
52
- {
53
- "accuracy": 0.648543,
54
- "f1": 0.624194,
55
- "f1_weighted": 0.628283
56
- },
57
- {
58
- "accuracy": 0.691143,
59
- "f1": 0.657925,
60
- "f1_weighted": 0.679795
61
- }
62
- ],
63
- "main_score": 0.67741,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 8.183109998703003,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveScenarioVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "a82e282d9f5aec1a8cf7d868ce40f70669c16b89",
3
- "task_name": "MassiveScenarioVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.728523,
9
- "f1": 0.701865,
10
- "f1_weighted": 0.724309,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.74217,
14
- "f1": 0.721876,
15
- "f1_weighted": 0.741215
16
- },
17
- {
18
- "accuracy": 0.746085,
19
- "f1": 0.724708,
20
- "f1_weighted": 0.744061
21
- },
22
- {
23
- "accuracy": 0.742729,
24
- "f1": 0.715626,
25
- "f1_weighted": 0.740294
26
- },
27
- {
28
- "accuracy": 0.725951,
29
- "f1": 0.704168,
30
- "f1_weighted": 0.71827
31
- },
32
- {
33
- "accuracy": 0.705257,
34
- "f1": 0.665866,
35
- "f1_weighted": 0.696431
36
- },
37
- {
38
- "accuracy": 0.722595,
39
- "f1": 0.691101,
40
- "f1_weighted": 0.719705
41
- },
42
- {
43
- "accuracy": 0.716443,
44
- "f1": 0.689115,
45
- "f1_weighted": 0.708131
46
- },
47
- {
48
- "accuracy": 0.71085,
49
- "f1": 0.69065,
50
- "f1_weighted": 0.710364
51
- },
52
- {
53
- "accuracy": 0.739374,
54
- "f1": 0.711083,
55
- "f1_weighted": 0.736213
56
- },
57
- {
58
- "accuracy": 0.733781,
59
- "f1": 0.704461,
60
- "f1_weighted": 0.728405
61
- }
62
- ],
63
- "main_score": 0.728523,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 5.51207423210144,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NFCorpus-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "a13d72fbb859be3dc19ab669d1ec9510407d2dcd",
3
- "mteb_dataset_name": "NFCorpus-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "NFCorpus-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 19.28,
10
- "map_at_1": 0.03605,
11
- "map_at_10": 0.08866,
12
- "map_at_100": 0.10878,
13
- "map_at_1000": 0.11959,
14
- "map_at_20": 0.09686,
15
- "map_at_3": 0.06829,
16
- "map_at_5": 0.07966,
17
- "mrr_at_1": 0.31928,
18
- "mrr_at_10": 0.41437,
19
- "mrr_at_100": 0.42156,
20
- "mrr_at_1000": 0.422,
21
- "mrr_at_20": 0.41738,
22
- "mrr_at_3": 0.39056,
23
- "mrr_at_5": 0.3993,
24
- "ndcg_at_1": 0.3012,
25
- "ndcg_at_10": 0.25389,
26
- "ndcg_at_100": 0.23107,
27
- "ndcg_at_1000": 0.31251,
28
- "ndcg_at_20": 0.23925,
29
- "ndcg_at_3": 0.28632,
30
- "ndcg_at_5": 0.27126,
31
- "precision_at_1": 0.31928,
32
- "precision_at_10": 0.18855,
33
- "precision_at_100": 0.06163,
34
- "precision_at_1000": 0.01931,
35
- "precision_at_20": 0.14217,
36
- "precision_at_3": 0.2751,
37
- "precision_at_5": 0.23494,
38
- "recall_at_1": 0.03605,
39
- "recall_at_10": 0.12335,
40
- "recall_at_100": 0.24212,
41
- "recall_at_1000": 0.52674,
42
- "recall_at_20": 0.14914,
43
- "recall_at_3": 0.0801,
44
- "recall_at_5": 0.09947,
45
- "hf_subset": "default",
46
- "main_score": 0.25389,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NQ-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "40a6d7f343b9c9f4855a426d8c431ad5f8aaf56b",
3
- "mteb_dataset_name": "NQ-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "NQ-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 8244.07,
10
- "map_at_1": 0.22333,
11
- "map_at_10": 0.35419,
12
- "map_at_100": 0.36617,
13
- "map_at_1000": 0.36666,
14
- "map_at_20": 0.36191,
15
- "map_at_3": 0.31386,
16
- "map_at_5": 0.3369,
17
- "mrr_at_1": 0.25158,
18
- "mrr_at_10": 0.37521,
19
- "mrr_at_100": 0.38499,
20
- "mrr_at_1000": 0.38535,
21
- "mrr_at_20": 0.38163,
22
- "mrr_at_3": 0.33939,
23
- "mrr_at_5": 0.36031,
24
- "ndcg_at_1": 0.25158,
25
- "ndcg_at_10": 0.42611,
26
- "ndcg_at_100": 0.47922,
27
- "ndcg_at_1000": 0.49185,
28
- "ndcg_at_20": 0.45171,
29
- "ndcg_at_3": 0.34799,
30
- "ndcg_at_5": 0.38727,
31
- "precision_at_1": 0.25158,
32
- "precision_at_10": 0.07363,
33
- "precision_at_100": 0.01035,
34
- "precision_at_1000": 0.00115,
35
- "precision_at_20": 0.04271,
36
- "precision_at_3": 0.16298,
37
- "precision_at_5": 0.11979,
38
- "recall_at_1": 0.22333,
39
- "recall_at_10": 0.62671,
40
- "recall_at_100": 0.86105,
41
- "recall_at_1000": 0.95798,
42
- "recall_at_20": 0.72228,
43
- "recall_at_3": 0.42149,
44
- "recall_at_5": 0.51241,
45
- "hf_subset": "default",
46
- "main_score": 0.42611,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Quora-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "3363d81e41b67c1032bf3b234882a03d271e2289",
3
- "mteb_dataset_name": "Quora-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "Quora-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 1180.22,
10
- "map_at_1": 0.42135,
11
- "map_at_10": 0.54669,
12
- "map_at_100": 0.55931,
13
- "map_at_1000": 0.55994,
14
- "map_at_20": 0.55423,
15
- "map_at_3": 0.51137,
16
- "map_at_5": 0.53213,
17
- "mrr_at_1": 0.49979,
18
- "mrr_at_10": 0.59915,
19
- "mrr_at_100": 0.60569,
20
- "mrr_at_1000": 0.60589,
21
- "mrr_at_20": 0.60345,
22
- "mrr_at_3": 0.57472,
23
- "mrr_at_5": 0.58942,
24
- "ndcg_at_1": 0.49979,
25
- "ndcg_at_10": 0.61001,
26
- "ndcg_at_100": 0.65467,
27
- "ndcg_at_1000": 0.66382,
28
- "ndcg_at_20": 0.63036,
29
- "ndcg_at_3": 0.55641,
30
- "ndcg_at_5": 0.58314,
31
- "precision_at_1": 0.49979,
32
- "precision_at_10": 0.10162,
33
- "precision_at_100": 0.01399,
34
- "precision_at_1000": 0.00154,
35
- "precision_at_20": 0.05794,
36
- "precision_at_3": 0.2526,
37
- "precision_at_5": 0.1746,
38
- "recall_at_1": 0.42135,
39
- "recall_at_10": 0.74453,
40
- "recall_at_100": 0.92748,
41
- "recall_at_1000": 0.98654,
42
- "recall_at_20": 0.81644,
43
- "recall_at_3": 0.59369,
44
- "recall_at_5": 0.66411,
45
- "hf_subset": "default",
46
- "main_score": 0.61001,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClustering-VN.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "dataset_revision": "7f7d4097979633181b2df3f73905218f74c4665d",
3
- "task_name": "RedditClustering-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "v_measure": 0.438902,
9
- "v_measure_std": 0.044274,
10
- "v_measures": [
11
- 0.39263,
12
- 0.5316,
13
- 0.399559,
14
- 0.425985,
15
- 0.427061,
16
- 0.446164,
17
- 0.512027,
18
- 0.424049,
19
- 0.426086,
20
- 0.40386
21
- ],
22
- "main_score": 0.438902,
23
- "hf_subset": "default",
24
- "languages": [
25
- "vie-Latn"
26
- ]
27
- }
28
- ]
29
- },
30
- "evaluation_time": 125.31211185455322,
31
- "kg_co2_emissions": null
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClusteringP2P-VN.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "dataset_revision": "841856dcb82496f1f2f59356e4798ce15baeb200",
3
- "task_name": "RedditClusteringP2P-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "v_measure": 0.561615,
9
- "v_measure_std": 0.118839,
10
- "v_measures": [
11
- 0.605948,
12
- 0.611291,
13
- 0.601361,
14
- 0.345424,
15
- 0.620675,
16
- 0.572602,
17
- 0.315689,
18
- 0.660052,
19
- 0.610597,
20
- 0.672515
21
- ],
22
- "main_score": 0.561615,
23
- "hf_subset": "default",
24
- "languages": [
25
- "vie-Latn"
26
- ]
27
- }
28
- ]
29
- },
30
- "evaluation_time": 1739.9407813549042,
31
- "kg_co2_emissions": null
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SCIDOCS-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "724cddfa9d328a193f303a0a9b7789468ac79f26",
3
- "mteb_dataset_name": "SCIDOCS-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "SCIDOCS-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 96.99,
10
- "map_at_1": 0.03771,
11
- "map_at_10": 0.07617,
12
- "map_at_100": 0.08657,
13
- "map_at_1000": 0.08887,
14
- "map_at_20": 0.08085,
15
- "map_at_3": 0.06048,
16
- "map_at_5": 0.06859,
17
- "mrr_at_1": 0.1825,
18
- "mrr_at_10": 0.2542,
19
- "mrr_at_100": 0.26523,
20
- "mrr_at_1000": 0.26606,
21
- "mrr_at_20": 0.25994,
22
- "mrr_at_3": 0.23333,
23
- "mrr_at_5": 0.24621,
24
- "ndcg_at_1": 0.1825,
25
- "ndcg_at_10": 0.13034,
26
- "ndcg_at_100": 0.18324,
27
- "ndcg_at_1000": 0.23392,
28
- "ndcg_at_20": 0.14611,
29
- "ndcg_at_3": 0.13596,
30
- "ndcg_at_5": 0.11389,
31
- "precision_at_1": 0.1825,
32
- "precision_at_10": 0.062,
33
- "precision_at_100": 0.01407,
34
- "precision_at_1000": 0.00263,
35
- "precision_at_20": 0.04013,
36
- "precision_at_3": 0.1225,
37
- "precision_at_5": 0.094,
38
- "recall_at_1": 0.03771,
39
- "recall_at_10": 0.12642,
40
- "recall_at_100": 0.28675,
41
- "recall_at_1000": 0.5355,
42
- "recall_at_20": 0.16375,
43
- "recall_at_3": 0.07496,
44
- "recall_at_5": 0.09571,
45
- "hf_subset": "default",
46
- "main_score": 0.13034,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SICK-R-VN.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "dataset_revision": "bc89f0401983c456b609f7fb324278f346b2cccf",
3
- "task_name": "SICK-R-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "pearson": 0.80717,
9
- "spearman": 0.771026,
10
- "cosine_pearson": 0.80717,
11
- "cosine_spearman": 0.771062,
12
- "manhattan_pearson": 0.768392,
13
- "manhattan_spearman": 0.770402,
14
- "euclidean_pearson": 0.768774,
15
- "euclidean_spearman": 0.771054,
16
- "main_score": 0.771062,
17
- "hf_subset": "default",
18
- "languages": [
19
- "vie-Latn"
20
- ]
21
- }
22
- ]
23
- },
24
- "evaluation_time": 25.608262062072754,
25
- "kg_co2_emissions": null
26
- }