This view is limited to 50 files because it contains too many changes.  See the raw diff here.
Files changed (50) hide show
  1. .codecarbon.lock +0 -0
  2. .github/pull_request_template.md +15 -8
  3. .github/workflows/generate_cached_results.yml +0 -149
  4. .github/workflows/model-results-comparison.yaml +0 -82
  5. .github/workflows/stale_pr.yml +0 -20
  6. .github/workflows/test.yml +3 -13
  7. .gitignore +0 -6
  8. README.md +2 -1
  9. load_external.py +253 -0
  10. makefile +2 -2
  11. paths.json +0 -0
  12. pyproject.toml +2 -13
  13. results.py +0 -3
  14. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonCounterfactualVNClassification.json +0 -95
  15. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonPolarityVNClassification.json +0 -95
  16. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonReviewsVNClassification.json +0 -73
  17. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ArguAna-VN.json +0 -158
  18. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AskUbuntuDupQuestions-VN.json +0 -26
  19. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/BIOSSES-VN.json +0 -26
  20. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Banking77VNClassification.json +0 -73
  21. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackAndroid-VN.json +0 -54
  22. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackGis-VN.json +0 -54
  23. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackMathematica-VN.json +0 -54
  24. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackPhysics-VN.json +0 -54
  25. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackProgrammers-VN.json +0 -54
  26. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackStats-VN.json +0 -54
  27. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackTex-VN.json +0 -54
  28. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackUnix-VN.json +0 -54
  29. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWebmasters-VN.json +0 -54
  30. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWordpress-VN.json +0 -54
  31. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ClimateFEVER-VN.json +0 -54
  32. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/DBPedia-VN.json +0 -54
  33. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/EmotionVNClassification.json +0 -137
  34. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FEVER-VN.json +0 -54
  35. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FiQA2018-VN.json +0 -54
  36. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/GreenNodeTableMarkdownRetrieval.json +0 -158
  37. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/HotpotQA-VN.json +0 -54
  38. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ImdbVNClassification.json +0 -95
  39. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MSMARCO-VN.json +0 -54
  40. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPDomainVNClassification.json +0 -73
  41. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPIntentVNClassification.json +0 -73
  42. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveIntentVNClassification.json +0 -73
  43. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveScenarioVNClassification.json +0 -73
  44. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NFCorpus-VN.json +0 -54
  45. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NQ-VN.json +0 -54
  46. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Quora-VN.json +0 -54
  47. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClustering-VN.json +0 -32
  48. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClusteringP2P-VN.json +0 -32
  49. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SCIDOCS-VN.json +0 -54
  50. results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SICK-R-VN.json +0 -26
.codecarbon.lock ADDED
File without changes
.github/pull_request_template.md CHANGED
@@ -1,11 +1,18 @@
1
- <!-- Description of the PR goes here -->
2
 
3
- ### Checklist
4
 
5
- - [ ] My model has a model sheet, report, or similar
6
- - [ ] My model has a reference implementation in [`mteb/models/model_implementations/`](https://github.com/embeddings-benchmark/mteb/tree/main/mteb/models/model_implementations), this can be as an API. Instruction on how to add a model can be found [here](https://embeddings-benchmark.github.io/mteb/contributing/adding_a_model/)
7
- - [ ] No, but there is an existing PR ___
8
- - [ ] The results submitted are obtained using the reference implementation
9
- - [ ] My model is available, either as a publicly accessible API or publicly on e.g., Huggingface
10
- - [ ] I *solemnly swear* that for all results submitted I have not trained on the evaluation dataset including training splits. If I have, I have disclosed it clearly.
11
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- If you are submitting a dataset or a model for the model registry, please use the corresponding checklists below. Otherwise, feel free to remove them. -->
2
 
3
+ <!-- add additional description, question etc. related to the new dataset -->
4
 
 
 
 
 
 
 
5
 
6
+ ## Checklist
7
+ <!-- Please do not delete this -->
8
+
9
+ - [ ] Run tests locally to make sure nothing is broken using `make test`.
10
+ - [ ] Run the results files checker `make pre-push`.
11
+
12
+ ### Adding a model checklist
13
+ <!--
14
+ When adding a model to the model registry
15
+ see also https://github.com/embeddings-benchmark/mteb/blob/main/docs/reproducible_workflow.md
16
+ -->
17
+
18
+ - [ ] I have added model implementation to `mteb/models/` [directory](https://github.com/embeddings-benchmark/mteb/tree/main/mteb/models). Instruction to add a model can be found [here](https://github.com/embeddings-benchmark/mteb/blob/main/docs/reproducible_workflow.md) in the following PR ____
.github/workflows/generate_cached_results.yml DELETED
@@ -1,149 +0,0 @@
1
- name: Generate Cached Results
2
-
3
- on:
4
- push:
5
- branches: [main]
6
- # Allow manual trigger for testing
7
- workflow_dispatch:
8
-
9
- jobs:
10
- generate-cache:
11
- runs-on: ubuntu-latest
12
- steps:
13
- - name: Free disk space
14
- run: |
15
- sudo rm -rf /usr/share/dotnet
16
- sudo rm -rf /opt/ghc
17
- sudo rm -rf /usr/local/share/boost
18
- docker system prune -af
19
-
20
- - name: Checkout repository
21
- uses: actions/checkout@v4
22
- with:
23
- fetch-depth: 0
24
- token: ${{ secrets.GITHUB_TOKEN }}
25
- lfs: true
26
-
27
- - name: Setup Python
28
- uses: actions/setup-python@v5
29
- with:
30
- python-version: '3.10'
31
- cache: 'pip'
32
-
33
- - name: Install dependencies
34
- run: |
35
- pip install git+https://github.com/embeddings-benchmark/mteb.git
36
-
37
- - name: Generate cached results
38
- run: |
39
- python scripts/generate_cached_results.py
40
- env:
41
- PYTHONUNBUFFERED: 1
42
-
43
- - name: Install Git LFS
44
- run: |
45
- sudo apt-get update
46
- sudo apt-get install -y git-lfs
47
- git lfs install
48
- # Explicitly pull LFS files to work around checkout action issues
49
- git lfs pull || true
50
-
51
- - name: Configure Git
52
- run: |
53
- git config --global user.name "github-actions[bot]"
54
- git config --global user.email "github-actions[bot]@users.noreply.github.com"
55
-
56
- - name: Update cached-data branch
57
- run: |
58
- # Check if __cached_results.json.gz was created
59
- if [ ! -f "__cached_results.json.gz" ]; then
60
- echo "❌ Cached results file not found"
61
- exit 1
62
- fi
63
-
64
- # Get file size for logging
65
- FILE_SIZE=$(stat -f%z __cached_results.json.gz 2>/dev/null || stat -c%s __cached_results.json.gz)
66
- echo "📦 Generated cache file: $(echo "scale=1; $FILE_SIZE/1024/1024" | bc -l)MB"
67
-
68
- # Temporarily move the cache file to avoid checkout conflicts
69
- if [ -f "__cached_results.json.gz" ]; then
70
- mv __cached_results.json.gz __cached_results.json.gz.tmp
71
- fi
72
-
73
- # Check if cached-data branch exists
74
- if git show-ref --verify --quiet refs/remotes/origin/cached-data; then
75
- echo "📋 Switching to existing cached-data branch"
76
- git checkout cached-data
77
- git pull origin cached-data
78
- # Ensure LFS files are available after branch checkout
79
- git lfs pull || true
80
- else
81
- echo "🆕 Creating new cached-data branch"
82
- git checkout --orphan cached-data
83
- # Remove all files from staging area when creating orphan branch
84
- git rm -rf . 2>/dev/null || true
85
- fi
86
-
87
- # Restore the cache file
88
- if [ -f "__cached_results.json.gz.tmp" ]; then
89
- mv __cached_results.json.gz.tmp __cached_results.json.gz
90
- fi
91
-
92
- # Setup Git LFS tracking for the cache file (if not already tracked)
93
- if ! git lfs ls-files | grep -q "__cached_results.json.gz"; then
94
- git lfs track "__cached_results.json.gz"
95
- fi
96
-
97
- # Ensure we only have the files we want
98
- # Remove all tracked files except README.md and .gitattributes
99
- if [ -f "README.md" ]; then
100
- git ls-files | grep -v "README.md" | grep -v ".gitattributes" | xargs -r git rm 2>/dev/null || true
101
- else
102
- git ls-files | grep -v ".gitattributes" | xargs -r git rm 2>/dev/null || true
103
- fi
104
-
105
- # Add the cached results file (will be tracked by LFS)
106
- git add __cached_results.json.gz
107
-
108
- # Preserve README.md if it exists in the working directory
109
- if [ -f "README.md" ]; then
110
- git add README.md
111
- fi
112
-
113
- # Add .gitattributes file (required for LFS tracking)
114
- if [ -f ".gitattributes" ]; then
115
- git add .gitattributes
116
- fi
117
-
118
- # Check if there are changes to commit
119
- if git diff --staged --quiet; then
120
- echo "✅ No changes in cached results, skipping commit"
121
- else
122
- # Verify we're not committing too many files (safety check)
123
- STAGED_FILES=$(git diff --staged --name-only | wc -l)
124
- if [ "$STAGED_FILES" -gt 10 ]; then
125
- echo "❌ ERROR: Too many files staged ($STAGED_FILES). Expected only 1-3 files (__cached_results.json.gz, README.md, .gitattributes)."
126
- echo "Staged files:"
127
- git diff --staged --name-only
128
- exit 1
129
- fi
130
-
131
- # Commit with timestamp and file size
132
- TIMESTAMP=$(date -u '+%Y-%m-%d %H:%M:%S UTC')
133
- COMMIT_MSG="Update cached results - $TIMESTAMP ($(echo "scale=1; $FILE_SIZE/1024/1024" | bc -l)MB)"
134
- git commit -m "$COMMIT_MSG"
135
-
136
- # Push to remote (LFS will handle the large file automatically)
137
- git push origin cached-data
138
- echo "✅ Successfully updated cached-data branch"
139
- fi
140
-
141
- - name: Report status
142
- if: always()
143
- run: |
144
- if [ -f "__cached_results.json.gz" ]; then
145
- FILE_SIZE=$(stat -f%z __cached_results.json.gz 2>/dev/null || stat -c%s __cached_results.json.gz)
146
- echo "✅ Workflow completed. Cache file size: $(echo "scale=1; $FILE_SIZE/1024/1024" | bc -l)MB"
147
- else
148
- echo "❌ Workflow failed - no cache file generated"
149
- fi
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.github/workflows/model-results-comparison.yaml DELETED
@@ -1,82 +0,0 @@
1
- name: Model Results Comparison
2
-
3
- on:
4
- pull_request_target:
5
- types: [opened, synchronize, edited]
6
- paths:
7
- - 'results/**/*.json'
8
- workflow_dispatch:
9
- inputs:
10
- reference_models:
11
- description: 'Space-separated list of reference models for comparison'
12
- required: true
13
- type: string
14
- default: 'intfloat/multilingual-e5-large google/gemini-embedding-001'
15
- pull_request_number:
16
- description: 'The pull request number to comment on (required if triggered manually)'
17
- required: false
18
- type: string
19
-
20
- permissions:
21
- contents: read
22
- pull-requests: write
23
-
24
- jobs:
25
- compare-results:
26
- runs-on: ubuntu-latest
27
-
28
- steps:
29
- - name: Free disk space
30
- run: |
31
- sudo rm -rf /usr/share/dotnet
32
- sudo rm -rf /opt/ghc
33
- sudo rm -rf /usr/local/share/boost
34
- docker system prune -af
35
-
36
- - name: Checkout code
37
- uses: actions/checkout@v4
38
- with:
39
- # IMPORTANT: For pull_request_target, check out the PR branch explicitly
40
- ref: ${{ github.event.pull_request.head.sha }}
41
- fetch-depth: 0
42
-
43
- - name: Fetch origin main
44
- run: git fetch origin main
45
-
46
- - name: Set up Python
47
- uses: actions/setup-python@v5
48
- with:
49
- python-version: '3.10'
50
-
51
- - name: Install dependencies
52
- run: |
53
- pip install git+https://github.com/embeddings-benchmark/mteb.git tabulate
54
-
55
- - name: Generate model comparison
56
- env:
57
- REFERENCE_MODELS: ${{ github.event.inputs.reference_models || 'intfloat/multilingual-e5-large google/gemini-embedding-001' }}
58
- run: |
59
- python scripts/create_pr_results_comment.py --reference-models $REFERENCE_MODELS --output model-comparison.md
60
-
61
- - name: Upload comparison report
62
- uses: actions/upload-artifact@v4
63
- with:
64
- name: model-comparison
65
- path: model-comparison.md
66
-
67
- - name: Determine PR Number
68
- id: pr_info
69
- run: |
70
- if [ "${{ github.event_name }}" == "pull_request_target" ]; then
71
- echo "pr_number=${{ github.event.number }}" >> $GITHUB_OUTPUT
72
- elif [ "${{ github.event_name }}" == "workflow_dispatch" ] && [ -n "${{ github.event.inputs.pull_request_number }}" ]; then
73
- echo "pr_number=${{ github.event.inputs.pull_request_number }}" >> $GITHUB_OUTPUT
74
- else
75
- echo "pr_number=" >> $GITHUB_OUTPUT
76
- fi
77
-
78
- - name: Post PR comment
79
- if: steps.pr_info.outputs.pr_number != ''
80
- env:
81
- GITHUB_TOKEN: ${{ github.token }}
82
- run: gh pr comment ${{ steps.pr_info.outputs.pr_number }} --body-file model-comparison.md --create-if-none --edit-last
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.github/workflows/stale_pr.yml DELETED
@@ -1,20 +0,0 @@
1
- name: Label and Close stale PRs
2
- on:
3
- schedule:
4
- - cron: '30 1 * * *' # runs daily at 1:30 AM
5
- jobs:
6
- stale:
7
- runs-on: ubuntu-latest
8
- permissions:
9
- actions: write
10
- issues: write
11
- pull-requests: write
12
- steps:
13
- - uses: actions/stale@v9.1.0
14
- with:
15
- stale-pr-label: 'stale'
16
- stale-issue-label: 'stale'
17
- days-before-pr-stale: 14
18
- days-before-pr-close: 7
19
- stale-pr-message: 'This pull request has been automatically marked as stale due to inactivity.'
20
- close-pr-message: 'This pull request has been automatically closed due to inactivity.'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.github/workflows/test.yml CHANGED
@@ -2,6 +2,7 @@
2
  # 1) install Python dependencies
3
  # 2) run make test
4
 
 
5
  name: Test
6
  on:
7
  push:
@@ -15,22 +16,13 @@ jobs:
15
  fail-fast: false
16
  matrix:
17
  os: [ubuntu-latest] #, macos-latest, windows-latest]
18
- python-version: ["3.10"]
19
  steps:
20
- - name: Free disk space
21
- run: |
22
- sudo rm -rf /usr/share/dotnet
23
- sudo rm -rf /opt/ghc
24
- sudo rm -rf /usr/local/share/boost
25
- docker system prune -af
26
-
27
  - uses: actions/checkout@v3
28
- with:
29
- fetch-depth: 0 # to allow us to examine the git diff
30
 
31
  - name: Setup Python ${{ matrix.python-version }}
32
  uses: actions/setup-python@v4
33
-
34
  - name: Install dependencies
35
  shell: bash
36
  run: |
@@ -38,7 +30,5 @@ jobs:
38
 
39
  - name: Run tests
40
  shell: bash
41
- env:
42
- PR_BASE_SHA: ${{ github.event.pull_request.base.sha }}
43
  run: |
44
  make test
 
2
  # 1) install Python dependencies
3
  # 2) run make test
4
 
5
+
6
  name: Test
7
  on:
8
  push:
 
16
  fail-fast: false
17
  matrix:
18
  os: [ubuntu-latest] #, macos-latest, windows-latest]
19
+ python-version: ["3.9"]
20
  steps:
 
 
 
 
 
 
 
21
  - uses: actions/checkout@v3
 
 
22
 
23
  - name: Setup Python ${{ matrix.python-version }}
24
  uses: actions/setup-python@v4
25
+
26
  - name: Install dependencies
27
  shell: bash
28
  run: |
 
30
 
31
  - name: Run tests
32
  shell: bash
 
 
33
  run: |
34
  make test
.gitignore CHANGED
@@ -20,9 +20,3 @@ tmp.py
20
 
21
  # UV
22
  uv.lock
23
-
24
- build/
25
- results.egg-info/
26
-
27
- # codecarbon
28
- .codecarbon.lock
 
20
 
21
  # UV
22
  uv.lock
 
 
 
 
 
 
README.md CHANGED
@@ -5,7 +5,8 @@ submission_name: MTEB
5
  ---
6
 
7
  > [!NOTE]
8
- > Previously, it was possible to submit model results to MTEB by adding them to the metadata of the model card on huggingface. However, this is no longer possible as we want to ensure that we can match the results with the model implementation. If you want to add your model, please follow the [guide](https://embeddings-benchmark.github.io/mteb/contributing/adding_a_model/) on how to do so.
 
9
 
10
  This repository contains the results of the embedding benchmark evaluated using the package `mteb`.
11
 
 
5
  ---
6
 
7
  > [!NOTE]
8
+ > Previously it was possible to submit models results to MTEB by adding the results to the model metadata. This is no longer an option as we want to ensure high quality metadata.
9
+
10
 
11
  This repository contains the results of the embedding benchmark evaluated using the package `mteb`.
12
 
load_external.py ADDED
@@ -0,0 +1,253 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __future__ import annotations
2
+
3
+ import json
4
+ import logging
5
+ import math
6
+ import re
7
+ from pathlib import Path
8
+ from typing import Any
9
+
10
+ from huggingface_hub import HfApi, get_hf_file_metadata, hf_hub_download, hf_hub_url
11
+ from huggingface_hub.errors import NotASafetensorsRepoError
12
+ from huggingface_hub.hf_api import ModelInfo
13
+ from huggingface_hub.repocard import metadata_load
14
+ from mteb import ModelMeta, get_task
15
+
16
+ API = HfApi()
17
+ logger = logging.getLogger(__name__)
18
+
19
+
20
+ library_mapping = {
21
+ "sentence-transformers": "Sentence Transformers",
22
+ }
23
+
24
+
25
+ def get_model_dir(model_id: str) -> Path:
26
+ external_result_dir = Path("results") / model_id.replace("/", "__") / "external"
27
+ return external_result_dir
28
+
29
+
30
+ renamed_tasks = {
31
+ "NorwegianParliament": "NorwegianParliamentClassification",
32
+ "CMedQAv2": "CMedQAv2-reranking",
33
+ "CMedQAv1": "CMedQAv1-reranking",
34
+ "8TagsClustering": "EightTagsClustering",
35
+ "PPC": "PpcPC",
36
+ "PawsX": "PawsXParaphraseIdentification",
37
+ }
38
+
39
+
40
+ def simplify_dataset_name(name: str) -> str:
41
+ task_name = name.replace("MTEB ", "").split()[0]
42
+ return renamed_tasks.get(task_name, task_name)
43
+
44
+
45
+ def get_model_parameters_memory(model_info: ModelInfo) -> tuple[int| None, float|None]:
46
+ try:
47
+ safetensors = API.get_safetensors_metadata(model_info.id)
48
+ num_parameters = sum(safetensors.parameter_count.values())
49
+ return num_parameters, round(num_parameters * 4 / 1024 ** 3, 2)
50
+ except NotASafetensorsRepoError as e:
51
+ logger.info(f"Could not find SafeTensors metadata for {model_info.id}")
52
+
53
+ filenames = [sib.rfilename for sib in model_info.siblings]
54
+ if "pytorch_model.bin" in filenames:
55
+ url = hf_hub_url(model_info.id, filename="pytorch_model.bin")
56
+ meta = get_hf_file_metadata(url)
57
+ bytes_per_param = 4
58
+ num_params = round(meta.size / bytes_per_param)
59
+ size_gb = round(meta.size * (4 / bytes_per_param) / 1024 ** 3, 2)
60
+ return num_params, size_gb
61
+ if "pytorch_model.bin.index.json" in filenames:
62
+ index_path = hf_hub_download(model_info.id, filename="pytorch_model.bin.index.json")
63
+ size = json.load(open(index_path))
64
+ bytes_per_param = 4
65
+ if "metadata" in size and "total_size" in size["metadata"]:
66
+ return round(size["metadata"]["total_size"] / bytes_per_param), round(size["metadata"]["total_size"] / 1024 ** 3, 2)
67
+ logger.info(f"Could not find the model parameters for {model_info.id}")
68
+ return None, None
69
+
70
+
71
+ def get_dim_seq_size(model: ModelInfo) -> tuple[str | None, str | None, int, float]:
72
+ siblings = model.siblings or []
73
+ filenames = [sib.rfilename for sib in siblings]
74
+ dim, seq = None, None
75
+ for filename in filenames:
76
+ if re.match(r"\d+_Pooling/config.json", filename):
77
+ st_config_path = hf_hub_download(model.id, filename=filename)
78
+ dim = json.load(open(st_config_path)).get("word_embedding_dimension", None)
79
+ break
80
+ for filename in filenames:
81
+ if re.match(r"\d+_Dense/config.json", filename):
82
+ st_config_path = hf_hub_download(model.id, filename=filename)
83
+ dim = json.load(open(st_config_path)).get("out_features", dim)
84
+ if "config.json" in filenames:
85
+ config_path = hf_hub_download(model.id, filename="config.json")
86
+ config = json.load(open(config_path))
87
+ if not dim:
88
+ dim = config.get("hidden_dim", config.get("hidden_size", config.get("d_model", None)))
89
+ seq = config.get("n_positions", config.get("max_position_embeddings", config.get("n_ctx", config.get("seq_length", None))))
90
+
91
+ parameters, memory = get_model_parameters_memory(model)
92
+ return dim, seq, parameters, memory
93
+
94
+
95
+ def create_model_meta(model_info: ModelInfo) -> ModelMeta | None:
96
+ readme_path = hf_hub_download(model_info.id, filename="README.md", etag_timeout=30)
97
+ meta = metadata_load(readme_path)
98
+ dim, seq, parameters, memory = None, None, None, None
99
+ try:
100
+ dim, seq, parameters, memory = get_dim_seq_size(model_info)
101
+ except Exception as e:
102
+ logger.error(f"Error getting model parameters for {model_info.id}, {e}")
103
+
104
+ release_date = str(model_info.created_at.date()) if model_info.created_at else ""
105
+ library = [library_mapping[model_info.library_name]] if model_info.library_name in library_mapping else []
106
+ languages = meta.get("language", [])
107
+ if not isinstance(languages, list) and isinstance(languages, str):
108
+ languages = [languages]
109
+ # yaml transforms norwegian `no` to False
110
+ for i in range(len(languages)):
111
+ if languages[i] is False:
112
+ languages[i] = "no"
113
+
114
+ model_meta = ModelMeta(
115
+ name=model_info.id,
116
+ revision=model_info.sha,
117
+ release_date=release_date,
118
+ open_weights=True,
119
+ framework=library,
120
+ license=meta.get("license", None),
121
+ embed_dim=dim,
122
+ max_tokens=seq,
123
+ n_parameters=parameters,
124
+ languages=languages,
125
+ )
126
+ return model_meta
127
+
128
+
129
+ def parse_readme(model_info: ModelInfo) -> dict[str, dict[str, Any]] | None:
130
+ model_id = model_info.id
131
+ try:
132
+ readme_path = hf_hub_download(model_info.id, filename="README.md", etag_timeout=30)
133
+ except Exception:
134
+ logger.warning(f"ERROR: Could not fetch metadata for {model_id}, trying again")
135
+ readme_path = hf_hub_download(model_id, filename="README.md", etag_timeout=30)
136
+ meta = metadata_load(readme_path)
137
+ if "model-index" not in meta:
138
+ logger.info(f"Could not find model-index in {model_id}")
139
+ return
140
+ model_index = meta["model-index"][0]
141
+ model_name_from_readme = model_index.get("name", None)
142
+ orgs = ["Alibaba-NLP", "HIT-TMG", "McGill-NLP", "Snowflake", "facebook", "jinaai", "nomic-ai"]
143
+ is_org = any([model_id.startswith(org) for org in orgs])
144
+ # There a lot of reuploads with tunes, quantization, etc. We only want the original model
145
+ # to prevent this most of the time we can check if the model name from the readme is the same as the model id
146
+ # but some orgs have a different naming in their readme
147
+ if model_name_from_readme and not model_info.id.endswith(model_name_from_readme) and not is_org:
148
+ logger.warning(f"Model name mismatch: {model_info.id} vs {model_name_from_readme}")
149
+ return
150
+ results = model_index.get("results", [])
151
+ model_results = {}
152
+ for result in results:
153
+ dataset = result["dataset"]
154
+ dataset_type = simplify_dataset_name(dataset["name"])
155
+
156
+ if dataset_type not in model_results:
157
+ output_dict = {
158
+ "dataset_revision": dataset.get("revision", ""),
159
+ "task_name": simplify_dataset_name(dataset["name"]),
160
+ "evaluation_time": None,
161
+ "mteb_version": None,
162
+ "scores": {},
163
+ }
164
+ else:
165
+ output_dict = model_results[dataset_type]
166
+
167
+ try:
168
+ mteb_task = get_task(output_dict["task_name"])
169
+ except Exception:
170
+ logger.warning(f"Error getting task for {model_id} {output_dict['task_name']}")
171
+ continue
172
+
173
+ mteb_task_metadata = mteb_task.metadata
174
+ mteb_task_eval_languages = mteb_task_metadata.eval_langs
175
+
176
+ scores_dict = output_dict["scores"]
177
+ current_split = dataset["split"]
178
+ current_config = dataset.get("config", "")
179
+ cur_split_metrics = {
180
+ "hf_subset": current_config,
181
+ "languages": mteb_task_eval_languages if isinstance(mteb_task_eval_languages, list) else mteb_task_eval_languages.get(current_config, ["None"]),
182
+ }
183
+ for metric in result["metrics"]:
184
+ if isinstance(metric["value"], (float, int)):
185
+ cur_split_metrics[metric["type"]] = metric["value"] / 100
186
+ else:
187
+ cur_split_metrics[metric["type"]] = metric["value"]
188
+
189
+ main_score_str = "main_score"
190
+ if main_score_str not in cur_split_metrics:
191
+ # old sts and sum_eval have cos_sim_pearson, but in model_meta cosine_spearman is main_score
192
+ for old_metric, new_metric in zip(["cos_sim_pearson", "cos_sim_spearman"], ["cosine_pearson", "cosine_spearman"]):
193
+ if old_metric in cur_split_metrics:
194
+ cur_split_metrics[new_metric] = cur_split_metrics[old_metric]
195
+
196
+ if mteb_task.metadata.main_score not in cur_split_metrics:
197
+ logger.warning(f"Could not find main score for {model_id} {output_dict['task_name']}, mteb task {mteb_task.metadata.name}. Main score: {mteb_task.metadata.main_score}. Metrics: {cur_split_metrics}, result {result['metrics']}")
198
+ continue
199
+
200
+ cur_split_metrics[main_score_str] = cur_split_metrics.get(mteb_task.metadata.main_score, None)
201
+ split_metrics = scores_dict.get(current_split, [])
202
+ split_metrics.append(cur_split_metrics)
203
+ scores_dict[current_split] = split_metrics
204
+ model_results[dataset_type] = output_dict
205
+ return model_results
206
+
207
+
208
+ def get_mteb_data() -> None:
209
+ models = sorted(list(API.list_models(filter="mteb", full=True)), key=lambda x: x.id)
210
+ # models = [model for model in models if model.id == "ai-forever/ru-en-RoSBERTa"]
211
+ for i, model_info in enumerate(models, start=1):
212
+ logger.info(f"[{i}/{len(models)}] Processing {model_info.id}")
213
+ model_path = get_model_dir(model_info.id)
214
+ if (model_path / "model_meta.json").exists() and len(list(model_path.glob("*.json"))) > 1:
215
+ logger.info(f"Model meta already exists for {model_info.id}")
216
+ # continue
217
+ if model_info.id.lower().endswith("gguf"):
218
+ logger.info(f"Skipping {model_info.id} GGUF model")
219
+ continue
220
+
221
+ spam_users = ["ILKT", "fine-tuned", "mlx-community"]
222
+ is_spam = False
223
+ for spam_user in spam_users:
224
+ if model_info.id.startswith(spam_user):
225
+ logger.info(f"Skipping {model_info.id}")
226
+ is_spam = True
227
+ continue
228
+ if is_spam:
229
+ continue
230
+ model_meta = create_model_meta(model_info)
231
+ model_results = parse_readme(model_info)
232
+
233
+ if not model_meta or not model_results:
234
+ logger.warning(f"Could not get model meta or results for {model_info.id}")
235
+ continue
236
+
237
+ if not model_path.exists():
238
+ model_path.mkdir(parents=True, exist_ok=True)
239
+
240
+ model_meta_path = model_path / "model_meta.json"
241
+ with model_meta_path.open("w") as f:
242
+ json.dump(model_meta.model_dump(), f, indent=4)
243
+
244
+ for model_result in model_results:
245
+ task_name = model_results[model_result]["task_name"]
246
+ result_file = model_path / f"{task_name}.json"
247
+ with result_file.open("w") as f:
248
+ json.dump(model_results[model_result], f, indent=4)
249
+
250
+
251
+ if __name__ == "__main__":
252
+ logging.basicConfig(level=logging.INFO)
253
+ get_mteb_data()
makefile CHANGED
@@ -1,7 +1,7 @@
1
  install-for-tests:
2
  @echo "--- Installing dependencies for tests ---"
3
  pip install pip --upgrade
4
- pip install . --group dev
5
 
6
  test:
7
  @echo "--- Running tests ---"
@@ -9,4 +9,4 @@ test:
9
 
10
  pre-push:
11
  @echo "--- Running pre-push commands ---"
12
- python reduce_large_json_files.py
 
1
  install-for-tests:
2
  @echo "--- Installing dependencies for tests ---"
3
  pip install pip --upgrade
4
+ pip install .
5
 
6
  test:
7
  @echo "--- Running tests ---"
 
9
 
10
  pre-push:
11
  @echo "--- Running pre-push commands ---"
12
+ python reduce_large_json_files.py
paths.json CHANGED
The diff for this file is too large to render. See raw diff
 
pyproject.toml CHANGED
@@ -3,16 +3,5 @@ name = "results"
3
  version = "0.1.0"
4
  description = "The result repository for mteb"
5
  readme = "README.md"
6
- requires-python = ">=3.10"
7
- dependencies = ["mteb>=2.0"]
8
-
9
- [dependency-groups]
10
- dev = [
11
- "pytest>=8.3.4",
12
- ]
13
- lint = [
14
- "ruff>=0.14.9",
15
- ]
16
- pr-comment = [
17
- "tabulate>=0.9.0",
18
- ]
 
3
  version = "0.1.0"
4
  description = "The result repository for mteb"
5
  readme = "README.md"
6
+ requires-python = ">=3.9"
7
+ dependencies = ["mteb[dev]>=1.13.0"]
 
 
 
 
 
 
 
 
 
 
 
results.py CHANGED
@@ -84,7 +84,6 @@ EVAL_LANGS = [
84
  "est-eng",
85
  "eus-eng",
86
  "fa",
87
- "fas-Arab",
88
  "fao-eng",
89
  "fi",
90
  "fin-eng",
@@ -241,7 +240,6 @@ VALIDATION_SPLIT = [
241
  "LEMBSummScreenFDRetrieval",
242
  "MSMARCO",
243
  "MSMARCO-PL",
244
- "MSMARCO-Fa",
245
  "MultilingualSentiment",
246
  "Ocnli",
247
  "TNews",
@@ -257,7 +255,6 @@ DEV_SPLIT = [
257
  "MMarcoRetrieval",
258
  "MSMARCO",
259
  "MSMARCO-PL",
260
- "MSMARCO-Fa",
261
  "T2Reranking",
262
  "T2Retrieval",
263
  "VideoRetrieval",
 
84
  "est-eng",
85
  "eus-eng",
86
  "fa",
 
87
  "fao-eng",
88
  "fi",
89
  "fin-eng",
 
240
  "LEMBSummScreenFDRetrieval",
241
  "MSMARCO",
242
  "MSMARCO-PL",
 
243
  "MultilingualSentiment",
244
  "Ocnli",
245
  "TNews",
 
255
  "MMarcoRetrieval",
256
  "MSMARCO",
257
  "MSMARCO-PL",
 
258
  "T2Reranking",
259
  "T2Retrieval",
260
  "VideoRetrieval",
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonCounterfactualVNClassification.json DELETED
@@ -1,95 +0,0 @@
1
- {
2
- "dataset_revision": "b48bc27d383cfca5b6a47135a52390fa5f66b253",
3
- "task_name": "AmazonCounterfactualVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.619742,
9
- "f1": 0.561867,
10
- "f1_weighted": 0.660984,
11
- "ap": 0.257592,
12
- "ap_weighted": 0.257592,
13
- "scores_per_experiment": [
14
- {
15
- "accuracy": 0.587983,
16
- "f1": 0.555485,
17
- "f1_weighted": 0.631829,
18
- "ap": 0.276246,
19
- "ap_weighted": 0.276246
20
- },
21
- {
22
- "accuracy": 0.736052,
23
- "f1": 0.655812,
24
- "f1_weighted": 0.761372,
25
- "ap": 0.318313,
26
- "ap_weighted": 0.318313
27
- },
28
- {
29
- "accuracy": 0.652361,
30
- "f1": 0.594521,
31
- "f1_weighted": 0.691796,
32
- "ap": 0.280129,
33
- "ap_weighted": 0.280129
34
- },
35
- {
36
- "accuracy": 0.51073,
37
- "f1": 0.482958,
38
- "f1_weighted": 0.559073,
39
- "ap": 0.224991,
40
- "ap_weighted": 0.224991
41
- },
42
- {
43
- "accuracy": 0.622318,
44
- "f1": 0.539112,
45
- "f1_weighted": 0.6635,
46
- "ap": 0.219724,
47
- "ap_weighted": 0.219724
48
- },
49
- {
50
- "accuracy": 0.583691,
51
- "f1": 0.53557,
52
- "f1_weighted": 0.630528,
53
- "ap": 0.241011,
54
- "ap_weighted": 0.241011
55
- },
56
- {
57
- "accuracy": 0.690987,
58
- "f1": 0.626203,
59
- "f1_weighted": 0.725049,
60
- "ap": 0.302755,
61
- "ap_weighted": 0.302755
62
- },
63
- {
64
- "accuracy": 0.594421,
65
- "f1": 0.549602,
66
- "f1_weighted": 0.639849,
67
- "ap": 0.255337,
68
- "ap_weighted": 0.255337
69
- },
70
- {
71
- "accuracy": 0.618026,
72
- "f1": 0.556143,
73
- "f1_weighted": 0.661415,
74
- "ap": 0.244741,
75
- "ap_weighted": 0.244741
76
- },
77
- {
78
- "accuracy": 0.600858,
79
- "f1": 0.523266,
80
- "f1_weighted": 0.645433,
81
- "ap": 0.212676,
82
- "ap_weighted": 0.212676
83
- }
84
- ],
85
- "main_score": 0.619742,
86
- "hf_subset": "default",
87
- "languages": [
88
- "vie-Latn"
89
- ]
90
- }
91
- ]
92
- },
93
- "evaluation_time": 1.7541632652282715,
94
- "kg_co2_emissions": null
95
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonPolarityVNClassification.json DELETED
@@ -1,95 +0,0 @@
1
- {
2
- "dataset_revision": "4e9a0d6e6bd97ab32f23c50c043d751eed2a5f8a",
3
- "task_name": "AmazonPolarityVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.887758,
9
- "f1": 0.887555,
10
- "f1_weighted": 0.88755,
11
- "ap": 0.85126,
12
- "ap_weighted": 0.85126,
13
- "scores_per_experiment": [
14
- {
15
- "accuracy": 0.887756,
16
- "f1": 0.887564,
17
- "f1_weighted": 0.88753,
18
- "ap": 0.86203,
19
- "ap_weighted": 0.86203
20
- },
21
- {
22
- "accuracy": 0.886623,
23
- "f1": 0.886583,
24
- "f1_weighted": 0.886599,
25
- "ap": 0.840929,
26
- "ap_weighted": 0.840929
27
- },
28
- {
29
- "accuracy": 0.887771,
30
- "f1": 0.887457,
31
- "f1_weighted": 0.887413,
32
- "ap": 0.86644,
33
- "ap_weighted": 0.86644
34
- },
35
- {
36
- "accuracy": 0.879607,
37
- "f1": 0.879353,
38
- "f1_weighted": 0.879312,
39
- "ap": 0.852779,
40
- "ap_weighted": 0.852779
41
- },
42
- {
43
- "accuracy": 0.894125,
44
- "f1": 0.894014,
45
- "f1_weighted": 0.893989,
46
- "ap": 0.86726,
47
- "ap_weighted": 0.86726
48
- },
49
- {
50
- "accuracy": 0.870804,
51
- "f1": 0.870803,
52
- "f1_weighted": 0.870801,
53
- "ap": 0.827334,
54
- "ap_weighted": 0.827334
55
- },
56
- {
57
- "accuracy": 0.896821,
58
- "f1": 0.896763,
59
- "f1_weighted": 0.896744,
60
- "ap": 0.86776,
61
- "ap_weighted": 0.86776
62
- },
63
- {
64
- "accuracy": 0.901696,
65
- "f1": 0.901691,
66
- "f1_weighted": 0.901696,
67
- "ap": 0.863676,
68
- "ap_weighted": 0.863676
69
- },
70
- {
71
- "accuracy": 0.90081,
72
- "f1": 0.900772,
73
- "f1_weighted": 0.900787,
74
- "ap": 0.858558,
75
- "ap_weighted": 0.858558
76
- },
77
- {
78
- "accuracy": 0.871574,
79
- "f1": 0.870546,
80
- "f1_weighted": 0.870631,
81
- "ap": 0.805833,
82
- "ap_weighted": 0.805833
83
- }
84
- ],
85
- "main_score": 0.887758,
86
- "hf_subset": "default",
87
- "languages": [
88
- "vie-Latn"
89
- ]
90
- }
91
- ]
92
- },
93
- "evaluation_time": 1072.2194910049438,
94
- "kg_co2_emissions": null
95
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AmazonReviewsVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "27da94deb6d4f44af789a3d70750fa506b79f189",
3
- "task_name": "AmazonReviewsVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.444801,
9
- "f1": 0.426392,
10
- "f1_weighted": 0.432322,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.450058,
14
- "f1": 0.42758,
15
- "f1_weighted": 0.434702
16
- },
17
- {
18
- "accuracy": 0.470502,
19
- "f1": 0.448032,
20
- "f1_weighted": 0.455157
21
- },
22
- {
23
- "accuracy": 0.436624,
24
- "f1": 0.42657,
25
- "f1_weighted": 0.430991
26
- },
27
- {
28
- "accuracy": 0.451811,
29
- "f1": 0.435001,
30
- "f1_weighted": 0.441204
31
- },
32
- {
33
- "accuracy": 0.462325,
34
- "f1": 0.433835,
35
- "f1_weighted": 0.440433
36
- },
37
- {
38
- "accuracy": 0.439836,
39
- "f1": 0.413124,
40
- "f1_weighted": 0.419804
41
- },
42
- {
43
- "accuracy": 0.418808,
44
- "f1": 0.415917,
45
- "f1_weighted": 0.420269
46
- },
47
- {
48
- "accuracy": 0.463785,
49
- "f1": 0.457309,
50
- "f1_weighted": 0.461625
51
- },
52
- {
53
- "accuracy": 0.439252,
54
- "f1": 0.428232,
55
- "f1_weighted": 0.433021
56
- },
57
- {
58
- "accuracy": 0.415012,
59
- "f1": 0.378321,
60
- "f1_weighted": 0.386011
61
- }
62
- ],
63
- "main_score": 0.444801,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 11.328449964523315,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ArguAna-VN.json DELETED
@@ -1,158 +0,0 @@
1
- {
2
- "dataset_revision": "2a5133a05d7430e6f353497b1624a6e73148105b",
3
- "task_name": "ArguAna-VN",
4
- "mteb_version": "1.38.39",
5
- "scores": {
6
- "test": [
7
- {
8
- "ndcg_at_1": 0.00154,
9
- "ndcg_at_3": 0.25091,
10
- "ndcg_at_5": 0.3202,
11
- "ndcg_at_10": 0.37932,
12
- "ndcg_at_20": 0.40551,
13
- "ndcg_at_100": 0.42317,
14
- "ndcg_at_1000": 0.42594,
15
- "map_at_1": 0.00154,
16
- "map_at_3": 0.18906,
17
- "map_at_5": 0.22748,
18
- "map_at_10": 0.25228,
19
- "map_at_20": 0.25969,
20
- "map_at_100": 0.26227,
21
- "map_at_1000": 0.26239,
22
- "recall_at_1": 0.00154,
23
- "recall_at_3": 0.43012,
24
- "recall_at_5": 0.59846,
25
- "recall_at_10": 0.77915,
26
- "recall_at_20": 0.88108,
27
- "recall_at_100": 0.97452,
28
- "recall_at_1000": 0.99537,
29
- "precision_at_1": 0.00154,
30
- "precision_at_3": 0.14337,
31
- "precision_at_5": 0.11969,
32
- "precision_at_10": 0.07792,
33
- "precision_at_20": 0.04405,
34
- "precision_at_100": 0.00975,
35
- "precision_at_1000": 0.001,
36
- "mrr_at_1": 0.001544,
37
- "mrr_at_3": 0.18906,
38
- "mrr_at_5": 0.227477,
39
- "mrr_at_10": 0.252276,
40
- "mrr_at_20": 0.259688,
41
- "mrr_at_100": 0.262271,
42
- "mrr_at_1000": 0.262394,
43
- "nauc_ndcg_at_1_max": -0.15791,
44
- "nauc_ndcg_at_1_std": -0.309824,
45
- "nauc_ndcg_at_1_diff1": -0.573698,
46
- "nauc_ndcg_at_3_max": 0.050466,
47
- "nauc_ndcg_at_3_std": -0.047425,
48
- "nauc_ndcg_at_3_diff1": -0.053972,
49
- "nauc_ndcg_at_5_max": 0.051704,
50
- "nauc_ndcg_at_5_std": -0.034976,
51
- "nauc_ndcg_at_5_diff1": -0.067507,
52
- "nauc_ndcg_at_10_max": 0.051058,
53
- "nauc_ndcg_at_10_std": 0.002573,
54
- "nauc_ndcg_at_10_diff1": -0.115774,
55
- "nauc_ndcg_at_20_max": 0.060309,
56
- "nauc_ndcg_at_20_std": -0.002941,
57
- "nauc_ndcg_at_20_diff1": -0.112784,
58
- "nauc_ndcg_at_100_max": 0.064026,
59
- "nauc_ndcg_at_100_std": -0.001192,
60
- "nauc_ndcg_at_100_diff1": -0.102481,
61
- "nauc_ndcg_at_1000_max": 0.055098,
62
- "nauc_ndcg_at_1000_std": -0.009017,
63
- "nauc_ndcg_at_1000_diff1": -0.099237,
64
- "nauc_map_at_1_max": -0.15791,
65
- "nauc_map_at_1_std": -0.309824,
66
- "nauc_map_at_1_diff1": -0.573698,
67
- "nauc_map_at_3_max": 0.049703,
68
- "nauc_map_at_3_std": -0.048378,
69
- "nauc_map_at_3_diff1": -0.054478,
70
- "nauc_map_at_5_max": 0.049706,
71
- "nauc_map_at_5_std": -0.041634,
72
- "nauc_map_at_5_diff1": -0.062748,
73
- "nauc_map_at_10_max": 0.049333,
74
- "nauc_map_at_10_std": -0.023585,
75
- "nauc_map_at_10_diff1": -0.08462,
76
- "nauc_map_at_20_max": 0.052529,
77
- "nauc_map_at_20_std": -0.026038,
78
- "nauc_map_at_20_diff1": -0.082893,
79
- "nauc_map_at_100_max": 0.052676,
80
- "nauc_map_at_100_std": -0.025392,
81
- "nauc_map_at_100_diff1": -0.081601,
82
- "nauc_map_at_1000_max": 0.05229,
83
- "nauc_map_at_1000_std": -0.025648,
84
- "nauc_map_at_1000_diff1": -0.081523,
85
- "nauc_recall_at_1_max": -0.15791,
86
- "nauc_recall_at_1_std": -0.309824,
87
- "nauc_recall_at_1_diff1": -0.573698,
88
- "nauc_recall_at_3_max": 0.051543,
89
- "nauc_recall_at_3_std": -0.046082,
90
- "nauc_recall_at_3_diff1": -0.053542,
91
- "nauc_recall_at_5_max": 0.055409,
92
- "nauc_recall_at_5_std": -0.022324,
93
- "nauc_recall_at_5_diff1": -0.077261,
94
- "nauc_recall_at_10_max": 0.055321,
95
- "nauc_recall_at_10_std": 0.07644,
96
- "nauc_recall_at_10_diff1": -0.207529,
97
- "nauc_recall_at_20_max": 0.098502,
98
- "nauc_recall_at_20_std": 0.101036,
99
- "nauc_recall_at_20_diff1": -0.255702,
100
- "nauc_recall_at_100_max": 0.346732,
101
- "nauc_recall_at_100_std": 0.450867,
102
- "nauc_recall_at_100_diff1": -0.448661,
103
- "nauc_recall_at_1000_max": -0.18723,
104
- "nauc_recall_at_1000_std": 0.764955,
105
- "nauc_recall_at_1000_diff1": -1.251363,
106
- "nauc_precision_at_1_max": -0.15791,
107
- "nauc_precision_at_1_std": -0.309824,
108
- "nauc_precision_at_1_diff1": -0.573698,
109
- "nauc_precision_at_3_max": 0.051543,
110
- "nauc_precision_at_3_std": -0.046082,
111
- "nauc_precision_at_3_diff1": -0.053542,
112
- "nauc_precision_at_5_max": 0.055409,
113
- "nauc_precision_at_5_std": -0.022324,
114
- "nauc_precision_at_5_diff1": -0.077261,
115
- "nauc_precision_at_10_max": 0.055321,
116
- "nauc_precision_at_10_std": 0.07644,
117
- "nauc_precision_at_10_diff1": -0.207529,
118
- "nauc_precision_at_20_max": 0.098502,
119
- "nauc_precision_at_20_std": 0.101036,
120
- "nauc_precision_at_20_diff1": -0.255702,
121
- "nauc_precision_at_100_max": 0.346732,
122
- "nauc_precision_at_100_std": 0.450867,
123
- "nauc_precision_at_100_diff1": -0.448661,
124
- "nauc_precision_at_1000_max": -0.18723,
125
- "nauc_precision_at_1000_std": 0.764955,
126
- "nauc_precision_at_1000_diff1": -1.251363,
127
- "nauc_mrr_at_1_max": -0.15791,
128
- "nauc_mrr_at_1_std": -0.309824,
129
- "nauc_mrr_at_1_diff1": -0.573698,
130
- "nauc_mrr_at_3_max": 0.049703,
131
- "nauc_mrr_at_3_std": -0.048378,
132
- "nauc_mrr_at_3_diff1": -0.054478,
133
- "nauc_mrr_at_5_max": 0.049706,
134
- "nauc_mrr_at_5_std": -0.041634,
135
- "nauc_mrr_at_5_diff1": -0.062748,
136
- "nauc_mrr_at_10_max": 0.049333,
137
- "nauc_mrr_at_10_std": -0.023585,
138
- "nauc_mrr_at_10_diff1": -0.08462,
139
- "nauc_mrr_at_20_max": 0.052529,
140
- "nauc_mrr_at_20_std": -0.026038,
141
- "nauc_mrr_at_20_diff1": -0.082893,
142
- "nauc_mrr_at_100_max": 0.052676,
143
- "nauc_mrr_at_100_std": -0.025392,
144
- "nauc_mrr_at_100_diff1": -0.081601,
145
- "nauc_mrr_at_1000_max": 0.05229,
146
- "nauc_mrr_at_1000_std": -0.025648,
147
- "nauc_mrr_at_1000_diff1": -0.081523,
148
- "main_score": 0.37932,
149
- "hf_subset": "default",
150
- "languages": [
151
- "vie-Latn"
152
- ]
153
- }
154
- ]
155
- },
156
- "evaluation_time": 56.45738124847412,
157
- "kg_co2_emissions": null
158
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/AskUbuntuDupQuestions-VN.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "dataset_revision": "5cfaa5c07252d30c37302bfc056f0d85884971a1",
3
- "task_name": "AskUbuntuDupQuestions-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "map": 0.618748,
9
- "mrr": 0.729745,
10
- "nAUC_map_max": 0.263989,
11
- "nAUC_map_std": 0.233967,
12
- "nAUC_map_diff1": 0.170011,
13
- "nAUC_mrr_max": 0.350929,
14
- "nAUC_mrr_std": 0.29905,
15
- "nAUC_mrr_diff1": 0.207218,
16
- "main_score": 0.618748,
17
- "hf_subset": "default",
18
- "languages": [
19
- "vie-Latn"
20
- ]
21
- }
22
- ]
23
- },
24
- "evaluation_time": 6.192819356918335,
25
- "kg_co2_emissions": null
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/BIOSSES-VN.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "dataset_revision": "1dae4a6df91c0852680cd4ab48c8c1d8a9ed49b2",
3
- "task_name": "BIOSSES-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "pearson": 0.799628,
9
- "spearman": 0.781415,
10
- "cosine_pearson": 0.799628,
11
- "cosine_spearman": 0.781415,
12
- "manhattan_pearson": 0.777069,
13
- "manhattan_spearman": 0.777819,
14
- "euclidean_pearson": 0.780537,
15
- "euclidean_spearman": 0.781415,
16
- "main_score": 0.781415,
17
- "hf_subset": "default",
18
- "languages": [
19
- "vie-Latn"
20
- ]
21
- }
22
- ]
23
- },
24
- "evaluation_time": 0.6975255012512207,
25
- "kg_co2_emissions": null
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Banking77VNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "42541b07c25a49604be129fba6d70b752be229c1",
3
- "task_name": "Banking77VNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.792094,
9
- "f1": 0.777269,
10
- "f1_weighted": 0.78413,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.784693,
14
- "f1": 0.766683,
15
- "f1_weighted": 0.774623
16
- },
17
- {
18
- "accuracy": 0.795627,
19
- "f1": 0.776761,
20
- "f1_weighted": 0.784439
21
- },
22
- {
23
- "accuracy": 0.796468,
24
- "f1": 0.781336,
25
- "f1_weighted": 0.787568
26
- },
27
- {
28
- "accuracy": 0.799832,
29
- "f1": 0.791148,
30
- "f1_weighted": 0.79406
31
- },
32
- {
33
- "accuracy": 0.77418,
34
- "f1": 0.752046,
35
- "f1_weighted": 0.764798
36
- },
37
- {
38
- "accuracy": 0.785114,
39
- "f1": 0.775827,
40
- "f1_weighted": 0.77941
41
- },
42
- {
43
- "accuracy": 0.799832,
44
- "f1": 0.785842,
45
- "f1_weighted": 0.792011
46
- },
47
- {
48
- "accuracy": 0.801514,
49
- "f1": 0.788659,
50
- "f1_weighted": 0.795154
51
- },
52
- {
53
- "accuracy": 0.80614,
54
- "f1": 0.795751,
55
- "f1_weighted": 0.800618
56
- },
57
- {
58
- "accuracy": 0.777544,
59
- "f1": 0.758633,
60
- "f1_weighted": 0.768624
61
- }
62
- ],
63
- "main_score": 0.792094,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 12.62467908859253,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackAndroid-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "4a022e7213ccc05ee970a176abd0293b3a0a2da0",
3
- "mteb_dataset_name": "CQADupstackAndroid-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackAndroid-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 44.89,
10
- "map_at_1": 0.26147,
11
- "map_at_10": 0.35997,
12
- "map_at_100": 0.3737,
13
- "map_at_1000": 0.37509,
14
- "map_at_20": 0.36694,
15
- "map_at_3": 0.32705,
16
- "map_at_5": 0.34598,
17
- "mrr_at_1": 0.32474,
18
- "mrr_at_10": 0.41319,
19
- "mrr_at_100": 0.42197,
20
- "mrr_at_1000": 0.42238,
21
- "mrr_at_20": 0.41806,
22
- "mrr_at_3": 0.38431,
23
- "mrr_at_5": 0.4008,
24
- "ndcg_at_1": 0.32474,
25
- "ndcg_at_10": 0.41928,
26
- "ndcg_at_100": 0.47686,
27
- "ndcg_at_1000": 0.49942,
28
- "ndcg_at_20": 0.43829,
29
- "ndcg_at_3": 0.36824,
30
- "ndcg_at_5": 0.39274,
31
- "precision_at_1": 0.32474,
32
- "precision_at_10": 0.08041,
33
- "precision_at_100": 0.01424,
34
- "precision_at_1000": 0.00203,
35
- "precision_at_20": 0.04828,
36
- "precision_at_3": 0.17468,
37
- "precision_at_5": 0.12921,
38
- "recall_at_1": 0.26147,
39
- "recall_at_10": 0.53732,
40
- "recall_at_100": 0.7876,
41
- "recall_at_1000": 0.93133,
42
- "recall_at_20": 0.60459,
43
- "recall_at_3": 0.38587,
44
- "recall_at_5": 0.4548,
45
- "hf_subset": "default",
46
- "main_score": 0.41928,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackGis-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "755156d548a8288efdb29b80bad302750ab33977",
3
- "mteb_dataset_name": "CQADupstackGis-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackGis-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 199.51,
10
- "map_at_1": 0.20556,
11
- "map_at_10": 0.2769,
12
- "map_at_100": 0.28656,
13
- "map_at_1000": 0.28748,
14
- "map_at_20": 0.28257,
15
- "map_at_3": 0.25622,
16
- "map_at_5": 0.26698,
17
- "mrr_at_1": 0.22308,
18
- "mrr_at_10": 0.2932,
19
- "mrr_at_100": 0.30271,
20
- "mrr_at_1000": 0.3034,
21
- "mrr_at_20": 0.29919,
22
- "mrr_at_3": 0.27256,
23
- "mrr_at_5": 0.28372,
24
- "ndcg_at_1": 0.22308,
25
- "ndcg_at_10": 0.31907,
26
- "ndcg_at_100": 0.36814,
27
- "ndcg_at_1000": 0.39334,
28
- "ndcg_at_20": 0.33973,
29
- "ndcg_at_3": 0.27709,
30
- "ndcg_at_5": 0.29583,
31
- "precision_at_1": 0.22308,
32
- "precision_at_10": 0.05015,
33
- "precision_at_100": 0.00785,
34
- "precision_at_1000": 0.00105,
35
- "precision_at_20": 0.02969,
36
- "precision_at_3": 0.11949,
37
- "precision_at_5": 0.08308,
38
- "recall_at_1": 0.20556,
39
- "recall_at_10": 0.43315,
40
- "recall_at_100": 0.66017,
41
- "recall_at_1000": 0.85128,
42
- "recall_at_20": 0.51194,
43
- "recall_at_3": 0.31881,
44
- "recall_at_5": 0.36374,
45
- "hf_subset": "default",
46
- "main_score": 0.31907,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackMathematica-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "d0cc9b60ba66faa3fb21cb9a54ef969af548b312",
3
- "mteb_dataset_name": "CQADupstackMathematica-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackMathematica-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 111.08,
10
- "map_at_1": 0.11446,
11
- "map_at_10": 0.17372,
12
- "map_at_100": 0.18428,
13
- "map_at_1000": 0.18553,
14
- "map_at_20": 0.17952,
15
- "map_at_3": 0.15416,
16
- "map_at_5": 0.16377,
17
- "mrr_at_1": 0.14499,
18
- "mrr_at_10": 0.20717,
19
- "mrr_at_100": 0.21714,
20
- "mrr_at_1000": 0.21797,
21
- "mrr_at_20": 0.21261,
22
- "mrr_at_3": 0.18799,
23
- "mrr_at_5": 0.19737,
24
- "ndcg_at_1": 0.14499,
25
- "ndcg_at_10": 0.21439,
26
- "ndcg_at_100": 0.26912,
27
- "ndcg_at_1000": 0.30206,
28
- "ndcg_at_20": 0.23465,
29
- "ndcg_at_3": 0.17684,
30
- "ndcg_at_5": 0.19131,
31
- "precision_at_1": 0.14499,
32
- "precision_at_10": 0.04009,
33
- "precision_at_100": 0.00755,
34
- "precision_at_1000": 0.00117,
35
- "precision_at_20": 0.02516,
36
- "precision_at_3": 0.08529,
37
- "precision_at_5": 0.06226,
38
- "recall_at_1": 0.11446,
39
- "recall_at_10": 0.30977,
40
- "recall_at_100": 0.55949,
41
- "recall_at_1000": 0.8006,
42
- "recall_at_20": 0.38463,
43
- "recall_at_3": 0.20025,
44
- "recall_at_5": 0.2377,
45
- "hf_subset": "default",
46
- "main_score": 0.21439,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackPhysics-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "8b6b68b59933cc72985f674f76c80a678c27d6be",
3
- "mteb_dataset_name": "CQADupstackPhysics-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackPhysics-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 190.76,
10
- "map_at_1": 0.21216,
11
- "map_at_10": 0.29794,
12
- "map_at_100": 0.31226,
13
- "map_at_1000": 0.31362,
14
- "map_at_20": 0.30679,
15
- "map_at_3": 0.26528,
16
- "map_at_5": 0.28309,
17
- "mrr_at_1": 0.27629,
18
- "mrr_at_10": 0.36405,
19
- "mrr_at_100": 0.37387,
20
- "mrr_at_1000": 0.37455,
21
- "mrr_at_20": 0.37036,
22
- "mrr_at_3": 0.34158,
23
- "mrr_at_5": 0.35354,
24
- "ndcg_at_1": 0.27629,
25
- "ndcg_at_10": 0.35518,
26
- "ndcg_at_100": 0.4152,
27
- "ndcg_at_1000": 0.44354,
28
- "ndcg_at_20": 0.38148,
29
- "ndcg_at_3": 0.30811,
30
- "ndcg_at_5": 0.32809,
31
- "precision_at_1": 0.27629,
32
- "precision_at_10": 0.07072,
33
- "precision_at_100": 0.01196,
34
- "precision_at_1000": 0.00166,
35
- "precision_at_20": 0.04381,
36
- "precision_at_3": 0.15258,
37
- "precision_at_5": 0.11134,
38
- "recall_at_1": 0.21216,
39
- "recall_at_10": 0.46074,
40
- "recall_at_100": 0.71265,
41
- "recall_at_1000": 0.9034,
42
- "recall_at_20": 0.55265,
43
- "recall_at_3": 0.32342,
44
- "recall_at_5": 0.38045,
45
- "hf_subset": "default",
46
- "main_score": 0.35518,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackProgrammers-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "1a628c4e61f71ffdb7707d6d4024d25cfe68215a",
3
- "mteb_dataset_name": "CQADupstackProgrammers-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackProgrammers-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 166.63,
10
- "map_at_1": 0.19431,
11
- "map_at_10": 0.27708,
12
- "map_at_100": 0.2901,
13
- "map_at_1000": 0.29131,
14
- "map_at_20": 0.28417,
15
- "map_at_3": 0.25104,
16
- "map_at_5": 0.26674,
17
- "mrr_at_1": 0.23328,
18
- "mrr_at_10": 0.31688,
19
- "mrr_at_100": 0.32747,
20
- "mrr_at_1000": 0.32809,
21
- "mrr_at_20": 0.32311,
22
- "mrr_at_3": 0.29201,
23
- "mrr_at_5": 0.30685,
24
- "ndcg_at_1": 0.23328,
25
- "ndcg_at_10": 0.3271,
26
- "ndcg_at_100": 0.38812,
27
- "ndcg_at_1000": 0.41372,
28
- "ndcg_at_20": 0.35064,
29
- "ndcg_at_3": 0.28162,
30
- "ndcg_at_5": 0.30513,
31
- "precision_at_1": 0.23328,
32
- "precision_at_10": 0.06101,
33
- "precision_at_100": 0.01064,
34
- "precision_at_1000": 0.00145,
35
- "precision_at_20": 0.03695,
36
- "precision_at_3": 0.13431,
37
- "precision_at_5": 0.09886,
38
- "recall_at_1": 0.19431,
39
- "recall_at_10": 0.43628,
40
- "recall_at_100": 0.70198,
41
- "recall_at_1000": 0.87329,
42
- "recall_at_20": 0.52302,
43
- "recall_at_3": 0.31336,
44
- "recall_at_5": 0.37245,
45
- "hf_subset": "default",
46
- "main_score": 0.3271,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackStats-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "6b8164f3af61f3bb7728724229ba36213fb46c25",
3
- "mteb_dataset_name": "CQADupstackStats-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackStats-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 234.79,
10
- "map_at_1": 0.16556,
11
- "map_at_10": 0.22848,
12
- "map_at_100": 0.23963,
13
- "map_at_1000": 0.24038,
14
- "map_at_20": 0.23554,
15
- "map_at_3": 0.20422,
16
- "map_at_5": 0.21873,
17
- "mrr_at_1": 0.18874,
18
- "mrr_at_10": 0.2507,
19
- "mrr_at_100": 0.26037,
20
- "mrr_at_1000": 0.26091,
21
- "mrr_at_20": 0.25655,
22
- "mrr_at_3": 0.22572,
23
- "mrr_at_5": 0.24078,
24
- "ndcg_at_1": 0.18874,
25
- "ndcg_at_10": 0.26859,
26
- "ndcg_at_100": 0.32125,
27
- "ndcg_at_1000": 0.34208,
28
- "ndcg_at_20": 0.29217,
29
- "ndcg_at_3": 0.22063,
30
- "ndcg_at_5": 0.2463,
31
- "precision_at_1": 0.18874,
32
- "precision_at_10": 0.04603,
33
- "precision_at_100": 0.00768,
34
- "precision_at_1000": 0.00102,
35
- "precision_at_20": 0.02881,
36
- "precision_at_3": 0.09492,
37
- "precision_at_5": 0.07219,
38
- "recall_at_1": 0.16556,
39
- "recall_at_10": 0.3683,
40
- "recall_at_100": 0.60969,
41
- "recall_at_1000": 0.76575,
42
- "recall_at_20": 0.45836,
43
- "recall_at_3": 0.24264,
44
- "recall_at_5": 0.30273,
45
- "hf_subset": "default",
46
- "main_score": 0.26859,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackTex-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "aec43e5ae40451526528b3fc80dd5983ec388e21",
3
- "mteb_dataset_name": "CQADupstackTex-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackTex-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 490.61,
10
- "map_at_1": 0.14818,
11
- "map_at_10": 0.20724,
12
- "map_at_100": 0.21763,
13
- "map_at_1000": 0.2188,
14
- "map_at_20": 0.21326,
15
- "map_at_3": 0.18685,
16
- "map_at_5": 0.19831,
17
- "mrr_at_1": 0.18582,
18
- "mrr_at_10": 0.24641,
19
- "mrr_at_100": 0.2559,
20
- "mrr_at_1000": 0.25671,
21
- "mrr_at_20": 0.25211,
22
- "mrr_at_3": 0.2268,
23
- "mrr_at_5": 0.23826,
24
- "ndcg_at_1": 0.18582,
25
- "ndcg_at_10": 0.24781,
26
- "ndcg_at_100": 0.29847,
27
- "ndcg_at_1000": 0.32928,
28
- "ndcg_at_20": 0.26771,
29
- "ndcg_at_3": 0.21233,
30
- "ndcg_at_5": 0.22894,
31
- "precision_at_1": 0.18582,
32
- "precision_at_10": 0.04523,
33
- "precision_at_100": 0.00827,
34
- "precision_at_1000": 0.00126,
35
- "precision_at_20": 0.02799,
36
- "precision_at_3": 0.10025,
37
- "precision_at_5": 0.07364,
38
- "recall_at_1": 0.14818,
39
- "recall_at_10": 0.33285,
40
- "recall_at_100": 0.56243,
41
- "recall_at_1000": 0.78623,
42
- "recall_at_20": 0.40797,
43
- "recall_at_3": 0.22972,
44
- "recall_at_5": 0.27396,
45
- "hf_subset": "default",
46
- "main_score": 0.24781,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackUnix-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "f8b884697871cb38901139f2435c273135f83a3f",
3
- "mteb_dataset_name": "CQADupstackUnix-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackUnix-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 248.55,
10
- "map_at_1": 0.22331,
11
- "map_at_10": 0.29694,
12
- "map_at_100": 0.30728,
13
- "map_at_1000": 0.30838,
14
- "map_at_20": 0.30234,
15
- "map_at_3": 0.27118,
16
- "map_at_5": 0.28451,
17
- "mrr_at_1": 0.26481,
18
- "mrr_at_10": 0.33574,
19
- "mrr_at_100": 0.34457,
20
- "mrr_at_1000": 0.34526,
21
- "mrr_at_20": 0.34089,
22
- "mrr_at_3": 0.31177,
23
- "mrr_at_5": 0.32465,
24
- "ndcg_at_1": 0.26481,
25
- "ndcg_at_10": 0.34515,
26
- "ndcg_at_100": 0.39566,
27
- "ndcg_at_1000": 0.42141,
28
- "ndcg_at_20": 0.36389,
29
- "ndcg_at_3": 0.29735,
30
- "ndcg_at_5": 0.31759,
31
- "precision_at_1": 0.26481,
32
- "precision_at_10": 0.05925,
33
- "precision_at_100": 0.00946,
34
- "precision_at_1000": 0.00128,
35
- "precision_at_20": 0.03458,
36
- "precision_at_3": 0.13301,
37
- "precision_at_5": 0.09432,
38
- "recall_at_1": 0.22331,
39
- "recall_at_10": 0.45476,
40
- "recall_at_100": 0.67928,
41
- "recall_at_1000": 0.86037,
42
- "recall_at_20": 0.52341,
43
- "recall_at_3": 0.32383,
44
- "recall_at_5": 0.37485,
45
- "hf_subset": "default",
46
- "main_score": 0.34515,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWebmasters-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "482d6e560d977b137e435d33379c5a8049e70e8d",
3
- "mteb_dataset_name": "CQADupstackWebmasters-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackWebmasters-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 73.56,
10
- "map_at_1": 0.18814,
11
- "map_at_10": 0.26309,
12
- "map_at_100": 0.27633,
13
- "map_at_1000": 0.2786,
14
- "map_at_20": 0.26934,
15
- "map_at_3": 0.23491,
16
- "map_at_5": 0.25194,
17
- "mrr_at_1": 0.225,
18
- "mrr_at_10": 0.30061,
19
- "mrr_at_100": 0.31017,
20
- "mrr_at_1000": 0.31098,
21
- "mrr_at_20": 0.30585,
22
- "mrr_at_3": 0.27083,
23
- "mrr_at_5": 0.29083,
24
- "ndcg_at_1": 0.225,
25
- "ndcg_at_10": 0.31666,
26
- "ndcg_at_100": 0.36984,
27
- "ndcg_at_1000": 0.40278,
28
- "ndcg_at_20": 0.33293,
29
- "ndcg_at_3": 0.26741,
30
- "ndcg_at_5": 0.29431,
31
- "precision_at_1": 0.225,
32
- "precision_at_10": 0.062,
33
- "precision_at_100": 0.01322,
34
- "precision_at_1000": 0.00234,
35
- "precision_at_20": 0.039,
36
- "precision_at_3": 0.125,
37
- "precision_at_5": 0.097,
38
- "recall_at_1": 0.18814,
39
- "recall_at_10": 0.42675,
40
- "recall_at_100": 0.66891,
41
- "recall_at_1000": 0.88531,
42
- "recall_at_20": 0.49,
43
- "recall_at_3": 0.28062,
44
- "recall_at_5": 0.35206,
45
- "hf_subset": "default",
46
- "main_score": 0.31666,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/CQADupstackWordpress-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "2230f80e1baf42aa005731ca86577621c566fcd7",
3
- "mteb_dataset_name": "CQADupstackWordpress-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "CQADupstackWordpress-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 283.45,
10
- "map_at_1": 0.13534,
11
- "map_at_10": 0.20314,
12
- "map_at_100": 0.21388,
13
- "map_at_1000": 0.21491,
14
- "map_at_20": 0.20936,
15
- "map_at_3": 0.18317,
16
- "map_at_5": 0.19319,
17
- "mrr_at_1": 0.15183,
18
- "mrr_at_10": 0.22404,
19
- "mrr_at_100": 0.23378,
20
- "mrr_at_1000": 0.23451,
21
- "mrr_at_20": 0.22996,
22
- "mrr_at_3": 0.20201,
23
- "mrr_at_5": 0.21366,
24
- "ndcg_at_1": 0.15183,
25
- "ndcg_at_10": 0.24744,
26
- "ndcg_at_100": 0.29839,
27
- "ndcg_at_1000": 0.32591,
28
- "ndcg_at_20": 0.2689,
29
- "ndcg_at_3": 0.20599,
30
- "ndcg_at_5": 0.22338,
31
- "precision_at_1": 0.15183,
32
- "precision_at_10": 0.04162,
33
- "precision_at_100": 0.0073,
34
- "precision_at_1000": 0.00109,
35
- "precision_at_20": 0.02592,
36
- "precision_at_3": 0.09511,
37
- "precision_at_5": 0.06702,
38
- "recall_at_1": 0.13534,
39
- "recall_at_10": 0.36562,
40
- "recall_at_100": 0.59327,
41
- "recall_at_1000": 0.80055,
42
- "recall_at_20": 0.44546,
43
- "recall_at_3": 0.24908,
44
- "recall_at_5": 0.29228,
45
- "hf_subset": "default",
46
- "main_score": 0.24744,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ClimateFEVER-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "42328bf787e17b1ad1a88be4f5e87ea9fb668511",
3
- "mteb_dataset_name": "ClimateFEVER-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "ClimateFEVER-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 16316.19,
10
- "map_at_1": 0.04967,
11
- "map_at_10": 0.08868,
12
- "map_at_100": 0.09661,
13
- "map_at_1000": 0.0982,
14
- "map_at_20": 0.09205,
15
- "map_at_3": 0.07457,
16
- "map_at_5": 0.08228,
17
- "mrr_at_1": 0.11353,
18
- "mrr_at_10": 0.18208,
19
- "mrr_at_100": 0.19152,
20
- "mrr_at_1000": 0.1925,
21
- "mrr_at_20": 0.18658,
22
- "mrr_at_3": 0.15849,
23
- "mrr_at_5": 0.17107,
24
- "ndcg_at_1": 0.11353,
25
- "ndcg_at_10": 0.13249,
26
- "ndcg_at_100": 0.17814,
27
- "ndcg_at_1000": 0.21712,
28
- "ndcg_at_20": 0.14482,
29
- "ndcg_at_3": 0.10497,
30
- "ndcg_at_5": 0.1154,
31
- "precision_at_1": 0.11353,
32
- "precision_at_10": 0.04223,
33
- "precision_at_100": 0.00927,
34
- "precision_at_1000": 0.00164,
35
- "precision_at_20": 0.02643,
36
- "precision_at_3": 0.08053,
37
- "precision_at_5": 0.06267,
38
- "recall_at_1": 0.04967,
39
- "recall_at_10": 0.16757,
40
- "recall_at_100": 0.33663,
41
- "recall_at_1000": 0.5625,
42
- "recall_at_20": 0.20262,
43
- "recall_at_3": 0.09955,
44
- "recall_at_5": 0.12846,
45
- "hf_subset": "default",
46
- "main_score": 0.13249,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/DBPedia-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "c3e20179fbcee16217ef9461a14a54b7faca9b63",
3
- "mteb_dataset_name": "DBPedia-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "DBPedia-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 4419.42,
10
- "map_at_1": 0.07486,
11
- "map_at_10": 0.15539,
12
- "map_at_100": 0.21047,
13
- "map_at_1000": 0.22274,
14
- "map_at_20": 0.17743,
15
- "map_at_3": 0.11208,
16
- "map_at_5": 0.13146,
17
- "mrr_at_1": 0.53106,
18
- "mrr_at_10": 0.6375,
19
- "mrr_at_100": 0.64307,
20
- "mrr_at_1000": 0.64321,
21
- "mrr_at_20": 0.64143,
22
- "mrr_at_3": 0.61439,
23
- "mrr_at_5": 0.62805,
24
- "ndcg_at_1": 0.43634,
25
- "ndcg_at_10": 0.34204,
26
- "ndcg_at_100": 0.38478,
27
- "ndcg_at_1000": 0.45509,
28
- "ndcg_at_20": 0.34051,
29
- "ndcg_at_3": 0.37097,
30
- "ndcg_at_5": 0.35581,
31
- "precision_at_1": 0.53106,
32
- "precision_at_10": 0.26553,
33
- "precision_at_100": 0.08469,
34
- "precision_at_1000": 0.01695,
35
- "precision_at_20": 0.2028,
36
- "precision_at_3": 0.39648,
37
- "precision_at_5": 0.34472,
38
- "recall_at_1": 0.07486,
39
- "recall_at_10": 0.21316,
40
- "recall_at_100": 0.44686,
41
- "recall_at_1000": 0.67695,
42
- "recall_at_20": 0.27749,
43
- "recall_at_3": 0.12701,
44
- "recall_at_5": 0.16045,
45
- "hf_subset": "default",
46
- "main_score": 0.34204,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/EmotionVNClassification.json DELETED
@@ -1,137 +0,0 @@
1
- {
2
- "dataset_revision": "797a93c0dd755ebf5818fbf54d0e0a024df9216d",
3
- "task_name": "EmotionVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "validation": [
7
- {
8
- "accuracy": 0.44375,
9
- "f1": 0.390648,
10
- "f1_weighted": 0.463387,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.472656,
14
- "f1": 0.404343,
15
- "f1_weighted": 0.498363
16
- },
17
- {
18
- "accuracy": 0.43125,
19
- "f1": 0.354818,
20
- "f1_weighted": 0.446879
21
- },
22
- {
23
- "accuracy": 0.473438,
24
- "f1": 0.402721,
25
- "f1_weighted": 0.496464
26
- },
27
- {
28
- "accuracy": 0.422656,
29
- "f1": 0.384927,
30
- "f1_weighted": 0.440412
31
- },
32
- {
33
- "accuracy": 0.461719,
34
- "f1": 0.414061,
35
- "f1_weighted": 0.476512
36
- },
37
- {
38
- "accuracy": 0.405469,
39
- "f1": 0.381572,
40
- "f1_weighted": 0.421386
41
- },
42
- {
43
- "accuracy": 0.478906,
44
- "f1": 0.391987,
45
- "f1_weighted": 0.487998
46
- },
47
- {
48
- "accuracy": 0.396875,
49
- "f1": 0.378969,
50
- "f1_weighted": 0.422998
51
- },
52
- {
53
- "accuracy": 0.463281,
54
- "f1": 0.413623,
55
- "f1_weighted": 0.487997
56
- },
57
- {
58
- "accuracy": 0.43125,
59
- "f1": 0.379461,
60
- "f1_weighted": 0.454856
61
- }
62
- ],
63
- "main_score": 0.44375,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ],
70
- "test": [
71
- {
72
- "accuracy": 0.446822,
73
- "f1": 0.380728,
74
- "f1_weighted": 0.468056,
75
- "scores_per_experiment": [
76
- {
77
- "accuracy": 0.481395,
78
- "f1": 0.401385,
79
- "f1_weighted": 0.508444
80
- },
81
- {
82
- "accuracy": 0.432558,
83
- "f1": 0.352439,
84
- "f1_weighted": 0.446656
85
- },
86
- {
87
- "accuracy": 0.475969,
88
- "f1": 0.394022,
89
- "f1_weighted": 0.49485
90
- },
91
- {
92
- "accuracy": 0.413953,
93
- "f1": 0.362915,
94
- "f1_weighted": 0.439652
95
- },
96
- {
97
- "accuracy": 0.455039,
98
- "f1": 0.399925,
99
- "f1_weighted": 0.475066
100
- },
101
- {
102
- "accuracy": 0.397674,
103
- "f1": 0.362269,
104
- "f1_weighted": 0.412897
105
- },
106
- {
107
- "accuracy": 0.490698,
108
- "f1": 0.389544,
109
- "f1_weighted": 0.498402
110
- },
111
- {
112
- "accuracy": 0.402326,
113
- "f1": 0.361925,
114
- "f1_weighted": 0.434939
115
- },
116
- {
117
- "accuracy": 0.473643,
118
- "f1": 0.405323,
119
- "f1_weighted": 0.496273
120
- },
121
- {
122
- "accuracy": 0.444961,
123
- "f1": 0.377529,
124
- "f1_weighted": 0.473386
125
- }
126
- ],
127
- "main_score": 0.446822,
128
- "hf_subset": "default",
129
- "languages": [
130
- "vie-Latn"
131
- ]
132
- }
133
- ]
134
- },
135
- "evaluation_time": 16.90190815925598,
136
- "kg_co2_emissions": null
137
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FEVER-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "a543dd8b98aed3603110c01d26db05ba39b87d49",
3
- "mteb_dataset_name": "FEVER-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "FEVER-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 7374.25,
10
- "map_at_1": 0.31332,
11
- "map_at_10": 0.42736,
12
- "map_at_100": 0.43506,
13
- "map_at_1000": 0.43552,
14
- "map_at_20": 0.432,
15
- "map_at_3": 0.3997,
16
- "map_at_5": 0.41616,
17
- "mrr_at_1": 0.33676,
18
- "mrr_at_10": 0.45185,
19
- "mrr_at_100": 0.45905,
20
- "mrr_at_1000": 0.45943,
21
- "mrr_at_20": 0.45621,
22
- "mrr_at_3": 0.42454,
23
- "mrr_at_5": 0.44056,
24
- "ndcg_at_1": 0.33676,
25
- "ndcg_at_10": 0.4881,
26
- "ndcg_at_100": 0.52495,
27
- "ndcg_at_1000": 0.53726,
28
- "ndcg_at_20": 0.50418,
29
- "ndcg_at_3": 0.43293,
30
- "ndcg_at_5": 0.46137,
31
- "precision_at_1": 0.33676,
32
- "precision_at_10": 0.0715,
33
- "precision_at_100": 0.00916,
34
- "precision_at_1000": 0.00104,
35
- "precision_at_20": 0.03934,
36
- "precision_at_3": 0.1826,
37
- "precision_at_5": 0.12488,
38
- "recall_at_1": 0.31332,
39
- "recall_at_10": 0.6531,
40
- "recall_at_100": 0.8205,
41
- "recall_at_1000": 0.91405,
42
- "recall_at_20": 0.71409,
43
- "recall_at_3": 0.50359,
44
- "recall_at_5": 0.57214,
45
- "hf_subset": "default",
46
- "main_score": 0.4881,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/FiQA2018-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "6c3cdf6f102151dbbbbc1d2cf999b305eba44dae",
3
- "mteb_dataset_name": "FiQA2018-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "FiQA2018-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 149.94,
10
- "map_at_1": 0.15915,
11
- "map_at_10": 0.23933,
12
- "map_at_100": 0.25351,
13
- "map_at_1000": 0.25561,
14
- "map_at_20": 0.24686,
15
- "map_at_3": 0.21014,
16
- "map_at_5": 0.22637,
17
- "mrr_at_1": 0.29024,
18
- "mrr_at_10": 0.3643,
19
- "mrr_at_100": 0.37292,
20
- "mrr_at_1000": 0.37358,
21
- "mrr_at_20": 0.36883,
22
- "mrr_at_3": 0.34301,
23
- "mrr_at_5": 0.35396,
24
- "ndcg_at_1": 0.29024,
25
- "ndcg_at_10": 0.29939,
26
- "ndcg_at_100": 0.35992,
27
- "ndcg_at_1000": 0.39932,
28
- "ndcg_at_20": 0.32076,
29
- "ndcg_at_3": 0.27282,
30
- "ndcg_at_5": 0.27997,
31
- "precision_at_1": 0.29024,
32
- "precision_at_10": 0.08153,
33
- "precision_at_100": 0.01478,
34
- "precision_at_1000": 0.00216,
35
- "precision_at_20": 0.05013,
36
- "precision_at_3": 0.17326,
37
- "precision_at_5": 0.12823,
38
- "recall_at_1": 0.15915,
39
- "recall_at_10": 0.35196,
40
- "recall_at_100": 0.57857,
41
- "recall_at_1000": 0.81862,
42
- "recall_at_20": 0.41776,
43
- "recall_at_3": 0.24858,
44
- "recall_at_5": 0.29204,
45
- "hf_subset": "default",
46
- "main_score": 0.29939,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/GreenNodeTableMarkdownRetrieval.json DELETED
@@ -1,158 +0,0 @@
1
- {
2
- "dataset_revision": "d86a4dad9fd7c70359f617d86984395ea89be1c5",
3
- "task_name": "GreenNodeTableMarkdownRetrieval",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "ndcg_at_1": 0.30846,
9
- "ndcg_at_3": 0.36243,
10
- "ndcg_at_5": 0.37957,
11
- "ndcg_at_10": 0.39719,
12
- "ndcg_at_20": 0.412,
13
- "ndcg_at_100": 0.43624,
14
- "ndcg_at_1000": 0.45896,
15
- "map_at_1": 0.30846,
16
- "map_at_3": 0.3491,
17
- "map_at_5": 0.35861,
18
- "map_at_10": 0.36591,
19
- "map_at_20": 0.36997,
20
- "map_at_100": 0.3732,
21
- "map_at_1000": 0.37392,
22
- "recall_at_1": 0.30846,
23
- "recall_at_3": 0.40102,
24
- "recall_at_5": 0.44265,
25
- "recall_at_10": 0.49702,
26
- "recall_at_20": 0.55553,
27
- "recall_at_100": 0.68819,
28
- "recall_at_1000": 0.87447,
29
- "precision_at_1": 0.30846,
30
- "precision_at_3": 0.13367,
31
- "precision_at_5": 0.08853,
32
- "precision_at_10": 0.0497,
33
- "precision_at_20": 0.02778,
34
- "precision_at_100": 0.00688,
35
- "precision_at_1000": 0.00087,
36
- "mrr_at_1": 0.308457,
37
- "mrr_at_3": 0.349096,
38
- "mrr_at_5": 0.35861,
39
- "mrr_at_10": 0.365907,
40
- "mrr_at_20": 0.369975,
41
- "mrr_at_100": 0.373196,
42
- "mrr_at_1000": 0.373924,
43
- "nauc_ndcg_at_1_max": 0.139786,
44
- "nauc_ndcg_at_1_std": -0.039386,
45
- "nauc_ndcg_at_1_diff1": 0.733187,
46
- "nauc_ndcg_at_3_max": 0.09619,
47
- "nauc_ndcg_at_3_std": -0.017705,
48
- "nauc_ndcg_at_3_diff1": 0.676034,
49
- "nauc_ndcg_at_5_max": 0.08391,
50
- "nauc_ndcg_at_5_std": -0.006374,
51
- "nauc_ndcg_at_5_diff1": 0.667409,
52
- "nauc_ndcg_at_10_max": 0.071841,
53
- "nauc_ndcg_at_10_std": 0.004193,
54
- "nauc_ndcg_at_10_diff1": 0.657513,
55
- "nauc_ndcg_at_20_max": 0.06409,
56
- "nauc_ndcg_at_20_std": 0.014012,
57
- "nauc_ndcg_at_20_diff1": 0.653131,
58
- "nauc_ndcg_at_100_max": 0.059821,
59
- "nauc_ndcg_at_100_std": 0.028492,
60
- "nauc_ndcg_at_100_diff1": 0.649499,
61
- "nauc_ndcg_at_1000_max": 0.066935,
62
- "nauc_ndcg_at_1000_std": 0.028327,
63
- "nauc_ndcg_at_1000_diff1": 0.655572,
64
- "nauc_map_at_1_max": 0.139786,
65
- "nauc_map_at_1_std": -0.039386,
66
- "nauc_map_at_1_diff1": 0.733187,
67
- "nauc_map_at_3_max": 0.106754,
68
- "nauc_map_at_3_std": -0.023135,
69
- "nauc_map_at_3_diff1": 0.689956,
70
- "nauc_map_at_5_max": 0.099963,
71
- "nauc_map_at_5_std": -0.016825,
72
- "nauc_map_at_5_diff1": 0.685233,
73
- "nauc_map_at_10_max": 0.09492,
74
- "nauc_map_at_10_std": -0.012515,
75
- "nauc_map_at_10_diff1": 0.681245,
76
- "nauc_map_at_20_max": 0.092859,
77
- "nauc_map_at_20_std": -0.00995,
78
- "nauc_map_at_20_diff1": 0.68014,
79
- "nauc_map_at_100_max": 0.092295,
80
- "nauc_map_at_100_std": -0.008175,
81
- "nauc_map_at_100_diff1": 0.67967,
82
- "nauc_map_at_1000_max": 0.092486,
83
- "nauc_map_at_1000_std": -0.008127,
84
- "nauc_map_at_1000_diff1": 0.679852,
85
- "nauc_recall_at_1_max": 0.139786,
86
- "nauc_recall_at_1_std": -0.039386,
87
- "nauc_recall_at_1_diff1": 0.733187,
88
- "nauc_recall_at_3_max": 0.066333,
89
- "nauc_recall_at_3_std": -0.002289,
90
- "nauc_recall_at_3_diff1": 0.636658,
91
- "nauc_recall_at_5_max": 0.036325,
92
- "nauc_recall_at_5_std": 0.025109,
93
- "nauc_recall_at_5_diff1": 0.615216,
94
- "nauc_recall_at_10_max": -0.001472,
95
- "nauc_recall_at_10_std": 0.058617,
96
- "nauc_recall_at_10_diff1": 0.583252,
97
- "nauc_recall_at_20_max": -0.037741,
98
- "nauc_recall_at_20_std": 0.102985,
99
- "nauc_recall_at_20_diff1": 0.560586,
100
- "nauc_recall_at_100_max": -0.089231,
101
- "nauc_recall_at_100_std": 0.224089,
102
- "nauc_recall_at_100_diff1": 0.515154,
103
- "nauc_recall_at_1000_max": -0.101471,
104
- "nauc_recall_at_1000_std": 0.430692,
105
- "nauc_recall_at_1000_diff1": 0.493278,
106
- "nauc_precision_at_1_max": 0.139786,
107
- "nauc_precision_at_1_std": -0.039386,
108
- "nauc_precision_at_1_diff1": 0.733187,
109
- "nauc_precision_at_3_max": 0.066333,
110
- "nauc_precision_at_3_std": -0.002289,
111
- "nauc_precision_at_3_diff1": 0.636658,
112
- "nauc_precision_at_5_max": 0.036325,
113
- "nauc_precision_at_5_std": 0.025109,
114
- "nauc_precision_at_5_diff1": 0.615216,
115
- "nauc_precision_at_10_max": -0.001472,
116
- "nauc_precision_at_10_std": 0.058617,
117
- "nauc_precision_at_10_diff1": 0.583252,
118
- "nauc_precision_at_20_max": -0.037741,
119
- "nauc_precision_at_20_std": 0.102985,
120
- "nauc_precision_at_20_diff1": 0.560586,
121
- "nauc_precision_at_100_max": -0.089231,
122
- "nauc_precision_at_100_std": 0.224089,
123
- "nauc_precision_at_100_diff1": 0.515154,
124
- "nauc_precision_at_1000_max": -0.101471,
125
- "nauc_precision_at_1000_std": 0.430692,
126
- "nauc_precision_at_1000_diff1": 0.493278,
127
- "nauc_mrr_at_1_max": 0.139786,
128
- "nauc_mrr_at_1_std": -0.039386,
129
- "nauc_mrr_at_1_diff1": 0.733187,
130
- "nauc_mrr_at_3_max": 0.106754,
131
- "nauc_mrr_at_3_std": -0.023135,
132
- "nauc_mrr_at_3_diff1": 0.689956,
133
- "nauc_mrr_at_5_max": 0.099963,
134
- "nauc_mrr_at_5_std": -0.016825,
135
- "nauc_mrr_at_5_diff1": 0.685233,
136
- "nauc_mrr_at_10_max": 0.09492,
137
- "nauc_mrr_at_10_std": -0.012515,
138
- "nauc_mrr_at_10_diff1": 0.681245,
139
- "nauc_mrr_at_20_max": 0.092859,
140
- "nauc_mrr_at_20_std": -0.00995,
141
- "nauc_mrr_at_20_diff1": 0.68014,
142
- "nauc_mrr_at_100_max": 0.092295,
143
- "nauc_mrr_at_100_std": -0.008175,
144
- "nauc_mrr_at_100_diff1": 0.67967,
145
- "nauc_mrr_at_1000_max": 0.092486,
146
- "nauc_mrr_at_1000_std": -0.008127,
147
- "nauc_mrr_at_1000_diff1": 0.679852,
148
- "main_score": 0.39719,
149
- "hf_subset": "default",
150
- "languages": [
151
- "vie-Latn"
152
- ]
153
- }
154
- ]
155
- },
156
- "evaluation_time": 716.343478679657,
157
- "kg_co2_emissions": null
158
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/HotpotQA-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "8a5220c7af5084f0d5d2afeb74f9c2b41b759ff0",
3
- "mteb_dataset_name": "HotpotQA-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "HotpotQA-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 13567.89,
10
- "map_at_1": 0.39226,
11
- "map_at_10": 0.6178,
12
- "map_at_100": 0.62675,
13
- "map_at_1000": 0.62738,
14
- "map_at_20": 0.6231,
15
- "map_at_3": 0.58315,
16
- "map_at_5": 0.60581,
17
- "mrr_at_1": 0.78453,
18
- "mrr_at_10": 0.84187,
19
- "mrr_at_100": 0.84371,
20
- "mrr_at_1000": 0.84377,
21
- "mrr_at_20": 0.8431,
22
- "mrr_at_3": 0.8326,
23
- "mrr_at_5": 0.83842,
24
- "ndcg_at_1": 0.78453,
25
- "ndcg_at_10": 0.70074,
26
- "ndcg_at_100": 0.73145,
27
- "ndcg_at_1000": 0.74324,
28
- "ndcg_at_20": 0.71379,
29
- "ndcg_at_3": 0.65249,
30
- "ndcg_at_5": 0.6806,
31
- "precision_at_1": 0.78453,
32
- "precision_at_10": 0.14613,
33
- "precision_at_100": 0.01701,
34
- "precision_at_1000": 0.00186,
35
- "precision_at_20": 0.07725,
36
- "precision_at_3": 0.4166,
37
- "precision_at_5": 0.27211,
38
- "recall_at_1": 0.39226,
39
- "recall_at_10": 0.73066,
40
- "recall_at_100": 0.85026,
41
- "recall_at_1000": 0.92794,
42
- "recall_at_20": 0.7725,
43
- "recall_at_3": 0.6249,
44
- "recall_at_5": 0.68027,
45
- "hf_subset": "default",
46
- "main_score": 0.70074,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/ImdbVNClassification.json DELETED
@@ -1,95 +0,0 @@
1
- {
2
- "dataset_revision": "0dccb383ee26c90c99d03c8674cf40de642f099a",
3
- "task_name": "ImdbVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.830619,
9
- "f1": 0.830218,
10
- "f1_weighted": 0.830237,
11
- "ap": 0.775496,
12
- "ap_weighted": 0.775496,
13
- "scores_per_experiment": [
14
- {
15
- "accuracy": 0.845523,
16
- "f1": 0.845006,
17
- "f1_weighted": 0.844953,
18
- "ap": 0.811288,
19
- "ap_weighted": 0.811288
20
- },
21
- {
22
- "accuracy": 0.812101,
23
- "f1": 0.810477,
24
- "f1_weighted": 0.810581,
25
- "ap": 0.740588,
26
- "ap_weighted": 0.740588
27
- },
28
- {
29
- "accuracy": 0.837824,
30
- "f1": 0.83732,
31
- "f1_weighted": 0.837373,
32
- "ap": 0.77416,
33
- "ap_weighted": 0.77416
34
- },
35
- {
36
- "accuracy": 0.860559,
37
- "f1": 0.860469,
38
- "f1_weighted": 0.860448,
39
- "ap": 0.820455,
40
- "ap_weighted": 0.820455
41
- },
42
- {
43
- "accuracy": 0.839409,
44
- "f1": 0.839408,
45
- "f1_weighted": 0.839411,
46
- "ap": 0.787281,
47
- "ap_weighted": 0.787281
48
- },
49
- {
50
- "accuracy": 0.816856,
51
- "f1": 0.816719,
52
- "f1_weighted": 0.81669,
53
- "ap": 0.767843,
54
- "ap_weighted": 0.767843
55
- },
56
- {
57
- "accuracy": 0.81826,
58
- "f1": 0.817479,
59
- "f1_weighted": 0.81755,
60
- "ap": 0.751218,
61
- "ap_weighted": 0.751218
62
- },
63
- {
64
- "accuracy": 0.826729,
65
- "f1": 0.82647,
66
- "f1_weighted": 0.82651,
67
- "ap": 0.765153,
68
- "ap_weighted": 0.765153
69
- },
70
- {
71
- "accuracy": 0.816403,
72
- "f1": 0.816309,
73
- "f1_weighted": 0.816333,
74
- "ap": 0.756783,
75
- "ap_weighted": 0.756783
76
- },
77
- {
78
- "accuracy": 0.832526,
79
- "f1": 0.832525,
80
- "f1_weighted": 0.832524,
81
- "ap": 0.780192,
82
- "ap_weighted": 0.780192
83
- }
84
- ],
85
- "main_score": 0.830619,
86
- "hf_subset": "default",
87
- "languages": [
88
- "vie-Latn"
89
- ]
90
- }
91
- ]
92
- },
93
- "evaluation_time": 162.1266748905182,
94
- "kg_co2_emissions": null
95
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MSMARCO-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "85d1ad4cc9070b8d019d65f5af1631a2ab91e294",
3
- "mteb_dataset_name": "MSMARCO-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "MSMARCO-VN",
6
- "scores": {
7
- "dev": [
8
- {
9
- "evaluation_time": 22839.72,
10
- "map_at_1": 0.14118,
11
- "map_at_10": 0.24393,
12
- "map_at_100": 0.25563,
13
- "map_at_1000": 0.25629,
14
- "map_at_20": 0.25091,
15
- "map_at_3": 0.20812,
16
- "map_at_5": 0.22766,
17
- "mrr_at_1": 0.14551,
18
- "mrr_at_10": 0.24818,
19
- "mrr_at_100": 0.25954,
20
- "mrr_at_1000": 0.26016,
21
- "mrr_at_20": 0.25497,
22
- "mrr_at_3": 0.21241,
23
- "mrr_at_5": 0.23198,
24
- "ndcg_at_1": 0.14551,
25
- "ndcg_at_10": 0.30497,
26
- "ndcg_at_100": 0.36328,
27
- "ndcg_at_1000": 0.38075,
28
- "ndcg_at_20": 0.33003,
29
- "ndcg_at_3": 0.2309,
30
- "ndcg_at_5": 0.26603,
31
- "precision_at_1": 0.14551,
32
- "precision_at_10": 0.05166,
33
- "precision_at_100": 0.00808,
34
- "precision_at_1000": 0.00095,
35
- "precision_at_20": 0.03099,
36
- "precision_at_3": 0.10149,
37
- "precision_at_5": 0.07867,
38
- "recall_at_1": 0.14118,
39
- "recall_at_10": 0.49585,
40
- "recall_at_100": 0.76965,
41
- "recall_at_1000": 0.90623,
42
- "recall_at_20": 0.59377,
43
- "recall_at_3": 0.29241,
44
- "recall_at_5": 0.37722,
45
- "hf_subset": "default",
46
- "main_score": 0.30497,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPDomainVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "6e1ec8c54c018151c77472d94b1c0765230cf6ca",
3
- "task_name": "MTOPDomainVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.855346,
9
- "f1": 0.853271,
10
- "f1_weighted": 0.854398,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.841622,
14
- "f1": 0.833588,
15
- "f1_weighted": 0.841051
16
- },
17
- {
18
- "accuracy": 0.845535,
19
- "f1": 0.84299,
20
- "f1_weighted": 0.845127
21
- },
22
- {
23
- "accuracy": 0.860658,
24
- "f1": 0.858643,
25
- "f1_weighted": 0.859362
26
- },
27
- {
28
- "accuracy": 0.865097,
29
- "f1": 0.860371,
30
- "f1_weighted": 0.864139
31
- },
32
- {
33
- "accuracy": 0.85351,
34
- "f1": 0.85694,
35
- "f1_weighted": 0.852418
36
- },
37
- {
38
- "accuracy": 0.851704,
39
- "f1": 0.850781,
40
- "f1_weighted": 0.850459
41
- },
42
- {
43
- "accuracy": 0.867279,
44
- "f1": 0.863823,
45
- "f1_weighted": 0.865732
46
- },
47
- {
48
- "accuracy": 0.868558,
49
- "f1": 0.862947,
50
- "f1_weighted": 0.866173
51
- },
52
- {
53
- "accuracy": 0.828681,
54
- "f1": 0.837955,
55
- "f1_weighted": 0.829348
56
- },
57
- {
58
- "accuracy": 0.870815,
59
- "f1": 0.864666,
60
- "f1_weighted": 0.870169
61
- }
62
- ],
63
- "main_score": 0.855346,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 9.854939937591553,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MTOPIntentVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "c4e81a5c9a813a0142d905e261e5a446cc6fbc4a",
3
- "task_name": "MTOPIntentVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.580114,
9
- "f1": 0.394653,
10
- "f1_weighted": 0.619897,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.558573,
14
- "f1": 0.394329,
15
- "f1_weighted": 0.598478
16
- },
17
- {
18
- "accuracy": 0.591528,
19
- "f1": 0.401524,
20
- "f1_weighted": 0.634855
21
- },
22
- {
23
- "accuracy": 0.580618,
24
- "f1": 0.387417,
25
- "f1_weighted": 0.616344
26
- },
27
- {
28
- "accuracy": 0.601685,
29
- "f1": 0.407789,
30
- "f1_weighted": 0.648841
31
- },
32
- {
33
- "accuracy": 0.569408,
34
- "f1": 0.394753,
35
- "f1_weighted": 0.606008
36
- },
37
- {
38
- "accuracy": 0.578963,
39
- "f1": 0.379034,
40
- "f1_weighted": 0.616886
41
- },
42
- {
43
- "accuracy": 0.58739,
44
- "f1": 0.400041,
45
- "f1_weighted": 0.622262
46
- },
47
- {
48
- "accuracy": 0.568731,
49
- "f1": 0.392647,
50
- "f1_weighted": 0.610166
51
- },
52
- {
53
- "accuracy": 0.562862,
54
- "f1": 0.393473,
55
- "f1_weighted": 0.60385
56
- },
57
- {
58
- "accuracy": 0.601384,
59
- "f1": 0.395527,
60
- "f1_weighted": 0.64128
61
- }
62
- ],
63
- "main_score": 0.580114,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 16.95852541923523,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveIntentVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "35c7ced69f958dbbaa24f792db4a9250e461866d",
3
- "task_name": "MassiveIntentVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.67741,
9
- "f1": 0.643766,
10
- "f1_weighted": 0.665189,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.679933,
14
- "f1": 0.670853,
15
- "f1_weighted": 0.66368
16
- },
17
- {
18
- "accuracy": 0.705717,
19
- "f1": 0.65224,
20
- "f1_weighted": 0.698288
21
- },
22
- {
23
- "accuracy": 0.675448,
24
- "f1": 0.627179,
25
- "f1_weighted": 0.662605
26
- },
27
- {
28
- "accuracy": 0.707399,
29
- "f1": 0.675465,
30
- "f1_weighted": 0.699657
31
- },
32
- {
33
- "accuracy": 0.659193,
34
- "f1": 0.624411,
35
- "f1_weighted": 0.641452
36
- },
37
- {
38
- "accuracy": 0.661996,
39
- "f1": 0.611837,
40
- "f1_weighted": 0.654907
41
- },
42
- {
43
- "accuracy": 0.679372,
44
- "f1": 0.646968,
45
- "f1_weighted": 0.669935
46
- },
47
- {
48
- "accuracy": 0.665359,
49
- "f1": 0.646588,
50
- "f1_weighted": 0.653284
51
- },
52
- {
53
- "accuracy": 0.648543,
54
- "f1": 0.624194,
55
- "f1_weighted": 0.628283
56
- },
57
- {
58
- "accuracy": 0.691143,
59
- "f1": 0.657925,
60
- "f1_weighted": 0.679795
61
- }
62
- ],
63
- "main_score": 0.67741,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 8.183109998703003,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/MassiveScenarioVNClassification.json DELETED
@@ -1,73 +0,0 @@
1
- {
2
- "dataset_revision": "a82e282d9f5aec1a8cf7d868ce40f70669c16b89",
3
- "task_name": "MassiveScenarioVNClassification",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "accuracy": 0.728523,
9
- "f1": 0.701865,
10
- "f1_weighted": 0.724309,
11
- "scores_per_experiment": [
12
- {
13
- "accuracy": 0.74217,
14
- "f1": 0.721876,
15
- "f1_weighted": 0.741215
16
- },
17
- {
18
- "accuracy": 0.746085,
19
- "f1": 0.724708,
20
- "f1_weighted": 0.744061
21
- },
22
- {
23
- "accuracy": 0.742729,
24
- "f1": 0.715626,
25
- "f1_weighted": 0.740294
26
- },
27
- {
28
- "accuracy": 0.725951,
29
- "f1": 0.704168,
30
- "f1_weighted": 0.71827
31
- },
32
- {
33
- "accuracy": 0.705257,
34
- "f1": 0.665866,
35
- "f1_weighted": 0.696431
36
- },
37
- {
38
- "accuracy": 0.722595,
39
- "f1": 0.691101,
40
- "f1_weighted": 0.719705
41
- },
42
- {
43
- "accuracy": 0.716443,
44
- "f1": 0.689115,
45
- "f1_weighted": 0.708131
46
- },
47
- {
48
- "accuracy": 0.71085,
49
- "f1": 0.69065,
50
- "f1_weighted": 0.710364
51
- },
52
- {
53
- "accuracy": 0.739374,
54
- "f1": 0.711083,
55
- "f1_weighted": 0.736213
56
- },
57
- {
58
- "accuracy": 0.733781,
59
- "f1": 0.704461,
60
- "f1_weighted": 0.728405
61
- }
62
- ],
63
- "main_score": 0.728523,
64
- "hf_subset": "default",
65
- "languages": [
66
- "vie-Latn"
67
- ]
68
- }
69
- ]
70
- },
71
- "evaluation_time": 5.51207423210144,
72
- "kg_co2_emissions": null
73
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NFCorpus-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "a13d72fbb859be3dc19ab669d1ec9510407d2dcd",
3
- "mteb_dataset_name": "NFCorpus-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "NFCorpus-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 19.28,
10
- "map_at_1": 0.03605,
11
- "map_at_10": 0.08866,
12
- "map_at_100": 0.10878,
13
- "map_at_1000": 0.11959,
14
- "map_at_20": 0.09686,
15
- "map_at_3": 0.06829,
16
- "map_at_5": 0.07966,
17
- "mrr_at_1": 0.31928,
18
- "mrr_at_10": 0.41437,
19
- "mrr_at_100": 0.42156,
20
- "mrr_at_1000": 0.422,
21
- "mrr_at_20": 0.41738,
22
- "mrr_at_3": 0.39056,
23
- "mrr_at_5": 0.3993,
24
- "ndcg_at_1": 0.3012,
25
- "ndcg_at_10": 0.25389,
26
- "ndcg_at_100": 0.23107,
27
- "ndcg_at_1000": 0.31251,
28
- "ndcg_at_20": 0.23925,
29
- "ndcg_at_3": 0.28632,
30
- "ndcg_at_5": 0.27126,
31
- "precision_at_1": 0.31928,
32
- "precision_at_10": 0.18855,
33
- "precision_at_100": 0.06163,
34
- "precision_at_1000": 0.01931,
35
- "precision_at_20": 0.14217,
36
- "precision_at_3": 0.2751,
37
- "precision_at_5": 0.23494,
38
- "recall_at_1": 0.03605,
39
- "recall_at_10": 0.12335,
40
- "recall_at_100": 0.24212,
41
- "recall_at_1000": 0.52674,
42
- "recall_at_20": 0.14914,
43
- "recall_at_3": 0.0801,
44
- "recall_at_5": 0.09947,
45
- "hf_subset": "default",
46
- "main_score": 0.25389,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/NQ-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "40a6d7f343b9c9f4855a426d8c431ad5f8aaf56b",
3
- "mteb_dataset_name": "NQ-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "NQ-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 8244.07,
10
- "map_at_1": 0.22333,
11
- "map_at_10": 0.35419,
12
- "map_at_100": 0.36617,
13
- "map_at_1000": 0.36666,
14
- "map_at_20": 0.36191,
15
- "map_at_3": 0.31386,
16
- "map_at_5": 0.3369,
17
- "mrr_at_1": 0.25158,
18
- "mrr_at_10": 0.37521,
19
- "mrr_at_100": 0.38499,
20
- "mrr_at_1000": 0.38535,
21
- "mrr_at_20": 0.38163,
22
- "mrr_at_3": 0.33939,
23
- "mrr_at_5": 0.36031,
24
- "ndcg_at_1": 0.25158,
25
- "ndcg_at_10": 0.42611,
26
- "ndcg_at_100": 0.47922,
27
- "ndcg_at_1000": 0.49185,
28
- "ndcg_at_20": 0.45171,
29
- "ndcg_at_3": 0.34799,
30
- "ndcg_at_5": 0.38727,
31
- "precision_at_1": 0.25158,
32
- "precision_at_10": 0.07363,
33
- "precision_at_100": 0.01035,
34
- "precision_at_1000": 0.00115,
35
- "precision_at_20": 0.04271,
36
- "precision_at_3": 0.16298,
37
- "precision_at_5": 0.11979,
38
- "recall_at_1": 0.22333,
39
- "recall_at_10": 0.62671,
40
- "recall_at_100": 0.86105,
41
- "recall_at_1000": 0.95798,
42
- "recall_at_20": 0.72228,
43
- "recall_at_3": 0.42149,
44
- "recall_at_5": 0.51241,
45
- "hf_subset": "default",
46
- "main_score": 0.42611,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/Quora-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "3363d81e41b67c1032bf3b234882a03d271e2289",
3
- "mteb_dataset_name": "Quora-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "Quora-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 1180.22,
10
- "map_at_1": 0.42135,
11
- "map_at_10": 0.54669,
12
- "map_at_100": 0.55931,
13
- "map_at_1000": 0.55994,
14
- "map_at_20": 0.55423,
15
- "map_at_3": 0.51137,
16
- "map_at_5": 0.53213,
17
- "mrr_at_1": 0.49979,
18
- "mrr_at_10": 0.59915,
19
- "mrr_at_100": 0.60569,
20
- "mrr_at_1000": 0.60589,
21
- "mrr_at_20": 0.60345,
22
- "mrr_at_3": 0.57472,
23
- "mrr_at_5": 0.58942,
24
- "ndcg_at_1": 0.49979,
25
- "ndcg_at_10": 0.61001,
26
- "ndcg_at_100": 0.65467,
27
- "ndcg_at_1000": 0.66382,
28
- "ndcg_at_20": 0.63036,
29
- "ndcg_at_3": 0.55641,
30
- "ndcg_at_5": 0.58314,
31
- "precision_at_1": 0.49979,
32
- "precision_at_10": 0.10162,
33
- "precision_at_100": 0.01399,
34
- "precision_at_1000": 0.00154,
35
- "precision_at_20": 0.05794,
36
- "precision_at_3": 0.2526,
37
- "precision_at_5": 0.1746,
38
- "recall_at_1": 0.42135,
39
- "recall_at_10": 0.74453,
40
- "recall_at_100": 0.92748,
41
- "recall_at_1000": 0.98654,
42
- "recall_at_20": 0.81644,
43
- "recall_at_3": 0.59369,
44
- "recall_at_5": 0.66411,
45
- "hf_subset": "default",
46
- "main_score": 0.61001,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClustering-VN.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "dataset_revision": "7f7d4097979633181b2df3f73905218f74c4665d",
3
- "task_name": "RedditClustering-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "v_measure": 0.438902,
9
- "v_measure_std": 0.044274,
10
- "v_measures": [
11
- 0.39263,
12
- 0.5316,
13
- 0.399559,
14
- 0.425985,
15
- 0.427061,
16
- 0.446164,
17
- 0.512027,
18
- 0.424049,
19
- 0.426086,
20
- 0.40386
21
- ],
22
- "main_score": 0.438902,
23
- "hf_subset": "default",
24
- "languages": [
25
- "vie-Latn"
26
- ]
27
- }
28
- ]
29
- },
30
- "evaluation_time": 125.31211185455322,
31
- "kg_co2_emissions": null
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/RedditClusteringP2P-VN.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "dataset_revision": "841856dcb82496f1f2f59356e4798ce15baeb200",
3
- "task_name": "RedditClusteringP2P-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "v_measure": 0.561615,
9
- "v_measure_std": 0.118839,
10
- "v_measures": [
11
- 0.605948,
12
- 0.611291,
13
- 0.601361,
14
- 0.345424,
15
- 0.620675,
16
- 0.572602,
17
- 0.315689,
18
- 0.660052,
19
- 0.610597,
20
- 0.672515
21
- ],
22
- "main_score": 0.561615,
23
- "hf_subset": "default",
24
- "languages": [
25
- "vie-Latn"
26
- ]
27
- }
28
- ]
29
- },
30
- "evaluation_time": 1739.9407813549042,
31
- "kg_co2_emissions": null
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SCIDOCS-VN.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "dataset_revision": "724cddfa9d328a193f303a0a9b7789468ac79f26",
3
- "mteb_dataset_name": "SCIDOCS-VN",
4
- "mteb_version": "1.38.41",
5
- "task_name": "SCIDOCS-VN",
6
- "scores": {
7
- "test": [
8
- {
9
- "evaluation_time": 96.99,
10
- "map_at_1": 0.03771,
11
- "map_at_10": 0.07617,
12
- "map_at_100": 0.08657,
13
- "map_at_1000": 0.08887,
14
- "map_at_20": 0.08085,
15
- "map_at_3": 0.06048,
16
- "map_at_5": 0.06859,
17
- "mrr_at_1": 0.1825,
18
- "mrr_at_10": 0.2542,
19
- "mrr_at_100": 0.26523,
20
- "mrr_at_1000": 0.26606,
21
- "mrr_at_20": 0.25994,
22
- "mrr_at_3": 0.23333,
23
- "mrr_at_5": 0.24621,
24
- "ndcg_at_1": 0.1825,
25
- "ndcg_at_10": 0.13034,
26
- "ndcg_at_100": 0.18324,
27
- "ndcg_at_1000": 0.23392,
28
- "ndcg_at_20": 0.14611,
29
- "ndcg_at_3": 0.13596,
30
- "ndcg_at_5": 0.11389,
31
- "precision_at_1": 0.1825,
32
- "precision_at_10": 0.062,
33
- "precision_at_100": 0.01407,
34
- "precision_at_1000": 0.00263,
35
- "precision_at_20": 0.04013,
36
- "precision_at_3": 0.1225,
37
- "precision_at_5": 0.094,
38
- "recall_at_1": 0.03771,
39
- "recall_at_10": 0.12642,
40
- "recall_at_100": 0.28675,
41
- "recall_at_1000": 0.5355,
42
- "recall_at_20": 0.16375,
43
- "recall_at_3": 0.07496,
44
- "recall_at_5": 0.09571,
45
- "hf_subset": "default",
46
- "main_score": 0.13034,
47
- "languages": [
48
- "vie-Latn"
49
- ]
50
- }
51
- ]
52
- },
53
- "evaluation_time": null
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
results/AITeamVN__Vietnamese_Embedding/fcbbb905e6c3757d421aaa5db6fd7c53d038f6fb/SICK-R-VN.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "dataset_revision": "bc89f0401983c456b609f7fb324278f346b2cccf",
3
- "task_name": "SICK-R-VN",
4
- "mteb_version": "1.38.41",
5
- "scores": {
6
- "test": [
7
- {
8
- "pearson": 0.80717,
9
- "spearman": 0.771026,
10
- "cosine_pearson": 0.80717,
11
- "cosine_spearman": 0.771062,
12
- "manhattan_pearson": 0.768392,
13
- "manhattan_spearman": 0.770402,
14
- "euclidean_pearson": 0.768774,
15
- "euclidean_spearman": 0.771054,
16
- "main_score": 0.771062,
17
- "hf_subset": "default",
18
- "languages": [
19
- "vie-Latn"
20
- ]
21
- }
22
- ]
23
- },
24
- "evaluation_time": 25.608262062072754,
25
- "kg_co2_emissions": null
26
- }