Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
24 commits
Select commit Hold shift + click to select a range
cd95405
add results of perplexity-ai/pplx-embed-v1-4b and perplexity-ai/pplx-…
bwang-pplx Mar 18, 2026
d487c57
[ADD] results COLQWEN3.5-4.5B (#448)
athrael-soju Mar 18, 2026
931e068
Results of F2LLM-v2 (#438)
Geralt-Targaryen Mar 18, 2026
5545039
[ADD] results for COLQWEN3.5-4.5B Vidore V3 (#451)
athrael-soju Mar 20, 2026
6c4d6c9
Add results for nanovdr/NanoVDR-S-Multi (ViDoRe v1/v2/v3) (#447)
Ryenhails Mar 20, 2026
e64ae3d
add colqwen (#452)
Samoed Mar 21, 2026
21ffdac
Add potion-base-32M and potion-retrieval-32M results (#454)
Pringled Mar 23, 2026
2d4261b
Adding BM25 missing results (#457)
fzoll Mar 23, 2026
93f1ecb
Create Results PR Comment for Results diff (#444)
ayush1298 Mar 24, 2026
9c3027c
Add Thai task results for 21 models (MTEB(tha, v1) coverage) (#458)
anusoft Mar 24, 2026
fe1b53f
add qihoo360/Zhinao-ChineseModernBert-Embedding. (#453)
gitabtion Mar 26, 2026
0023368
add bright results (#460)
Samoed Mar 26, 2026
6c19c31
Add MTEB(spa, v1) results for 14 multilingual models (#456)
Clemente-H Mar 27, 2026
4311450
Add results for microsoft/harrier-oss-v1 (0.6b, 270m, 27b) (#463)
intfloat Mar 31, 2026
dcb3f3f
Adding more BM25 results (#464)
fzoll Mar 31, 2026
8047390
Adding more minishlab/potion-multilingual-128M results (#466)
fzoll Mar 31, 2026
2965501
add bm25s (#467)
Samoed Mar 31, 2026
f644bbd
Adding more intfloat/multilingual-e5-small results (#469)
fzoll Mar 31, 2026
a4116da
Adding sentence-transformers/static-similarity-mrl-multilingual-v1 re…
fzoll Apr 1, 2026
ab78bb7
add bright results (#471)
Samoed Apr 1, 2026
ce37689
add bright results (#472)
Samoed Apr 1, 2026
6e02b2c
[ADD] results VN-MTEB
BaoLocPham Mar 5, 2026
025677d
[ADD] results VN-MTEB (remove duplicate)
BaoLocPham Mar 5, 2026
364a9c5
[DELETE] remove potential wrong results
BaoLocPham Apr 1, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
The diff you're trying to view is too large. We only load the first 3000 changed files.
Original file line number Diff line number Diff line change
@@ -0,0 +1,139 @@
{
"dataset_revision": "b48bc27d383cfca5b6a47135a52390fa5f66b253",
"task_name": "AmazonCounterfactualVNClassification",
"mteb_version": "2.8.1",
"scores": {
"test": [
{
"scores_per_experiment": [
{
"accuracy": 0.587983,
"f1": 0.555485,
"f1_weighted": 0.631829,
"precision": 0.619032,
"precision_weighted": 0.827906,
"recall": 0.697761,
"recall_weighted": 0.587983,
"ap": 0.276246,
"ap_weighted": 0.276246
},
{
"accuracy": 0.736052,
"f1": 0.655812,
"f1_weighted": 0.761372,
"precision": 0.647167,
"precision_weighted": 0.81801,
"recall": 0.719762,
"recall_weighted": 0.736052,
"ap": 0.318313,
"ap_weighted": 0.318313
},
{
"accuracy": 0.652361,
"f1": 0.594521,
"f1_weighted": 0.691796,
"precision": 0.615868,
"precision_weighted": 0.808758,
"recall": 0.691431,
"recall_weighted": 0.652361,
"ap": 0.280129,
"ap_weighted": 0.280129
},
{
"accuracy": 0.51073,
"f1": 0.482958,
"f1_weighted": 0.559073,
"precision": 0.567234,
"precision_weighted": 0.776301,
"recall": 0.609387,
"recall_weighted": 0.51073,
"ap": 0.224991,
"ap_weighted": 0.224991
},
{
"accuracy": 0.622318,
"f1": 0.539112,
"f1_weighted": 0.6635,
"precision": 0.556778,
"precision_weighted": 0.750774,
"recall": 0.590798,
"recall_weighted": 0.622318,
"ap": 0.219724,
"ap_weighted": 0.219724
},
{
"accuracy": 0.583691,
"f1": 0.53557,
"f1_weighted": 0.630528,
"precision": 0.580971,
"precision_weighted": 0.782264,
"recall": 0.635726,
"recall_weighted": 0.583691,
"ap": 0.241011,
"ap_weighted": 0.241011
},
{
"accuracy": 0.690987,
"f1": 0.626203,
"f1_weighted": 0.725049,
"precision": 0.633489,
"precision_weighted": 0.818484,
"recall": 0.715053,
"recall_weighted": 0.690987,
"ap": 0.302755,
"ap_weighted": 0.302755
},
{
"accuracy": 0.594421,
"f1": 0.549602,
"f1_weighted": 0.639849,
"precision": 0.595785,
"precision_weighted": 0.797782,
"recall": 0.660568,
"recall_weighted": 0.594421,
"ap": 0.255337,
"ap_weighted": 0.255337
},
{
"accuracy": 0.618026,
"f1": 0.556143,
"f1_weighted": 0.661415,
"precision": 0.583627,
"precision_weighted": 0.779435,
"recall": 0.638444,
"recall_weighted": 0.618026,
"ap": 0.244741,
"ap_weighted": 0.244741
},
{
"accuracy": 0.600858,
"f1": 0.523266,
"f1_weighted": 0.645433,
"precision": 0.547742,
"precision_weighted": 0.744271,
"recall": 0.577675,
"recall_weighted": 0.600858,
"ap": 0.212676,
"ap_weighted": 0.212676
}
],
"accuracy": 0.619742,
"f1": 0.561867,
"f1_weighted": 0.660984,
"precision": 0.594769,
"precision_weighted": 0.790399,
"recall": 0.653661,
"recall_weighted": 0.619742,
"ap": 0.257592,
"ap_weighted": 0.257592,
"main_score": 0.619742,
"hf_subset": "default",
"languages": [
"vie-Latn"
]
}
]
},
"evaluation_time": 7.614054918289185,
"kg_co2_emissions": null
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,139 @@
{
"dataset_revision": "4e9a0d6e6bd97ab32f23c50c043d751eed2a5f8a",
"task_name": "AmazonPolarityVNClassification",
"mteb_version": "2.8.1",
"scores": {
"test": [
{
"scores_per_experiment": [
{
"accuracy": 0.887756,
"f1": 0.887564,
"f1_weighted": 0.88753,
"precision": 0.891256,
"precision_weighted": 0.891543,
"recall": 0.88811,
"recall_weighted": 0.887756,
"ap": 0.86203,
"ap_weighted": 0.86203
},
{
"accuracy": 0.886623,
"f1": 0.886583,
"f1_weighted": 0.886599,
"precision": 0.88686,
"precision_weighted": 0.886802,
"recall": 0.886533,
"recall_weighted": 0.886623,
"ap": 0.840929,
"ap_weighted": 0.840929
},
{
"accuracy": 0.887771,
"f1": 0.887457,
"f1_weighted": 0.887413,
"precision": 0.893213,
"precision_weighted": 0.893569,
"recall": 0.88821,
"recall_weighted": 0.887771,
"ap": 0.86644,
"ap_weighted": 0.86644
},
{
"accuracy": 0.879607,
"f1": 0.879353,
"f1_weighted": 0.879312,
"precision": 0.883746,
"precision_weighted": 0.884055,
"recall": 0.879994,
"recall_weighted": 0.879607,
"ap": 0.852779,
"ap_weighted": 0.852779
},
{
"accuracy": 0.894125,
"f1": 0.894014,
"f1_weighted": 0.893989,
"precision": 0.896441,
"precision_weighted": 0.896679,
"recall": 0.894412,
"recall_weighted": 0.894125,
"ap": 0.86726,
"ap_weighted": 0.86726
},
{
"accuracy": 0.870804,
"f1": 0.870803,
"f1_weighted": 0.870801,
"precision": 0.870905,
"precision_weighted": 0.870966,
"recall": 0.870869,
"recall_weighted": 0.870804,
"ap": 0.827334,
"ap_weighted": 0.827334
},
{
"accuracy": 0.896821,
"f1": 0.896763,
"f1_weighted": 0.896744,
"precision": 0.898222,
"precision_weighted": 0.89841,
"recall": 0.897045,
"recall_weighted": 0.896821,
"ap": 0.86776,
"ap_weighted": 0.86776
},
{
"accuracy": 0.901696,
"f1": 0.901691,
"f1_weighted": 0.901696,
"precision": 0.901689,
"precision_weighted": 0.901697,
"recall": 0.901694,
"recall_weighted": 0.901696,
"ap": 0.863676,
"ap_weighted": 0.863676
},
{
"accuracy": 0.90081,
"f1": 0.900772,
"f1_weighted": 0.900787,
"precision": 0.901091,
"precision_weighted": 0.901025,
"recall": 0.900715,
"recall_weighted": 0.90081,
"ap": 0.858558,
"ap_weighted": 0.858558
},
{
"accuracy": 0.871571,
"f1": 0.870543,
"f1_weighted": 0.870628,
"precision": 0.882083,
"precision_weighted": 0.881628,
"recall": 0.87096,
"recall_weighted": 0.871571,
"ap": 0.805829,
"ap_weighted": 0.805829
}
],
"accuracy": 0.887758,
"f1": 0.887554,
"f1_weighted": 0.88755,
"precision": 0.890551,
"precision_weighted": 0.890637,
"recall": 0.887854,
"recall_weighted": 0.887758,
"ap": 0.851259,
"ap_weighted": 0.851259,
"main_score": 0.887758,
"hf_subset": "default",
"languages": [
"vie-Latn"
]
}
]
},
"evaluation_time": 5269.787292957306,
"kg_co2_emissions": null
}
Loading
Loading