Files
SGPT-1.3B-weightedmean-msma…/evaluation/mteb/TwitterURLCorpus.json

48 lines
1.3 KiB
JSON
Raw Normal View History

{
"dataset_version": null,
"mteb_version": "0.0.2",
"test": {
"cos_sim": {
"accuracy": 0.8797881010594947,
"accuracy_threshold": 0.6835847496986389,
"ap": 0.8378748636891035,
"f1": 0.7594113995691386,
"f1_threshold": 0.6381056308746338,
"precision": 0.7222029307590805,
"recall": 0.8006621496766245
},
"dot": {
"accuracy": 0.8569294058291613,
"accuracy_threshold": 2096.25537109375,
"ap": 0.7815363722278026,
"f1": 0.7208894926888564,
"f1_threshold": 1926.618408203125,
"precision": 0.6728959487419075,
"recall": 0.7762550046196489
},
"euclidean": {
"accuracy": 0.877362517949315,
"accuracy_threshold": 44.67449188232422,
"ap": 0.8319012184470559,
"f1": 0.7551480646234611,
"f1_threshold": 46.798240661621094,
"precision": 0.7263352535381552,
"recall": 0.786341238065907
},
"evaluation_time": 221.47,
"manhattan": {
"accuracy": 0.8774013272790779,
"accuracy_threshold": 1568.80908203125,
"ap": 0.8323305405113404,
"f1": 0.7563960775639608,
"f1_threshold": 1638.9019775390625,
"precision": 0.72563304569246,
"recall": 0.789882968894364
},
"max": {
"accuracy": 0.8797881010594947,
"ap": 0.8378748636891035,
"f1": 0.7594113995691386
}
}
}