AugTriever-TQGen-Topic / eval_results.txt
memray's picture
Upload 130 files
cdaafab
eval/beir-arguana_ndcg@10 = 0.38885
eval/beir-arguana_recall@100 = 0.93741
eval/beir-avg_ndcg@10 = 0.346578
eval/beir-avg_recall@10 = 0.40747999999999995
eval/beir-avg_recall@100 = 0.628204
eval/beir-avg_recall@20 = 0.47427600000000003
eval/beir-climate-fever_ndcg@10 = 0.15352
eval/beir-climate-fever_recall@10 = 0.19198
eval/beir-climate-fever_recall@100 = 0.41676
eval/beir-climate-fever_recall@20 = 0.24603
eval/beir-cqadupstack_ndcg@10 = 0.27400583333333334
eval/beir-cqadupstack_recall@100 = 0.6010533333333333
eval/beir-dbpedia-entity_ndcg@10 = 0.29868
eval/beir-dbpedia-entity_recall@10 = 0.19285
eval/beir-dbpedia-entity_recall@100 = 0.44014
eval/beir-dbpedia-entity_recall@20 = 0.25779
eval/beir-fever_ndcg@10 = 0.57055
eval/beir-fever_recall@10 = 0.75716
eval/beir-fever_recall@100 = 0.90338
eval/beir-fever_recall@20 = 0.81739
eval/beir-fiqa_ndcg@10 = 0.23933
eval/beir-fiqa_recall@100 = 0.56617
eval/beir-hotpotqa_ndcg@10 = 0.50295
eval/beir-hotpotqa_recall@10 = 0.53862
eval/beir-hotpotqa_recall@100 = 0.70223
eval/beir-hotpotqa_recall@20 = 0.58913
eval/beir-msmarco_ndcg@10 = 0.20719
eval/beir-msmarco_recall@10 = 0.35679
eval/beir-msmarco_recall@100 = 0.67851
eval/beir-msmarco_recall@20 = 0.46104
eval/beir-nfcorpus_ndcg@10 = 0.31333
eval/beir-nfcorpus_recall@100 = 0.29552
eval/beir-nq_ndcg@10 = 0.27242
eval/beir-nq_recall@100 = 0.79193
eval/beir-quora_ndcg@10 = 0.77732
eval/beir-quora_recall@100 = 0.97812
eval/beir-scidocs_ndcg@10 = 0.1498
eval/beir-scidocs_recall@100 = 0.36587
eval/beir-scifact_ndcg@10 = 0.66223
eval/beir-scifact_recall@100 = 0.90156
eval/beir-trec-covid_ndcg@10 = 0.58943
eval/beir-trec-covid_recall@100 = 0.10692
eval/beir-webis-touche2020_ndcg@10 = 0.18989
eval/beir-webis-touche2020_recall@100 = 0.45315
eval/qa-curatedtrec-test-acc@100 = 0.9279538904899135
eval/qa-curatedtrec-test-acc@20 = 0.8414985590778098
eval/qa-curatedtrec-test-acc@5 = 0.6412103746397695
eval/qa-entityqs-macro-acc@100 = 0.7589938692225383
eval/qa-entityqs-macro-acc@20 = 0.6390821603389343
eval/qa-entityqs-macro-acc@5 = 0.49924685926586126
eval/qa-nq-test-acc@100 = 0.8011080332409972
eval/qa-nq-test-acc@20 = 0.6518005540166205
eval/qa-nq-test-acc@5 = 0.4476454293628809
eval/qa-squad1-test-acc@100 = 0.7701040681173131
eval/qa-squad1-test-acc@20 = 0.6120151371807001
eval/qa-squad1-test-acc@5 = 0.41911069063386946
eval/qa-trivia-test-acc@100 = 0.8332891363917617
eval/qa-trivia-test-acc@20 = 0.7486077963404932
eval/qa-trivia-test-acc@5 = 0.6133651551312649
eval/qa-webq-test-acc@100 = 0.8105314960629921
eval/qa-webq-test-acc@20 = 0.6683070866141733
eval/qa-webq-test-acc@5 = 0.45570866141732286
eval/senteval-CR = 87.12
eval/senteval-MPQA = 89.04
eval/senteval-MR = 81.19
eval/senteval-MRPC = 71.69
eval/senteval-SICKRelatedness = 0.6909477744903625
eval/senteval-SST2 = 85.21
eval/senteval-STS12 = 0.5915463354166586
eval/senteval-STS13 = 0.7392805868780532
eval/senteval-STS14 = 0.670800547509326
eval/senteval-STS15 = 0.790296617367925
eval/senteval-STS16 = 0.7906179855205325
eval/senteval-STSBenchmark = 0.7534738635333297
eval/senteval-SUBJ = 95.61
eval/senteval-TREC = 80.69
eval/senteval-avg_sts_7 = 0.7181376729594552
eval/senteval-avg_transfer = 84.36428571428571
train/global_step = 100000