eval/beir-arguana_ndcg@10 = 0.38885 eval/beir-arguana_recall@100 = 0.93741 eval/beir-avg_ndcg@10 = 0.346578 eval/beir-avg_recall@10 = 0.40747999999999995 eval/beir-avg_recall@100 = 0.628204 eval/beir-avg_recall@20 = 0.47427600000000003 eval/beir-climate-fever_ndcg@10 = 0.15352 eval/beir-climate-fever_recall@10 = 0.19198 eval/beir-climate-fever_recall@100 = 0.41676 eval/beir-climate-fever_recall@20 = 0.24603 eval/beir-cqadupstack_ndcg@10 = 0.27400583333333334 eval/beir-cqadupstack_recall@100 = 0.6010533333333333 eval/beir-dbpedia-entity_ndcg@10 = 0.29868 eval/beir-dbpedia-entity_recall@10 = 0.19285 eval/beir-dbpedia-entity_recall@100 = 0.44014 eval/beir-dbpedia-entity_recall@20 = 0.25779 eval/beir-fever_ndcg@10 = 0.57055 eval/beir-fever_recall@10 = 0.75716 eval/beir-fever_recall@100 = 0.90338 eval/beir-fever_recall@20 = 0.81739 eval/beir-fiqa_ndcg@10 = 0.23933 eval/beir-fiqa_recall@100 = 0.56617 eval/beir-hotpotqa_ndcg@10 = 0.50295 eval/beir-hotpotqa_recall@10 = 0.53862 eval/beir-hotpotqa_recall@100 = 0.70223 eval/beir-hotpotqa_recall@20 = 0.58913 eval/beir-msmarco_ndcg@10 = 0.20719 eval/beir-msmarco_recall@10 = 0.35679 eval/beir-msmarco_recall@100 = 0.67851 eval/beir-msmarco_recall@20 = 0.46104 eval/beir-nfcorpus_ndcg@10 = 0.31333 eval/beir-nfcorpus_recall@100 = 0.29552 eval/beir-nq_ndcg@10 = 0.27242 eval/beir-nq_recall@100 = 0.79193 eval/beir-quora_ndcg@10 = 0.77732 eval/beir-quora_recall@100 = 0.97812 eval/beir-scidocs_ndcg@10 = 0.1498 eval/beir-scidocs_recall@100 = 0.36587 eval/beir-scifact_ndcg@10 = 0.66223 eval/beir-scifact_recall@100 = 0.90156 eval/beir-trec-covid_ndcg@10 = 0.58943 eval/beir-trec-covid_recall@100 = 0.10692 eval/beir-webis-touche2020_ndcg@10 = 0.18989 eval/beir-webis-touche2020_recall@100 = 0.45315 eval/qa-curatedtrec-test-acc@100 = 0.9279538904899135 eval/qa-curatedtrec-test-acc@20 = 0.8414985590778098 eval/qa-curatedtrec-test-acc@5 = 0.6412103746397695 eval/qa-entityqs-macro-acc@100 = 0.7589938692225383 eval/qa-entityqs-macro-acc@20 = 0.6390821603389343 eval/qa-entityqs-macro-acc@5 = 0.49924685926586126 eval/qa-nq-test-acc@100 = 0.8011080332409972 eval/qa-nq-test-acc@20 = 0.6518005540166205 eval/qa-nq-test-acc@5 = 0.4476454293628809 eval/qa-squad1-test-acc@100 = 0.7701040681173131 eval/qa-squad1-test-acc@20 = 0.6120151371807001 eval/qa-squad1-test-acc@5 = 0.41911069063386946 eval/qa-trivia-test-acc@100 = 0.8332891363917617 eval/qa-trivia-test-acc@20 = 0.7486077963404932 eval/qa-trivia-test-acc@5 = 0.6133651551312649 eval/qa-webq-test-acc@100 = 0.8105314960629921 eval/qa-webq-test-acc@20 = 0.6683070866141733 eval/qa-webq-test-acc@5 = 0.45570866141732286 eval/senteval-CR = 87.12 eval/senteval-MPQA = 89.04 eval/senteval-MR = 81.19 eval/senteval-MRPC = 71.69 eval/senteval-SICKRelatedness = 0.6909477744903625 eval/senteval-SST2 = 85.21 eval/senteval-STS12 = 0.5915463354166586 eval/senteval-STS13 = 0.7392805868780532 eval/senteval-STS14 = 0.670800547509326 eval/senteval-STS15 = 0.790296617367925 eval/senteval-STS16 = 0.7906179855205325 eval/senteval-STSBenchmark = 0.7534738635333297 eval/senteval-SUBJ = 95.61 eval/senteval-TREC = 80.69 eval/senteval-avg_sts_7 = 0.7181376729594552 eval/senteval-avg_transfer = 84.36428571428571 train/global_step = 100000