Model,Accuracy Qwen2-7B-Instruct,0.9346330275229358 Meta-Llama-3.1-8B-Instruct,0.8956422018348624 llama3-8b-cpt-sea-lionv2.1-instruct,0.9174311926605505 Qwen2_5_32B_Instruct,0.9472477064220184 Qwen2_5_7B_Instruct,0.9254587155963303 Qwen2_5_1_5B_Instruct,0.9231651376146789 Qwen2-72B-Instruct,0.9346330275229358 Sailor2-8B-Chat,0.9461009174311926 Meta-Llama-3-8B-Instruct,0.8784403669724771 Meta-Llama-3.1-70B-Instruct,0.9529816513761468 Qwen2_5_3B_Instruct,0.8245412844036697 llama3.1-70b-cpt-sea-lionv3-instruct,0.9529816513761468 SeaLLMs-v3-7B-Chat,0.9403669724770642 Qwen2_5_72B_Instruct,0.9334862385321101 meralion-merged-llama3-8b-sg-inst-avg-diff,0.8841743119266054 gemma-2-9b-it,0.9311926605504587 llama3.1-8b-cpt-sea-lionv3-instruct,0.9231651376146789 Meta-Llama-3-70B-Instruct,0.9495412844036697 Qwen2_5_14B_Instruct,0.9311926605504587 gemma2-9b-cpt-sea-lionv3-instruct,0.9311926605504587 gemma-2-2b-it,0.9243119266055045 llama3-8b-cpt-sea-lionv2-instruct,0.9128440366972477 cross_openhermes_llama3_8b_12288_inst,0.9288990825688074 Qwen2_5_0_5B_Instruct,0.7889908256880734 GPT4o_0513,0.9415137614678899