Skip to content

Latest commit

 

History

History
75 lines (48 loc) · 3.76 KB

_MMLU_redux.md

File metadata and controls

75 lines (48 loc) · 3.76 KB

View the table

python src/evaluation/mcqa_eval.py mmlu-redux

Run model inference

# export VLLM_WORKER_MULTIPROC_METHOD="FLASH_ATTN"
# VLLM_WORKER_MULTIPROC_METHOD=spawn 
bash zero_eval_local.sh -d mmlu-redux -m Qwen/Qwen2-1.5B-Instruct -p Qwen2-1.5B-Instruct -s 4 
bash zero_eval_local.sh -d mmlu-redux -m meta-llama/Meta-Llama-3-70B-Instruct -p Meta-Llama-3-70B-Instruct -s 1 
bash zero_eval_local.sh -d mmlu-redux -m meta-llama/Meta-Llama-3-8B-Instruct -p Meta-Llama-3-8B-Instruct -s 4
bash zero_eval_local.sh -d mmlu-redux -m Qwen/Qwen2-7B-Instruct -p Qwen2-7B-Instruct -s 4 
bash zero_eval_local.sh -d mmlu-redux -m Qwen/Qwen2-72B-Instruct -p Qwen2-72B-Instruct -s 1
bash zero_eval_local.sh -d mmlu-redux -m princeton-nlp/Llama-3-Instruct-8B-SimPO-v0.2 -p Llama-3-Instruct-8B-SimPO-v0.2 -s 4
bash zero_eval_local.sh -d mmlu-redux -m 01-ai/Yi-1.5-34B-Chat -p Yi-1.5-34B-Chat -s 1
bash zero_eval_api.sh -d mmlu-redux -f openai -m openai/gpt-3.5-turbo-0125 -p gpt-3.5-turbo-0125 -s 8

bash zero_eval_api.sh -d mmlu-redux -f openai -m openai/gpt-4o-mini-2024-07-18 -p gpt-4o-mini-2024-07-18 -s 8


bash zero_eval_api.sh -d mmlu-redux -f openai -m openai/gpt-4o-2024-05-13 -p gpt-4o-2024-05-13 -s 8 
bash zero_eval_api.sh -d mmlu-redux -f openai -m openai/gpt-4-0314 -p gpt-4-0314 -s 8
bash zero_eval_api.sh -d mmlu-redux -f openai -m openai/gpt-4-turbo-2024-04-09 -p gpt-4-turbo-2024-04-09 -s 8 

bash zero_eval_api.sh -d mmlu-redux -f anthropic -m anthropic/claude-3-5-sonnet-20240620 -p claude-3-5-sonnet-20240620 -s 8 
bash zero_eval_api.sh -d mmlu-redux -f anthropic -m anthropic/claude-3-haiku-20240307 -p claude-3-haiku-20240307 -s 8
bash zero_eval_api.sh -d mmlu-redux -f anthropic -m anthropic/claude-3-opus-20240229 -p claude-3-opus-20240229 -s 8
bash zero_eval_api.sh -d mmlu-redux -f anthropic -m anthropic/claude-3-sonnet-20240229 -p claude-3-sonnet-20240229 -s 8


bash zero_eval_api.sh -d mmlu-redux -f openai -m google/gemma-2-9b-it@nvidia -p gemma-2-9b-it@nvidia -s 8
bash zero_eval_api.sh -d mmlu-redux -f openai -m google/gemma-2-27b-it@nvidia -p gemma-2-27b-it@nvidia -s 8

bash zero_eval_api.sh -d mmlu-redux -f google -m google/-1.5-pro -p gemini-1.5-pro -s 8
bash zero_eval_api.sh -d mmlu-redux -f google -m google/gemini-1.5-flash -p gemini-1.5-flash -s 8



bash zero_eval_api.sh -d mmlu-redux -f openai -m deepseek-chat -p deepseek-chat -s 8
bash zero_eval_api.sh -d mmlu-redux -f openai -m deepseek-coder -p deepseek-coder -s 8


bash zero_eval_api.sh -d mmlu-redux -f reka -m reka/reka-flash-20240226 -p reka-flash-20240226 -s 8
bash zero_eval_api.sh -d mmlu-redux -f reka -m reka/reka-core-20240501 -p reka-core-20240501 -s 8
bash zero_eval_api.sh -d mmlu-redux -f openai -m yi-large -p yi-large -s 8
bash zero_eval_api.sh -d mmlu-redux -f openai -m yi-large-preview -p yi-large-preview -s 8

bash zero_eval_api.sh -d mmlu-redux -f together -m google/gemma-2-27b-it@together -p gemma-2-27b-it@together -s 8

# bash zero_eval_api.sh -d mmlu-redux -f together -m meta-llama/Llama-3-70b-chat-hf -p Llama-3-70b-chat-hf -s 8
# bash zero_eval_api.sh -d mmlu-redux -f together -m meta-llama/Llama-3-8B-chat-hf -p Llama-3-8B-chat-hf -s 8

bash zero_eval_api.sh -d mmlu-redux -f google -m google/gemini-1.5-pro-exp-0827 -p gemini-1.5-pro-exp-0827 -s 1