summaryrefslogtreecommitdiff
path: root/code_eval/test-humaneval-ckpt-list.sh
diff options
context:
space:
mode:
authorYuren Hao <yurenh2@timan108.cs.illinois.edu>2025-09-04 22:16:22 -0500
committerYuren Hao <yurenh2@timan108.cs.illinois.edu>2025-09-04 22:16:22 -0500
commitfc6d57ffb8d5ddb5820fcc00b5491a585c259ebc (patch)
treee9841f93a353e2107225cfc721d1ce57c0e594dc /code_eval/test-humaneval-ckpt-list.sh
Initial commit
Diffstat (limited to 'code_eval/test-humaneval-ckpt-list.sh')
-rw-r--r--code_eval/test-humaneval-ckpt-list.sh28
1 files changed, 28 insertions, 0 deletions
diff --git a/code_eval/test-humaneval-ckpt-list.sh b/code_eval/test-humaneval-ckpt-list.sh
new file mode 100644
index 0000000..03514bd
--- /dev/null
+++ b/code_eval/test-humaneval-ckpt-list.sh
@@ -0,0 +1,28 @@
+export VLLM_ENABLE_V1_MULTIPROCESSING=0
+
+# Declare an associative array to store model mappings
+# TODO:Replace with actual model name and path
+declare -A model_dict=(
+ ["model_name_1"]="/path/to/model1"
+ ["model_name_2"]="/path/to/model2"
+)
+
+# 外层循环:遍历model_dict
+for exp_name in "${!model_dict[@]}"; do
+ model="${model_dict[$exp_name]}"
+ echo "Evaluating model: $model"
+
+ python OpenCodeEval/main.py --model_name $model \
+ --task "HumanEval" \
+ --save "test/output_humaneval_${exp_name}" \
+ --num_gpus 1 \
+ --batch_size 164 \
+ --max_tokens 4096 \
+ --temperature 0.1 \
+ --seed 0 \
+ --prompt_type "Completion" \
+ --model_type "Chat" \
+ --prompt_prefix $'Please provide a self-contained Python script that solves the following problem in a markdown code block:\n```python\n' \
+ --prompt_suffix $'\n```\n' \
+
+done \ No newline at end of file