summaryrefslogtreecommitdiff
path: root/code_eval/README.md
diff options
context:
space:
mode:
author= <=>2025-06-04 11:49:37 +0800
committer= <=>2025-06-04 11:49:37 +0800
commit947d9dfdf16ae37109898111a5caacae7377b96d (patch)
treeff4e884020fb7d968a6192106f370b215647f569 /code_eval/README.md
parent5e163b529a78d528b745b8b57ba794b7b2bba97a (diff)
update code and kk eval
Diffstat (limited to 'code_eval/README.md')
-rw-r--r--code_eval/README.md27
1 files changed, 27 insertions, 0 deletions
diff --git a/code_eval/README.md b/code_eval/README.md
new file mode 100644
index 0000000..32daaa9
--- /dev/null
+++ b/code_eval/README.md
@@ -0,0 +1,27 @@
+## Quick Start
+
+1. Download benchmark datasets:
+
+```bash
+cd OpenCodeEval/data
+bash dataset.sh
+```
+
+2. Install dependencies:
+
+```bash
+pip install -e .
+```
+
+3. **Configure Evaluation Scripts**
+ - Replace placeholders in the evaluation scripts with the actual model name and path.
+ - Adjust any other necessary settings (e.g., evaluation parameters, output paths) to suit your requirements.
+
+4. Execute the evaluation script for your desired benchmark. For example, to evaluate using the `test-humaneval-ckpt-list.sh` script:
+
+Such as:
+```bash
+bash test-humaneval-ckpt-list.sh
+```
+
+ > **Note**: Ensure that all configurations are correctly set before running the script to avoid errors.