diff options
| author | = <=> | 2025-06-04 11:49:37 +0800 |
|---|---|---|
| committer | = <=> | 2025-06-04 11:49:37 +0800 |
| commit | 947d9dfdf16ae37109898111a5caacae7377b96d (patch) | |
| tree | ff4e884020fb7d968a6192106f370b215647f569 /code_eval/README.md | |
| parent | 5e163b529a78d528b745b8b57ba794b7b2bba97a (diff) | |
update code and kk eval
Diffstat (limited to 'code_eval/README.md')
| -rw-r--r-- | code_eval/README.md | 27 |
1 files changed, 27 insertions, 0 deletions
diff --git a/code_eval/README.md b/code_eval/README.md new file mode 100644 index 0000000..32daaa9 --- /dev/null +++ b/code_eval/README.md @@ -0,0 +1,27 @@ +## Quick Start + +1. Download benchmark datasets: + +```bash +cd OpenCodeEval/data +bash dataset.sh +``` + +2. Install dependencies: + +```bash +pip install -e . +``` + +3. **Configure Evaluation Scripts** + - Replace placeholders in the evaluation scripts with the actual model name and path. + - Adjust any other necessary settings (e.g., evaluation parameters, output paths) to suit your requirements. + +4. Execute the evaluation script for your desired benchmark. For example, to evaluate using the `test-humaneval-ckpt-list.sh` script: + +Such as: +```bash +bash test-humaneval-ckpt-list.sh +``` + + > **Note**: Ensure that all configurations are correctly set before running the script to avoid errors. |
