blob: 1e61ca683a2c49694046112ba7570e62206d03e4 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
|
## Quick Start
1. Download benchmark datasets:
```bash
cd OpenCodeEval/data
bash dataset.sh
```
2. Install dependencies:
```bash
pip install -e .
```
3. **Configure Evaluation Scripts**
- Replace placeholders in the evaluation scripts with the actual model name and path.
- Adjust any other necessary settings (e.g., evaluation parameters, output paths) to suit your requirements.
4. Execute the evaluation script for your desired benchmark. For example, to evaluate using the `test-humaneval-ckpt-list.sh` script:
```bash
bash test-humaneval-ckpt-list.sh
```
> **Note**: Ensure that all configurations are correctly set before running the script to avoid errors.
|