mirror of
https://github.com/All-Hands-AI/OpenHands.git
synced 2026-01-09 14:57:59 -05:00
* add draft dockerfile for build all * add rsync for build * add all-in-one docker * update prepare scripts * Update swe_env_box.py * Add swe_entry.sh (buggy now) * Parse the test command in swe_entry.sh * Update README for instance eval in sandbox * revert specialized config * replace run_as_devin as an init arg * set container & run_as_root via args * update swe entry script * update env * remove mounting * allow error after swe_entry * update swe_env_box * move file * update gitignore * get swe_env_box a working demo * support faking user response & provide sandox ahead of time; also return state for controller * tweak main to support adding controller kwargs * add module * initialize plugin for provided sandbox * add pip cache to plugin & fix jupyter kernel waiting * better print Observation output * add run infer scripts * update readme * add utility for getting diff patch * use get_diff_patch in infer * update readme * support cost tracking for codeact * add swe agent edit hack * disable color in git diff * fix git diff cmd * fix state return * support limit eval * increase t imeout and export pip cache * add eval limit config * return state when hit turn limit * save log to file; allow agent to give up * run eval with max 50 turns * add outputs to gitignore * save swe_instance & instruction * add uuid to swebench * add streamlit dep * fix save series * fix the issue where session id might be duplicated * allow setting temperature for llm (use 0 for eval) * Get report from agent running log * support evaluating task success right after inference. * remove extra log * comment out prompt for baseline * add visualizer for eval * use plaintext for instruction * reduce timeout for all; only increase timeout for init * reduce timeout for all; only increase timeout for init * ignore sid for swe env * close sandbox in each eval loop * update visualizer instruction * increase max chars * add finish action to history too * show test result in metrics * add sidebars for visualizer * also visualize swe_instance * cleanup browser when agent controller finish runinng * do not mount workspace for swe-eval to avoid accidentally overwrite files * Revert "do not mount workspace for swe-eval to avoid accidentally overwrite files" This reverts commit8ef7739054. * Revert "Revert "do not mount workspace for swe-eval to avoid accidentally overwrite files"" This reverts commit016cfbb9f0. * run jupyter command via copy to, instead of cp to mount * only print mixin output when failed * change ssh box logging * add visualizer for pass rate * add instance id to sandbox name * only remove container we created * use opendevin logger in main * support multi-processing infer * add back metadata, support keyboard interrupt * remove container with startswith * make pbar behave correctly * update instruction w/ multi-processing * show resolved rate by repo * rename tmp dir name * attempt to fix racing for copy to ssh_box * fix script * bump swe-bench-all version * fix ipython with self-contained commands * add jupyter demo to swe_env_box * make resolved count two column * increase height * do not add glob to url params * analyze obs length * print instance id prior to removal handler * add gold patch in visualizer * fix interactive git by adding a git --no-pager as alias * increase max_char to 10k to cover 98% of swe-bench obs cases * allow parsing note * prompt v2 * add iteration reminder * adjust user response * adjust order * fix return eval * fix typo * add reminder before logging * remove other resolve rate * re adjust to new folder structure * support adding eval note * fix eval note path * make sure first log of each instance is printed * add eval note * fix the display for visualizer * tweak visualizer for better git patch reading * exclude empty patch * add retry mechanism for swe_env_box start * fix ssh timeout issue * add stat field for apply test patch success * add visualization for fine-grained report * attempt to support monologue agent by constraining it to single thread * also log error msg when stopeed * save error as well * override WORKSPACE_MOUNT_PATH and WORKSPACE_BASE for monologue to work in mp * add retry mechanism for sshbox * remove retry for swe env box * try to handle loop state stopped * Add get report scripts * Add script to convert agent output to swe-bench format * Merge fine grained report for visualizer * Update eval readme * Update README.md * Add CodeAct gpt4-1106 output and eval logs on swe-bench-lite * Update the script to get model report * Update get_model_report.sh * Update get_agent_report.sh * Update report merge script * Add agent output conversion script * Update swe_lite_env_setup.sh * Add example swe-bench output files * Update eval readme * Remove redundant scripts * set iteration count down to false by default * fix: Issue where CodeAct agent was trying to log cost on local llm and throwing Undefined Model execption out of litellm (#1666) * fix: Issue where CodeAct agent was trying to log cost on local llm and throwing Undefined Model execption out of litellm * Review Feedback * Missing None Check * Review feedback and improved error handling --------- Co-authored-by: Robert Brennan <accounts@rbren.io> * fix prepare_swe_util scripts * update builder images * update setup script * remove swe-bench build workflow * update lock * remove experiments since they are moved to hf * remove visualizer (since it is moved to hf repo) * simply jupyter execution via heredoc * update ssh_box * add initial docker readme * add pkg-config as dependency * add script for swe_bench all-in-one docker * add rsync to builder * rename var * update commit * update readme * update lock * support specify timeout for long running tasks * fix path * separate building of all deps and files * support returning states at the end of controller * remove return None * support specify timeout for long running tasks * add timeout for all existing sandbox impl * fix swe_env_box for new codebase * update llm config in config.py * support pass sandbox in * remove force set * update eval script * fix issue of overriding final state * change default eval output to hf demo * change default eval output to hf demo * fix config * only close it when it is NOT external sandbox * add scripts * tweak config * only put in hostory when state has history attr * fix agent controller on the case of run out interaction budget * always assume state is always not none * remove print of final state * catch all exception when cannot compute completion cost * Update README.md * save source into json * fix path * update docker path * return the final state on close * merge AgentState with State * fix integration test * merge AgentState with State * fix integration test * add ChangeAgentStateAction to history in attempt to fix integration * add back set agent state * update tests * update tests * move scripts for setup * update script and readme for infer * do not reset logger when n processes == 1 * update eval_infer scripts and readme * simplify readme * copy over dir after eval * copy over dir after eval * directly return get state * update lock * fix output saving of infer * replace print with logger * update eval_infer script * add back the missing .close * increase timeout * copy all swe_bench_format file * attempt to fix output parsing * log git commit id as metadata * fix eval script * update lock * update unit tests * fix argparser unit test * fix lock * the deps are now lightweight enough to be incude in make build * add spaces for tests * add eval outputs to gitignore * remove git submodule * readme * tweak git email * update upload instruction * bump codeact version for eval --------- Co-authored-by: Bowen Li <libowen.ne@gmail.com> Co-authored-by: huybery <huybery@gmail.com> Co-authored-by: Bart Shappee <bshappee@gmail.com> Co-authored-by: Robert Brennan <accounts@rbren.io>
151 lines
5.8 KiB
Markdown
151 lines
5.8 KiB
Markdown
# SWE-Bench Evaluation with OpenDevin SWE-Bench Docker Image
|
|
|
|
|
|
This folder contains evaluation harness we built on top of the original [SWE-Bench benchmark](https://www.swebench.com/) ([paper](https://arxiv.org/abs/2310.06770)). We create [a fork of SWE-Bench](https://github.com/OpenDevin/OD-SWE-bench.git) mostly build on top of [the original repo](https://github.com/princeton-nlp/SWE-bench) and [containerized](#opendevin-swe-bench-docker-image) it for easy evaluation.
|
|
|
|
## OpenDevin SWE-Bench Docker Image
|
|
|
|
In [OpenDevin-SWE-Bench fork](https://github.com/OpenDevin/OD-SWE-bench.git) (mostly from [original repo](https://github.com/princeton-nlp/SWE-bench) with some fixes), we try to pre-build the **testbed** (i.e., code of the repository we want the agent to edit) AND the **conda environment**, so that in evaluation (inference) time, we can directly leverage existing environments for effecienct evaluation.
|
|
|
|
**We pack everything you need for SWE-Bench evaluation into one, gigantic, docker image.** To use it:
|
|
|
|
```bash
|
|
docker pull ghcr.io/opendevin/eval-swe-bench:full-v1.0
|
|
```
|
|
|
|
The Docker image contains several important directories:
|
|
- `/swe_util/OD-SWE-bench`: root directory for the OD-SWE-bench repository
|
|
- `/swe_util/eval_data`: director to eval data
|
|
- `/swe_util/eval_data/eval_logs/`: evaluation logs
|
|
- `/swe_util/eval_data/eval_temp/`: temporary folder for the evaluation process
|
|
- `/swe_util/eval_data/instances/`: swe-bench raw instances
|
|
- `/swe_util/eval_data/outputs/`: model or agent outputs
|
|
- `/swe_util/eval_data/testbed_logs/`: logs for testbed building
|
|
- `/swe_util/eval_data/testbeds/`: directory for all testbeds
|
|
- `/swe_util/miniforge3/`: directory for miniforge3
|
|
|
|
To reproduce how we pack the image, check [this doc](./BUILD_TESTBED_AND_ENV.md).
|
|
|
|
NOTE: We only support SWE-Bench lite for now. But modifying our existing scripts for full SWE-Bench should be quite straight forward.
|
|
|
|
## Test if your environment works
|
|
|
|
```bash
|
|
python3 evaluation/swe_bench/swe_env_box.py
|
|
```
|
|
|
|
If you get to the interactive shell successfully, it means success!
|
|
|
|
## Configure your LLM
|
|
|
|
Create a `config.toml` file if not exists at the root of workspace.
|
|
|
|
Add the following configurations:
|
|
|
|
```toml
|
|
[core]
|
|
max_iterations = 100
|
|
cache_dir = "/tmp/cache"
|
|
sandbox_container_image = "ghcr.io/opendevin/sandbox:latest"
|
|
sandbox_type = "ssh"
|
|
use_host_network = true
|
|
ssh_hostname = "localhost"
|
|
sandbox_timeout = 120
|
|
# eval specific
|
|
run_as_devin = false
|
|
|
|
# TODO: Change these to the model you want to evaluate
|
|
[eval_gpt4_1106_preview]
|
|
model = "gpt-4-1106-preview"
|
|
api_key = "XXX"
|
|
temperature = 0.0
|
|
|
|
[eval_some_openai_compatible_model]
|
|
model = "openai/MODEL_NAME"
|
|
base_url = "https://OPENAI_COMPATIBLE_URL/v1"
|
|
api_key = "XXX"
|
|
temperature = 0.0
|
|
```
|
|
|
|
## Run Inference on SWE-Bench Instances
|
|
|
|
```bash
|
|
./evaluation/swe_bench/scripts/run_infer.sh eval_gpt4_1106_preview
|
|
```
|
|
|
|
You can replace `eval_gpt4_1106_preview` with any model you setted up in `config.toml`.
|
|
|
|
|
|
## Evaluate Generated Patches
|
|
|
|
After running the inference described in the previous section, you will obtain a `output.jsonl` (by default it will save to `evaluation/evaluation_outputs`). Then you can run this one line script to evaluate generated patches, and produce a fine-grained report:
|
|
|
|
If you want to evaluate existing results, you should first run this to clone existing outputs
|
|
|
|
```bash
|
|
git clone https://huggingface.co/spaces/OpenDevin/evaluation evaluation/evaluation_outputs
|
|
```
|
|
|
|
Then you can run the following:
|
|
```bash
|
|
# ./evaluation/swe_bench/scripts/eval_infer.sh $YOUR_OUTPUT_JSONL
|
|
# For example:
|
|
./evaluation/swe_bench/scripts/eval_infer.sh evaluation/evaluation_outputs/outputs/swe_bench/CodeActAgent/gpt-4-1106-preview_maxiter_50_N_v1.0/output.jsonl
|
|
```
|
|
|
|
The final results will be saved to `evaluation/evaluation_outputs/outputs/swe_bench/CodeActAgent/gpt-4-1106-preview_maxiter_50_N_v1.0/output.merged.jsonl`.
|
|
|
|
It will contains an additional field `fine_grained_report` (see example below) compared to the `output.jsonl` from the previous inference stage.
|
|
|
|
```json
|
|
"fine_grained_report": {
|
|
"gold_tests": {
|
|
"FAIL_TO_PASS": "[\"tests/test_ext_viewcode.py::test_viewcode_epub_default\"]",
|
|
"PASS_TO_PASS": "[\"tests/test_ext_viewcode.py::test_viewcode_epub_enabled\", \"tests/test_ext_viewcode.py::test_linkcode\", \"tests/test_ext_viewcode.py::test_local_source_files\"]"
|
|
},
|
|
"generated": true,
|
|
"with_logs": true,
|
|
"applied": true,
|
|
"test_errored": false,
|
|
"test_timeout": false,
|
|
"resolved": true,
|
|
"log_parse": {
|
|
"tests/test_ext_viewcode.py::test_viewcode_epub_default": "PASSED",
|
|
"tests/test_ext_viewcode.py::test_viewcode_epub_enabled": "PASSED",
|
|
"tests/test_ext_viewcode.py::test_linkcode": "PASSED",
|
|
"tests/test_ext_viewcode.py::test_local_source_files": "PASSED",
|
|
"tests/test_ext_viewcode.py::test_viewcode": "FAILED"
|
|
},
|
|
"eval_report": {
|
|
"FAIL_TO_PASS": {
|
|
"success": [
|
|
"tests/test_ext_viewcode.py::test_viewcode_epub_default"
|
|
],
|
|
"failure": []
|
|
},
|
|
"PASS_TO_PASS": {
|
|
"success": [
|
|
"tests/test_ext_viewcode.py::test_viewcode_epub_enabled",
|
|
"tests/test_ext_viewcode.py::test_linkcode",
|
|
"tests/test_ext_viewcode.py::test_local_source_files"
|
|
],
|
|
"failure": []
|
|
},
|
|
"FAIL_TO_FAIL": {
|
|
"success": [],
|
|
"failure": []
|
|
},
|
|
"PASS_TO_FAIL": {
|
|
"success": [],
|
|
"failure": []
|
|
}
|
|
}
|
|
}
|
|
```
|
|
|
|
Please refer to [EVAL_PATCH.md](./EVAL_PATCH.md) if you want to learn more about how to evaluate patches that are already generated (e.g., not by OpenDevin).
|
|
|
|
## Submit your evaluation results
|
|
|
|
You can start your own fork of [our huggingface evaluation outputs](https://huggingface.co/spaces/OpenDevin/evaluation) and submit a PR of your evaluation results following the guide [here](https://huggingface.co/docs/hub/en/repositories-pull-requests-discussions#pull-requests-and-discussions).
|