Dataset Viewer
Auto-converted to Parquet
benchmark
string
check
string
score
int64
reason
string
SWE-Bench-Lancer
I.d.1
1
As discussed in Section 1 of the paper, the benchmark uses a set of test cases that are verified for correctness and quality by human experts.
SWE-Bench-Lancer
I.d.2
0
The benchmark does not use objective metrics to measure the quality of test cases.
SWE-Bench-Lancer
I.f.2
1
As discussed in Section 1, the end-to-end testing is designed to simulate the entire user workflow.
SWE-Bench-Lancer
I.f.3
0
The test cases use hard-coded timeouts, which may lead to non-deterministic results if the system is slow or unresponsive.
SWE-Bench-Lancer
II.1
1
The package dependencies are specified in the repository of each task.
SWE-Bench-Lancer
II.2
1
The benchmark does not require any external APIs.
SWE-Bench-Lancer
II.3
1
The benchmark does not require any external APIs.
SWE-Bench-Lancer
II.4
1
The benchmark uses docker containers to isolate the environment, and the state is cleared between runs.
SWE-Bench-Lancer
II.5
0
The agent can access the file system where the test cases are stored, which may lead to the agent accessing the ground truth information.
SWE-Bench-Lancer
II.6
1
The environment setup is static and does not change over time.
SWE-Bench-Lancer
II.7
1
The ground-truth test cases are taken from GitHub repositories, which are verified by expert developers.
SWE-Bench-Lancer
II.8
1
Each task represents a real-world software issue with a corresponding patch, which are solvable by the agent.
SWE-Bench-Lancer
II.9
1
The benchmark uses existing patches as ground truth, which can be considered as an Oracle solver.
SWE-Bench-Lancer
II.10
0
The benchmark does not handle the isolation between the agent and test cases properly. The test cases are stored not only in a file system that the agent can access, but also in a ZIP file that agent can read the directory structure and update files.
SWE-Bench-Lancer
III.1
1
The benchmark is open-sourced and available on GitHub.
SWE-Bench-Lancer
III.2
1
The benchmark provides an open-source evaluation harness for users.
SWE-Bench-Lancer
III.3
1
The benchmark maintains a private test set.
SWE-Bench-Lancer
III.4
0
The report does not discuss any measures or plans for consistent update.
SWE-Bench-Lancer
III.5
1
Such a relationship is clearly stated in Section 2 of the paper.
SWE-Bench-Lancer
III.6
1
As shown in Section 3, the benchmark is designed to evaluate the LLM model.
SWE-Bench-Lancer
III.7
1
The benchmark uses end-to-end testing to mitigate grader hacking.
SWE-Bench-Lancer
III.8
1
The benchmark discusses the potential impact of grader hacking in Section 1 and Appendix A.7.
SWE-Bench-Lancer
III.9
0
The benchmark does not include any quantitative analysis to assess the impact of grader hacking.
SWE-Bench-Lancer
III.10
0
The benchmark does not report any metrics about statistical significance.
SWE-Bench-Lancer
III.11
0
The benchmark does not provide any guidance on interpreting results with eval flaws.
SWE-Bench-Lancer
III.12
0
The benchmark does not report results of non-AI baselines.
SWE-Bench-Lancer
III.13
0
The benchmark does not report results of trivial agents.
Bird-Bench
I.d.1
1
As discussed in Section 3.4 of the paper, the validity of the database is verified by executing the ground-truth query.
Bird-Bench
I.d.2
0
The paper does not use objective metrics to measure the usefulness and completeness of the database or ground-truth queries.
Bird-Bench
I.f.2
0
The paper does not provide any information about the coverage of the database or ground-truth queries.
Bird-Bench
I.f.3
1
Executing SQL queries on a database is deterministic, and the paper does not mention any non-deterministic behavior.
Bird-Bench
II.1
1
The task instruction in Figure 9 speficies the SQL language is SQLite.
Bird-Bench
II.2
1
No external API is required for the evaluation of the benchmark.
Bird-Bench
II.3
1
No external API is required for the evaluation of the benchmark.
Bird-Bench
II.4
0
Databse file is neither opened in a read-only mode nor re-initialized between runs. This may lead to unexpected data manipulation by the agent.
Bird-Bench
II.5
1
Agent cannot access the host file system.
Bird-Bench
II.6
1
The environment setup is static and does not change over time.
Bird-Bench
II.7
0
As discussed in Section 3.4 of the paper, the correctness of the query is not fully verified, especially for the SQL queries that two annotators reach a concensus on.
Bird-Bench
II.8
0
The ambiguity of the SQL queries is not fully verified.
Bird-Bench
II.9
0
The Benchmark does not include an Oracle solver that can automatically solve all text-to-SQL tasks.
Bird-Bench
II.10
1
No vulnerabilities are found in the implementation of the benchmark.
Bird-Bench
III.1
1
The benchmark is open-sourced and available on GitHub.
Bird-Bench
III.2
1
The benchmark provides an open-source evaluation harness for users.
Bird-Bench
III.3
1
The benchmark has a private test set.
Bird-Bench
III.4
0
The benchmark does not discuss any measures or plans for consistent update.
Bird-Bench
III.5
1
It is clearly discussed in Section 2 of the paper.
Bird-Bench
III.6
1
It is clearly discussed in Section 2 of the paper.
Bird-Bench
III.7
0
No efforts are made to mitigate errors when both annotators make the same mistake.
Bird-Bench
III.8
0
The paper does not discuss the potential impact of annotation errors.
Bird-Bench
III.9
0
The paper does not analyze the quantitative impact of annotation errors.
Bird-Bench
III.10
0
The paper does not report any metrics about statistical significance.
Bird-Bench
III.11
0
The paper does not provide any guidance on interpreting results with eval flaws.
Bird-Bench
III.12
1
The paper reports the results of human experts.
Bird-Bench
III.13
0
The paper does not report the results of any trivial agents.
Cy-Bench
I.h.1
1
The specific format required for the answer is provided in the task description.
Cy-Bench
I.h.2
1
The ground truth is complex enough to prevent trivial guessing.
Cy-Bench
II.1
1
Agents are granted access to using all tools. The versions of tools can be checked by the agent.
Cy-Bench
II.2
1
The benchmark does not require any external APIs.
Cy-Bench
II.3
1
The benchmark does not require any external APIs.
Cy-Bench
II.4
1
The benchmark uses docker containers to isolate the environment, and the state is cleared between runs.
Cy-Bench
II.5
1
The agent cannot directly access the container which contains the ground truth.
Cy-Bench
II.6
1
The environment setup is static and does not change over time.
Cy-Bench
II.7
1
As shown in Section 3.3 of the paper, the ground truth is verified by human.
Cy-Bench
II.8
1
As shown in Section 3.3 of the paper, each task is verified to be solvable.
Cy-Bench
II.9
1
As shown in Section 3.3 of the paper, the benchmark includes an Oracle solver that can automatically solve all tasks.
Cy-Bench
II.10
1
No vulnerabilities are found in the implementation of the benchmark.
Cy-Bench
III.1
1
The benchmark is open-sourced and available on GitHub.
Cy-Bench
III.2
1
The benchmark provides an open-source evaluation harness for users.
Cy-Bench
III.3
0
The benchmark does not contain measures to prevent data contamination.
Cy-Bench
III.4
0
The report does not discuss plans to consistently update tasks over time.
Cy-Bench
III.5
1
Such a relationship is clearly stated in Section 1 of the paper.
Cy-Bench
III.6
1
As shown in Section 1, the benchmark is designed to evaluate both agent frameworks and LLM models.
Cy-Bench
III.7
1
Annotation flaws are mitigated by developing verifiable tasks.
Cy-Bench
III.8
1
No unavoidable flaws are identified in the benchmark.
Cy-Bench
III.9
1
No unavoidable flaws are identified in the benchmark.
Cy-Bench
III.10
0
The report does not include any metrics about statistical significance.
Cy-Bench
III.11
1
No evaluation flaws are identified in the benchmark.
Cy-Bench
III.12
1
Human peerformance is reported in Section 5 of the paper.
Cy-Bench
III.13
0
The report does not report results of trivial agents.
SWE-Bench-Verified
I.d.1
1
Test cases are directly taken from GitHub repositories, and the paper does not mention any verification process.
SWE-Bench-Verified
I.d.2
0
The paper does not use objective metrics to measure quality of test cases.
SWE-Bench-Verified
II.1
1
The versions of package dependencies are specified in the repository.
SWE-Bench-Verified
II.2
1
The benchmark does not require any external APIs.
SWE-Bench-Verified
II.3
1
The benchmark does not require any external APIs.
SWE-Bench-Verified
II.4
1
The benchmark uses docker containers to isolate the environment, and the state is cleared between runs.
SWE-Bench-Verified
II.5
1
The agent cannot access the host file system, and the ground truth is not accessible to the agent.
SWE-Bench-Verified
II.6
1
The environment setup is static and does not change over time.
SWE-Bench-Verified
II.7
1
The ground-truth patches are taken from GitHub repositories, which is verified by expert developers.
SWE-Bench-Verified
II.8
1
Each task represents a real-world GitHub issue and a corresponding pull request, which are solvable by the agent.
SWE-Bench-Verified
II.9
1
Pull requests from GitHub are used as ground truth, which can be considered as an Oracle solver.
SWE-Bench-Verified
II.10
1
No vulnerabilities are found in the implementation of the benchmark, and the evaluation process is secure.
SWE-Bench-Verified
III.1
1
The benchmark is open-sourced and available on GitHub.
SWE-Bench-Verified
III.2
1
The benchmark provides an open-source evaluation harness for users.
SWE-Bench-Verified
III.3
0
The benchmark does not discuss measures to prevent data contamination.
SWE-Bench-Verified
III.4
0
The benchmark does not discuss plans to consistently update tasks over time.
SWE-Bench-Verified
III.5
1
Such a relationship is clearly stated in Section 2 of the paper.
SWE-Bench-Verified
III.6
1
The benchmark is designed to evaluate both the model and the agent framework, as discussed in Section 5 of the paper.
SWE-Bench-Verified
III.7
0
The benchmark does not discuss any efforts to prevent, identify, and correct flaws.
SWE-Bench-Verified
III.8
0
The benchmark does not discuss the potential impact of unavoidable flaws.
SWE-Bench-Verified
III.9
0
The benchmark does not include quantitative analysis to assess the impact of unavoidable flaws.
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
8