Dataset Viewer
Auto-converted to Parquet Duplicate
suite_config
dict
split
stringclasses
1 value
results
listlengths
1
11
submission
dict
{ "name": "asta-bench", "version": "1.0.0", "splits": [ { "name": "validation", "tasks": [ { "name": "ArxivDIGESTables_Clean_validation", "path": "astabench/arxivdigestables_validation", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_validation", "path": "astabench/sqa_dev", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_validation", "path": "astabench/litqa2_validation", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "PaperFindingBench_validation", "path": "astabench/paper_finder_validation", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_validation", "path": "astabench/paper_finder_litqa2_validation", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "DiscoveryBench_validation", "path": "astabench/discoverybench_validation", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_validation", "path": "astabench/core_bench_validation", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_validation", "path": "astabench/ds1000_validation", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_validation", "path": "astabench/e2e_discovery_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_validation", "path": "astabench/e2e_discovery_hard_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_validation", "path": "astabench/super_validation", "primary_metric": "entrypoint/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_validation", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_validation", "weight": 0.5 } ] }, { "name": "test", "tasks": [ { "name": "PaperFindingBench_test", "path": "astabench/paper_finder_test", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_test", "path": "astabench/paper_finder_litqa2_test", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_test", "path": "astabench/sqa_test", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "ArxivDIGESTables_Clean_test", "path": "astabench/arxivdigestables_test", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_test", "path": "astabench/litqa2_test", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "DiscoveryBench_test", "path": "astabench/discoverybench_test", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_test", "path": "astabench/core_bench_test", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_test", "path": "astabench/ds1000_test", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_test", "path": "astabench/e2e_discovery_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_test", "path": "astabench/e2e_discovery_hard_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_test", "path": "astabench/super_test", "primary_metric": "output_match/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_test", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_test", "weight": 0.5 } ] } ] }
validation
[ { "task_name": "PaperFindingBench_validation", "eval_spec": { "solver": "astabench/solvers/search/paper_finder.py@ai2i_paper_finder", "solver_args": "{\"base_url\": \"http://35.247.123.160:8000\"}", "model": "openai/gpt-4o-mini", "model_args": "{}", "task_args": "{\"with_search...
{ "submit_time": "2025-08-07T19:06:21.862000", "username": "Ai2", "agent_name": "Asta Paper Finder", "agent_description": null, "agent_url": null, "logs_url": null, "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/aryeh_tiktinsky_ai2_Asta_Paper_Finder_2025-08-07T19-06-21", "summary_url": null, "openness": "Open source & closed weights", "tool_usage": "Custom interface" }
{ "name": "asta-bench", "version": "1.0.0", "splits": [ { "name": "validation", "tasks": [ { "name": "ArxivDIGESTables_Clean_validation", "path": "astabench/arxivdigestables_validation", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_validation", "path": "astabench/sqa_dev", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_validation", "path": "astabench/litqa2_validation", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "PaperFindingBench_validation", "path": "astabench/paper_finder_validation", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_validation", "path": "astabench/paper_finder_litqa2_validation", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "DiscoveryBench_validation", "path": "astabench/discoverybench_validation", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_validation", "path": "astabench/core_bench_validation", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_validation", "path": "astabench/ds1000_validation", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_validation", "path": "astabench/e2e_discovery_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_validation", "path": "astabench/e2e_discovery_hard_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_validation", "path": "astabench/super_validation", "primary_metric": "entrypoint/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_validation", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_validation", "weight": 0.5 } ] }, { "name": "test", "tasks": [ { "name": "PaperFindingBench_test", "path": "astabench/paper_finder_test", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_test", "path": "astabench/paper_finder_litqa2_test", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_test", "path": "astabench/sqa_test", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "ArxivDIGESTables_Clean_test", "path": "astabench/arxivdigestables_test", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_test", "path": "astabench/litqa2_test", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "DiscoveryBench_test", "path": "astabench/discoverybench_test", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_test", "path": "astabench/core_bench_test", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_test", "path": "astabench/ds1000_test", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_test", "path": "astabench/e2e_discovery_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_test", "path": "astabench/e2e_discovery_hard_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_test", "path": "astabench/super_test", "primary_metric": "output_match/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_test", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_test", "weight": 0.5 } ] } ] }
validation
[ { "task_name": "PaperFindingBench_validation", "eval_spec": { "solver": "/home/aryeht/PycharmProjects/asta-bench/astabench/solvers/search/youcom_search.py@youcom_solver", "solver_args": "{}", "model": "openai/gpt-4o-mini", "model_args": "{}", "task_args": "{\"with_search_tools\...
{ "submit_time": "2025-08-07T18:54:02.182000", "username": "Ai2", "agent_name": "You.com Search API", "agent_description": null, "agent_url": null, "logs_url": null, "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/aryeh_tiktinsky_ai2_You.com_Search_API_2025-08-07T18-54-02", "summary_url": null, "openness": "Closed source & API available", "tool_usage": "Fully custom" }
{ "name": "asta-bench", "version": "1.0.0", "splits": [ { "name": "validation", "tasks": [ { "name": "ArxivDIGESTables_Clean_validation", "path": "astabench/arxivdigestables_validation", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_validation", "path": "astabench/sqa_dev", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_validation", "path": "astabench/litqa2_validation", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "PaperFindingBench_validation", "path": "astabench/paper_finder_validation", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_validation", "path": "astabench/paper_finder_litqa2_validation", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "DiscoveryBench_validation", "path": "astabench/discoverybench_validation", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_validation", "path": "astabench/core_bench_validation", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_validation", "path": "astabench/ds1000_validation", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_validation", "path": "astabench/e2e_discovery_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_validation", "path": "astabench/e2e_discovery_hard_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_validation", "path": "astabench/super_validation", "primary_metric": "entrypoint/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_validation", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_validation", "weight": 0.5 } ] }, { "name": "test", "tasks": [ { "name": "PaperFindingBench_test", "path": "astabench/paper_finder_test", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_test", "path": "astabench/paper_finder_litqa2_test", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_test", "path": "astabench/sqa_test", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "ArxivDIGESTables_Clean_test", "path": "astabench/arxivdigestables_test", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_test", "path": "astabench/litqa2_test", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "DiscoveryBench_test", "path": "astabench/discoverybench_test", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_test", "path": "astabench/core_bench_test", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_test", "path": "astabench/ds1000_test", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_test", "path": "astabench/e2e_discovery_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_test", "path": "astabench/e2e_discovery_hard_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_test", "path": "astabench/super_test", "primary_metric": "output_match/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_test", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_test", "weight": 0.5 } ] } ] }
validation
[ { "task_name": "SUPER_Expert_validation", "eval_spec": { "solver": "astabench/solvers/code_agent/agent.py@code_agent", "solver_args": "{\"json_output\": 1, \"max_context_tokens\": 1000000, \"max_tries\": 200}", "model": "openai/gpt-5-mini-2025-08-07", "model_args": "{}", "task_...
{ "submit_time": "2025-08-14T19:05:15.098000", "username": "Ai2", "agent_name": "Asta Code", "agent_description": null, "agent_url": "https://github.com/allenai/asta-bench", "logs_url": null, "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/miked-ai_Asta_Code_GPT-5-mini_2025-08-14T19-05-15", "summary_url": null, "openness": "Open source & closed weights", "tool_usage": "Custom interface" }
{ "name": "asta-bench", "version": "1.0.0", "splits": [ { "name": "validation", "tasks": [ { "name": "ArxivDIGESTables_Clean_validation", "path": "astabench/arxivdigestables_validation", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_validation", "path": "astabench/sqa_dev", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_validation", "path": "astabench/litqa2_validation", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "PaperFindingBench_validation", "path": "astabench/paper_finder_validation", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_validation", "path": "astabench/paper_finder_litqa2_validation", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "DiscoveryBench_validation", "path": "astabench/discoverybench_validation", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_validation", "path": "astabench/core_bench_validation", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_validation", "path": "astabench/ds1000_validation", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_validation", "path": "astabench/e2e_discovery_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_validation", "path": "astabench/e2e_discovery_hard_validation", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_validation", "path": "astabench/super_validation", "primary_metric": "entrypoint/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_validation", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_validation", "weight": 0.5 } ] }, { "name": "test", "tasks": [ { "name": "PaperFindingBench_test", "path": "astabench/paper_finder_test", "primary_metric": "score_paper_finder/adjusted_f1_micro_avg", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_Search_test", "path": "astabench/paper_finder_litqa2_test", "primary_metric": "score_paper_finder/recall_at_30", "tags": [ "lit" ] }, { "name": "ScholarQA_CS2_test", "path": "astabench/sqa_test", "primary_metric": "global_avg/mean", "tags": [ "lit" ] }, { "name": "ArxivDIGESTables_Clean_test", "path": "astabench/arxivdigestables_test", "primary_metric": "score_tables/mean", "tags": [ "lit" ] }, { "name": "LitQA2_FullText_test", "path": "astabench/litqa2_test", "primary_metric": "is_correct/accuracy", "tags": [ "lit" ] }, { "name": "DiscoveryBench_test", "path": "astabench/discoverybench_test", "primary_metric": "score_discoverybench/mean", "tags": [ "data" ] }, { "name": "CORE_Bench_Hard_test", "path": "astabench/core_bench_test", "primary_metric": "score_with_stderr/accuracy", "tags": [ "code" ] }, { "name": "DS_1000_test", "path": "astabench/ds1000_test", "primary_metric": "ds1000_scorer/accuracy", "tags": [ "code" ] }, { "name": "E2E_Bench_test", "path": "astabench/e2e_discovery_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "E2E_Bench_Hard_test", "path": "astabench/e2e_discovery_hard_test", "primary_metric": "score_rubric/accuracy", "tags": [ "discovery" ] }, { "name": "SUPER_Expert_test", "path": "astabench/super_test", "primary_metric": "output_match/mean", "tags": [ "code" ] } ], "macro_average_weight_adjustments": [ { "tag": "lit", "task": "LitQA2_FullText_test", "weight": 0.5 }, { "tag": "lit", "task": "LitQA2_FullText_Search_test", "weight": 0.5 } ] } ] }
validation
[ { "task_name": "SUPER_Expert_validation", "eval_spec": { "solver": "astabench/solvers/code_agent/agent.py@code_agent", "solver_args": "{\"json_output\": 1, \"max_context_tokens\": 1000000, \"max_tries\": 200}", "model": "openai/gpt-5-2025-08-07", "model_args": "{}", "task_args"...
{ "submit_time": "2025-08-14T19:06:15.866000", "username": "Ai2", "agent_name": "Asta Code", "agent_description": null, "agent_url": "https://github.com/allenai/asta-bench", "logs_url": null, "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/miked-ai_Asta_Code_GPT-5_2025-08-14T19-06-15", "summary_url": null, "openness": "Open source & closed weights", "tool_usage": "Custom interface" }
{"name":"asta-bench","version":"1.0.0","splits":[{"name":"validation","tasks":[{"name":"ArxivDIGESTa(...TRUNCATED)
validation
[{"task_name":"SUPER_Expert_validation","eval_spec":{"solver":"agent_baselines/solvers/react/basic_a(...TRUNCATED)
{"submit_time":"2025-08-26T06:45:15.968802Z","username":"Ai2","agent_name":"ReAct","agent_descriptio(...TRUNCATED)
{"name":"asta-bench","version":"1.0.0","splits":[{"name":"validation","tasks":[{"name":"ArxivDIGESTa(...TRUNCATED)
validation
[{"task_name":"E2E_Bench_Hard_validation","eval_spec":{"solver":"astabench/evals/e2e_discovery/solve(...TRUNCATED)
{"submit_time":"2025-07-10T18:12:16.960575Z","username":"Ai2","agent_name":"Asta CodeScientist","age(...TRUNCATED)
{"name":"asta-bench","version":"1.0.0","splits":[{"name":"validation","tasks":[{"name":"ArxivDIGESTa(...TRUNCATED)
validation
[{"task_name":"DiscoveryBench_validation","eval_spec":{"solver":"astabench/solvers/datavoyager/agent(...TRUNCATED)
{"submit_time":"2025-08-14T04:16:47.886330Z","username":"Ai2","agent_name":"Asta DataVoyager","agent(...TRUNCATED)
{"name":"asta-bench","version":"1.0.0","splits":[{"name":"validation","tasks":[{"name":"ArxivDIGESTa(...TRUNCATED)
validation
[{"task_name":"DiscoveryBench_validation","eval_spec":{"solver":"astabench/solvers/datavoyager/agent(...TRUNCATED)
{"submit_time":"2025-08-14T19:32:30.253858Z","username":"Ai2","agent_name":"Asta DataVoyager","agent(...TRUNCATED)
{"name":"asta-bench","version":"1.0.0","splits":[{"name":"validation","tasks":[{"name":"ArxivDIGESTa(...TRUNCATED)
validation
[{"task_name":"DiscoveryBench_validation","eval_spec":{"solver":"astabench/solvers/datavoyager/agent(...TRUNCATED)
{"submit_time":"2025-08-14T19:42:20.164747Z","username":"Ai2","agent_name":"Asta DataVoyager","agent(...TRUNCATED)
{"name":"asta-bench","version":"1.0.0","splits":[{"name":"validation","tasks":[{"name":"ArxivDIGESTa(...TRUNCATED)
validation
[{"task_name":"DiscoveryBench_validation","eval_spec":{"solver":"astabench/solvers/datavoyager/agent(...TRUNCATED)
{"submit_time":"2025-08-14T21:44:09.404096Z","username":"Ai2","agent_name":"Asta DataVoyager","agent(...TRUNCATED)
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
40

Spaces using allenai/asta-bench-results 2