| suite_config
				 dict | split
				 stringclasses 1
				value | results
				 listlengths 1 11 | submission
				 dict | 
|---|---|---|---|
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "PaperFindingBench_validation",
    "eval_spec": {
      "solver": "astabench/solvers/search/paper_finder.py@ai2i_paper_finder",
      "solver_args": "{\"base_url\": \"http://35.247.123.160:8000\"}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{\"with_search_tools\": false}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "ca5b0ad"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_paper_finder/semantic_f1",
        "value": 0.3991389077954741
      },
      {
        "name": "score_paper_finder/specific_f1",
        "value": 0.2
      },
      {
        "name": "score_paper_finder/metadata_f1",
        "value": 0.4869762288477035
      },
      {
        "name": "score_paper_finder/adjusted_f1_micro_avg",
        "value": 0.3796132940146119
      },
      {
        "name": "score_paper_finder/stderr",
        "value": 0.029428016906952455
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3490,
            "output_tokens": 28,
            "total_tokens": 3518,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 168134,
            "output_tokens": 34180,
            "total_tokens": 202314,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3495,
            "output_tokens": 36,
            "total_tokens": 3531,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 339782,
            "output_tokens": 76846,
            "total_tokens": 416628,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3486,
            "output_tokens": 29,
            "total_tokens": 3515,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 560531,
            "output_tokens": 94799,
            "total_tokens": 655330,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3489,
            "output_tokens": 32,
            "total_tokens": 3521,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 421496,
            "output_tokens": 74287,
            "total_tokens": 495783,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3499,
            "output_tokens": 48,
            "total_tokens": 3547,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 608301,
            "output_tokens": 100811,
            "total_tokens": 709112,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3499,
            "output_tokens": 37,
            "total_tokens": 3536,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 715599,
            "output_tokens": 118139,
            "total_tokens": 833738,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3498,
            "output_tokens": 38,
            "total_tokens": 3536,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 481360,
            "output_tokens": 101699,
            "total_tokens": 583059,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3494,
            "output_tokens": 34,
            "total_tokens": 3528,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 674838,
            "output_tokens": 102341,
            "total_tokens": 777179,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3504,
            "output_tokens": 44,
            "total_tokens": 3548,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 443033,
            "output_tokens": 87187,
            "total_tokens": 530220,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3499,
            "output_tokens": 49,
            "total_tokens": 3548,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 136628,
            "output_tokens": 28219,
            "total_tokens": 164847,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3503,
            "output_tokens": 35,
            "total_tokens": 3538,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 277043,
            "output_tokens": 39367,
            "total_tokens": 316410,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3501,
            "output_tokens": 42,
            "total_tokens": 3543,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 823767,
            "output_tokens": 122200,
            "total_tokens": 945967,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3540,
            "output_tokens": 68,
            "total_tokens": 3608,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 767663,
            "output_tokens": 140144,
            "total_tokens": 907807,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3503,
            "output_tokens": 33,
            "total_tokens": 3536,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 168158,
            "output_tokens": 36183,
            "total_tokens": 204341,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 29,
            "total_tokens": 3517,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 263367,
            "output_tokens": 46650,
            "total_tokens": 310017,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3490,
            "output_tokens": 24,
            "total_tokens": 3514,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 253486,
            "output_tokens": 55352,
            "total_tokens": 308838,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3511,
            "output_tokens": 36,
            "total_tokens": 3547,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 223203,
            "output_tokens": 41457,
            "total_tokens": 264660,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 23,
            "total_tokens": 3511,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 156487,
            "output_tokens": 31058,
            "total_tokens": 187545,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3525,
            "output_tokens": 69,
            "total_tokens": 3594,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 575980,
            "output_tokens": 96635,
            "total_tokens": 672615,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3497,
            "output_tokens": 38,
            "total_tokens": 3535,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 751168,
            "output_tokens": 133345,
            "total_tokens": 884513,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3491,
            "output_tokens": 32,
            "total_tokens": 3523,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 564152,
            "output_tokens": 95719,
            "total_tokens": 659871,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3504,
            "output_tokens": 48,
            "total_tokens": 3552,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 460566,
            "output_tokens": 96436,
            "total_tokens": 557002,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3500,
            "output_tokens": 38,
            "total_tokens": 3538,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 608547,
            "output_tokens": 50993,
            "total_tokens": 659540,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 33,
            "total_tokens": 3521,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 552088,
            "output_tokens": 75059,
            "total_tokens": 627147,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3487,
            "output_tokens": 45,
            "total_tokens": 3532,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 317688,
            "output_tokens": 58623,
            "total_tokens": 376311,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3493,
            "output_tokens": 36,
            "total_tokens": 3529,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 723708,
            "output_tokens": 83932,
            "total_tokens": 807640,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3505,
            "output_tokens": 49,
            "total_tokens": 3554,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 417704,
            "output_tokens": 88948,
            "total_tokens": 506652,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 29,
            "total_tokens": 3513,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 731872,
            "output_tokens": 116456,
            "total_tokens": 848328,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3504,
            "output_tokens": 45,
            "total_tokens": 3549,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 348601,
            "output_tokens": 69961,
            "total_tokens": 418562,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3487,
            "output_tokens": 28,
            "total_tokens": 3515,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 345232,
            "output_tokens": 71406,
            "total_tokens": 416638,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3481,
            "output_tokens": 24,
            "total_tokens": 3505,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 122966,
            "output_tokens": 20184,
            "total_tokens": 143150,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3494,
            "output_tokens": 36,
            "total_tokens": 3530,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 776706,
            "output_tokens": 84228,
            "total_tokens": 860934,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 45,
            "total_tokens": 3533,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 639421,
            "output_tokens": 110616,
            "total_tokens": 750037,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3502,
            "output_tokens": 45,
            "total_tokens": 3547,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 659376,
            "output_tokens": 141704,
            "total_tokens": 801080,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 25,
            "total_tokens": 3509,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 317831,
            "output_tokens": 54628,
            "total_tokens": 372459,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3492,
            "output_tokens": 41,
            "total_tokens": 3533,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 315342,
            "output_tokens": 43774,
            "total_tokens": 359116,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 33,
            "total_tokens": 3521,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 644873,
            "output_tokens": 112266,
            "total_tokens": 757139,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 28,
            "total_tokens": 3512,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 521722,
            "output_tokens": 74452,
            "total_tokens": 596174,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3498,
            "output_tokens": 35,
            "total_tokens": 3533,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 606261,
            "output_tokens": 118499,
            "total_tokens": 724760,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3501,
            "output_tokens": 40,
            "total_tokens": 3541,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 388984,
            "output_tokens": 90648,
            "total_tokens": 479632,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 25,
            "total_tokens": 3509,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 539753,
            "output_tokens": 88238,
            "total_tokens": 627991,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 29,
            "total_tokens": 3517,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 490439,
            "output_tokens": 77388,
            "total_tokens": 567827,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 28,
            "total_tokens": 3512,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 342393,
            "output_tokens": 63366,
            "total_tokens": 405759,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3523,
            "output_tokens": 66,
            "total_tokens": 3589,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 509291,
            "output_tokens": 99194,
            "total_tokens": 608485,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3486,
            "output_tokens": 28,
            "total_tokens": 3514,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 165938,
            "output_tokens": 28813,
            "total_tokens": 194751,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3500,
            "output_tokens": 41,
            "total_tokens": 3541,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 295853,
            "output_tokens": 44789,
            "total_tokens": 340642,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3498,
            "output_tokens": 39,
            "total_tokens": 3537,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 660150,
            "output_tokens": 108437,
            "total_tokens": 768587,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3494,
            "output_tokens": 35,
            "total_tokens": 3529,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 789532,
            "output_tokens": 114391,
            "total_tokens": 903923,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3481,
            "output_tokens": 67,
            "total_tokens": 3548,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 31897,
            "output_tokens": 5003,
            "total_tokens": 36900,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 86,
            "total_tokens": 3570,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4841,
            "output_tokens": 284,
            "total_tokens": 5125,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 70,
            "total_tokens": 3554,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4976,
            "output_tokens": 311,
            "total_tokens": 5287,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3481,
            "output_tokens": 23,
            "total_tokens": 3504,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5424,
            "output_tokens": 377,
            "total_tokens": 5801,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3478,
            "output_tokens": 39,
            "total_tokens": 3517,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4859,
            "output_tokens": 297,
            "total_tokens": 5156,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3477,
            "output_tokens": 39,
            "total_tokens": 3516,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4846,
            "output_tokens": 333,
            "total_tokens": 5179,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3480,
            "output_tokens": 43,
            "total_tokens": 3523,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5312,
            "output_tokens": 362,
            "total_tokens": 5674,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3477,
            "output_tokens": 48,
            "total_tokens": 3525,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5178,
            "output_tokens": 274,
            "total_tokens": 5452,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3478,
            "output_tokens": 41,
            "total_tokens": 3519,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5345,
            "output_tokens": 366,
            "total_tokens": 5711,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3480,
            "output_tokens": 36,
            "total_tokens": 3516,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4872,
            "output_tokens": 333,
            "total_tokens": 5205,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3482,
            "output_tokens": 37,
            "total_tokens": 3519,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4337,
            "output_tokens": 366,
            "total_tokens": 4703,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3508,
            "output_tokens": 223,
            "total_tokens": 3731,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4980,
            "output_tokens": 329,
            "total_tokens": 5309,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3488,
            "output_tokens": 97,
            "total_tokens": 3585,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4447,
            "output_tokens": 334,
            "total_tokens": 4781,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3493,
            "output_tokens": 81,
            "total_tokens": 3574,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5223,
            "output_tokens": 260,
            "total_tokens": 5483,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3484,
            "output_tokens": 111,
            "total_tokens": 3595,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4997,
            "output_tokens": 363,
            "total_tokens": 5360,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 5583,
            "output_tokens": 273,
            "total_tokens": 5856,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5578,
            "output_tokens": 345,
            "total_tokens": 5923,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3489,
            "output_tokens": 54,
            "total_tokens": 3543,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 4703,
            "output_tokens": 302,
            "total_tokens": 5005,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3514,
            "output_tokens": 104,
            "total_tokens": 3618,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5271,
            "output_tokens": 300,
            "total_tokens": 5571,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.0394904,
      0.0738141,
      0.10297769999999999,
      0.0809069,
      0.110382,
      0.127933,
      0.09794059999999999,
      0.1174952,
      0.0883781,
      0.0341879,
      0.0525586,
      0.1404292,
      0.1423539,
      0.040376499999999996,
      0.054006700000000005,
      0.0564544,
      0.048040599999999996,
      0.037021899999999996,
      0.1057545,
      0.13757729999999999,
      0.1037503,
      0.09387099999999998,
      0.0903819,
      0.09428239999999999,
      0.06438550000000001,
      0.1150361,
      0.08660209999999999,
      0.12876959999999998,
      0.0720545,
      0.0720831,
      0.029312699999999997,
      0.12045679999999999,
      0.11735849999999999,
      0.1318242,
      0.06259429999999999,
      0.0581838,
      0.1184437,
      0.090943,
      0.11712070000000001,
      0.08431009999999999,
      0.0982305,
      0.0890091,
      0.0685757,
      0.1000742,
      0.037114,
      0.0566609,
      0.11852479999999999,
      0.1337946,
      0.0145634,
      0.0101677,
      0.010032000000000001,
      0.0096257,
      0.0096897,
      0.0097003,
      0.009806,
      0.0097999,
      0.009785900000000002,
      0.0096804,
      0.009655100000000002,
      0.011629599999999999,
      0.010268300000000001,
      0.0101688,
      0.010464900000000001,
      0.0173833,
      0.009853600000000002,
      0.0104721
    ]
  },
  {
    "task_name": "LitQA2_FullText_Search_validation",
    "eval_spec": {
      "solver": "astabench/solvers/search/paper_finder.py@ai2i_paper_finder",
      "solver_args": "{\"base_url\": \"http://35.247.123.160:8000\"}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{\"with_search_tools\": false}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "ca5b0ad"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_paper_finder/recall_at_30",
        "value": 0.8
      },
      {
        "name": "score_paper_finder/stderr",
        "value": 0.13333333333333333
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3495,
            "output_tokens": 39,
            "total_tokens": 3534,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 678272,
            "output_tokens": 119314,
            "total_tokens": 797586,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3517,
            "output_tokens": 65,
            "total_tokens": 3582,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 627944,
            "output_tokens": 97148,
            "total_tokens": 725092,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3506,
            "output_tokens": 44,
            "total_tokens": 3550,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 811496,
            "output_tokens": 66162,
            "total_tokens": 877658,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3540,
            "output_tokens": 71,
            "total_tokens": 3611,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 779809,
            "output_tokens": 108070,
            "total_tokens": 887879,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3491,
            "output_tokens": 35,
            "total_tokens": 3526,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 677015,
            "output_tokens": 65326,
            "total_tokens": 742341,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3498,
            "output_tokens": 41,
            "total_tokens": 3539,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 5160,
            "output_tokens": 395,
            "total_tokens": 5555,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3494,
            "output_tokens": 35,
            "total_tokens": 3529,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 735209,
            "output_tokens": 90043,
            "total_tokens": 825252,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3495,
            "output_tokens": 36,
            "total_tokens": 3531,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 695687,
            "output_tokens": 114082,
            "total_tokens": 809769,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3491,
            "output_tokens": 33,
            "total_tokens": 3524,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 706809,
            "output_tokens": 58363,
            "total_tokens": 765172,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4o-2024-08-06",
          "usage": {
            "input_tokens": 3496,
            "output_tokens": 40,
            "total_tokens": 3536,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "gemini/gemini-2.0-flash",
          "usage": {
            "input_tokens": 323996,
            "output_tokens": 63804,
            "total_tokens": 387800,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.12468029999999998,
      0.1110961,
      0.1168194,
      0.1307689,
      0.1029094,
      0.009829000000000001,
      0.1186231,
      0.12429899999999999,
      0.10308359999999998,
      0.0670612
    ]
  }
] | 
	{
  "submit_time": "2025-08-07T19:06:21.862673Z",
  "username": "Ai2",
  "agent_name": "Asta Paper Finder",
  "agent_description": null,
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/aryeh_tiktinsky_ai2_Asta_Paper_Finder_2025-08-07T19-06-21",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "PaperFindingBench_validation",
    "eval_spec": {
      "solver": "/home/aryeht/PycharmProjects/asta-bench/astabench/solvers/search/youcom_search.py@youcom_solver",
      "solver_args": "{}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{\"with_search_tools\": false}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "ce99d55"
      },
      "packages": "{\"inspect_ai\": \"0.3.106\"}"
    },
    "metrics": [
      {
        "name": "score_paper_finder/semantic_f1",
        "value": 0.030522991829073212
      },
      {
        "name": "score_paper_finder/specific_f1",
        "value": 0.286475468975469
      },
      {
        "name": "score_paper_finder/metadata_f1",
        "value": 0
      },
      {
        "name": "score_paper_finder/adjusted_f1_micro_avg",
        "value": 0.06560391359924551
      },
      {
        "name": "score_paper_finder/stderr",
        "value": 0.014752214406566292
      }
    ],
    "model_usages": [
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1422,
            "output_tokens": 179,
            "total_tokens": 1601,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1447,
            "output_tokens": 243,
            "total_tokens": 1690,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1358,
            "output_tokens": 224,
            "total_tokens": 1582,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1476,
            "output_tokens": 228,
            "total_tokens": 1704,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1436,
            "output_tokens": 202,
            "total_tokens": 1638,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1405,
            "output_tokens": 212,
            "total_tokens": 1617,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1404,
            "output_tokens": 222,
            "total_tokens": 1626,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1474,
            "output_tokens": 255,
            "total_tokens": 1729,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1360,
            "output_tokens": 210,
            "total_tokens": 1570,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1348,
            "output_tokens": 196,
            "total_tokens": 1544,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1428,
            "output_tokens": 186,
            "total_tokens": 1614,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1496,
            "output_tokens": 229,
            "total_tokens": 1725,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1416,
            "output_tokens": 228,
            "total_tokens": 1644,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1386,
            "output_tokens": 214,
            "total_tokens": 1600,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1496,
            "output_tokens": 214,
            "total_tokens": 1710,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1488,
            "output_tokens": 297,
            "total_tokens": 1785,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1449,
            "output_tokens": 223,
            "total_tokens": 1672,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 7254,
            "output_tokens": 1108,
            "total_tokens": 8362,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1419,
            "output_tokens": 262,
            "total_tokens": 1681,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1443,
            "output_tokens": 249,
            "total_tokens": 1692,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1392,
            "output_tokens": 184,
            "total_tokens": 1576,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1493,
            "output_tokens": 269,
            "total_tokens": 1762,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1475,
            "output_tokens": 212,
            "total_tokens": 1687,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1457,
            "output_tokens": 261,
            "total_tokens": 1718,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1469,
            "output_tokens": 201,
            "total_tokens": 1670,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1464,
            "output_tokens": 211,
            "total_tokens": 1675,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1519,
            "output_tokens": 274,
            "total_tokens": 1793,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1426,
            "output_tokens": 194,
            "total_tokens": 1620,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1441,
            "output_tokens": 194,
            "total_tokens": 1635,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1488,
            "output_tokens": 205,
            "total_tokens": 1693,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1499,
            "output_tokens": 225,
            "total_tokens": 1724,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1481,
            "output_tokens": 195,
            "total_tokens": 1676,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1483,
            "output_tokens": 211,
            "total_tokens": 1694,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1365,
            "output_tokens": 167,
            "total_tokens": 1532,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1351,
            "output_tokens": 216,
            "total_tokens": 1567,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1451,
            "output_tokens": 201,
            "total_tokens": 1652,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1340,
            "output_tokens": 165,
            "total_tokens": 1505,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1474,
            "output_tokens": 255,
            "total_tokens": 1729,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1384,
            "output_tokens": 224,
            "total_tokens": 1608,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1462,
            "output_tokens": 216,
            "total_tokens": 1678,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1432,
            "output_tokens": 244,
            "total_tokens": 1676,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1368,
            "output_tokens": 178,
            "total_tokens": 1546,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1444,
            "output_tokens": 236,
            "total_tokens": 1680,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1478,
            "output_tokens": 280,
            "total_tokens": 1758,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1441,
            "output_tokens": 207,
            "total_tokens": 1648,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1523,
            "output_tokens": 246,
            "total_tokens": 1769,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1438,
            "output_tokens": 204,
            "total_tokens": 1642,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 6844,
            "output_tokens": 1018,
            "total_tokens": 7862,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1434,
            "output_tokens": 257,
            "total_tokens": 1691,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1413,
            "output_tokens": 190,
            "total_tokens": 1603,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1516,
            "output_tokens": 240,
            "total_tokens": 1756,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1468,
            "output_tokens": 273,
            "total_tokens": 1741,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1499,
            "output_tokens": 211,
            "total_tokens": 1710,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1224,
            "output_tokens": 161,
            "total_tokens": 1385,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1392,
            "output_tokens": 230,
            "total_tokens": 1622,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1479,
            "output_tokens": 272,
            "total_tokens": 1751,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1458,
            "output_tokens": 157,
            "total_tokens": 1615,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1367,
            "output_tokens": 148,
            "total_tokens": 1515,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1281,
            "output_tokens": 136,
            "total_tokens": 1417,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1040,
            "output_tokens": 146,
            "total_tokens": 1186,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1252,
            "output_tokens": 171,
            "total_tokens": 1423,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1476,
            "output_tokens": 327,
            "total_tokens": 1803,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1334,
            "output_tokens": 158,
            "total_tokens": 1492,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1387,
            "output_tokens": 196,
            "total_tokens": 1583,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1399,
            "output_tokens": 193,
            "total_tokens": 1592,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1494,
            "output_tokens": 223,
            "total_tokens": 1717,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ]
    ],
    "model_costs": [
      0.0003207,
      0.0003628499999999999,
      0.00033810000000000003,
      0.00035820000000000003,
      0.0003366,
      0.00033795000000000006,
      0.0003438,
      0.00037410000000000004,
      0.00033,
      0.0003198,
      0.0003258,
      0.00036180000000000007,
      0.0003492,
      0.00033630000000000004,
      0.0003528,
      0.0004014,
      0.00035114999999999994,
      0.0017529000000000008,
      0.0003700499999999999,
      0.00036585,
      0.0003191999999999999,
      0.00038535,
      0.00034844999999999993,
      0.00037515,
      0.00034094999999999997,
      0.0003462,
      0.00039225000000000013,
      0.00033030000000000006,
      0.00033255,
      0.0003462,
      0.00035985000000000005,
      0.00033915,
      0.0003490500000000001,
      0.00030494999999999996,
      0.00033225,
      0.00033825,
      0.00030000000000000003,
      0.0003741,
      0.000342,
      0.0003489000000000001,
      0.0003612,
      0.000312,
      0.0003582,
      0.0003897,
      0.00034034999999999995,
      0.00037605,
      0.0003381,
      0.0016373999999999996,
      0.0003693,
      0.00032595,
      0.0003714,
      0.000384,
      0.00035145,
      0.0002802,
      0.0003468,
      0.00038505,
      0.00031289999999999996,
      0.00029385,
      0.00027374999999999996,
      0.0002436,
      0.0002904,
      0.00041759999999999996,
      0.0002949,
      0.00032565,
      0.0003256499999999999,
      0.00035789999999999997
    ]
  },
  {
    "task_name": "LitQA2_FullText_Search_validation",
    "eval_spec": {
      "solver": "/home/aryeht/PycharmProjects/asta-bench/astabench/solvers/search/youcom_search.py@youcom_solver",
      "solver_args": "{}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{\"with_search_tools\": false}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "ce99d55"
      },
      "packages": "{\"inspect_ai\": \"0.3.106\"}"
    },
    "metrics": [
      {
        "name": "score_paper_finder/recall_at_30",
        "value": 0.5
      },
      {
        "name": "score_paper_finder/stderr",
        "value": 0.16666666666666666
      }
    ],
    "model_usages": [
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1474,
            "output_tokens": 288,
            "total_tokens": 1762,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1486,
            "output_tokens": 328,
            "total_tokens": 1814,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1511,
            "output_tokens": 288,
            "total_tokens": 1799,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1373,
            "output_tokens": 288,
            "total_tokens": 1661,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1548,
            "output_tokens": 338,
            "total_tokens": 1886,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1558,
            "output_tokens": 365,
            "total_tokens": 1923,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1434,
            "output_tokens": 186,
            "total_tokens": 1620,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1517,
            "output_tokens": 319,
            "total_tokens": 1836,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1205,
            "output_tokens": 209,
            "total_tokens": 1414,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ],
      [
        {
          "model": "gpt-4o-mini-2024-07-18",
          "usage": {
            "input_tokens": 1519,
            "output_tokens": 400,
            "total_tokens": 1919,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": 0
          }
        }
      ]
    ],
    "model_costs": [
      0.0003938999999999999,
      0.0004197,
      0.00039944999999999987,
      0.0003787499999999999,
      0.000435,
      0.0004527,
      0.0003267,
      0.0004189500000000001,
      0.00030615000000000004,
      0.00046784999999999985
    ]
  }
] | 
	{
  "submit_time": "2025-08-07T18:54:02.182157Z",
  "username": "Ai2",
  "agent_name": "You.com Search API",
  "agent_description": null,
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/aryeh_tiktinsky_ai2_You.com_Search_API_2025-08-07T18-54-02",
  "summary_url": null,
  "openness": "Closed source & API available",
  "tool_usage": "Fully custom"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "SUPER_Expert_validation",
    "eval_spec": {
      "solver": "astabench/solvers/code_agent/agent.py@code_agent",
      "solver_args": "{\"json_output\": 1, \"max_context_tokens\": 1000000, \"max_tries\": 200}",
      "model": "openai/gpt-5-mini-2025-08-07",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "be252e2"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "submitted/mean",
        "value": 0.7
      },
      {
        "name": "submitted/stderr",
        "value": 0.06546536707079771
      },
      {
        "name": "entrypoint/mean",
        "value": 0.26
      },
      {
        "name": "entrypoint/stderr",
        "value": 0.06266203485560375
      }
    ],
    "model_usages": [
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 478880,
            "output_tokens": 35038,
            "total_tokens": 513918,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 253312,
            "reasoning_tokens": 24704
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 116247,
            "output_tokens": 25429,
            "total_tokens": 141676,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 53632,
            "reasoning_tokens": 13568
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 146547,
            "output_tokens": 18905,
            "total_tokens": 165452,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 82816,
            "reasoning_tokens": 9984
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 274905,
            "output_tokens": 21113,
            "total_tokens": 296018,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 166656,
            "reasoning_tokens": 5888
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 89930,
            "output_tokens": 12890,
            "total_tokens": 102820,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 39040,
            "reasoning_tokens": 6976
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 346542,
            "output_tokens": 35303,
            "total_tokens": 381845,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 215680,
            "reasoning_tokens": 19136
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 65933,
            "output_tokens": 9272,
            "total_tokens": 75205,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 35584,
            "reasoning_tokens": 4736
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 360892,
            "output_tokens": 32589,
            "total_tokens": 393481,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 188544,
            "reasoning_tokens": 19968
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 157839,
            "output_tokens": 18374,
            "total_tokens": 176213,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 65408,
            "reasoning_tokens": 12032
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 404249,
            "output_tokens": 42858,
            "total_tokens": 447107,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 183680,
            "reasoning_tokens": 17984
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 57806,
            "output_tokens": 11878,
            "total_tokens": 69684,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 24192,
            "reasoning_tokens": 7040
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 319958,
            "output_tokens": 46134,
            "total_tokens": 366092,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 201856,
            "reasoning_tokens": 21760
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 184530,
            "output_tokens": 9354,
            "total_tokens": 193884,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 110976,
            "reasoning_tokens": 5248
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 153897,
            "output_tokens": 15312,
            "total_tokens": 169209,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 82304,
            "reasoning_tokens": 7104
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 442739,
            "output_tokens": 40915,
            "total_tokens": 483654,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 280704,
            "reasoning_tokens": 21568
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 187181,
            "output_tokens": 20143,
            "total_tokens": 207324,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 88448,
            "reasoning_tokens": 6848
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 450944,
            "output_tokens": 37475,
            "total_tokens": 488419,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 289408,
            "reasoning_tokens": 15808
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 497282,
            "output_tokens": 31254,
            "total_tokens": 528536,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 297856,
            "reasoning_tokens": 18368
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 557613,
            "output_tokens": 42550,
            "total_tokens": 600163,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 366464,
            "reasoning_tokens": 24512
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 506368,
            "output_tokens": 54428,
            "total_tokens": 560796,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 329216,
            "reasoning_tokens": 29824
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 1028138,
            "output_tokens": 50655,
            "total_tokens": 1078793,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 769152,
            "reasoning_tokens": 27520
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 204715,
            "output_tokens": 20559,
            "total_tokens": 225274,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 110976,
            "reasoning_tokens": 9792
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 552201,
            "output_tokens": 30300,
            "total_tokens": 582501,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 355456,
            "reasoning_tokens": 13056
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 396381,
            "output_tokens": 27921,
            "total_tokens": 424302,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 259456,
            "reasoning_tokens": 9408
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 68144,
            "output_tokens": 9608,
            "total_tokens": 77752,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 27904,
            "reasoning_tokens": 4928
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 250268,
            "output_tokens": 29482,
            "total_tokens": 279750,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 139776,
            "reasoning_tokens": 15680
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 46727,
            "output_tokens": 8142,
            "total_tokens": 54869,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 25344,
            "reasoning_tokens": 3712
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 233282,
            "output_tokens": 27157,
            "total_tokens": 260439,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 119424,
            "reasoning_tokens": 15872
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 187143,
            "output_tokens": 17493,
            "total_tokens": 204636,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 115328,
            "reasoning_tokens": 10112
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 41857,
            "output_tokens": 11162,
            "total_tokens": 53019,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 19584,
            "reasoning_tokens": 5120
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 320612,
            "output_tokens": 27305,
            "total_tokens": 347917,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 190336,
            "reasoning_tokens": 14464
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 148426,
            "output_tokens": 16272,
            "total_tokens": 164698,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 82048,
            "reasoning_tokens": 6592
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 210069,
            "output_tokens": 18914,
            "total_tokens": 228983,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 134656,
            "reasoning_tokens": 10112
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 98063,
            "output_tokens": 19571,
            "total_tokens": 117634,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 46208,
            "reasoning_tokens": 11904
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 199299,
            "output_tokens": 22901,
            "total_tokens": 222200,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 122240,
            "reasoning_tokens": 10752
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 342733,
            "output_tokens": 20596,
            "total_tokens": 363329,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 211968,
            "reasoning_tokens": 9344
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 99684,
            "output_tokens": 17548,
            "total_tokens": 117232,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 45184,
            "reasoning_tokens": 9344
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 355034,
            "output_tokens": 42397,
            "total_tokens": 397431,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 194176,
            "reasoning_tokens": 17216
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 508419,
            "output_tokens": 28371,
            "total_tokens": 536790,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 311296,
            "reasoning_tokens": 13888
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 643578,
            "output_tokens": 41398,
            "total_tokens": 684976,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 447744,
            "reasoning_tokens": 21184
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 628441,
            "output_tokens": 46677,
            "total_tokens": 675118,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 422656,
            "reasoning_tokens": 25152
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 69632,
            "output_tokens": 6638,
            "total_tokens": 76270,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 19840,
            "reasoning_tokens": 2880
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 733639,
            "output_tokens": 62532,
            "total_tokens": 796171,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 520832,
            "reasoning_tokens": 35200
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 311594,
            "output_tokens": 17581,
            "total_tokens": 329175,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 182656,
            "reasoning_tokens": 8640
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 166737,
            "output_tokens": 14968,
            "total_tokens": 181705,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 89472,
            "reasoning_tokens": 7040
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 811449,
            "output_tokens": 56228,
            "total_tokens": 867677,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 547584,
            "reasoning_tokens": 31744
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 612253,
            "output_tokens": 38499,
            "total_tokens": 650752,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 366208,
            "reasoning_tokens": 18112
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 1019521,
            "output_tokens": 54688,
            "total_tokens": 1074209,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 821376,
            "reasoning_tokens": 34560
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 59900,
            "output_tokens": 6988,
            "total_tokens": 66888,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 16896,
            "reasoning_tokens": 3776
          }
        }
      ],
      [
        {
          "model": "gpt-5-mini-2025-08-07",
          "usage": {
            "input_tokens": 420775,
            "output_tokens": 34555,
            "total_tokens": 455330,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 277760,
            "reasoning_tokens": 16000
          }
        }
      ]
    ],
    "model_costs": [
      0.1328008,
      0.06785255000000001,
      0.05581315,
      0.07345465,
      0.0394785,
      0.10871349999999998,
      0.02702085,
      0.11297859999999996,
      0.061490949999999996,
      0.14545025,
      0.0327643,
      0.12683989999999998,
      0.0398709,
      0.05057985,
      0.12935635,
      0.06718044999999999,
      0.1225692,
      0.1198109,
      0.14204884999999995,
      0.1613744,
      0.18528529999999999,
      0.06732715,
      0.11867265000000002,
      0.09655964999999998,
      0.029973599999999996,
      0.0900814,
      0.02226335,
      0.0857641,
      0.05582295000000001,
      0.028381849999999997,
      0.09193739999999999,
      0.05118969999999999,
      0.06004764999999999,
      0.05326095,
      0.06812275,
      0.07918244999999999,
      0.0498506,
      0.12986289999999998,
      0.11380515000000002,
      0.1429481,
      0.15536665000000008,
      0.02622,
      0.19128655000000003,
      0.07196290000000001,
      0.05148905000000001,
      0.19211185,
      0.14766444999999997,
      0.17944665,
      0.0251494,
      0.11180775
    ]
  }
] | 
	{
  "submit_time": "2025-08-14T19:05:15.098359Z",
  "username": "Ai2",
  "agent_name": "Asta Code",
  "agent_description": null,
  "agent_url": "https://github.com/allenai/asta-bench",
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/miked-ai_Asta_Code_GPT-5-mini_2025-08-14T19-05-15",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "SUPER_Expert_validation",
    "eval_spec": {
      "solver": "astabench/solvers/code_agent/agent.py@code_agent",
      "solver_args": "{\"json_output\": 1, \"max_context_tokens\": 1000000, \"max_tries\": 200}",
      "model": "openai/gpt-5-2025-08-07",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "be252e2"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "submitted/mean",
        "value": 0.28
      },
      {
        "name": "submitted/stderr",
        "value": 0.06414269805898185
      },
      {
        "name": "entrypoint/mean",
        "value": 0.44
      },
      {
        "name": "entrypoint/stderr",
        "value": 0.07091242083423346
      }
    ],
    "model_usages": [
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 136744,
            "output_tokens": 15302,
            "total_tokens": 152046,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 31488,
            "reasoning_tokens": 11456
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 28421,
            "output_tokens": 4989,
            "total_tokens": 33410,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 5888,
            "reasoning_tokens": 2432
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 199653,
            "output_tokens": 24212,
            "total_tokens": 223865,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 69632,
            "reasoning_tokens": 15168
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 53152,
            "output_tokens": 5740,
            "total_tokens": 58892,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 5120,
            "reasoning_tokens": 1728
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 363554,
            "output_tokens": 46459,
            "total_tokens": 410013,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 130944,
            "reasoning_tokens": 34944
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 167188,
            "output_tokens": 15063,
            "total_tokens": 182251,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 41600,
            "reasoning_tokens": 9216
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 64583,
            "output_tokens": 16745,
            "total_tokens": 81328,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 19840,
            "reasoning_tokens": 9088
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 1037980,
            "output_tokens": 88855,
            "total_tokens": 1126835,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 503936,
            "reasoning_tokens": 60864
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 197761,
            "output_tokens": 19316,
            "total_tokens": 217077,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 64640,
            "reasoning_tokens": 12992
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 356870,
            "output_tokens": 45234,
            "total_tokens": 402104,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 199168,
            "reasoning_tokens": 28416
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 170834,
            "output_tokens": 30811,
            "total_tokens": 201645,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 81920,
            "reasoning_tokens": 19712
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 869050,
            "output_tokens": 83980,
            "total_tokens": 953030,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 501504,
            "reasoning_tokens": 50432
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 162257,
            "output_tokens": 22929,
            "total_tokens": 185186,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 58752,
            "reasoning_tokens": 15296
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 752015,
            "output_tokens": 66987,
            "total_tokens": 819002,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 325760,
            "reasoning_tokens": 32960
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 149359,
            "output_tokens": 18055,
            "total_tokens": 167414,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 59264,
            "reasoning_tokens": 10368
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 857791,
            "output_tokens": 72454,
            "total_tokens": 930245,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 440192,
            "reasoning_tokens": 41088
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 352618,
            "output_tokens": 39802,
            "total_tokens": 392420,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 198272,
            "reasoning_tokens": 27840
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 503670,
            "output_tokens": 60344,
            "total_tokens": 564014,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 276352,
            "reasoning_tokens": 41920
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 128348,
            "output_tokens": 18194,
            "total_tokens": 146542,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 36992,
            "reasoning_tokens": 11264
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 229914,
            "output_tokens": 42563,
            "total_tokens": 272477,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 99072,
            "reasoning_tokens": 24000
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 198515,
            "output_tokens": 35487,
            "total_tokens": 234002,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 94080,
            "reasoning_tokens": 17920
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 107084,
            "output_tokens": 20327,
            "total_tokens": 127411,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 19328,
            "reasoning_tokens": 15360
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 187068,
            "output_tokens": 18419,
            "total_tokens": 205487,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 43648,
            "reasoning_tokens": 8128
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 97533,
            "output_tokens": 19805,
            "total_tokens": 117338,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 30208,
            "reasoning_tokens": 15808
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 123425,
            "output_tokens": 28354,
            "total_tokens": 151779,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 47360,
            "reasoning_tokens": 17280
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 1009044,
            "output_tokens": 84204,
            "total_tokens": 1093248,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 589696,
            "reasoning_tokens": 51200
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 57288,
            "output_tokens": 12961,
            "total_tokens": 70249,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 30464,
            "reasoning_tokens": 8576
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 183468,
            "output_tokens": 24496,
            "total_tokens": 207964,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 59904,
            "reasoning_tokens": 18560
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 134643,
            "output_tokens": 18605,
            "total_tokens": 153248,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 70144,
            "reasoning_tokens": 11392
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 570590,
            "output_tokens": 62953,
            "total_tokens": 633543,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 329216,
            "reasoning_tokens": 45248
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 462648,
            "output_tokens": 54029,
            "total_tokens": 516677,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 154880,
            "reasoning_tokens": 27264
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 308014,
            "output_tokens": 34842,
            "total_tokens": 342856,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 140416,
            "reasoning_tokens": 22080
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 454892,
            "output_tokens": 40897,
            "total_tokens": 495789,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 156416,
            "reasoning_tokens": 27648
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 373897,
            "output_tokens": 52511,
            "total_tokens": 426408,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 171008,
            "reasoning_tokens": 34880
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 19526,
            "output_tokens": 11915,
            "total_tokens": 31441,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 8576,
            "reasoning_tokens": 6400
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 99013,
            "output_tokens": 12502,
            "total_tokens": 111515,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 23808,
            "reasoning_tokens": 7808
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 136398,
            "output_tokens": 19103,
            "total_tokens": 155501,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 54656,
            "reasoning_tokens": 12864
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 468171,
            "output_tokens": 45514,
            "total_tokens": 513685,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 191360,
            "reasoning_tokens": 26112
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 467080,
            "output_tokens": 47081,
            "total_tokens": 514161,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 210048,
            "reasoning_tokens": 28096
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 114990,
            "output_tokens": 12189,
            "total_tokens": 127179,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 27648,
            "reasoning_tokens": 6080
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 156958,
            "output_tokens": 21860,
            "total_tokens": 178818,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 40704,
            "reasoning_tokens": 11712
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 77007,
            "output_tokens": 12943,
            "total_tokens": 89950,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 21120,
            "reasoning_tokens": 6464
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 114500,
            "output_tokens": 16263,
            "total_tokens": 130763,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 41856,
            "reasoning_tokens": 9536
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 174796,
            "output_tokens": 23511,
            "total_tokens": 198307,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 41600,
            "reasoning_tokens": 11840
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 87910,
            "output_tokens": 14166,
            "total_tokens": 102076,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 24064,
            "reasoning_tokens": 11008
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 432589,
            "output_tokens": 44533,
            "total_tokens": 477122,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 225408,
            "reasoning_tokens": 27264
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 301286,
            "output_tokens": 42348,
            "total_tokens": 343634,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 118144,
            "reasoning_tokens": 26048
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 393636,
            "output_tokens": 45497,
            "total_tokens": 439133,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 195840,
            "reasoning_tokens": 22400
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 430452,
            "output_tokens": 44435,
            "total_tokens": 474887,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 223104,
            "reasoning_tokens": 28480
          }
        }
      ],
      [
        {
          "model": "gpt-5-2025-08-07",
          "usage": {
            "input_tokens": 274698,
            "output_tokens": 46035,
            "total_tokens": 320733,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": 96896,
            "reasoning_tokens": 32256
          }
        }
      ]
    ],
    "model_costs": [
      0.288526,
      0.07879225000000001,
      0.41335025000000003,
      0.11808,
      0.7717205,
      0.31281500000000007,
      0.22585875000000002,
      1.619097,
      0.3676412499999999,
      0.6743635000000001,
      0.4294925000000001,
      1.3619205000000003,
      0.36601525,
      1.2434087500000002,
      0.30057675,
      1.3015627499999998,
      0.6157364999999998,
      0.9221315000000002,
      0.30075900000000005,
      0.6015665,
      0.49717375,
      0.315381,
      0.36892100000000005,
      0.28598225000000005,
      0.38454125,
      1.4399369999999998,
      0.16694799999999999,
      0.40690300000000007,
      0.27544175,
      0.9723995,
      0.9443600000000001,
      0.5754695000000001,
      0.801617,
      0.8000972500000001,
      0.13390950000000001,
      0.22200225000000004,
      0.30003950000000007,
      0.82507375,
      0.8183560000000001,
      0.23452350000000002,
      0.3690055,
      0.20192875,
      0.25866700000000004,
      0.40680499999999997,
      0.22447550000000002,
      0.7324822500000001,
      0.6671754999999999,
      0.7266950000000001,
      0.7314230000000002,
      0.6947145
    ]
  }
] | 
	{
  "submit_time": "2025-08-14T19:06:15.866442Z",
  "username": "Ai2",
  "agent_name": "Asta Code",
  "agent_description": null,
  "agent_url": "https://github.com/allenai/asta-bench",
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/miked-ai_Asta_Code_GPT-5_2025-08-14T19-06-15",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "SUPER_Expert_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "submitted/mean",
        "value": 0.08
      },
      {
        "name": "submitted/stderr",
        "value": 0.038756171332144394
      },
      {
        "name": "entrypoint/mean",
        "value": 0.24
      },
      {
        "name": "entrypoint/stderr",
        "value": 0.061011875725893214
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2164,
            "output_tokens": 24283,
            "total_tokens": 2399992,
            "input_tokens_cache_write": 66899,
            "input_tokens_cache_read": 2306646,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 851,
            "output_tokens": 9687,
            "total_tokens": 589904,
            "input_tokens_cache_write": 47160,
            "input_tokens_cache_read": 532206,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1937,
            "output_tokens": 7545,
            "total_tokens": 205042,
            "input_tokens_cache_write": 28889,
            "input_tokens_cache_read": 166671,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1875,
            "output_tokens": 3234,
            "total_tokens": 57989,
            "input_tokens_cache_write": 14401,
            "input_tokens_cache_read": 38479,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1839,
            "output_tokens": 2983,
            "total_tokens": 154286,
            "input_tokens_cache_write": 35607,
            "input_tokens_cache_read": 113857,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2064,
            "output_tokens": 6433,
            "total_tokens": 239089,
            "input_tokens_cache_write": 55166,
            "input_tokens_cache_read": 175426,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 778,
            "output_tokens": 1842,
            "total_tokens": 43260,
            "input_tokens_cache_write": 28120,
            "input_tokens_cache_read": 12520,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1929,
            "output_tokens": 5775,
            "total_tokens": 556097,
            "input_tokens_cache_write": 38895,
            "input_tokens_cache_read": 509498,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2179,
            "output_tokens": 9674,
            "total_tokens": 536351,
            "input_tokens_cache_write": 89450,
            "input_tokens_cache_read": 435048,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4506,
            "output_tokens": 2525,
            "total_tokens": 26155,
            "input_tokens_cache_write": 10268,
            "input_tokens_cache_read": 8856,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2087,
            "output_tokens": 22262,
            "total_tokens": 978041,
            "input_tokens_cache_write": 79000,
            "input_tokens_cache_read": 874692,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2606,
            "output_tokens": 6287,
            "total_tokens": 80874,
            "input_tokens_cache_write": 24823,
            "input_tokens_cache_read": 47158,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1831,
            "output_tokens": 2548,
            "total_tokens": 41923,
            "input_tokens_cache_write": 9953,
            "input_tokens_cache_read": 27591,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 773,
            "output_tokens": 2907,
            "total_tokens": 84806,
            "input_tokens_cache_write": 20560,
            "input_tokens_cache_read": 60566,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1846,
            "output_tokens": 2167,
            "total_tokens": 262826,
            "input_tokens_cache_write": 81057,
            "input_tokens_cache_read": 177756,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1856,
            "output_tokens": 2077,
            "total_tokens": 114244,
            "input_tokens_cache_write": 17734,
            "input_tokens_cache_read": 92577,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 842,
            "output_tokens": 5900,
            "total_tokens": 248302,
            "input_tokens_cache_write": 34492,
            "input_tokens_cache_read": 207068,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1887,
            "output_tokens": 4367,
            "total_tokens": 199836,
            "input_tokens_cache_write": 25850,
            "input_tokens_cache_read": 167732,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 772,
            "output_tokens": 1686,
            "total_tokens": 78953,
            "input_tokens_cache_write": 76495,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 785,
            "output_tokens": 2661,
            "total_tokens": 112843,
            "input_tokens_cache_write": 58040,
            "input_tokens_cache_read": 51357,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1841,
            "output_tokens": 5240,
            "total_tokens": 120653,
            "input_tokens_cache_write": 48945,
            "input_tokens_cache_read": 64627,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2284,
            "output_tokens": 3645,
            "total_tokens": 54357,
            "input_tokens_cache_write": 37166,
            "input_tokens_cache_read": 11262,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2362,
            "output_tokens": 3338,
            "total_tokens": 120349,
            "input_tokens_cache_write": 24598,
            "input_tokens_cache_read": 90051,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 776,
            "output_tokens": 8000,
            "total_tokens": 115198,
            "input_tokens_cache_write": 27067,
            "input_tokens_cache_read": 79355,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1936,
            "output_tokens": 1865,
            "total_tokens": 16326,
            "input_tokens_cache_write": 8767,
            "input_tokens_cache_read": 3758,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1846,
            "output_tokens": 4355,
            "total_tokens": 118204,
            "input_tokens_cache_write": 16152,
            "input_tokens_cache_read": 95851,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1807,
            "output_tokens": 3650,
            "total_tokens": 73841,
            "input_tokens_cache_write": 20194,
            "input_tokens_cache_read": 48190,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4945,
            "output_tokens": 3216,
            "total_tokens": 87233,
            "input_tokens_cache_write": 20145,
            "input_tokens_cache_read": 58927,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3731,
            "output_tokens": 5363,
            "total_tokens": 403398,
            "input_tokens_cache_write": 70497,
            "input_tokens_cache_read": 323807,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 760,
            "output_tokens": 4217,
            "total_tokens": 131685,
            "input_tokens_cache_write": 36973,
            "input_tokens_cache_read": 89735,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1810,
            "output_tokens": 5174,
            "total_tokens": 296955,
            "input_tokens_cache_write": 87067,
            "input_tokens_cache_read": 202904,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2059,
            "output_tokens": 2278,
            "total_tokens": 27073,
            "input_tokens_cache_write": 11935,
            "input_tokens_cache_read": 10801,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1832,
            "output_tokens": 2027,
            "total_tokens": 75116,
            "input_tokens_cache_write": 33810,
            "input_tokens_cache_read": 37447,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1858,
            "output_tokens": 4337,
            "total_tokens": 122232,
            "input_tokens_cache_write": 51043,
            "input_tokens_cache_read": 64994,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1787,
            "output_tokens": 4429,
            "total_tokens": 36290,
            "input_tokens_cache_write": 12560,
            "input_tokens_cache_read": 17514,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1879,
            "output_tokens": 4412,
            "total_tokens": 181518,
            "input_tokens_cache_write": 65928,
            "input_tokens_cache_read": 109299,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1776,
            "output_tokens": 3526,
            "total_tokens": 74038,
            "input_tokens_cache_write": 18753,
            "input_tokens_cache_read": 49983,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1830,
            "output_tokens": 9545,
            "total_tokens": 102194,
            "input_tokens_cache_write": 21044,
            "input_tokens_cache_read": 69775,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2056,
            "output_tokens": 1616,
            "total_tokens": 18759,
            "input_tokens_cache_write": 9709,
            "input_tokens_cache_read": 5378,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 813,
            "output_tokens": 4681,
            "total_tokens": 121039,
            "input_tokens_cache_write": 19256,
            "input_tokens_cache_read": 96289,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1800,
            "output_tokens": 2038,
            "total_tokens": 34640,
            "input_tokens_cache_write": 11485,
            "input_tokens_cache_read": 19317,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1864,
            "output_tokens": 6143,
            "total_tokens": 452829,
            "input_tokens_cache_write": 70061,
            "input_tokens_cache_read": 374761,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1863,
            "output_tokens": 2429,
            "total_tokens": 36035,
            "input_tokens_cache_write": 11221,
            "input_tokens_cache_read": 20522,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 855,
            "output_tokens": 12201,
            "total_tokens": 331291,
            "input_tokens_cache_write": 31582,
            "input_tokens_cache_read": 286653,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 820,
            "output_tokens": 6826,
            "total_tokens": 191326,
            "input_tokens_cache_write": 25804,
            "input_tokens_cache_read": 157876,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1977,
            "output_tokens": 3736,
            "total_tokens": 82480,
            "input_tokens_cache_write": 23185,
            "input_tokens_cache_read": 53582,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1845,
            "output_tokens": 5823,
            "total_tokens": 326935,
            "input_tokens_cache_write": 87809,
            "input_tokens_cache_read": 231458,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 986,
            "output_tokens": 23539,
            "total_tokens": 2053547,
            "input_tokens_cache_write": 81054,
            "input_tokens_cache_read": 1947968,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3343,
            "output_tokens": 6707,
            "total_tokens": 183933,
            "input_tokens_cache_write": 26142,
            "input_tokens_cache_read": 147741,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2013,
            "output_tokens": 4796,
            "total_tokens": 81814,
            "input_tokens_cache_write": 19794,
            "input_tokens_cache_read": 55211,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.35029387999999995,
      0.12916528,
      0.07395228,
      0.03191532,
      0.05811876,
      0.09658328000000001,
      0.037112,
      0.10429804000000001,
      0.16469304,
      0.024681279999999996,
      0.27617936000000004,
      0.05582844,
      0.02381708,
      0.03765168000000001,
      0.10542228000000001,
      0.03493296,
      0.07533103999999999,
      0.05824615999999999,
      0.08385659999999999,
      0.07342056,
      0.07654796,
      0.05447416,
      0.04704368,
      0.0660362,
      0.01807644,
      0.04271688,
      0.0400948,
      0.04167916,
      0.15175435999999998,
      0.061627799999999996,
      0.12544332,
      0.02355828,
      0.046379359999999994,
      0.07507691999999999,
      0.03310672,
      0.09382311999999998,
      0.038276439999999995,
      0.06627,
      0.01824804,
      0.04633351999999999,
      0.022622359999999998,
      0.12610508,
      0.02406916,
      0.10400224,
      0.06639408,
      0.04399716,
      0.13109364,
      0.33183624,
      0.06746368,
      0.045005279999999995
    ]
  },
  {
    "task_name": "ScholarQA_CS2_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "global_avg/mean",
        "value": 0.5898521066552724
      },
      {
        "name": "global_avg/stderr",
        "value": 0.01559875386280728
      },
      {
        "name": "ingredient_recall/mean",
        "value": 0.38029118969135295
      },
      {
        "name": "ingredient_recall/stderr",
        "value": 0.01846878463234233
      },
      {
        "name": "answer_precision/mean",
        "value": 0.7842142857142856
      },
      {
        "name": "answer_precision/stderr",
        "value": 0.024290468533694147
      },
      {
        "name": "citation_precision/mean",
        "value": 0.7108134018759018
      },
      {
        "name": "citation_precision/stderr",
        "value": 0.021078653771274374
      },
      {
        "name": "citation_recall/mean",
        "value": 0.4840895493395493
      },
      {
        "name": "citation_recall/stderr",
        "value": 0.02514827030410554
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1890,
            "total_tokens": 44086,
            "input_tokens_cache_write": 11195,
            "input_tokens_cache_read": 30985,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1604,
            "total_tokens": 42268,
            "input_tokens_cache_write": 8852,
            "input_tokens_cache_read": 31796,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1735,
            "total_tokens": 41995,
            "input_tokens_cache_write": 12657,
            "input_tokens_cache_read": 27587,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1427,
            "total_tokens": 36478,
            "input_tokens_cache_write": 10680,
            "input_tokens_cache_read": 24355,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1624,
            "total_tokens": 40304,
            "input_tokens_cache_write": 8139,
            "input_tokens_cache_read": 30525,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1785,
            "total_tokens": 41505,
            "input_tokens_cache_write": 12363,
            "input_tokens_cache_read": 27341,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1441,
            "total_tokens": 29421,
            "input_tokens_cache_write": 8080,
            "input_tokens_cache_read": 19887,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1329,
            "total_tokens": 23744,
            "input_tokens_cache_write": 5551,
            "input_tokens_cache_read": 16851,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1101,
            "total_tokens": 20291,
            "input_tokens_cache_write": 3625,
            "input_tokens_cache_read": 15552,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2566,
            "total_tokens": 32851,
            "input_tokens_cache_write": 10220,
            "input_tokens_cache_read": 20052,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2303,
            "total_tokens": 43283,
            "input_tokens_cache_write": 8848,
            "input_tokens_cache_read": 32116,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 9,
            "output_tokens": 2264,
            "total_tokens": 63480,
            "input_tokens_cache_write": 52669,
            "input_tokens_cache_read": 8538,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1824,
            "total_tokens": 33602,
            "input_tokens_cache_write": 8495,
            "input_tokens_cache_read": 23267,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1809,
            "total_tokens": 43359,
            "input_tokens_cache_write": 9108,
            "input_tokens_cache_read": 32426,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1635,
            "total_tokens": 43543,
            "input_tokens_cache_write": 9393,
            "input_tokens_cache_read": 32499,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 19,
            "output_tokens": 1767,
            "total_tokens": 47593,
            "input_tokens_cache_write": 13127,
            "input_tokens_cache_read": 32680,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 1869,
            "total_tokens": 55177,
            "input_tokens_cache_write": 9222,
            "input_tokens_cache_read": 44068,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2411,
            "total_tokens": 47087,
            "input_tokens_cache_write": 10379,
            "input_tokens_cache_read": 34281,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1493,
            "total_tokens": 32688,
            "input_tokens_cache_write": 5679,
            "input_tokens_cache_read": 25500,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1615,
            "total_tokens": 41762,
            "input_tokens_cache_write": 8560,
            "input_tokens_cache_read": 31571,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1638,
            "total_tokens": 24390,
            "input_tokens_cache_write": 6129,
            "input_tokens_cache_read": 16610,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1838,
            "total_tokens": 29952,
            "input_tokens_cache_write": 8120,
            "input_tokens_cache_read": 19981,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1629,
            "total_tokens": 49883,
            "input_tokens_cache_write": 17290,
            "input_tokens_cache_read": 30948,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1693,
            "total_tokens": 42315,
            "input_tokens_cache_write": 8802,
            "input_tokens_cache_read": 31804,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2049,
            "total_tokens": 31344,
            "input_tokens_cache_write": 9410,
            "input_tokens_cache_read": 19872,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1719,
            "total_tokens": 30423,
            "input_tokens_cache_write": 9616,
            "input_tokens_cache_read": 19075,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1847,
            "total_tokens": 43884,
            "input_tokens_cache_write": 9319,
            "input_tokens_cache_read": 32702,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1847,
            "total_tokens": 42590,
            "input_tokens_cache_write": 8788,
            "input_tokens_cache_read": 31939,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 1708,
            "total_tokens": 54203,
            "input_tokens_cache_write": 9199,
            "input_tokens_cache_read": 43278,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1879,
            "total_tokens": 30754,
            "input_tokens_cache_write": 8633,
            "input_tokens_cache_read": 20229,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2259,
            "total_tokens": 43203,
            "input_tokens_cache_write": 8941,
            "input_tokens_cache_read": 31987,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 1707,
            "total_tokens": 57409,
            "input_tokens_cache_write": 10206,
            "input_tokens_cache_read": 45478,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1508,
            "total_tokens": 60645,
            "input_tokens_cache_write": 18537,
            "input_tokens_cache_read": 40584,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 2355,
            "total_tokens": 53650,
            "input_tokens_cache_write": 8588,
            "input_tokens_cache_read": 42689,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1665,
            "total_tokens": 27685,
            "input_tokens_cache_write": 10089,
            "input_tokens_cache_read": 15918,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1926,
            "total_tokens": 31048,
            "input_tokens_cache_write": 8766,
            "input_tokens_cache_read": 20343,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1541,
            "total_tokens": 35210,
            "input_tokens_cache_write": 13935,
            "input_tokens_cache_read": 19721,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1631,
            "total_tokens": 26469,
            "input_tokens_cache_write": 8065,
            "input_tokens_cache_read": 16760,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2036,
            "total_tokens": 45278,
            "input_tokens_cache_write": 9653,
            "input_tokens_cache_read": 33573,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2034,
            "total_tokens": 28216,
            "input_tokens_cache_write": 7182,
            "input_tokens_cache_read": 18987,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 2304,
            "total_tokens": 57535,
            "input_tokens_cache_write": 18039,
            "input_tokens_cache_read": 37174,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1918,
            "total_tokens": 48869,
            "input_tokens_cache_write": 13149,
            "input_tokens_cache_read": 33786,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2007,
            "total_tokens": 41815,
            "input_tokens_cache_write": 8489,
            "input_tokens_cache_read": 31303,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1793,
            "total_tokens": 42869,
            "input_tokens_cache_write": 8986,
            "input_tokens_cache_read": 32074,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 1518,
            "total_tokens": 56126,
            "input_tokens_cache_write": 9798,
            "input_tokens_cache_read": 44792,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2075,
            "total_tokens": 40755,
            "input_tokens_cache_write": 8209,
            "input_tokens_cache_read": 30455,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 19,
            "output_tokens": 1767,
            "total_tokens": 61090,
            "input_tokens_cache_write": 11335,
            "input_tokens_cache_read": 47969,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1522,
            "total_tokens": 43589,
            "input_tokens_cache_write": 20269,
            "input_tokens_cache_read": 21785,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1775,
            "total_tokens": 44430,
            "input_tokens_cache_write": 9499,
            "input_tokens_cache_read": 33140,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1334,
            "total_tokens": 38632,
            "input_tokens_cache_write": 7664,
            "input_tokens_cache_read": 29618,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1489,
            "total_tokens": 26858,
            "input_tokens_cache_write": 6720,
            "input_tokens_cache_read": 18636,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1729,
            "total_tokens": 65884,
            "input_tokens_cache_write": 30399,
            "input_tokens_cache_read": 33743,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1724,
            "total_tokens": 29257,
            "input_tokens_cache_write": 7879,
            "input_tokens_cache_read": 19641,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1825,
            "total_tokens": 39610,
            "input_tokens_cache_write": 7839,
            "input_tokens_cache_read": 29930,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1664,
            "total_tokens": 41673,
            "input_tokens_cache_write": 8656,
            "input_tokens_cache_read": 31337,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1838,
            "total_tokens": 40311,
            "input_tokens_cache_write": 8083,
            "input_tokens_cache_read": 30374,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 2290,
            "total_tokens": 55668,
            "input_tokens_cache_write": 16409,
            "input_tokens_cache_read": 36951,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 2198,
            "total_tokens": 72459,
            "input_tokens_cache_write": 13299,
            "input_tokens_cache_read": 56944,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1658,
            "total_tokens": 28790,
            "input_tokens_cache_write": 8448,
            "input_tokens_cache_read": 18671,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2297,
            "total_tokens": 31579,
            "input_tokens_cache_write": 9448,
            "input_tokens_cache_read": 19821,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2186,
            "total_tokens": 30913,
            "input_tokens_cache_write": 9261,
            "input_tokens_cache_read": 19453,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2024,
            "total_tokens": 30907,
            "input_tokens_cache_write": 9279,
            "input_tokens_cache_read": 19591,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1782,
            "total_tokens": 33189,
            "input_tokens_cache_write": 10565,
            "input_tokens_cache_read": 20829,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1537,
            "total_tokens": 36717,
            "input_tokens_cache_write": 7469,
            "input_tokens_cache_read": 27695,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1894,
            "total_tokens": 31030,
            "input_tokens_cache_write": 9350,
            "input_tokens_cache_read": 19773,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1306,
            "total_tokens": 31293,
            "input_tokens_cache_write": 9222,
            "input_tokens_cache_read": 20752,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1703,
            "total_tokens": 39270,
            "input_tokens_cache_write": 10054,
            "input_tokens_cache_read": 27497,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2088,
            "total_tokens": 30476,
            "input_tokens_cache_write": 9212,
            "input_tokens_cache_read": 19163,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1392,
            "total_tokens": 20474,
            "input_tokens_cache_write": 3847,
            "input_tokens_cache_read": 15222,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1746,
            "total_tokens": 30913,
            "input_tokens_cache_write": 9444,
            "input_tokens_cache_read": 19710,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1846,
            "total_tokens": 32551,
            "input_tokens_cache_write": 10567,
            "input_tokens_cache_read": 20125,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1636,
            "total_tokens": 65633,
            "input_tokens_cache_write": 16568,
            "input_tokens_cache_read": 47413,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1395,
            "total_tokens": 48745,
            "input_tokens_cache_write": 18630,
            "input_tokens_cache_read": 28707,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1826,
            "total_tokens": 35074,
            "input_tokens_cache_write": 6286,
            "input_tokens_cache_read": 26946,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1800,
            "total_tokens": 32967,
            "input_tokens_cache_write": 9707,
            "input_tokens_cache_read": 21447,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2366,
            "total_tokens": 36613,
            "input_tokens_cache_write": 14123,
            "input_tokens_cache_read": 20111,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 19,
            "output_tokens": 1922,
            "total_tokens": 34692,
            "input_tokens_cache_write": 9277,
            "input_tokens_cache_read": 23474,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2101,
            "total_tokens": 29744,
            "input_tokens_cache_write": 12609,
            "input_tokens_cache_read": 15021,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1569,
            "total_tokens": 42429,
            "input_tokens_cache_write": 8838,
            "input_tokens_cache_read": 32006,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 9,
            "output_tokens": 1741,
            "total_tokens": 36766,
            "input_tokens_cache_write": 26483,
            "input_tokens_cache_read": 8533,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1494,
            "total_tokens": 29122,
            "input_tokens_cache_write": 7931,
            "input_tokens_cache_read": 19684,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1393,
            "total_tokens": 45860,
            "input_tokens_cache_write": 10885,
            "input_tokens_cache_read": 33566,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1543,
            "total_tokens": 30222,
            "input_tokens_cache_write": 9365,
            "input_tokens_cache_read": 19301,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2670,
            "total_tokens": 30605,
            "input_tokens_cache_write": 9017,
            "input_tokens_cache_read": 18905,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 2499,
            "total_tokens": 61310,
            "input_tokens_cache_write": 11285,
            "input_tokens_cache_read": 47508,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 19,
            "output_tokens": 1820,
            "total_tokens": 57795,
            "input_tokens_cache_write": 13591,
            "input_tokens_cache_read": 42365,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1289,
            "total_tokens": 33132,
            "input_tokens_cache_write": 5878,
            "input_tokens_cache_read": 25949,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1846,
            "total_tokens": 40488,
            "input_tokens_cache_write": 8172,
            "input_tokens_cache_read": 30454,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 1232,
            "total_tokens": 59497,
            "input_tokens_cache_write": 17550,
            "input_tokens_cache_read": 40697,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2068,
            "total_tokens": 29934,
            "input_tokens_cache_write": 8040,
            "input_tokens_cache_read": 19813,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1612,
            "total_tokens": 39714,
            "input_tokens_cache_write": 7841,
            "input_tokens_cache_read": 30245,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2323,
            "total_tokens": 32451,
            "input_tokens_cache_write": 10145,
            "input_tokens_cache_read": 19970,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1889,
            "total_tokens": 25409,
            "input_tokens_cache_write": 6943,
            "input_tokens_cache_read": 16564,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2324,
            "total_tokens": 29727,
            "input_tokens_cache_write": 12481,
            "input_tokens_cache_read": 14909,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2564,
            "total_tokens": 34191,
            "input_tokens_cache_write": 16009,
            "input_tokens_cache_read": 15605,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2495,
            "total_tokens": 45081,
            "input_tokens_cache_write": 9368,
            "input_tokens_cache_read": 33202,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2232,
            "total_tokens": 32266,
            "input_tokens_cache_write": 9933,
            "input_tokens_cache_read": 20088,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 2330,
            "total_tokens": 29358,
            "input_tokens_cache_write": 7620,
            "input_tokens_cache_read": 19395,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 2004,
            "total_tokens": 40320,
            "input_tokens_cache_write": 7777,
            "input_tokens_cache_read": 30523,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1959,
            "total_tokens": 31615,
            "input_tokens_cache_write": 9660,
            "input_tokens_cache_read": 19983,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.0212466,
      0.01782448,
      0.021816759999999998,
      0.018349200000000003,
      0.0170898,
      0.02170308,
      0.01544536,
      0.01222548,
      0.00928356,
      0.022098560000000003,
      0.02064208,
      0.06241524,
      0.01766516,
      0.018950880000000003,
      0.01854572,
      0.0228246,
      0.02023784,
      0.022778279999999998,
      0.013703799999999999,
      0.01755848,
      0.014020199999999998,
      0.01708088,
      0.026294639999999998,
      0.01813112,
      0.01920616,
      0.0180284,
      0.01933596,
      0.018743919999999997,
      0.01950764,
      0.017777719999999997,
      0.02054876,
      0.02068664,
      0.027828520000000002,
      0.02143752,
      0.01803284,
      0.01810784,
      0.021687079999999997,
      0.015940199999999998,
      0.020495640000000002,
      0.01684736,
      0.030243319999999997,
      0.02353668,
      0.01903404,
      0.01873672,
      0.01946776,
      0.018958199999999998,
      0.02225572,
      0.028110200000000002,
      0.019263000000000002,
      0.015382239999999998,
      0.014177279999999999,
      0.04002483999999999,
      0.01635668,
      0.0175462,
      0.01783176,
      0.01787772,
      0.02853948,
      0.026660919999999998,
      0.01658408,
      0.02023208,
      0.01957164,
      0.01895268,
      0.01936972,
      0.015845400000000003,
      0.01851824,
      0.016116560000000002,
      0.01907856,
      0.01910744,
      0.010643159999999999,
      0.018015200000000002,
      0.0195714,
      0.02691784,
      0.02651696,
      0.01575848,
      0.01863316,
      0.025206279999999998,
      0.01885812,
      0.022225079999999998,
      0.01768728,
      0.034136839999999995,
      0.01549212,
      0.019155079999999998,
      0.01709148,
      0.0212198,
      0.02509604,
      0.024275399999999996,
      0.01312272,
      0.01800512,
      0.02574816,
      0.01790744,
      0.0167214,
      0.021045,
      0.01583452,
      0.02298012,
      0.0275238,
      0.02201696,
      0.020478439999999997,
      0.018501999999999998,
      0.018247640000000002,
      0.019105039999999997
    ]
  },
  {
    "task_name": "PaperFindingBench_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{\"with_search_tools\": true}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_paper_finder/semantic_f1",
        "value": 0.030075109311941118
      },
      {
        "name": "score_paper_finder/specific_f1",
        "value": 0.05
      },
      {
        "name": "score_paper_finder/metadata_f1",
        "value": 0
      },
      {
        "name": "score_paper_finder/adjusted_f1_micro_avg",
        "value": 0.029448564348078392
      },
      {
        "name": "score_paper_finder/stderr",
        "value": 0.009582510067831838
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 657,
            "total_tokens": 60197,
            "input_tokens_cache_write": 24625,
            "input_tokens_cache_read": 34905,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 765,
            "total_tokens": 28276,
            "input_tokens_cache_write": 8711,
            "input_tokens_cache_read": 18790,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 531,
            "total_tokens": 50844,
            "input_tokens_cache_write": 12091,
            "input_tokens_cache_read": 38210,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 509,
            "total_tokens": 38142,
            "input_tokens_cache_write": 12006,
            "input_tokens_cache_read": 25617,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 564,
            "total_tokens": 37296,
            "input_tokens_cache_write": 12898,
            "input_tokens_cache_read": 23824,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 455,
            "total_tokens": 24298,
            "input_tokens_cache_write": 15591,
            "input_tokens_cache_read": 8245,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 599,
            "total_tokens": 52237,
            "input_tokens_cache_write": 12629,
            "input_tokens_cache_read": 38997,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 627,
            "total_tokens": 32765,
            "input_tokens_cache_write": 15115,
            "input_tokens_cache_read": 17013,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 810,
            "total_tokens": 99780,
            "input_tokens_cache_write": 28307,
            "input_tokens_cache_read": 70651,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 619,
            "total_tokens": 12253,
            "input_tokens_cache_write": 3382,
            "input_tokens_cache_read": 8245,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 834,
            "total_tokens": 28016,
            "input_tokens_cache_write": 9312,
            "input_tokens_cache_read": 17860,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 658,
            "total_tokens": 54074,
            "input_tokens_cache_write": 21363,
            "input_tokens_cache_read": 32043,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 688,
            "total_tokens": 32839,
            "input_tokens_cache_write": 23856,
            "input_tokens_cache_read": 8288,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 809,
            "total_tokens": 62749,
            "input_tokens_cache_write": 26397,
            "input_tokens_cache_read": 35533,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 877,
            "total_tokens": 30415,
            "input_tokens_cache_write": 9763,
            "input_tokens_cache_read": 19765,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 646,
            "total_tokens": 26078,
            "input_tokens_cache_write": 17193,
            "input_tokens_cache_read": 8232,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 1001,
            "total_tokens": 31943,
            "input_tokens_cache_write": 11226,
            "input_tokens_cache_read": 19706,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 590,
            "total_tokens": 12422,
            "input_tokens_cache_write": 3595,
            "input_tokens_cache_read": 8230,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 15,
            "output_tokens": 1104,
            "total_tokens": 51311,
            "input_tokens_cache_write": 8323,
            "input_tokens_cache_read": 41869,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 787,
            "total_tokens": 45969,
            "input_tokens_cache_write": 16832,
            "input_tokens_cache_read": 28340,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 781,
            "total_tokens": 29774,
            "input_tokens_cache_write": 9549,
            "input_tokens_cache_read": 19434,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 510,
            "total_tokens": 46555,
            "input_tokens_cache_write": 17754,
            "input_tokens_cache_read": 28281,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 477,
            "total_tokens": 54932,
            "input_tokens_cache_write": 21756,
            "input_tokens_cache_read": 32689,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 835,
            "total_tokens": 29449,
            "input_tokens_cache_write": 9138,
            "input_tokens_cache_read": 19466,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 754,
            "total_tokens": 211350,
            "input_tokens_cache_write": 49603,
            "input_tokens_cache_read": 160980,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 690,
            "total_tokens": 34476,
            "input_tokens_cache_write": 25544,
            "input_tokens_cache_read": 8235,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 680,
            "total_tokens": 44015,
            "input_tokens_cache_write": 35078,
            "input_tokens_cache_read": 8250,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 565,
            "total_tokens": 16345,
            "input_tokens_cache_write": 7547,
            "input_tokens_cache_read": 8226,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 866,
            "total_tokens": 56563,
            "input_tokens_cache_write": 23097,
            "input_tokens_cache_read": 32590,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 727,
            "total_tokens": 52688,
            "input_tokens_cache_write": 21367,
            "input_tokens_cache_read": 30584,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 657,
            "total_tokens": 13078,
            "input_tokens_cache_write": 4192,
            "input_tokens_cache_read": 8222,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 651,
            "total_tokens": 33364,
            "input_tokens_cache_write": 24469,
            "input_tokens_cache_read": 8237,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 669,
            "total_tokens": 43489,
            "input_tokens_cache_write": 16691,
            "input_tokens_cache_read": 26119,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 668,
            "total_tokens": 55628,
            "input_tokens_cache_write": 22526,
            "input_tokens_cache_read": 32424,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 580,
            "total_tokens": 14683,
            "input_tokens_cache_write": 5869,
            "input_tokens_cache_read": 8227,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 1043,
            "total_tokens": 54846,
            "input_tokens_cache_write": 30419,
            "input_tokens_cache_read": 23374,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 886,
            "total_tokens": 63445,
            "input_tokens_cache_write": 26397,
            "input_tokens_cache_read": 36152,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 563,
            "total_tokens": 38416,
            "input_tokens_cache_write": 14235,
            "input_tokens_cache_read": 23608,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 765,
            "total_tokens": 57567,
            "input_tokens_cache_write": 23683,
            "input_tokens_cache_read": 33109,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 744,
            "total_tokens": 65017,
            "input_tokens_cache_write": 16893,
            "input_tokens_cache_read": 47368,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 549,
            "total_tokens": 28536,
            "input_tokens_cache_write": 9067,
            "input_tokens_cache_read": 18910,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 737,
            "total_tokens": 15963,
            "input_tokens_cache_write": 6990,
            "input_tokens_cache_read": 8229,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 619,
            "total_tokens": 15642,
            "input_tokens_cache_write": 6788,
            "input_tokens_cache_read": 8228,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 754,
            "total_tokens": 79236,
            "input_tokens_cache_write": 34670,
            "input_tokens_cache_read": 43802,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 932,
            "total_tokens": 46299,
            "input_tokens_cache_write": 18419,
            "input_tokens_cache_read": 26938,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 627,
            "total_tokens": 27915,
            "input_tokens_cache_write": 19040,
            "input_tokens_cache_read": 8241,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 759,
            "total_tokens": 54281,
            "input_tokens_cache_write": 22819,
            "input_tokens_cache_read": 30693,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 806,
            "total_tokens": 50028,
            "input_tokens_cache_write": 19464,
            "input_tokens_cache_read": 29748,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 606,
            "total_tokens": 48649,
            "input_tokens_cache_write": 12270,
            "input_tokens_cache_read": 35761,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 598,
            "total_tokens": 55096,
            "input_tokens_cache_write": 25607,
            "input_tokens_cache_read": 28879,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 15,
            "output_tokens": 616,
            "total_tokens": 28671,
            "input_tokens_cache_write": 4445,
            "input_tokens_cache_read": 23595,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 553,
            "total_tokens": 24674,
            "input_tokens_cache_write": 6933,
            "input_tokens_cache_read": 17178,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 704,
            "total_tokens": 21519,
            "input_tokens_cache_write": 4708,
            "input_tokens_cache_read": 16097,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 924,
            "total_tokens": 23876,
            "input_tokens_cache_write": 6993,
            "input_tokens_cache_read": 15949,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 719,
            "total_tokens": 20001,
            "input_tokens_cache_write": 3764,
            "input_tokens_cache_read": 15508,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 927,
            "total_tokens": 23864,
            "input_tokens_cache_write": 7105,
            "input_tokens_cache_read": 15822,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 690,
            "total_tokens": 42719,
            "input_tokens_cache_write": 5822,
            "input_tokens_cache_read": 36189,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 522,
            "total_tokens": 54915,
            "input_tokens_cache_write": 21896,
            "input_tokens_cache_read": 32487,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 704,
            "total_tokens": 138776,
            "input_tokens_cache_write": 43172,
            "input_tokens_cache_read": 94887,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1021,
            "total_tokens": 187709,
            "input_tokens_cache_write": 54478,
            "input_tokens_cache_read": 132197,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 819,
            "total_tokens": 170374,
            "input_tokens_cache_write": 50297,
            "input_tokens_cache_read": 119245,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 884,
            "total_tokens": 183847,
            "input_tokens_cache_write": 75852,
            "input_tokens_cache_read": 107093,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 19,
            "output_tokens": 1211,
            "total_tokens": 121168,
            "input_tokens_cache_write": 32634,
            "input_tokens_cache_read": 87304,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 28,
            "output_tokens": 1344,
            "total_tokens": 102855,
            "input_tokens_cache_write": 22547,
            "input_tokens_cache_read": 78936,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 790,
            "total_tokens": 383838,
            "input_tokens_cache_write": 88277,
            "input_tokens_cache_read": 294753,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 1014,
            "total_tokens": 219867,
            "input_tokens_cache_write": 96228,
            "input_tokens_cache_read": 122612,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.030053399999999997,
      0.013282200000000001,
      0.0172814,
      0.01609936,
      0.01706792,
      0.0180762,
      0.01815436,
      0.018992039999999998,
      0.03720868,
      0.0065232,
      0.0140848,
      0.02656644,
      0.02727664,
      0.03248364,
      0.014860199999999999,
      0.020441159999999996,
      0.01681448,
      0.006619000000000001,
      0.01610052,
      0.0222552,
      0.01423572,
      0.022064479999999997,
      0.026287119999999994,
      0.01404328,
      0.06550779999999999,
      0.028968399999999995,
      0.0384636,
      0.01047068,
      0.0291762,
      0.02672972,
      0.00748336,
      0.027737559999999998,
      0.02146452,
      0.02779992,
      0.00885276,
      0.03646892,
      0.03284116,
      0.01838364,
      0.02939972,
      0.023668039999999998,
      0.012783799999999998,
      0.010601919999999999,
      0.00992784,
      0.04119816,
      0.024310039999999998,
      0.022212879999999997,
      0.028318439999999997,
      0.025075840000000002,
      0.01756448,
      0.03031892,
      0.0088086,
      0.01052724,
      0.00881976,
      0.01197292,
      0.007888639999999999,
      0.012086759999999998,
      0.011491519999999998,
      0.02659096,
      0.05358936,
      0.06994016,
      0.063123,
      0.08796984,
      0.05692072000000001,
      0.04357068,
      0.12956764,
      0.11010336
    ]
  },
  {
    "task_name": "ArxivDIGESTables_Clean_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{\"args\": [], \"kwargs\": {}}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_tables/mean",
        "value": 0.20944852865388583
      },
      {
        "name": "score_tables/stderr",
        "value": 0.0203418258216058
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2014,
            "total_tokens": 6836,
            "input_tokens_cache_write": 4818,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1538,
            "total_tokens": 4779,
            "input_tokens_cache_write": 3237,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1211,
            "total_tokens": 4345,
            "input_tokens_cache_write": 3130,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1190,
            "total_tokens": 3943,
            "input_tokens_cache_write": 2749,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 804,
            "total_tokens": 4115,
            "input_tokens_cache_write": 3307,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 733,
            "total_tokens": 4050,
            "input_tokens_cache_write": 3313,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1557,
            "total_tokens": 5397,
            "input_tokens_cache_write": 3836,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2043,
            "total_tokens": 6559,
            "input_tokens_cache_write": 4512,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 887,
            "total_tokens": 3630,
            "input_tokens_cache_write": 2739,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2940,
            "total_tokens": 6693,
            "input_tokens_cache_write": 3749,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1236,
            "total_tokens": 4277,
            "input_tokens_cache_write": 3037,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 3972,
            "total_tokens": 10080,
            "input_tokens_cache_write": 6104,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1405,
            "total_tokens": 4470,
            "input_tokens_cache_write": 3061,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2131,
            "total_tokens": 5338,
            "input_tokens_cache_write": 3203,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2945,
            "total_tokens": 7404,
            "input_tokens_cache_write": 4455,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1226,
            "total_tokens": 4787,
            "input_tokens_cache_write": 3557,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1172,
            "total_tokens": 4274,
            "input_tokens_cache_write": 3098,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2159,
            "total_tokens": 4776,
            "input_tokens_cache_write": 2613,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 3056,
            "total_tokens": 9360,
            "input_tokens_cache_write": 6300,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 1641,
            "total_tokens": 9528,
            "input_tokens_cache_write": 4053,
            "input_tokens_cache_read": 3824,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1236,
            "total_tokens": 3902,
            "input_tokens_cache_write": 2662,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 568,
            "total_tokens": 3287,
            "input_tokens_cache_write": 2715,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1851,
            "total_tokens": 5230,
            "input_tokens_cache_write": 3375,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 3913,
            "total_tokens": 11794,
            "input_tokens_cache_write": 4061,
            "input_tokens_cache_read": 3810,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 3266,
            "total_tokens": 9126,
            "input_tokens_cache_write": 5856,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1806,
            "total_tokens": 5109,
            "input_tokens_cache_write": 3299,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1476,
            "total_tokens": 4989,
            "input_tokens_cache_write": 3509,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1261,
            "total_tokens": 4191,
            "input_tokens_cache_write": 2926,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1875,
            "total_tokens": 5077,
            "input_tokens_cache_write": 3198,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1401,
            "total_tokens": 4872,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 3467,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 604,
            "total_tokens": 3716,
            "input_tokens_cache_write": 3108,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2619,
            "total_tokens": 7346,
            "input_tokens_cache_write": 4723,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1843,
            "total_tokens": 4815,
            "input_tokens_cache_write": 2968,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2665,
            "total_tokens": 7550,
            "input_tokens_cache_write": 4881,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1908,
            "total_tokens": 6302,
            "input_tokens_cache_write": 4390,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1042,
            "total_tokens": 5417,
            "input_tokens_cache_write": 4371,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 624,
            "total_tokens": 4962,
            "input_tokens_cache_write": 4334,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1396,
            "total_tokens": 4322,
            "input_tokens_cache_write": 2922,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2482,
            "total_tokens": 6767,
            "input_tokens_cache_write": 4281,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1250,
            "total_tokens": 4471,
            "input_tokens_cache_write": 3217,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2043,
            "total_tokens": 7481,
            "input_tokens_cache_write": 5434,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 14,
            "output_tokens": 2689,
            "total_tokens": 13804,
            "input_tokens_cache_write": 3971,
            "input_tokens_cache_read": 7130,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2237,
            "total_tokens": 9694,
            "input_tokens_cache_write": 7453,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1561,
            "total_tokens": 4908,
            "input_tokens_cache_write": 3343,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 747,
            "total_tokens": 4405,
            "input_tokens_cache_write": 3654,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1460,
            "total_tokens": 4550,
            "input_tokens_cache_write": 3086,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1206,
            "total_tokens": 4288,
            "input_tokens_cache_write": 3078,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2032,
            "total_tokens": 6016,
            "input_tokens_cache_write": 3980,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 3625,
            "total_tokens": 9818,
            "input_tokens_cache_write": 6189,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 3354,
            "total_tokens": 13399,
            "input_tokens_cache_write": 5131,
            "input_tokens_cache_read": 4904,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1357,
            "total_tokens": 5275,
            "input_tokens_cache_write": 3914,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1637,
            "total_tokens": 5123,
            "input_tokens_cache_write": 3482,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1313,
            "total_tokens": 4552,
            "input_tokens_cache_write": 3235,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 14,
            "output_tokens": 1954,
            "total_tokens": 13279,
            "input_tokens_cache_write": 4026,
            "input_tokens_cache_read": 7285,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1375,
            "total_tokens": 4448,
            "input_tokens_cache_write": 3069,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 766,
            "total_tokens": 4863,
            "input_tokens_cache_write": 4093,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1523,
            "total_tokens": 5222,
            "input_tokens_cache_write": 3695,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2553,
            "total_tokens": 7586,
            "input_tokens_cache_write": 5029,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2202,
            "total_tokens": 7269,
            "input_tokens_cache_write": 5063,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 14,
            "output_tokens": 2131,
            "total_tokens": 13839,
            "input_tokens_cache_write": 4152,
            "input_tokens_cache_read": 7542,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2437,
            "total_tokens": 6116,
            "input_tokens_cache_write": 3675,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1757,
            "total_tokens": 6307,
            "input_tokens_cache_write": 4546,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2013,
            "total_tokens": 5790,
            "input_tokens_cache_write": 3773,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2357,
            "total_tokens": 6274,
            "input_tokens_cache_write": 3913,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1334,
            "total_tokens": 4461,
            "input_tokens_cache_write": 3123,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1476,
            "total_tokens": 5065,
            "input_tokens_cache_write": 3585,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 2323,
            "total_tokens": 6600,
            "input_tokens_cache_write": 4273,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 3556,
            "total_tokens": 8864,
            "input_tokens_cache_write": 5304,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1944,
            "total_tokens": 9219,
            "input_tokens_cache_write": 7271,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4,
            "output_tokens": 1776,
            "total_tokens": 5033,
            "input_tokens_cache_write": 3253,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.012877199999999998,
      0.0093922,
      0.0079772,
      0.007512199999999999,
      0.006526199999999999,
      0.0062482,
      0.0100672,
      0.0126872,
      0.006290199999999999,
      0.0155122,
      0.0079842,
      0.0219952,
      0.0086842,
      0.0117302,
      0.016238199999999998,
      0.0084642,
      0.0077892,
      0.0112522,
      0.0185272,
      0.01093092,
      0.0076092,
      0.0049902,
      0.0107822,
      0.020025799999999996,
      0.0189232,
      0.0105262,
      0.0094162,
      0.0079732,
      0.0107012,
      0.00588456,
      0.0055271999999999995,
      0.015202199999999999,
      0.0103432,
      0.0155442,
      0.0120252,
      0.0085422,
      0.0068332,
      0.0085092,
      0.0142122,
      0.0082202,
      0.0136092,
      0.0153086,
      0.0164042,
      0.0095902,
      0.0066451999999999995,
      0.0089292,
      0.007905200000000001,
      0.012111199999999999,
      0.0206922,
      0.01894732,
      0.0093452,
      0.010033199999999999,
      0.0084902,
      0.012436,
      0.008572199999999999,
      0.007160199999999999,
      0.009790199999999999,
      0.0152442,
      0.0138742,
      0.01329056,
      0.0134262,
      0.0115772,
      0.0118282,
      0.013344199999999999,
      0.0084622,
      0.0094922,
      0.013568199999999999,
      0.0195312,
      0.0150502,
      0.0103602
    ]
  },
  {
    "task_name": "CORE_Bench_Hard_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_with_stderr/accuracy",
        "value": 0.02857142857142857
      },
      {
        "name": "score_with_stderr/stderr",
        "value": 0.028571428571428574
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 6933,
            "output_tokens": 2952,
            "total_tokens": 96752,
            "input_tokens_cache_write": 31844,
            "input_tokens_cache_read": 55023,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2779,
            "output_tokens": 7173,
            "total_tokens": 419317,
            "input_tokens_cache_write": 31087,
            "input_tokens_cache_read": 378278,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 5376,
            "output_tokens": 1705,
            "total_tokens": 56264,
            "input_tokens_cache_write": 10089,
            "input_tokens_cache_read": 39094,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3618,
            "output_tokens": 1559,
            "total_tokens": 152492,
            "input_tokens_cache_write": 23141,
            "input_tokens_cache_read": 124174,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2887,
            "output_tokens": 2312,
            "total_tokens": 222341,
            "input_tokens_cache_write": 82014,
            "input_tokens_cache_read": 135128,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4328,
            "output_tokens": 6734,
            "total_tokens": 409396,
            "input_tokens_cache_write": 33031,
            "input_tokens_cache_read": 365303,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 5418,
            "output_tokens": 886,
            "total_tokens": 31339,
            "input_tokens_cache_write": 8566,
            "input_tokens_cache_read": 16469,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 812,
            "output_tokens": 2095,
            "total_tokens": 58238,
            "input_tokens_cache_write": 14322,
            "input_tokens_cache_read": 41009,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 6622,
            "output_tokens": 2085,
            "total_tokens": 139763,
            "input_tokens_cache_write": 69042,
            "input_tokens_cache_read": 62014,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4479,
            "output_tokens": 5118,
            "total_tokens": 221808,
            "input_tokens_cache_write": 51649,
            "input_tokens_cache_read": 160562,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 886,
            "output_tokens": 3919,
            "total_tokens": 98247,
            "input_tokens_cache_write": 15792,
            "input_tokens_cache_read": 77650,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 5828,
            "output_tokens": 1890,
            "total_tokens": 146343,
            "input_tokens_cache_write": 45442,
            "input_tokens_cache_read": 93183,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2796,
            "output_tokens": 1322,
            "total_tokens": 92551,
            "input_tokens_cache_write": 20877,
            "input_tokens_cache_read": 67556,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3256,
            "output_tokens": 2800,
            "total_tokens": 78868,
            "input_tokens_cache_write": 15801,
            "input_tokens_cache_read": 57011,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2853,
            "output_tokens": 2872,
            "total_tokens": 50757,
            "input_tokens_cache_write": 7963,
            "input_tokens_cache_read": 37069,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4093,
            "output_tokens": 2312,
            "total_tokens": 47912,
            "input_tokens_cache_write": 7110,
            "input_tokens_cache_read": 34397,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1767,
            "output_tokens": 1220,
            "total_tokens": 114131,
            "input_tokens_cache_write": 18211,
            "input_tokens_cache_read": 92933,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2671,
            "output_tokens": 1936,
            "total_tokens": 34945,
            "input_tokens_cache_write": 5975,
            "input_tokens_cache_read": 24363,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 5915,
            "output_tokens": 2225,
            "total_tokens": 103614,
            "input_tokens_cache_write": 11740,
            "input_tokens_cache_read": 83734,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3018,
            "output_tokens": 2757,
            "total_tokens": 112813,
            "input_tokens_cache_write": 25339,
            "input_tokens_cache_read": 81699,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 5676,
            "output_tokens": 2658,
            "total_tokens": 183111,
            "input_tokens_cache_write": 41176,
            "input_tokens_cache_read": 133601,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3863,
            "output_tokens": 2310,
            "total_tokens": 91079,
            "input_tokens_cache_write": 40979,
            "input_tokens_cache_read": 43927,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 6730,
            "output_tokens": 1219,
            "total_tokens": 85897,
            "input_tokens_cache_write": 77948,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4830,
            "output_tokens": 2709,
            "total_tokens": 101418,
            "input_tokens_cache_write": 86192,
            "input_tokens_cache_read": 7687,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3026,
            "output_tokens": 2279,
            "total_tokens": 252801,
            "input_tokens_cache_write": 128893,
            "input_tokens_cache_read": 118603,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4163,
            "output_tokens": 2868,
            "total_tokens": 180908,
            "input_tokens_cache_write": 36518,
            "input_tokens_cache_read": 137359,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1757,
            "output_tokens": 1560,
            "total_tokens": 156608,
            "input_tokens_cache_write": 97972,
            "input_tokens_cache_read": 55319,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2785,
            "output_tokens": 2520,
            "total_tokens": 288046,
            "input_tokens_cache_write": 230753,
            "input_tokens_cache_read": 51988,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4797,
            "output_tokens": 5895,
            "total_tokens": 183446,
            "input_tokens_cache_write": 26252,
            "input_tokens_cache_read": 146502,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3046,
            "output_tokens": 9339,
            "total_tokens": 1000387,
            "input_tokens_cache_write": 84902,
            "input_tokens_cache_read": 903100,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 6517,
            "output_tokens": 2524,
            "total_tokens": 86853,
            "input_tokens_cache_write": 22120,
            "input_tokens_cache_read": 55692,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2928,
            "output_tokens": 3010,
            "total_tokens": 349319,
            "input_tokens_cache_write": 214243,
            "input_tokens_cache_read": 129138,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4707,
            "output_tokens": 2155,
            "total_tokens": 141900,
            "input_tokens_cache_write": 42215,
            "input_tokens_cache_read": 92823,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 835,
            "output_tokens": 1703,
            "total_tokens": 79666,
            "input_tokens_cache_write": 11438,
            "input_tokens_cache_read": 65690,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2905,
            "output_tokens": 2422,
            "total_tokens": 86803,
            "input_tokens_cache_write": 59619,
            "input_tokens_cache_read": 21857,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.05360023999999999,
      0.09226443999999999,
      0.02433732,
      0.04220531999999999,
      0.10438184,
      0.10075044,
      0.01776192,
      0.026632320000000004,
      0.08764072,
      0.08854915999999999,
      0.038388799999999994,
      0.06511904,
      0.03380628,
      0.03416668,
      0.02469892,
      0.02238416,
      0.03193924,
      0.01780484,
      0.03207072,
      0.045317319999999994,
      0.06703688000000001,
      0.056823559999999995,
      0.088208,
      0.10150695999999998,
      0.14991804,
      0.06230912,
      0.11004312,
      0.24722003999999997,
      0.06538975999999999,
      0.2502724,
      0.04188496,
      0.23895644,
      0.06202644,
      0.0241732,
      0.07337955999999998
    ]
  },
  {
    "task_name": "DS_1000_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "ds1000_scorer/accuracy",
        "value": 0.61
      },
      {
        "name": "ds1000_scorer/stderr",
        "value": 0.04902071300001973
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2044,
            "output_tokens": 464,
            "total_tokens": 2508,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4230,
            "output_tokens": 378,
            "total_tokens": 4608,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2775,
            "output_tokens": 695,
            "total_tokens": 3470,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2475,
            "output_tokens": 576,
            "total_tokens": 3051,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2053,
            "output_tokens": 471,
            "total_tokens": 2524,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1271,
            "output_tokens": 1258,
            "total_tokens": 5037,
            "input_tokens_cache_write": 2508,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1902,
            "output_tokens": 483,
            "total_tokens": 2385,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1002,
            "output_tokens": 1527,
            "total_tokens": 13483,
            "input_tokens_cache_write": 5787,
            "input_tokens_cache_read": 5167,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2208,
            "output_tokens": 520,
            "total_tokens": 2728,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 718,
            "output_tokens": 506,
            "total_tokens": 4032,
            "input_tokens_cache_write": 2808,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1087,
            "output_tokens": 1578,
            "total_tokens": 5138,
            "input_tokens_cache_write": 2473,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2849,
            "output_tokens": 524,
            "total_tokens": 3373,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3912,
            "output_tokens": 664,
            "total_tokens": 4576,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1621,
            "output_tokens": 1596,
            "total_tokens": 15517,
            "input_tokens_cache_write": 6468,
            "input_tokens_cache_read": 5832,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1031,
            "output_tokens": 433,
            "total_tokens": 1464,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2693,
            "output_tokens": 1058,
            "total_tokens": 3751,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3691,
            "output_tokens": 948,
            "total_tokens": 4639,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2878,
            "output_tokens": 1088,
            "total_tokens": 6322,
            "input_tokens_cache_write": 2356,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2875,
            "output_tokens": 614,
            "total_tokens": 3489,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2555,
            "output_tokens": 538,
            "total_tokens": 3093,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1778,
            "output_tokens": 439,
            "total_tokens": 2217,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2439,
            "output_tokens": 738,
            "total_tokens": 3177,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1753,
            "output_tokens": 447,
            "total_tokens": 2200,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2206,
            "output_tokens": 471,
            "total_tokens": 2677,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3366,
            "output_tokens": 559,
            "total_tokens": 3925,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2550,
            "output_tokens": 655,
            "total_tokens": 3205,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2431,
            "output_tokens": 558,
            "total_tokens": 2989,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3141,
            "output_tokens": 2091,
            "total_tokens": 10932,
            "input_tokens_cache_write": 3135,
            "input_tokens_cache_read": 2565,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2666,
            "output_tokens": 474,
            "total_tokens": 3140,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2380,
            "output_tokens": 452,
            "total_tokens": 2832,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2383,
            "output_tokens": 611,
            "total_tokens": 2994,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2078,
            "output_tokens": 1476,
            "total_tokens": 23420,
            "input_tokens_cache_write": 7453,
            "input_tokens_cache_read": 12413,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4265,
            "output_tokens": 875,
            "total_tokens": 5140,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4356,
            "output_tokens": 1005,
            "total_tokens": 5361,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2507,
            "output_tokens": 685,
            "total_tokens": 3192,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 709,
            "output_tokens": 345,
            "total_tokens": 1054,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 6051,
            "output_tokens": 1676,
            "total_tokens": 15848,
            "input_tokens_cache_write": 3182,
            "input_tokens_cache_read": 4939,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2590,
            "output_tokens": 654,
            "total_tokens": 3244,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2385,
            "output_tokens": 747,
            "total_tokens": 3132,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2641,
            "output_tokens": 387,
            "total_tokens": 3028,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4122,
            "output_tokens": 734,
            "total_tokens": 4856,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2382,
            "output_tokens": 559,
            "total_tokens": 2941,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2901,
            "output_tokens": 566,
            "total_tokens": 3467,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1993,
            "output_tokens": 426,
            "total_tokens": 2419,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1901,
            "output_tokens": 460,
            "total_tokens": 2361,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2575,
            "output_tokens": 912,
            "total_tokens": 3487,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2093,
            "output_tokens": 560,
            "total_tokens": 2653,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2017,
            "output_tokens": 397,
            "total_tokens": 2414,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3206,
            "output_tokens": 434,
            "total_tokens": 3640,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 761,
            "output_tokens": 441,
            "total_tokens": 1202,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1578,
            "output_tokens": 1081,
            "total_tokens": 5215,
            "input_tokens_cache_write": 2556,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1918,
            "output_tokens": 389,
            "total_tokens": 2307,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2261,
            "output_tokens": 504,
            "total_tokens": 2765,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2220,
            "output_tokens": 469,
            "total_tokens": 2689,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2393,
            "output_tokens": 891,
            "total_tokens": 3284,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3164,
            "output_tokens": 902,
            "total_tokens": 4066,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2656,
            "output_tokens": 705,
            "total_tokens": 3361,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3858,
            "output_tokens": 729,
            "total_tokens": 4587,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4288,
            "output_tokens": 887,
            "total_tokens": 5175,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2811,
            "output_tokens": 663,
            "total_tokens": 3474,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2864,
            "output_tokens": 588,
            "total_tokens": 3452,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2345,
            "output_tokens": 812,
            "total_tokens": 3157,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2623,
            "output_tokens": 725,
            "total_tokens": 3348,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1832,
            "output_tokens": 514,
            "total_tokens": 2346,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3158,
            "output_tokens": 778,
            "total_tokens": 3936,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2068,
            "output_tokens": 484,
            "total_tokens": 2552,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2064,
            "output_tokens": 562,
            "total_tokens": 2626,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1784,
            "output_tokens": 396,
            "total_tokens": 2180,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2543,
            "output_tokens": 694,
            "total_tokens": 3237,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2809,
            "output_tokens": 2522,
            "total_tokens": 28987,
            "input_tokens_cache_write": 12758,
            "input_tokens_cache_read": 10898,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2299,
            "output_tokens": 721,
            "total_tokens": 3020,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 803,
            "output_tokens": 729,
            "total_tokens": 1532,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2446,
            "output_tokens": 659,
            "total_tokens": 3105,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2214,
            "output_tokens": 533,
            "total_tokens": 2747,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2030,
            "output_tokens": 654,
            "total_tokens": 2684,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4310,
            "output_tokens": 818,
            "total_tokens": 5128,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2485,
            "output_tokens": 510,
            "total_tokens": 2995,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2510,
            "output_tokens": 713,
            "total_tokens": 3223,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2135,
            "output_tokens": 480,
            "total_tokens": 2615,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2240,
            "output_tokens": 481,
            "total_tokens": 2721,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4325,
            "output_tokens": 793,
            "total_tokens": 5118,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2243,
            "output_tokens": 577,
            "total_tokens": 2820,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1969,
            "output_tokens": 453,
            "total_tokens": 2422,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1349,
            "output_tokens": 855,
            "total_tokens": 4309,
            "input_tokens_cache_write": 2105,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2289,
            "output_tokens": 585,
            "total_tokens": 2874,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2707,
            "output_tokens": 564,
            "total_tokens": 3271,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1469,
            "output_tokens": 683,
            "total_tokens": 2152,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 767,
            "output_tokens": 298,
            "total_tokens": 1065,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 973,
            "output_tokens": 622,
            "total_tokens": 1595,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3909,
            "output_tokens": 609,
            "total_tokens": 4518,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3025,
            "output_tokens": 611,
            "total_tokens": 3636,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1856,
            "output_tokens": 629,
            "total_tokens": 2485,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2515,
            "output_tokens": 1358,
            "total_tokens": 12258,
            "input_tokens_cache_write": 4382,
            "input_tokens_cache_read": 4003,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2351,
            "output_tokens": 575,
            "total_tokens": 2926,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1753,
            "output_tokens": 381,
            "total_tokens": 2134,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2616,
            "output_tokens": 1224,
            "total_tokens": 3840,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1812,
            "output_tokens": 325,
            "total_tokens": 2137,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1906,
            "output_tokens": 455,
            "total_tokens": 2361,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 4541,
            "output_tokens": 902,
            "total_tokens": 5443,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2412,
            "output_tokens": 665,
            "total_tokens": 3077,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.0034912,
      0.004895999999999999,
      0.004999999999999999,
      0.004284,
      0.0035263999999999994,
      0.0085568,
      0.0034536,
      0.01310996,
      0.0038464,
      0.0054064,
      0.0096546,
      0.004375199999999999,
      0.0057856,
      0.014615359999999997,
      0.0025567999999999997,
      0.0063863999999999995,
      0.0067448,
      0.0090104,
      0.004756,
      0.004196,
      0.0031783999999999996,
      0.0049032,
      0.0031904,
      0.0036488,
      0.0049288,
      0.00466,
      0.0041768,
      0.014217,
      0.004028800000000001,
      0.003712,
      0.0043504,
      0.01601244,
      0.006911999999999999,
      0.0075048,
      0.0047456,
      0.0019472,
      0.015121919999999999,
      0.0046879999999999995,
      0.004896,
      0.0036608,
      0.0062336,
      0.0041416,
      0.0045848,
      0.0032984,
      0.0033607999999999997,
      0.0057079999999999995,
      0.0039144,
      0.0032015999999999998,
      0.0043008,
      0.0023728,
      0.0081424,
      0.0030904,
      0.0038247999999999997,
      0.003652,
      0.0054784,
      0.0061392,
      0.004944799999999999,
      0.0060024,
      0.0069784,
      0.004900799999999999,
      0.0046432,
      0.005124,
      0.0049984,
      0.0035215999999999997,
      0.0056384,
      0.0035903999999999997,
      0.0038992,
      0.0030112,
      0.004810399999999999,
      0.025965039999999995,
      0.0047232,
      0.0035583999999999998,
      0.004592799999999999,
      0.0039032,
      0.00424,
      0.00672,
      0.004028,
      0.00486,
      0.0036279999999999997,
      0.0037159999999999997,
      0.006632,
      0.0041024,
      0.0033872,
      0.0066042,
      0.0041712,
      0.0044216,
      0.0039071999999999996,
      0.0018055999999999999,
      0.0032663999999999996,
      0.0055632,
      0.004864,
      0.0040008000000000005,
      0.012146239999999999,
      0.0041808,
      0.0029264,
      0.0069888,
      0.0027496,
      0.0033447999999999998,
      0.0072407999999999995,
      0.004589599999999999
    ]
  },
  {
    "task_name": "LitQA2_FullText_Search_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{\"with_search_tools\": true}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_paper_finder/recall_at_30",
        "value": 0.1
      },
      {
        "name": "score_paper_finder/stderr",
        "value": 0.09999999999999999
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 692,
            "total_tokens": 56902,
            "input_tokens_cache_write": 22434,
            "input_tokens_cache_read": 33766,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 443,
            "total_tokens": 61923,
            "input_tokens_cache_write": 25542,
            "input_tokens_cache_read": 35928,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 546,
            "total_tokens": 63611,
            "input_tokens_cache_write": 25955,
            "input_tokens_cache_read": 37100,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 7,
            "output_tokens": 530,
            "total_tokens": 15392,
            "input_tokens_cache_write": 6571,
            "input_tokens_cache_read": 8284,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 466,
            "total_tokens": 59936,
            "input_tokens_cache_write": 24121,
            "input_tokens_cache_read": 35339,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 729,
            "total_tokens": 32265,
            "input_tokens_cache_write": 12082,
            "input_tokens_cache_read": 19444,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 422,
            "total_tokens": 57260,
            "input_tokens_cache_write": 22773,
            "input_tokens_cache_read": 34055,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 588,
            "total_tokens": 142006,
            "input_tokens_cache_write": 59907,
            "input_tokens_cache_read": 81499,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 10,
            "output_tokens": 498,
            "total_tokens": 35790,
            "input_tokens_cache_write": 12076,
            "input_tokens_cache_read": 23206,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 12,
            "output_tokens": 506,
            "total_tokens": 77345,
            "input_tokens_cache_write": 21307,
            "input_tokens_cache_read": 55520,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.027911279999999997,
      0.03019624,
      0.031115000000000004,
      0.009359319999999999,
      0.028820119999999998,
      0.01656152,
      0.0271934,
      0.06878851999999999,
      0.01593248,
      0.027782199999999996
    ]
  },
  {
    "task_name": "DiscoveryBench_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_discoverybench/mean",
        "value": 0.22055555555555553
      },
      {
        "name": "score_discoverybench/stderr",
        "value": 0.06690639767154347
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 1394,
            "total_tokens": 21123,
            "input_tokens_cache_write": 7647,
            "input_tokens_cache_read": 12066,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 20,
            "output_tokens": 1772,
            "total_tokens": 32099,
            "input_tokens_cache_write": 8373,
            "input_tokens_cache_read": 21934,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3129,
            "output_tokens": 929,
            "total_tokens": 6722,
            "input_tokens_cache_write": 2664,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2841,
            "output_tokens": 987,
            "total_tokens": 6415,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 2587,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3080,
            "output_tokens": 927,
            "total_tokens": 6635,
            "input_tokens_cache_write": 2628,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3066,
            "output_tokens": 942,
            "total_tokens": 6627,
            "input_tokens_cache_write": 2619,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1479,
            "output_tokens": 906,
            "total_tokens": 7037,
            "input_tokens_cache_write": 2560,
            "input_tokens_cache_read": 2092,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1478,
            "output_tokens": 881,
            "total_tokens": 4480,
            "input_tokens_cache_write": 2121,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1503,
            "output_tokens": 1960,
            "total_tokens": 27585,
            "input_tokens_cache_write": 5590,
            "input_tokens_cache_read": 18532,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1471,
            "output_tokens": 863,
            "total_tokens": 4443,
            "input_tokens_cache_write": 2109,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1477,
            "output_tokens": 1169,
            "total_tokens": 5475,
            "input_tokens_cache_write": 2829,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1494,
            "output_tokens": 1015,
            "total_tokens": 4984,
            "input_tokens_cache_write": 2475,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1495,
            "output_tokens": 1051,
            "total_tokens": 5236,
            "input_tokens_cache_write": 2690,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3492,
            "output_tokens": 1342,
            "total_tokens": 9636,
            "input_tokens_cache_write": 2589,
            "input_tokens_cache_read": 2213,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1469,
            "output_tokens": 938,
            "total_tokens": 4857,
            "input_tokens_cache_write": 2450,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1488,
            "output_tokens": 1111,
            "total_tokens": 4899,
            "input_tokens_cache_write": 2300,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1068,
            "output_tokens": 911,
            "total_tokens": 4057,
            "input_tokens_cache_write": 2078,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2763,
            "output_tokens": 1426,
            "total_tokens": 7133,
            "input_tokens_cache_write": 2944,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3008,
            "output_tokens": 874,
            "total_tokens": 3882,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2889,
            "output_tokens": 666,
            "total_tokens": 3555,
            "input_tokens_cache_write": 0,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2991,
            "output_tokens": 1376,
            "total_tokens": 7059,
            "input_tokens_cache_write": 2692,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1065,
            "output_tokens": 919,
            "total_tokens": 4034,
            "input_tokens_cache_write": 2050,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3048,
            "output_tokens": 1122,
            "total_tokens": 7064,
            "input_tokens_cache_write": 2894,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 3051,
            "output_tokens": 1195,
            "total_tokens": 6754,
            "input_tokens_cache_write": 2508,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 1062,
            "output_tokens": 1117,
            "total_tokens": 4496,
            "input_tokens_cache_write": 2317,
            "input_tokens_cache_read": 0,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.01420108,
      0.01723172,
      0.008883199999999999,
      0.006427759999999999,
      0.0088,
      0.008839799999999998,
      0.00753456,
      0.006827400000000001,
      0.016114959999999998,
      0.0067377999999999995,
      0.0086866,
      0.0077302,
      0.00809,
      0.010927639999999999,
      0.0073771999999999996,
      0.0079344,
      0.0065764,
      0.0108584,
      0.0059023999999999995,
      0.0049752,
      0.010588799999999999,
      0.006578,
      0.0098204,
      0.0097288,
      0.0076346
    ]
  },
  {
    "task_name": "E2E_Bench_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.08802308802308803
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.03245724277116994
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 28,
            "output_tokens": 7886,
            "total_tokens": 36104,
            "input_tokens_cache_write": 7732,
            "input_tokens_cache_read": 20458,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 49,
            "output_tokens": 14886,
            "total_tokens": 245920,
            "input_tokens_cache_write": 127588,
            "input_tokens_cache_read": 103397,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 2600,
            "total_tokens": 18064,
            "input_tokens_cache_write": 5465,
            "input_tokens_cache_read": 9981,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 28,
            "output_tokens": 4711,
            "total_tokens": 29851,
            "input_tokens_cache_write": 6471,
            "input_tokens_cache_read": 18641,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 4458,
            "total_tokens": 22571,
            "input_tokens_cache_write": 7991,
            "input_tokens_cache_read": 10104,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 23,
            "output_tokens": 5098,
            "total_tokens": 26588,
            "input_tokens_cache_write": 6853,
            "input_tokens_cache_read": 14614,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 23,
            "output_tokens": 4593,
            "total_tokens": 26000,
            "input_tokens_cache_write": 6995,
            "input_tokens_cache_read": 14389,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 33,
            "output_tokens": 4250,
            "total_tokens": 27122,
            "input_tokens_cache_write": 4796,
            "input_tokens_cache_read": 18043,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 2062,
            "output_tokens": 3320,
            "total_tokens": 28840,
            "input_tokens_cache_write": 11012,
            "input_tokens_cache_read": 12446,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 23,
            "output_tokens": 3023,
            "total_tokens": 20309,
            "input_tokens_cache_write": 4726,
            "input_tokens_cache_read": 12537,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.04093504,
      0.19544295999999997,
      0.01667788,
      0.026828679999999997,
      0.026645719999999998,
      0.02843252,
      0.026536519999999997,
      0.02326584,
      0.02693728,
      0.01783936
    ]
  },
  {
    "task_name": "E2E_Bench_Hard_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.05769230769230769
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.03520132967557451
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 28,
            "output_tokens": 7409,
            "total_tokens": 38701,
            "input_tokens_cache_write": 8171,
            "input_tokens_cache_read": 23093,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 53,
            "output_tokens": 4217,
            "total_tokens": 64594,
            "input_tokens_cache_write": 8001,
            "input_tokens_cache_read": 52323,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 26,
            "output_tokens": 5545,
            "total_tokens": 58534,
            "input_tokens_cache_write": 14153,
            "input_tokens_cache_read": 38810,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 41,
            "output_tokens": 10366,
            "total_tokens": 107711,
            "input_tokens_cache_write": 26162,
            "input_tokens_cache_read": 71142,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 18,
            "output_tokens": 3720,
            "total_tokens": 22327,
            "input_tokens_cache_write": 5904,
            "input_tokens_cache_read": 12685,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 48,
            "output_tokens": 10958,
            "total_tokens": 144016,
            "input_tokens_cache_write": 28409,
            "input_tokens_cache_read": 104601,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 3429,
            "total_tokens": 16983,
            "input_tokens_cache_write": 5588,
            "input_tokens_cache_read": 7953,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 4431,
            "total_tokens": 19693,
            "input_tokens_cache_write": 7059,
            "input_tokens_cache_read": 8190,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 33,
            "output_tokens": 6888,
            "total_tokens": 60560,
            "input_tokens_cache_write": 11822,
            "input_tokens_cache_read": 41817,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 23,
            "output_tokens": 4873,
            "total_tokens": 54453,
            "input_tokens_cache_write": 16764,
            "input_tokens_cache_read": 32793,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.03967684,
      0.029097239999999996,
      0.039458599999999996,
      0.07335016,
      0.021813199999999998,
      0.08064748000000001,
      0.01995064,
      0.025448600000000002,
      0.04274576,
      0.03889784
    ]
  },
  {
    "task_name": "LitQA2_FullText_validation",
    "eval_spec": {
      "solver": "agent_baselines/solvers/react/basic_agent.py@instantiated_basic_agent",
      "solver_args": "{\"max_steps\": 100, \"with_report_editor\": 0, \"with_search_tools\": 0, \"with_stateful_python\": 0, \"with_table_editor\": 0, \"with_thinking_tool\": 0}",
      "model": "anthropic/claude-3-5-haiku-20241022",
      "model_args": "{}",
      "task_args": "{\"litqa_args\": [], \"litqa_kwargs\": {}}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/agent-baselines.git",
        "commit": "898367f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_litqa2/astabench/precision",
        "value": 0.6666666666666666
      },
      {
        "name": "score_litqa2/astabench/coverage",
        "value": 0.6
      },
      {
        "name": "is_correct/accuracy",
        "value": 0.4
      },
      {
        "name": "is_correct/stderr",
        "value": 0.16329931618554522
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 496,
            "total_tokens": 36586,
            "input_tokens_cache_write": 12830,
            "input_tokens_cache_read": 23244,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 358,
            "total_tokens": 22796,
            "input_tokens_cache_write": 6048,
            "input_tokens_cache_read": 16377,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 9,
            "output_tokens": 338,
            "total_tokens": 32411,
            "input_tokens_cache_write": 27859,
            "input_tokens_cache_read": 4205,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 400,
            "total_tokens": 19540,
            "input_tokens_cache_write": 3992,
            "input_tokens_cache_read": 15135,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 13,
            "output_tokens": 369,
            "total_tokens": 27746,
            "input_tokens_cache_write": 14022,
            "input_tokens_cache_read": 13342,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 9,
            "output_tokens": 374,
            "total_tokens": 31938,
            "input_tokens_cache_write": 23392,
            "input_tokens_cache_read": 8163,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 563,
            "total_tokens": 60227,
            "input_tokens_cache_write": 30987,
            "input_tokens_cache_read": 28661,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 9,
            "output_tokens": 464,
            "total_tokens": 13568,
            "input_tokens_cache_write": 8903,
            "input_tokens_cache_read": 4192,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 9,
            "output_tokens": 423,
            "total_tokens": 29683,
            "input_tokens_cache_write": 21095,
            "input_tokens_cache_read": 8156,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-3-5-haiku-20241022",
          "usage": {
            "input_tokens": 16,
            "output_tokens": 740,
            "total_tokens": 89665,
            "input_tokens_cache_write": 43806,
            "input_tokens_cache_read": 45103,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.01668632,
      0.00880056,
      0.0295546,
      0.006813199999999999,
      0.016575760000000002,
      0.02554824,
      0.035544679999999995,
      0.01110156,
      0.02344668,
      0.050387039999999994
    ]
  }
] | 
	{
  "submit_time": "2025-08-26T06:45:15.968802Z",
  "username": "Ai2",
  "agent_name": "ReAct",
  "agent_description": "Simple ReAct agent using built-in LLM tool-calling. This variant uses Anthropic Claude 3.5 Haiku as the base model.",
  "agent_url": "https://github.com/allenai/agent-baselines",
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0/validation/miked-ai_ReAct_2025-08-26T06-45-15",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Standard"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "E2E_Bench_Hard_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/codescientist/codescientist_cached.py@codescientist_cached_solver",
      "solver_args": "{}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "52aeb97"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.5513869463869464
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.03947954041934493
      }
    ],
    "model_usages": [
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 540557,
            "output_tokens": 97124,
            "total_tokens": 637681,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 417540,
            "output_tokens": 72348,
            "total_tokens": 489888,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 278295,
            "output_tokens": 58621,
            "total_tokens": 336916,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 674488,
            "output_tokens": 131804,
            "total_tokens": 806292,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 867919,
            "output_tokens": 201933,
            "total_tokens": 1069852,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 157569,
            "output_tokens": 25728,
            "total_tokens": 183297,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 492367,
            "output_tokens": 123474,
            "total_tokens": 615841,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 866423,
            "output_tokens": 169792,
            "total_tokens": 1036215,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 280497,
            "output_tokens": 63881,
            "total_tokens": 344378,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 466881,
            "output_tokens": 103393,
            "total_tokens": 570274,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      3.078531,
      2.33784,
      1.7142,
      4.000524,
      5.632752,
      0.858627,
      3.329211,
      5.146149,
      1.799706,
      2.951538
    ]
  },
  {
    "task_name": "E2E_Bench_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/codescientist/codescientist_cached.py@codescientist_cached_solver",
      "solver_args": "{}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "52aeb97"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.5128102453102453
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.10005590594924381
      }
    ],
    "model_usages": [
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 204613,
            "output_tokens": 41330,
            "total_tokens": 245943,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 1719381,
            "output_tokens": 339408,
            "total_tokens": 2058789,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 314427,
            "output_tokens": 46719,
            "total_tokens": 361146,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 486175,
            "output_tokens": 71972,
            "total_tokens": 558147,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 295133,
            "output_tokens": 59863,
            "total_tokens": 354996,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 464306,
            "output_tokens": 86149,
            "total_tokens": 550455,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 297609,
            "output_tokens": 45422,
            "total_tokens": 343031,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 1603155,
            "output_tokens": 310461,
            "total_tokens": 1913616,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 314184,
            "output_tokens": 47371,
            "total_tokens": 361555,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-3-7-sonnet-20250219",
          "usage": {
            "input_tokens": 353966,
            "output_tokens": 64999,
            "total_tokens": 418965,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      1.233789,
      10.249263,
      1.644066,
      2.538105,
      1.783344,
      2.685153,
      1.574157,
      9.466380000000001,
      1.653117,
      2.036883
    ]
  }
] | 
	{
  "submit_time": "2025-07-10T18:12:16.960575Z",
  "username": "Ai2",
  "agent_name": "Asta CodeScientist",
  "agent_description": null,
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_CodeScientist_2025-07-10T18-12-16",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Fully custom"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "DiscoveryBench_validation",
    "eval_spec": {
      "solver": "astabench/solvers/datavoyager/agent.py@datavoyager_solver",
      "solver_args": "{\"config_file\": \"astabench/solvers/datavoyager/dv_core/config/datavoyager_modal_deployment_magentic_one_config_20250617_o3.yaml\"}",
      "model": "openai/o3",
      "model_args": "{}",
      "task_args": "{}",
      "revision": null,
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_discoverybench/mean",
        "value": 0.20766666666666667
      },
      {
        "name": "score_discoverybench/stderr",
        "value": 0.05925837962310103
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 102807,
            "output_tokens": 10931,
            "total_tokens": 113738,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 157529,
            "output_tokens": 11709,
            "total_tokens": 169238,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 586169,
            "output_tokens": 64426,
            "total_tokens": 650595,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 51946,
            "output_tokens": 7287,
            "total_tokens": 59233,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 27470,
            "output_tokens": 6595,
            "total_tokens": 34065,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 53912,
            "output_tokens": 7522,
            "total_tokens": 61434,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 24063,
            "output_tokens": 4108,
            "total_tokens": 28171,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 22585,
            "output_tokens": 4583,
            "total_tokens": 27168,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 43863,
            "output_tokens": 10145,
            "total_tokens": 54008,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 34423,
            "output_tokens": 5965,
            "total_tokens": 40388,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 38940,
            "output_tokens": 7659,
            "total_tokens": 46599,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 240328,
            "output_tokens": 32955,
            "total_tokens": 273283,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 49240,
            "output_tokens": 7105,
            "total_tokens": 56345,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 121201,
            "output_tokens": 21533,
            "total_tokens": 142734,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 25115,
            "output_tokens": 4386,
            "total_tokens": 29501,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 24243,
            "output_tokens": 4342,
            "total_tokens": 28585,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 35819,
            "output_tokens": 6970,
            "total_tokens": 42789,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 28079,
            "output_tokens": 6073,
            "total_tokens": 34152,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 24297,
            "output_tokens": 5013,
            "total_tokens": 29310,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 20390,
            "output_tokens": 3843,
            "total_tokens": 24233,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 20825,
            "output_tokens": 4935,
            "total_tokens": 25760,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 20246,
            "output_tokens": 5028,
            "total_tokens": 25274,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 23177,
            "output_tokens": 5543,
            "total_tokens": 28720,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 28424,
            "output_tokens": 5570,
            "total_tokens": 33994,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/o3",
          "usage": {
            "input_tokens": 25934,
            "output_tokens": 6061,
            "total_tokens": 31995,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.293062,
      0.40873,
      1.6877459999999997,
      0.162188,
      0.10769999999999999,
      0.16799999999999998,
      0.08098999999999999,
      0.08183399999999999,
      0.16888599999999998,
      0.11656599999999999,
      0.139152,
      0.744296,
      0.15531999999999999,
      0.414666,
      0.085318,
      0.08322199999999999,
      0.12739799999999998,
      0.104742,
      0.088698,
      0.07152399999999999,
      0.08113000000000001,
      0.080716,
      0.090698,
      0.101408,
      0.100356
    ]
  }
] | 
	{
  "submit_time": "2025-08-14T04:16:47.886330Z",
  "username": "Ai2",
  "agent_name": "Asta DataVoyager",
  "agent_description": "o3",
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_DataVoyager_2025-08-14T04-16-47",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "DiscoveryBench_validation",
    "eval_spec": {
      "solver": "astabench/solvers/datavoyager/agent.py@datavoyager_solver",
      "solver_args": "{\"config_file\": \"astabench/solvers/datavoyager/dv_core/config/datavoyager_modal_deployment_magentic_one_config_20250617_sonnet_4.yaml\"}",
      "model": "anthropic/claude-sonnet-4",
      "model_args": "{}",
      "task_args": "{}",
      "revision": null,
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_discoverybench/mean",
        "value": 0.23424486624486623
      },
      {
        "name": "score_discoverybench/stderr",
        "value": 0.06863816396339442
      }
    ],
    "model_usages": [
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 281607,
            "output_tokens": 15624,
            "total_tokens": 297231,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 375233,
            "output_tokens": 15165,
            "total_tokens": 390398,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 107407,
            "output_tokens": 10776,
            "total_tokens": 118183,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 140661,
            "output_tokens": 10750,
            "total_tokens": 151411,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 224732,
            "output_tokens": 13612,
            "total_tokens": 238344,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 88276,
            "output_tokens": 9672,
            "total_tokens": 97948,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 80563,
            "output_tokens": 7030,
            "total_tokens": 87593,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 45416,
            "output_tokens": 6544,
            "total_tokens": 51960,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 64224,
            "output_tokens": 6636,
            "total_tokens": 70860,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 40905,
            "output_tokens": 5151,
            "total_tokens": 46056,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 219354,
            "output_tokens": 15883,
            "total_tokens": 235237,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 60318,
            "output_tokens": 6928,
            "total_tokens": 67246,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 141870,
            "output_tokens": 12666,
            "total_tokens": 154536,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 131567,
            "output_tokens": 12748,
            "total_tokens": 144315,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 60084,
            "output_tokens": 6355,
            "total_tokens": 66439,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 53105,
            "output_tokens": 6850,
            "total_tokens": 59955,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 77448,
            "output_tokens": 9446,
            "total_tokens": 86894,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 62480,
            "output_tokens": 9125,
            "total_tokens": 71605,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 60573,
            "output_tokens": 9888,
            "total_tokens": 70461,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 67121,
            "output_tokens": 7963,
            "total_tokens": 75084,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 120951,
            "output_tokens": 10734,
            "total_tokens": 131685,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 49993,
            "output_tokens": 7969,
            "total_tokens": 57962,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 73847,
            "output_tokens": 11058,
            "total_tokens": 84905,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 83525,
            "output_tokens": 9691,
            "total_tokens": 93216,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 53757,
            "output_tokens": 9813,
            "total_tokens": 63570,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      1.0791810000000002,
      1.353174,
      0.48386100000000004,
      0.583233,
      0.878376,
      0.40990800000000005,
      0.34713900000000003,
      0.234408,
      0.292212,
      0.19998,
      0.8963070000000001,
      0.284874,
      0.6155999999999999,
      0.585921,
      0.275577,
      0.262065,
      0.374034,
      0.324315,
      0.33003899999999997,
      0.32080800000000004,
      0.5238630000000001,
      0.26951400000000003,
      0.38741100000000006,
      0.39593999999999996,
      0.308466
    ]
  }
] | 
	{
  "submit_time": "2025-08-14T19:32:30.253858Z",
  "username": "Ai2",
  "agent_name": "Asta DataVoyager",
  "agent_description": "claude-sonnet-4",
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_DataVoyager_2025-08-14T19-32-30",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "DiscoveryBench_validation",
    "eval_spec": {
      "solver": "astabench/solvers/datavoyager/agent.py@datavoyager_solver",
      "solver_args": "{\"config_file\": \"astabench/solvers/datavoyager/dv_core/config/datavoyager_modal_deployment_magentic_one_config_20250617_gpt5_minimal.yaml\"}",
      "model": "openai/gpt-5",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/openlocus/asta-bench.git",
        "commit": "ae8b47f"
      },
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_discoverybench/mean",
        "value": 0.20444444444444443
      },
      {
        "name": "score_discoverybench/stderr",
        "value": 0.06078701940706636
      }
    ],
    "model_usages": [
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 132545,
            "output_tokens": 13830,
            "total_tokens": 146375,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 200069,
            "output_tokens": 23179,
            "total_tokens": 223248,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 103723,
            "output_tokens": 16478,
            "total_tokens": 120201,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 116019,
            "output_tokens": 16784,
            "total_tokens": 132803,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 789629,
            "output_tokens": 42124,
            "total_tokens": 831753,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 38490,
            "output_tokens": 10413,
            "total_tokens": 48903,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 34256,
            "output_tokens": 6637,
            "total_tokens": 40893,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 53934,
            "output_tokens": 13038,
            "total_tokens": 66972,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 31610,
            "output_tokens": 6734,
            "total_tokens": 38344,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 61598,
            "output_tokens": 15816,
            "total_tokens": 77414,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 59625,
            "output_tokens": 8562,
            "total_tokens": 68187,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 79244,
            "output_tokens": 13714,
            "total_tokens": 92958,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 74873,
            "output_tokens": 12085,
            "total_tokens": 86958,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 47965,
            "output_tokens": 5864,
            "total_tokens": 53829,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 62679,
            "output_tokens": 9112,
            "total_tokens": 71791,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 17066,
            "output_tokens": 4590,
            "total_tokens": 21656,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 71580,
            "output_tokens": 19006,
            "total_tokens": 90586,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 45178,
            "output_tokens": 9997,
            "total_tokens": 55175,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 32561,
            "output_tokens": 8504,
            "total_tokens": 41065,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 46984,
            "output_tokens": 8259,
            "total_tokens": 55243,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 34930,
            "output_tokens": 9222,
            "total_tokens": 44152,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 25696,
            "output_tokens": 7795,
            "total_tokens": 33491,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 62512,
            "output_tokens": 7342,
            "total_tokens": 69854,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 35140,
            "output_tokens": 7943,
            "total_tokens": 43083,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-5",
          "usage": {
            "input_tokens": 33609,
            "output_tokens": 9083,
            "total_tokens": 42692,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.30398125,
      0.48187625,
      0.29443375000000005,
      0.31286375,
      1.4082762500000001,
      0.1522425,
      0.10919000000000001,
      0.19779750000000001,
      0.10685250000000002,
      0.23515750000000002,
      0.16015125000000002,
      0.23619500000000002,
      0.21444125000000003,
      0.11859625000000001,
      0.16946875,
      0.0672325,
      0.27953500000000003,
      0.1564425,
      0.12574125,
      0.14132,
      0.13588250000000002,
      0.11007,
      0.15156,
      0.123355,
      0.13284125000000002
    ]
  }
] | 
	{
  "submit_time": "2025-08-14T19:42:20.164747Z",
  "username": "Ai2",
  "agent_name": "Asta DataVoyager",
  "agent_description": "gpt-5/minimal_reasoning",
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_DataVoyager_2025-08-14T19-42-20",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "DiscoveryBench_validation",
    "eval_spec": {
      "solver": "astabench/solvers/datavoyager/agent.py@datavoyager_solver",
      "solver_args": "{\"config_file\": \"astabench/solvers/datavoyager/dv_core/config/datavoyager_modal_deployment_magentic_one_config_20250617_gpt-41.yaml\"}",
      "model": "openai/gpt-4.1",
      "model_args": "{}",
      "task_args": "{}",
      "revision": null,
      "packages": "{\"inspect_ai\": \"0.3.114\"}"
    },
    "metrics": [
      {
        "name": "score_discoverybench/mean",
        "value": 0.254
      },
      {
        "name": "score_discoverybench/stderr",
        "value": 0.07399499482572629
      }
    ],
    "model_usages": [
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 134348,
            "output_tokens": 5427,
            "total_tokens": 139775,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 95790,
            "output_tokens": 4467,
            "total_tokens": 100257,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 559011,
            "output_tokens": 80687,
            "total_tokens": 639698,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 18485,
            "output_tokens": 2768,
            "total_tokens": 21253,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 24862,
            "output_tokens": 4014,
            "total_tokens": 28876,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 82973,
            "output_tokens": 7049,
            "total_tokens": 90022,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 31545,
            "output_tokens": 2709,
            "total_tokens": 34254,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 26437,
            "output_tokens": 2999,
            "total_tokens": 29436,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 29629,
            "output_tokens": 2937,
            "total_tokens": 32566,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 25561,
            "output_tokens": 3154,
            "total_tokens": 28715,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 25925,
            "output_tokens": 2705,
            "total_tokens": 28630,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 31575,
            "output_tokens": 2822,
            "total_tokens": 34397,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 29464,
            "output_tokens": 3662,
            "total_tokens": 33126,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 30249,
            "output_tokens": 2923,
            "total_tokens": 33172,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 39180,
            "output_tokens": 3140,
            "total_tokens": 42320,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 27918,
            "output_tokens": 2929,
            "total_tokens": 30847,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 23518,
            "output_tokens": 3058,
            "total_tokens": 26576,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 22706,
            "output_tokens": 3243,
            "total_tokens": 25949,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 38259,
            "output_tokens": 5505,
            "total_tokens": 43764,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 21096,
            "output_tokens": 3338,
            "total_tokens": 24434,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 21001,
            "output_tokens": 2744,
            "total_tokens": 23745,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 27384,
            "output_tokens": 3821,
            "total_tokens": 31205,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 34066,
            "output_tokens": 3662,
            "total_tokens": 37728,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 31185,
            "output_tokens": 4047,
            "total_tokens": 35232,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "gpt-4.1",
          "usage": {
            "input_tokens": 21920,
            "output_tokens": 2799,
            "total_tokens": 24719,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        },
        {
          "model": "openai/gpt-4o",
          "usage": {
            "input_tokens": 0,
            "output_tokens": 0,
            "total_tokens": 0,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.312112,
      0.227316,
      1.763518,
      0.059114,
      0.08183599999999999,
      0.22233799999999998,
      0.08476199999999999,
      0.07686599999999999,
      0.082754,
      0.076354,
      0.07349,
      0.085726,
      0.088224,
      0.083882,
      0.10348,
      0.07926799999999999,
      0.0715,
      0.071356,
      0.120558,
      0.068896,
      0.063954,
      0.085336,
      0.097428,
      0.094746,
      0.066232
    ]
  }
] | 
	{
  "submit_time": "2025-08-14T21:44:09.404096Z",
  "username": "Ai2",
  "agent_name": "Asta DataVoyager",
  "agent_description": "gpt-4.1",
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_DataVoyager_2025-08-14T21-44-09",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Custom interface"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "E2E_Bench_Hard_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/autoasta/autoasta_cached.py@autoasta_cached_solver",
      "solver_args": "{\"model\": \"claude-sonnet-4-20250514\"}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "c45780e"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.6551864801864802
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.04073052321238961
      }
    ],
    "model_usages": [
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3280995,
            "output_tokens": 33825,
            "total_tokens": 3314820,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3521468,
            "output_tokens": 36304,
            "total_tokens": 3557772,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 4010432,
            "output_tokens": 41345,
            "total_tokens": 4051777,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3812115,
            "output_tokens": 39300,
            "total_tokens": 3851415,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3236062,
            "output_tokens": 33361,
            "total_tokens": 3269423,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 4453517,
            "output_tokens": 45913,
            "total_tokens": 4499430,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3447692,
            "output_tokens": 35543,
            "total_tokens": 3483235,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 4509323,
            "output_tokens": 46488,
            "total_tokens": 4555811,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 4629942,
            "output_tokens": 47731,
            "total_tokens": 4677673,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 4926602,
            "output_tokens": 50790,
            "total_tokens": 4977392,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      10.35036,
      11.108964,
      12.651471,
      12.025845,
      10.208601,
      14.049246,
      10.876221,
      14.225289,
      14.605791000000002,
      15.541656000000001
    ]
  },
  {
    "task_name": "E2E_Bench_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/autoasta/autoasta_cached.py@autoasta_cached_solver",
      "solver_args": "{\"model\": \"claude-sonnet-4-20250514\"}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "c45780e"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.7012157287157288
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.04558707785029943
      }
    ],
    "model_usages": [
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3552370,
            "output_tokens": 36622,
            "total_tokens": 3588992,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 2491874,
            "output_tokens": 25689,
            "total_tokens": 2517563,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3466598,
            "output_tokens": 35738,
            "total_tokens": 3502336,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3938399,
            "output_tokens": 40602,
            "total_tokens": 3979001,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3937066,
            "output_tokens": 40588,
            "total_tokens": 3977654,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 2915908,
            "output_tokens": 30061,
            "total_tokens": 2945969,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3173130,
            "output_tokens": 32713,
            "total_tokens": 3205843,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3078577,
            "output_tokens": 31738,
            "total_tokens": 3110315,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3910341,
            "output_tokens": 40313,
            "total_tokens": 3950654,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "anthropic/claude-sonnet-4-20250514",
          "usage": {
            "input_tokens": 3810555,
            "output_tokens": 39284,
            "total_tokens": 3849839,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      11.206439999999999,
      7.860957000000001,
      10.935864,
      12.424227,
      12.420018,
      9.198639,
      10.010085,
      9.711801,
      12.335718,
      12.020925
    ]
  }
] | 
	{
  "submit_time": "2025-08-03T22:39:22.125992Z",
  "username": "Ai2",
  "agent_name": "Asta Panda",
  "agent_description": "v1.4.7-Claude4",
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_Panda_2025-08-03T22-39-22",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Fully custom"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "E2E_Bench_Hard_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/autoasta/autoasta_cached.py@autoasta_cached_solver",
      "solver_args": "{\"model\": \"gpt-4.1\"}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "c45780e"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.39762237762237757
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.07140205972509524
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 8474086,
            "output_tokens": 50441,
            "total_tokens": 8524527,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 3164674,
            "output_tokens": 18837,
            "total_tokens": 3183511,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 3608594,
            "output_tokens": 21480,
            "total_tokens": 3630074,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 3166107,
            "output_tokens": 18846,
            "total_tokens": 3184953,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2923433,
            "output_tokens": 17401,
            "total_tokens": 2940834,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2984588,
            "output_tokens": 17765,
            "total_tokens": 3002353,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 4034821,
            "output_tokens": 24017,
            "total_tokens": 4058838,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 4218543,
            "output_tokens": 25110,
            "total_tokens": 4243653,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 4557428,
            "output_tokens": 27128,
            "total_tokens": 4584556,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 3450999,
            "output_tokens": 20542,
            "total_tokens": 3471541,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      17.3517,
      6.4800439999999995,
      7.389028,
      6.482982,
      5.9860739999999995,
      6.111296,
      8.261778,
      8.637965999999999,
      9.33188,
      7.0663339999999994
    ]
  },
  {
    "task_name": "E2E_Bench_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/autoasta/autoasta_cached.py@autoasta_cached_solver",
      "solver_args": "{\"model\": \"gpt-4.1\"}",
      "model": "openai/gpt-4o-mini",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "c45780e"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.4430050505050505
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.07155502223526382
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1702109,
            "output_tokens": 10132,
            "total_tokens": 1712241,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2834355,
            "output_tokens": 16871,
            "total_tokens": 2851226,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 6078476,
            "output_tokens": 36181,
            "total_tokens": 6114657,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 13280246,
            "output_tokens": 79049,
            "total_tokens": 13359295,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 9897267,
            "output_tokens": 58912,
            "total_tokens": 9956179,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1707756,
            "output_tokens": 10165,
            "total_tokens": 1717921,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1894741,
            "output_tokens": 11278,
            "total_tokens": 1906019,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1687937,
            "output_tokens": 10047,
            "total_tokens": 1697984,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2009082,
            "output_tokens": 11959,
            "total_tokens": 2021041,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2093867,
            "output_tokens": 12463,
            "total_tokens": 2106330,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      3.4852739999999995,
      5.803678,
      12.446399999999999,
      27.192884,
      20.265829999999998,
      3.4968319999999995,
      3.879706,
      3.45625,
      4.113836,
      4.287438
    ]
  }
] | 
	{
  "submit_time": "2025-08-04T03:31:48.698423Z",
  "username": "Ai2",
  "agent_name": "Asta Panda",
  "agent_description": "v1.4.7-gpt41",
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Asta_Panda_2025-08-04T03-31-48",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Fully custom"
} | 
| 
	{
  "name": "asta-bench",
  "version": "1.0.0",
  "splits": [
    {
      "name": "validation",
      "tasks": [
        {
          "name": "ArxivDIGESTables_Clean_validation",
          "path": "astabench/arxivdigestables_validation",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_validation",
          "path": "astabench/sqa_dev",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_validation",
          "path": "astabench/litqa2_validation",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "PaperFindingBench_validation",
          "path": "astabench/paper_finder_validation",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_validation",
          "path": "astabench/paper_finder_litqa2_validation",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_validation",
          "path": "astabench/discoverybench_validation",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_validation",
          "path": "astabench/core_bench_validation",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_validation",
          "path": "astabench/ds1000_validation",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_validation",
          "path": "astabench/e2e_discovery_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_validation",
          "path": "astabench/e2e_discovery_hard_validation",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_validation",
          "path": "astabench/super_validation",
          "primary_metric": "entrypoint/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_validation",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_validation",
          "weight": 0.5
        }
      ]
    },
    {
      "name": "test",
      "tasks": [
        {
          "name": "PaperFindingBench_test",
          "path": "astabench/paper_finder_test",
          "primary_metric": "score_paper_finder/adjusted_f1_micro_avg",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_Search_test",
          "path": "astabench/paper_finder_litqa2_test",
          "primary_metric": "score_paper_finder/recall_at_30",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ScholarQA_CS2_test",
          "path": "astabench/sqa_test",
          "primary_metric": "global_avg/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "ArxivDIGESTables_Clean_test",
          "path": "astabench/arxivdigestables_test",
          "primary_metric": "score_tables/mean",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "LitQA2_FullText_test",
          "path": "astabench/litqa2_test",
          "primary_metric": "is_correct/accuracy",
          "tags": [
            "lit"
          ]
        },
        {
          "name": "DiscoveryBench_test",
          "path": "astabench/discoverybench_test",
          "primary_metric": "score_discoverybench/mean",
          "tags": [
            "data"
          ]
        },
        {
          "name": "CORE_Bench_Hard_test",
          "path": "astabench/core_bench_test",
          "primary_metric": "score_with_stderr/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "DS_1000_test",
          "path": "astabench/ds1000_test",
          "primary_metric": "ds1000_scorer/accuracy",
          "tags": [
            "code"
          ]
        },
        {
          "name": "E2E_Bench_test",
          "path": "astabench/e2e_discovery_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "E2E_Bench_Hard_test",
          "path": "astabench/e2e_discovery_hard_test",
          "primary_metric": "score_rubric/accuracy",
          "tags": [
            "discovery"
          ]
        },
        {
          "name": "SUPER_Expert_test",
          "path": "astabench/super_test",
          "primary_metric": "output_match/mean",
          "tags": [
            "code"
          ]
        }
      ],
      "macro_average_weight_adjustments": [
        {
          "tag": "lit",
          "task": "LitQA2_FullText_test",
          "weight": 0.5
        },
        {
          "tag": "lit",
          "task": "LitQA2_FullText_Search_test",
          "weight": 0.5
        }
      ]
    }
  ]
} | 
	validation | 
	[
  {
    "task_name": "E2E_Bench_Hard_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/faker/faker_cached.py@faker_cached_solver",
      "solver_args": "{}",
      "model": "openai/gpt-4.1",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "52aeb97"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.17221445221445225
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.033189482458013564
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2376,
            "output_tokens": 2794,
            "total_tokens": 5170,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2443,
            "output_tokens": 2275,
            "total_tokens": 4718,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2571,
            "output_tokens": 2361,
            "total_tokens": 4932,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2548,
            "output_tokens": 3730,
            "total_tokens": 6278,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2417,
            "output_tokens": 2982,
            "total_tokens": 5399,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2654,
            "output_tokens": 2904,
            "total_tokens": 5558,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2362,
            "output_tokens": 2643,
            "total_tokens": 5005,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2658,
            "output_tokens": 3255,
            "total_tokens": 5913,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2682,
            "output_tokens": 2830,
            "total_tokens": 5512,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 2197,
            "output_tokens": 3095,
            "total_tokens": 5292,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.027104,
      0.023086000000000002,
      0.02403,
      0.034935999999999995,
      0.02869,
      0.02854,
      0.025868,
      0.031355999999999995,
      0.028004,
      0.029153999999999996
    ]
  },
  {
    "task_name": "E2E_Bench_validation",
    "eval_spec": {
      "solver": "astabench/evals/e2e_discovery/solvers/faker/faker_cached.py@faker_cached_solver",
      "solver_args": "{}",
      "model": "openai/gpt-4.1",
      "model_args": "{}",
      "task_args": "{}",
      "revision": {
        "type": "git",
        "origin": "https://github.com/allenai/asta-bench.git",
        "commit": "52aeb97"
      },
      "packages": "{\"inspect_ai\": \"0.3.105\"}"
    },
    "metrics": [
      {
        "name": "score_rubric/accuracy",
        "value": 0.4008874458874459
      },
      {
        "name": "score_rubric/stderr",
        "value": 0.0676127442019462
      }
    ],
    "model_usages": [
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1476,
            "output_tokens": 2525,
            "total_tokens": 4001,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1454,
            "output_tokens": 3311,
            "total_tokens": 4765,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1599,
            "output_tokens": 3101,
            "total_tokens": 4700,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1546,
            "output_tokens": 3101,
            "total_tokens": 4647,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1450,
            "output_tokens": 3983,
            "total_tokens": 5433,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1656,
            "output_tokens": 2929,
            "total_tokens": 4585,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1381,
            "output_tokens": 2564,
            "total_tokens": 3945,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1416,
            "output_tokens": 2490,
            "total_tokens": 3906,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1397,
            "output_tokens": 2169,
            "total_tokens": 3566,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ],
      [
        {
          "model": "openai/gpt-4.1",
          "usage": {
            "input_tokens": 1779,
            "output_tokens": 2614,
            "total_tokens": 4393,
            "input_tokens_cache_write": null,
            "input_tokens_cache_read": null,
            "reasoning_tokens": null
          }
        }
      ]
    ],
    "model_costs": [
      0.023152,
      0.029396,
      0.028006,
      0.0279,
      0.034763999999999996,
      0.026743999999999997,
      0.023274,
      0.022752,
      0.020146,
      0.02447
    ]
  }
] | 
	{
  "submit_time": "2025-07-10T18:14:56.434410Z",
  "username": "Ai2",
  "agent_name": "Faker",
  "agent_description": null,
  "agent_url": null,
  "logs_url": null,
  "logs_url_public": "hf://datasets/allenai/asta-bench-submissions/1.0.0-dev1/validation/pclark425_Faker_2025-07-10T18-14-56",
  "summary_url": null,
  "openness": "Open source & closed weights",
  "tool_usage": "Standard"
} | 
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.