Skip to content

[ACL 2025 main] SeedBench: A Multi-task Benchmark for Evaluating Large Language Models in Seed Science🌾

License

Notifications You must be signed in to change notification settings

open-sciencelab/SeedBench

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

51 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SeedBench: A Multi-task Benchmark for Evaluating Large Language Models in Seed Science

SeedBench is the first multi-task benchmark designed to evaluate large language models (LLMs) in seed science, focusing on seed breeding. This repository includes the dataset, evaluation code, and documentation to support research in this domain. Here is the usage.


🌾 Overview

SeedBench assesses LLMs across three core seed breeding stages:

  • Gene Information Retrieval
  • Gene Function and Regulation Analysis
  • Variety Breeding with Agronomic Trait Optimization

Breeding Workflow
Breeding Expert Workflow Framework

Built with domain experts, SeedBench features 2,264 expert-validated questions across 11 task types and 10 subcategories, initially targeting rice breeding. Future updates will include other crops like maize, soybean, and wheat.

🔎 Dataset Details

  • Corpus: 308,727 publications cleaned to 1.1 billion tokens; 279 segments from 113 documents.

  • Questions: 2,264 across 11 task types, bilingual (English/Chinese), expert-validated.

  • Focus: Rice breeding as a representative case.

    Types and metrics:

    Type ID Question Type Metric Count
    Q&A
    QA-1 Multiple Choice Accuracy 200
    QA-2 Multiple Answer Macro-F1 187
    QA-3 Fill-in-the-Blank ROUGE-L 224
    QA-4 Generation ROUGE-L 242
    Summarization
    SUM-1 Simple Summarization ROUGE-L 225
    SUM-2 Key Information Extraction ROUGE-L 225
    Reading Comprehension
    RC-1 Multiple Choice Accuracy 113
    RC-2 Multiple Answer Macro-F1 108
    RC-3 Fill-in-the-Blank ROUGE-L 221
    RC-4 Generation ROUGE-L 240
    RC-5 Subcategory Classification Accuracy 279

    Taxonomy Distribution:

    Taxonomy Distribution

☀️ Key Results

We evaluated 26 LLMs, including proprietary, open-source, and domain-specific models. Highlights:

Performance by Question Type

  • Top Performers: DeepSeek-V3 (68.37), GPT-4 (67.88).

    Proprietary LLM Radar

    Open-Source LLM Radar

Performance by Task Types

Model QA-1 QA-2 QA-3 QA-4 SUM-1 SUM-2 RC-1 RC-2 RC-3 RC-4 RC-5 Avg
GPT-4 60.50 73.87 21.35 36.07 58.73 62.89 100.00 96.44 87.86 62.29 86.74 67.88
DeepSeek-V3 72.50 79.84 29.29 40.63 48.06 54.67 100.00 97.22 87.89 55.19 86.74 68.37
Qwen2-72B 59.50 75.98 19.55 31.62 31.08 63.09 99.12 94.24 72.20 51.58 89.96 62.54

Performance by Subcategory

Model C1 C2 C3 C4 C5 C6 C7 C8 C9 C10 Avg
GPT-4 59.59 60.55 76.32 61.16 56.34 59.35 63.67 64.74 60.65 67.66 62.06
DeepSeek-V3-671B 56.03 62.42 74.81 63.17 55.23 58.84 68.23 69.04 66.46 68.48 63.30
Qwen2-72B 51.16 58.10 74.07 59.72 51.58 57.76 58.85 61.63 56.69 59.11 57.62
  • Top Performers: DeepSeek-V3-671B (63.30), GPT-4 (62.06).

🐝 Repository Contents

  • base_model_eval/: Used to test base models without dialogue capabilities, i.e., evaluating performance after pretraining.
  • sft_model_eval/: Used to test SFT (Supervised Fine-Tuning) models, with a total of 2,264 questions covering 10 subcategories (see Fig 2).
    • one-shot/: Organized by 11 task types (see Tab 1).
    • zero-shot/: Organized by 11 task types (see Tab 1).
  • corpus/: 279 high-quality text segments and low-quality questions discarded after expert validation.
  • README.md: This file.

📬 Cite

Open an issue on this repository for questions or contributions.

@inproceedings{ying-etal-2025-seedbench,
  title = "{S}eed{B}ench: A Multi-task Benchmark for Evaluating Large Language Models in Seed Science",
  author = "Ying, Jie  and
    Chen, Zihong  and
    Wang, Zhefan  and
    Jiang, Wanli  and
    Wang, Chenyang  and
    Yuan, Zhonghang  and
    Su, Haoyang  and
    Kong, Huanjun  and
    Yang, Fan  and
    Dong, Nanqing",
  editor = "Che, Wanxiang  and
    Nabende, Joyce  and
    Shutova, Ekaterina  and
    Pilehvar, Mohammad Taher",
  booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
  month = jul,
  year = "2025",
  address = "Vienna, Austria",
  publisher = "Association for Computational Linguistics",
  url = "https://aclanthology.org/2025.acl-long.1516/",
  pages = "31395--31449",
  ISBN = "979-8-89176-251-0",
  abstract = "Seed science is essential for modern agriculture, directly influencing crop yields and global food security. However, challenges such as interdisciplinary complexity and high costs with limited returns hinder progress, leading to a shortage of experts and insufficient technological support. While large language models (LLMs) have shown promise across various fields, their application in seed science remains limited due to the scarcity of digital resources, complex gene-trait relationships, and the lack of standardized benchmarks. To address this gap, we introduce SeedBench{---}the first multi-task benchmark specifically designed for seed science. Developed in collaboration with domain experts, SeedBench focuses on seed breeding and simulates key aspects of modern breeding processes. We conduct a comprehensive evaluation of 26 leading LLMs, encompassing proprietary, open-source, and domain-specific fine-tuned models. Our findings not only highlight the substantial gaps between the power of LLMs and the real-world seed science problems, but also make a foundational step for research on LLMs for seed design."
}

About

[ACL 2025 main] SeedBench: A Multi-task Benchmark for Evaluating Large Language Models in Seed Science🌾

Topics

Resources

License

Stars

Watchers

Forks

Contributors 5

Languages