菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-21
📄 Abstract - Orchestrating LLM Agents for Scientific Research: A Pilot Study of Multiple Choice Question (MCQ) Generation and Evaluation

Advances in large language models (LLMs) are rapidly transforming scientific work, yet empirical evidence on how these systems reshape research activities remains limited. We report a mixed-methods pilot evaluation of an AI-orchestrated research workflow in which a human researcher coordinated multiple LLM-based agents to perform data extraction, corpus construction, artifact generation, and artifact evaluation. Using the generation and assessment of multiple-choice questions (MCQs) as a testbed, we collected 1,071 SAT Math MCQs and employed LLM agents to extract questions from PDFs, retrieve and convert open textbooks into structured representations, align each MCQ with relevant textbook content, generate new MCQs under specified difficulty and cognitive levels, and evaluate both original and generated MCQs using a 24-criterion quality framework. Across all evaluations, average MCQ quality was high. However, criterion-level analysis and equivalence testing show that generated MCQs are not fully comparable to expert-vetted baseline questions. Strict similarity (24/24 criteria equivalent) was never achieved. Persistent gaps concentrated in skill\ depth, cognitive engagement, difficulty calibration, and metadata alignment, while surface-level qualities, such as {grammar fluency}, {clarity options}, {no duplicates}, were consistently strong. Beyond MCQ outcomes, the study documents a labor shift. The researcher's work moved from ``authoring items'' toward {specification, orchestration, verification}, and {governance}. Formalizing constraints, designing rubrics, building validation loops, recovering from tool failures, and auditing provenance constituted the primary activities. We discuss implications for the future of scientific work, including emerging ``AI research operations'' skills required for AI-empowered research pipelines.

顶级标签: llm agents model evaluation
详细标签: multi-agent systems workflow orchestration question generation automated evaluation scientific research 或 搜索:

为科学研究编排大型语言模型智能体:一项关于多项选择题生成与评估的试点研究 / Orchestrating LLM Agents for Scientific Research: A Pilot Study of Multiple Choice Question (MCQ) Generation and Evaluation


1️⃣ 一句话总结

这项试点研究表明,通过人类研究者协调多个大型语言模型智能体,可以构建一个自动化生成和评估多项选择题的AI研究流程,虽然生成的题目在表面质量上表现优异,但在深度技能和认知参与度等方面仍与专家题目存在差距,同时研究者的角色也从直接创作转向了对整个AI工作流程的规范、编排与监督。

源自 arXiv: 2602.18891