中华医学教育杂志 ›› 2024, Vol. 44 ›› Issue (8): 561-569.DOI: 10.3760/cma.j.cn115259-20240524-00520

• 专论 •    下一篇

生成式大语言模型在医学考试题库建设中的实践探索

江哲涵1, 奉世聪2, 王维民1   

  1. 1北京大学医学教育研究所,北京 100191;
    2北京大学教育学院2023级医学教育专业硕士研究生,北京 100871
  • 收稿日期:2024-05-24 出版日期:2024-08-01 发布日期:2024-07-31
  • 通讯作者: 王维民, Email: wwm@bjmu.edu.cn
  • 基金资助:
    国家卫生健康委员会人才交流服务中心项目(202110-335);国家自然科学基金委员会青年科学基金项目(72104006);国家卫生健康委员会国家医学考试中心“十四五”改革重点项目(2022-21)

Exploratory practice of generative large language models in the construction of medical item banks

Jiang Zhehan1, Feng Shicong2, Wang Weimin1   

  1. 1Institute of Medical Education, Peking University, Beijing 100191, China;
    2Master Degree Candidate, Medical Education Major, Enrolled in 2023, Graduate School of Education, Peking University, Beijing 100871, China
  • Received:2024-05-24 Online:2024-08-01 Published:2024-07-31
  • Contact: Wang Weimin, Email: wwm@bjmu.edu.cn
  • Supported by:
    Project of the Health Human Resources Development Center, National Health Commission, P.R.China(202110-335);National Natural Science Foundation Youth Project(72104006);Key Reform Project of the National Medical Examinations Centre, National Health Commission, P.R.China during the 14th Five-Year Plan Period(2022-21)

摘要: 传统的医学考试题库建设耗时长且依赖于命题专家资源,而大语言模型为题库建设带来了新方式,其试题生成质量很大程度上取决于提示词的设计。为了提高医学试题质量,帮助医学教师有效利用大语言模型开展命题工作,本文介绍了大语言模型中常用的提示工程,并以“术后胆漏”试题生成为例,探索了零样本、少样本、思维链、自洽性思维链、思维树提示工程策略的命题效果。分析结果显示,零样本和少样本提示操作简便,但在试题多样性和深度上存在一定局限。通过增加思维成分的提示策略,可以引导大语言模型执行草稿、打磨、比较和确定等命题过程,从而提高试题质量。同时,虽然通过改进提示词可以有效提高命题效果,但其具体实施与设计仍有极大的挖掘空间,需要进一步的研究和探索。

关键词: 人工智能, 生成式大语言模型, 提示工程, 医学试题, 题库建设, 考试命题

Abstract: Item development in healthcare profession education is time-consuming and heavily reliant on content experts. While large language models (LLMs) introduce a new approach to reduce the burdens, the output quality is largely contingent upon the prompt. This article aims to guide educators in effectively leveraging LLMs for item development, enhancing the quality through prompt engineering. Using ″postoperative bile leakage″ as an example, the paper demonstrates the effectiveness of various prompt engineering strategies, including Zero-shot, Few-shot, Chain of Thought (CoT), CoT with Self-Consistency (CoT-SC), and Tree of Thoughts (ToT). It is found that while Zero-shot and Few-shot methods are straightforward, they have certain limitations in terms of item diversity and depth. Conversely, prompt strategies incorporating ″Thought″ elements can navigate the LLMs through stages of drafting, refining, comparing, and finalizing, thereby elevating question quality. Although refining prompts indeed leads to notable improvements in question formulation efficacy, there remains substantial room for exploring and optimizing prompt formulations and strategies to further augment the quality of generated questions. The pursuit of advancing prompt engineering techniques holds the promise of significantly elevating the standards of question bank development within medical education.

Key words: Artificial intelligence, Generative large language models, Prompt engineering, Medical test questions, Item bank construction, Assessment development

中图分类号: