<?xml version="1.0" encoding="iso-8859-1" standalone="no"?>
<!DOCTYPE GmsArticle SYSTEM "http://www.egms.de/dtd/2.0.34/GmsArticle.dtd">
<GmsArticle xmlns:xlink="http://www.w3.org/1999/xlink">
  <MetaData>
    <Identifier>mibe000287</Identifier>
    <IdentifierDoi>10.3205/mibe000287</IdentifierDoi>
    <IdentifierUrn>urn:nbn:de:0183-mibe0002878</IdentifierUrn>
    <ArticleType>Research Article</ArticleType>
    <TitleGroup>
      <Title language="en">Performance evaluation of Meditron3-70B in medical coding: Current limitations and integration perspectives for clinical practice</Title>
      <TitleTranslated language="de">Evaluation von Meditron3-70B f&#252;r die medizinische Kodierung: Derzeitige Einschr&#228;nkungen und Perspektiven f&#252;r eine Integration in die klinische Praxis</TitleTranslated>
    </TitleGroup>
    <CreatorList>
      <Creator>
        <PersonNames>
          <Lastname>Galland-Decker</Lastname>
          <LastnameHeading>Galland-Decker</LastnameHeading>
          <Firstname>Coralie</Firstname>
          <Initials>C</Initials>
        </PersonNames>
        <Address>Medical Informatics, Lausanne University Hospital, 1011 Lausanne, Switzerland<Affiliation>Medical informatics, Lausanne University Hospital, Lausanne, Switzerland</Affiliation></Address>
        <Email>Coralie.Galland&#64;chuv.ch</Email>
        <Creatorrole corresponding="yes" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Usenbacher</Lastname>
          <LastnameHeading>Usenbacher</LastnameHeading>
          <Firstname>Muaziza</Firstname>
          <Initials>M</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Medical informatics, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Nunes</Lastname>
          <LastnameHeading>Nunes</LastnameHeading>
          <Firstname>Christophe</Firstname>
          <Initials>C</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Medical informatics, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Bouche</Lastname>
          <LastnameHeading>Bouche</LastnameHeading>
          <Firstname>Fran&#231;ois</Firstname>
          <Initials>F</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Medical informatics, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Carra</Lastname>
          <LastnameHeading>Carra</LastnameHeading>
          <Firstname>Giorgia</Firstname>
          <Initials>G</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Biomedical Data Science Center, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
          <Affiliation>Infectious Diseases Service, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Boillat-Blanco</Lastname>
          <LastnameHeading>Boillat-Blanco</LastnameHeading>
          <Firstname>No&#233;mie</Firstname>
          <Initials>N</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Infectious Diseases Service, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Hartley</Lastname>
          <LastnameHeading>Hartley</LastnameHeading>
          <Firstname>Mary-Anne</Firstname>
          <Initials>MA</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Laboratory for Intelligent Global Health and Humanitarian Response Technologies (LiGHT), EPFL, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Sallinen</Lastname>
          <LastnameHeading>Sallinen</LastnameHeading>
          <Firstname>Alexandre</Firstname>
          <Initials>A</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Laboratory for Intelligent Global Health and Humanitarian Response Technologies (LiGHT), EPFL, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Raisaro</Lastname>
          <LastnameHeading>Raisaro</LastnameHeading>
          <Firstname>Jean Louis</Firstname>
          <Initials>JL</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Biomedical Data Science Center, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
      <Creator>
        <PersonNames>
          <Lastname>Bastardot</Lastname>
          <LastnameHeading>Bastardot</LastnameHeading>
          <Firstname>Fran&#231;ois</Firstname>
          <Initials>F</Initials>
        </PersonNames>
        <Address>
          <Affiliation>Medical informatics, Lausanne University Hospital, Lausanne, Switzerland</Affiliation>
        </Address>
        <Creatorrole corresponding="no" presenting="no">author</Creatorrole>
      </Creator>
    </CreatorList>
    <PublisherList>
      <Publisher>
        <Corporation>
          <Corporatename>German Medical Science GMS Publishing House</Corporatename>
        </Corporation>
        <Address>D&#252;sseldorf</Address>
      </Publisher>
    </PublisherList>
    <SubjectGroup>
      <SubjectheadingDDB>610</SubjectheadingDDB>
      <Keyword language="en">large language models</Keyword>
      <Keyword language="en">emergency room visits</Keyword>
      <Keyword language="en">clinical coding</Keyword>
      <Keyword language="en">SNOMED CT</Keyword>
      <Keyword language="en">ICD-10</Keyword>
      <Keyword language="de">Large Language Models</Keyword>
      <Keyword language="de">Notaufnahme</Keyword>
      <Keyword language="de">klinische Kodierung</Keyword>
      <Keyword language="de">SNOMED CT</Keyword>
      <Keyword language="de">ICD-10</Keyword>
      <SectionHeading language="en">EFMI STC 2025</SectionHeading>
    </SubjectGroup>
    <DatePublishedList>
      <DatePublished>20251017</DatePublished>
    </DatePublishedList>
    <Language>engl</Language>
    <License license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
      <AltText language="en">This is an Open Access article distributed under the terms of the Creative Commons Attribution 4.0 License.</AltText>
      <AltText language="de">Dieser Artikel ist ein Open-Access-Artikel und steht unter den Lizenzbedingungen der Creative Commons Attribution 4.0 License (Namensnennung).</AltText>
    </License>
    <SourceGroup>
      <Journal>
        <ISSN>1860-9171</ISSN>
        <Volume>21</Volume>
        <JournalTitle>GMS Medizinische Informatik, Biometrie und Epidemiologie</JournalTitle>
        <JournalTitleAbbr>GMS Med Inform Biom Epidemiol</JournalTitleAbbr>
      </Journal>
    </SourceGroup>
    <ArticleNo>15</ArticleNo>
  </MetaData>
  <OrigData>
    <Abstract language="de" linked="yes"><Pgraph>Das Aufkommen von Large Language Models (LLMs) stellt eine Herausforderung f&#252;r deren Integration in die klinische Praxis, insbesondere f&#252;r die medizinische Kodierung, dar. Diese Studie evaluierte die Leistung von Meditron3-70B, einem aktuellen Open-Source-LLM f&#252;r den medizinischen Bereich, bei der Erzeugung von SNOMED CT- und ICD-10-Codes anhand von 200 fiktiven Konsultationsvignetten aus der Notaufnahme. Experten aus der Medizin bewerteten die Genauigkeit der Ergebnisse. Obwohl Meditron bei Standard-Benchmarks wie MedQA gute Ergebnisse erzielte, wurden erhebliche M&#228;ngel hinsichtlich der Relevanz und Vollst&#228;ndigkeit der erzeugten Diagnosecodes festgestellt, wobei nur 2&#37; der Antworten als akzeptabel eingesch&#228;tzt wurden. LLMs sind zwar vielversprechend f&#252;r die Unterst&#252;tzung der klinischen Entscheidungsfindung, ihre derzeitige F&#228;higkeit, genaue und umfassende medizinische Codes zu erstellen, ist jedoch noch begrenzt. Die Integration spezialisierter Retrieval-Tools durch hybride Ans&#228;tze k&#246;nnte die Kodierungsgenauigkeit verbessern und rechtfertigt weitere Untersuchungen in realen klinischen Umgebungen.</Pgraph></Abstract>
    <Abstract language="en" linked="yes"><Pgraph>The emergence of large language models (LLMs) presents challenges for their integration into clinical practice, particularly for medical coding. This study evaluated the performance of Meditron3-70B, a state-of-the-art open-source medical LLM, in generating SNOMED CT and ICD-10 codes using 200 fictitious emergency department consultation vignettes. Healthcare professionals assessed the accuracy of the outputs. Although Meditron performed well on standard benchmarks such as MedQA, significant shortcomings were observed in the relevance and completeness of the generated diagnostic codes, with only 2&#37; of responses deemed acceptable. While LLMs show promise for supporting clinical decision-making, their current ability to produce accurate and comprehensive medical codes remains limited. Integrating specialized retrieval tools through hybrid approaches could enhance coding accuracy and warrants further investigation in real-world clinical settings.</Pgraph></Abstract>
    <TextBlock name="Introduction" linked="yes">
      <MainHeadline>Introduction</MainHeadline><Pgraph>Medical coding involves converting clinical information from electronic health records (EHRs) &#8211; often unstructured free text &#8211; into standardized codes according to established classification systems. This process is crucial for administrative and public health purposes, such as statistical reporting, reimbursement, and epidemiological surveillance <TextLink reference="1"></TextLink>. However, it imposes a significant documentary burden on health providers, thereby contributing to professional fatigue and dissatisfaction. Meditron3-70B <TextLink reference="2"></TextLink> is a large language model (LLM) specifically fine-tuned on various biomedical and clinical datasets, aiming to support various healthcare-related natural language processing tasks.</Pgraph><Pgraph>Our study investigates a central research question: <Mark2>How well does Meditron3-70B perform in real-world medical coding tasks in SNOMED CT</Mark2> <TextLink reference="3"></TextLink> <Mark2>and ICD-10</Mark2> <TextLink reference="4"></TextLink><Mark2>,</Mark2> <Mark2>based on emergency department (ED) anamnesis&#63;</Mark2> We focus on its ability to assign specific, accurate codes aligned with current classification standards.  </Pgraph></TextBlock>
    <TextBlock name="Methodology" linked="yes">
      <MainHeadline>Methodology</MainHeadline><Pgraph>This study was conducted at Lausanne University Hospital (CHUV) between September and December 2024. We generated 200 fictitious clinical vignettes reflecting common presenting complaints in the emergency department. Each vignette simulated a pre-admission scenario, and the model was prompted to assign specific SNOMED CT and ICD-10 codes, per upcoming requirements for administrative coding of entry diagnoses used to determine reimbursement categories and care package allocations in the Swiss outpatient system. The outputs were evaluated by two physicians and three nurses (from internal medicine, pediatrics, psychiatry, and emergency medicine) using a 9-item evaluation grid based on a 5-point Likert scale.</Pgraph><Pgraph>To elicit the model&#8217;s output, a standardized prompt was used (Figure 1 <ImgLink imgNo="1" imgType="figure" />). </Pgraph></TextBlock>
    <TextBlock name="Results" linked="yes">
      <MainHeadline>Results</MainHeadline><Pgraph>The evaluation of the model&#8217;s performance in identifying SNOMED CT and ICD-10 codes revealed heterogeneous results across the assessment criteria (Figure 2 <ImgLink imgNo="2" imgType="figure" />).</Pgraph><Pgraph>The model demonstrated significant shortcomings in the relevance and completeness of its responses, with only 2&#37; rated as acceptable (4&#8211;5 points on the Likert scale). Confidence in the model was moderate, with only 19&#37; of responses considered satisfactory (4&#8211;5 points). In contrast, the model performed well in question understanding (76&#37;) and contextual awareness (86&#37;). Finally, it achieved excellent results regarding fairness (100&#37;) and absence of harm (98&#37;). </Pgraph></TextBlock>
    <TextBlock name="Discussion" linked="yes">
      <MainHeadline>Discussion</MainHeadline><Pgraph>Our results show that the model struggles to produce relevant and complete diagnostic codes based solely on patient anamnesis despite a good general understanding of the clinical questions and context. Medical coding is a complex task that requires surface-level comprehension, nuanced clinical reasoning and the ability to synthesize information. The limited quality of the generated codes explains the moderate confidence reported by healthcare professionals, and highlights a key barrier integrating LLMs in real-world coding workflows.</Pgraph><Pgraph>These findings are consistent with recent studies, which also report that current LLMs (e.g. ChatGPT-4.5), often fall short in tasks requiring high precision and domain-specific reasoning <TextLink reference="5"></TextLink>. While prompt engineering can help clarify expectations, it does not sufficiently compensate for the model&#8217;s limited access to up-to-date medical knowledge. Hybrid approaches such as Retrieval-Augmented Generation (RAG), which enable dynamic access to curated external sources during inference, appear particularly promising <TextLink reference="6"></TextLink>, <TextLink reference="7"></TextLink>, <TextLink reference="8"></TextLink>. They could help improve the specificity and accuracy of generated codes and better align model outputs with clinical documentation requirements.</Pgraph><Pgraph>Systematic comparisons with general-purpose models such as ChatGPT are needed to better characterize the strengths and limitations of specialized versus broadly trained language models. This study also highlights ethical and legal concerns inherent to generative AI in clinical settings. These include transparency of model outputs, accountability for errors or omissions, data privacy, and bias mitigation.</Pgraph></TextBlock>
    <TextBlock name="Conclusion" linked="yes">
      <MainHeadline>Conclusion</MainHeadline><Pgraph>Meditron3-70B showed apparent limitations in generating relevant and comprehensive diagnostic codes from emergency department anamnesis alone. These shortcomings, consistent with other recent findings, suggest that current LLMs are not yet reliable for standalone use in complex medical coding tasks. Future research should focus on hybrid systems that combine LLMs with structured retrieval tools to enhance performance and increase trust in AI-assisted documentation within clinical settings. </Pgraph></TextBlock>
    <TextBlock name="Notes" linked="yes">
      <MainHeadline>Notes</MainHeadline><SubHeadline>Authors&#8217; ORCIDs</SubHeadline><Pgraph><UnorderedList><ListItem level="1">Coralie Galland-Decker: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0000-0001-8897-8473">0000-0001-8897-8473</Hyperlink></ListItem><ListItem level="1">Giorgia Carra: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0000-0001-8002-224X">0000-0001-8002-224X</Hyperlink></ListItem><ListItem level="1">No&#233;mie Boillat-Blanco: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0000-0002-2490-8174">0000-0002-2490-8174</Hyperlink></ListItem><ListItem level="1">Mary-Anne Hartley: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0000-0002-8826-3870">0000-0002-8826-3870</Hyperlink></ListItem><ListItem level="1">Alexandre Sallinen: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0009-0005-1776-8539">0009-0005-1776-8539</Hyperlink></ListItem><ListItem level="1">Jean-Louis Raisaro: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0000-0003-2052-6133">0000-0003-2052-6133</Hyperlink></ListItem><ListItem level="1">Fran&#231;ois Bastardot: <Hyperlink href="https:&#47;&#47;orcid.org&#47;0000-0003-4060-0353">0000-0003-4060-0353</Hyperlink></ListItem></UnorderedList></Pgraph><SubHeadline>Competing interests</SubHeadline><Pgraph>The authors declare that they have no competing interests.</Pgraph></TextBlock>
    <References linked="yes">
      <Reference refNo="1">
        <RefAuthor>Dong H</RefAuthor>
        <RefAuthor>Falis M</RefAuthor>
        <RefAuthor>Whiteley W</RefAuthor>
        <RefAuthor>Alex B</RefAuthor>
        <RefAuthor>Matterson J</RefAuthor>
        <RefAuthor>Ji S</RefAuthor>
        <RefAuthor>Chen J</RefAuthor>
        <RefAuthor>Wu H</RefAuthor>
        <RefTitle>Automated clinical coding: what, why, and where we are&#63;</RefTitle>
        <RefYear>2022</RefYear>
        <RefJournal>NPJ Digit Med</RefJournal>
        <RefPage>159</RefPage>
        <RefTotal>Dong H, Falis M, Whiteley W, Alex B, Matterson J, Ji S, Chen J, Wu H. Automated clinical coding: what, why, and where we are&#63; NPJ Digit Med. 2022 Oct;5(1):159. DOI: 10.1038&#47;s41746-022-00705-7</RefTotal>
        <RefLink>https:&#47;&#47;doi.org&#47;10.1038&#47;s41746-022-00705-7</RefLink>
      </Reference>
      <Reference refNo="2">
        <RefAuthor>Chen Z</RefAuthor>
        <RefAuthor>Cano AH</RefAuthor>
        <RefAuthor>Romanou A</RefAuthor>
        <RefAuthor>Bonnet A</RefAuthor>
        <RefAuthor>Matoba K</RefAuthor>
        <RefAuthor>Salvi F</RefAuthor>
        <RefAuthor>Pagliardini M</RefAuthor>
        <RefAuthor>Fan S</RefAuthor>
        <RefAuthor>K&#246;pf A</RefAuthor>
        <RefAuthor>Mohtashami A</RefAuthor>
        <RefAuthor>Sallinen A</RefAuthor>
        <RefAuthor>Sakhaeirad A</RefAuthor>
        <RefAuthor>Swamy V</RefAuthor>
        <RefAuthor>Krawczuk I</RefAuthor>
        <RefAuthor>Bayazit D</RefAuthor>
        <RefAuthor>Marmet A</RefAuthor>
        <RefAuthor>Montariol S</RefAuthor>
        <RefAuthor>Hartley MA</RefAuthor>
        <RefAuthor>Jaggi M</RefAuthor>
        <RefAuthor>Antoine Bosselut A</RefAuthor>
        <RefTitle>MEDITRON-70B: Scaling Medical Pretraining for Large Language Models &#91;Preprint&#93;</RefTitle>
        <RefYear>2023</RefYear>
        <RefJournal>arXiv</RefJournal>
        <RefPage></RefPage>
        <RefTotal>Chen Z, Cano AH, Romanou A, Bonnet A, Matoba K, Salvi F, Pagliardini M, Fan S, K&#246;pf A, Mohtashami A, Sallinen A, Sakhaeirad A, Swamy V, Krawczuk I, Bayazit D, Marmet A, Montariol S, Hartley MA, Jaggi M, Antoine Bosselut A. MEDITRON-70B: Scaling Medical Pretraining for Large Language Models &#91;Preprint&#93;. arXiv. 2023 Nov 27. DOI: 10.48550&#47;arXiv.2311.16079</RefTotal>
        <RefLink>https:&#47;&#47;doi.org&#47;10.48550&#47;arXiv.2311.16079</RefLink>
      </Reference>
      <Reference refNo="5">
        <RefAuthor>Soroush A</RefAuthor>
        <RefAuthor>Glicksberg BS</RefAuthor>
        <RefAuthor>Zimlichman E</RefAuthor>
        <RefAuthor>Barash Y</RefAuthor>
        <RefAuthor>Freeman R</RefAuthor>
        <RefAuthor>Charney AW</RefAuthor>
        <RefAuthor>Nadkarni GN</RefAuthor>
        <RefAuthor>Klang E</RefAuthor>
        <RefTitle>Large Language Models Are Poor Medical Coders &#8211; Benchmarking of Medical Code Querying</RefTitle>
        <RefYear>2024</RefYear>
        <RefJournal>NEJM AI</RefJournal>
        <RefPage>AIdbp2300040</RefPage>
        <RefTotal>Soroush A, Glicksberg BS, Zimlichman E, Barash Y, Freeman R, Charney AW, Nadkarni GN, Klang E. Large Language Models Are Poor Medical Coders &#8211; Benchmarking of Medical Code Querying. NEJM AI. 2024;1(5):AIdbp2300040. 
DOI: 10.1056&#47;AIdbp2300040</RefTotal>
        <RefLink>https:&#47;&#47;doi.org&#47;10.1056&#47;AIdbp2300040</RefLink>
      </Reference>
      <Reference refNo="6">
        <RefAuthor>Ng KKY</RefAuthor>
        <RefAuthor>Matsuba I</RefAuthor>
        <RefAuthor>Zhang PC</RefAuthor>
        <RefTitle>RAG in Health Care: A Novel Framework for Improving Communication and Decision-Making by Addressing LLM Limitations</RefTitle>
        <RefYear>2025</RefYear>
        <RefJournal>NEJM AI</RefJournal>
        <RefPage>AIra2400380</RefPage>
        <RefTotal>Ng KKY, Matsuba I, Zhang PC. RAG in Health Care: A Novel Framework for Improving Communication and Decision-Making by Addressing LLM Limitations. NEJM AI. 2025;2(1):AIra2400380. DOI: 10.1056&#47;AIra2400380</RefTotal>
        <RefLink>https:&#47;&#47;doi.org&#47;10.1056&#47;AIra2400380</RefLink>
      </Reference>
      <Reference refNo="7">
        <RefAuthor>Puts S</RefAuthor>
        <RefAuthor>Zegers CML</RefAuthor>
        <RefAuthor>Dekker A</RefAuthor>
        <RefAuthor>Bermejo I</RefAuthor>
        <RefTitle>Developing an ICD-10 Coding Assistant: Pilot Study Using RoBERTa and GPT-4 for Term Extraction and Description-Based Code Selection</RefTitle>
        <RefYear>2025</RefYear>
        <RefJournal>JMIR Form Res</RefJournal>
        <RefPage>e60095</RefPage>
        <RefTotal>Puts S, Zegers CML, Dekker A, Bermejo I. Developing an ICD-10 Coding Assistant: Pilot Study Using RoBERTa and GPT-4 for Term Extraction and Description-Based Code Selection. JMIR Form Res. 2025 Feb;9:e60095. DOI: 10.2196&#47;60095</RefTotal>
        <RefLink>https:&#47;&#47;doi.org&#47;10.2196&#47;60095</RefLink>
      </Reference>
      <Reference refNo="8">
        <RefAuthor>Kwan K</RefAuthor>
        <RefTitle>Large language models are good medical coders, if provided with tools &#91;Preprint&#93;</RefTitle>
        <RefYear>2024</RefYear>
        <RefJournal>arXiv</RefJournal>
        <RefPage></RefPage>
        <RefTotal>Kwan K. Large language models are good medical coders, if provided with tools &#91;Preprint&#93;. arXiv. 2024 Jul 6. 
DOI: 10.48550&#47;arXiv.2407.12849</RefTotal>
        <RefLink>https:&#47;&#47;doi.org&#47;10.48550&#47;arXiv.2407.12849</RefLink>
      </Reference>
      <Reference refNo="3">
        <RefAuthor>SNOMED Intenational</RefAuthor>
        <RefTitle></RefTitle>
        <RefYear></RefYear>
        <RefBookTitle>SNOMED-CT</RefBookTitle>
        <RefPage></RefPage>
        <RefTotal>SNOMED Intenational. SNOMED-CT. International Edition.</RefTotal>
      </Reference>
      <Reference refNo="4">
        <RefAuthor>World Health Organisation</RefAuthor>
        <RefTitle></RefTitle>
        <RefYear>2019</RefYear>
        <RefBookTitle>ICD-10. International Statistical Classification of Diseases and Related Health Problems. 10th Revision</RefBookTitle>
        <RefPage></RefPage>
        <RefTotal>World Health Organisation. ICD-10. International Statistical Classification of Diseases and Related Health Problems. 10th Revision. WHO; 2019.</RefTotal>
      </Reference>
    </References>
    <Media>
      <Tables>
        <NoOfTables>0</NoOfTables>
      </Tables>
      <Figures>
        <Figure width="541" height="544" format="png">
          <MediaNo>1</MediaNo>
          <MediaID>1</MediaID>
          <Caption><Pgraph><Mark1>Figure 1: Standardized prompt</Mark1></Pgraph></Caption>
        </Figure>
        <Figure width="854" height="668" format="png">
          <MediaNo>2</MediaNo>
          <MediaID>2</MediaID>
          <Caption><Pgraph><Mark1>Figure 2: Evaluation criteria: Unsatisfactory in red (1&#8211;2 points on the Likert scale), satisfactory in yellow (3 points), and highly satisfactory in green (4&#8211;5 points)</Mark1></Pgraph></Caption>
        </Figure>
        <NoOfPictures>2</NoOfPictures>
      </Figures>
      <InlineFigures>
        <NoOfPictures>0</NoOfPictures>
      </InlineFigures>
      <Attachments>
        <NoOfAttachments>0</NoOfAttachments>
      </Attachments>
    </Media>
  </OrigData>
</GmsArticle>