<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="letter"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Aging</journal-id><journal-id journal-id-type="publisher-id">aging</journal-id><journal-id journal-id-type="index">31</journal-id><journal-title>JMIR Aging</journal-title><abbrev-journal-title>JMIR Aging</abbrev-journal-title><issn pub-type="epub">2561-7605</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v9i1e77591</article-id><article-id pub-id-type="doi">10.2196/77591</article-id><article-categories><subj-group subj-group-type="heading"><subject>Research Letter</subject></subj-group></article-categories><title-group><article-title>Leveraging Multimodal Large Language Models for Fall Risk Reduction in Older Adults in the Home: Proposed Model Design</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Do</surname><given-names>Justin</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Suresh</surname><given-names>Vivaswat</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Lily</given-names></name><degrees>BA</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Chavre</surname><given-names>Bharvi M</given-names></name><degrees>BA</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Cha</surname><given-names>Jeremy</given-names></name><degrees>BA</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Pugliese</surname><given-names>Robert</given-names></name><degrees>PharmD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib></contrib-group><aff id="aff1"><institution>Sidney Kimmel Medical College, Thomas Jefferson University</institution><addr-line>925 Chestnut St., Basement Vault</addr-line><addr-line>Philadelphia</addr-line><addr-line>PA</addr-line><country>United States</country></aff><aff id="aff2"><institution>Health Design Lab, Thomas Jefferson University</institution><addr-line>Philadelphia</addr-line><addr-line>PA</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Wang</surname><given-names>Jinjiao</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Azizan</surname><given-names>Azliyana</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Menychtas</surname><given-names>Dimitrios</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Justin Do, BS, Sidney Kimmel Medical College, Thomas Jefferson University, 925 Chestnut St., Basement Vault, Philadelphia, PA, 19107, United States, 1 4084718003, 1 2154841949; <email>jhd012@students.jefferson.edu</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>13</day><month>5</month><year>2026</year></pub-date><volume>9</volume><elocation-id>e77591</elocation-id><history><date date-type="received"><day>21</day><month>05</month><year>2025</year></date><date date-type="rev-recd"><day>01</day><month>03</month><year>2026</year></date><date date-type="accepted"><day>16</day><month>03</month><year>2026</year></date></history><copyright-statement>&#x00A9; Justin Do, Vivaswat Suresh, Lily Zhang, Bharvi M Chavre, Jeremy Cha, Robert Pugliese. Originally published in JMIR Aging (<ext-link ext-link-type="uri" xlink:href="https://aging.jmir.org">https://aging.jmir.org</ext-link>), 13.5.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Aging, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://aging.jmir.org">https://aging.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://aging.jmir.org/2026/1/e77591"/><abstract><p>This research letter proposes a novel model design leveraging natively multimodal large language models to identify fall risks and generate visualizations of recommended home environmental modifications, aiming to improve the accessibility and impact of personalized fall prevention advice for older adults. Through a pilot rating study, this work demonstrates that multimodal large language models can generate safe and actionable advice to reduce fall risk in lived spaces of older adults, and also generate realistic edits based on original images. While this concept needs further testing and clinical comparison, it highlights a promising avenue for further innovation of fall prevention tactics.</p></abstract><kwd-group><kwd>LLM</kwd><kwd>multimodal</kwd><kwd>image generation</kwd><kwd>fall risk</kwd><kwd>older adults</kwd><kwd>large language model</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Falls among older adults cause significant mortality and increased health care costs [<xref ref-type="bibr" rid="ref1">1</xref>]. Current literature has identified combined behavioral and exercise interventions as effective preventions for fall risks, improving balance performance, and reducing fear of falls [<xref ref-type="bibr" rid="ref2">2</xref>], while limited evidence exists regarding medication-induced falls [<xref ref-type="bibr" rid="ref3">3</xref>]. Home environmental intervention is also effective: safety assessments have been shown to reduce fall rate by 23%&#x2010;36% [<xref ref-type="bibr" rid="ref1">1</xref>], with applied home modifications contributing a 7% risk reduction [<xref ref-type="bibr" rid="ref4">4</xref>]. However, both external (insurance) and self-imposed (ie, the perception that safety assessments are invasive) barriers impede widespread implementation [<xref ref-type="bibr" rid="ref5">5</xref>]. While research on frailty assessments is robust, gaps remain in technology-enabled interventions [<xref ref-type="bibr" rid="ref6">6</xref>]. Prior studies have shown acceptability by older adults to embrace digital and electronic tools [<xref ref-type="bibr" rid="ref5">5</xref>]. Existing remote home assessment protocols rely on caregiver camera operation, written instructions comprehension, and professional review of footage [<xref ref-type="bibr" rid="ref7">7</xref>], while telehealth occupational therapy (OT) assessments may require insurance authorization, creating both obstacles and delays. Multimodal large language models (LLM) can fuse visual and text information, offering a scalable alternative while preventing encroachment on user values. This study aims to evaluate the ability of LLMs to produce safe, clinically useful, and actionable outputs that identify fall risks from user-provided home imagery and uniquely generate visualizations of the recommended environmental modifications.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><p>We selected Google&#x2019;s Gemini family due to its strong visual reasoning performance supported by validated benchmarks [<xref ref-type="bibr" rid="ref8">8</xref>]. We designed our framework to focus on providing reliable output by employing a low model temperature (0.15), in-context learning through grounding responses in evidence-based CDC STEADI patient materials (<xref ref-type="fig" rid="figure1">Figure 1</xref>), and structured XML prompts iterated using artificial intelligence (AI)-driven prompt engineering. The core innovation of this study used the gemini-2.0-flash-exp-image-generation model to directly modify the uploaded images with the model&#x2019;s suggested changes. The model leverages a two-shot prompting system (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>, <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>), in which the primary LLM generates textual recommendations based on an input image/video, then directs the image generation model to visually render these changes (eg, adding grab bars, removing hazards) onto the original image, iterating until the generated image reflects the proposed modifications. We conducted a formative, blinded, paired comparison of outputs generated from 27 publicly licensed &#x201C;lived-in&#x201D; home interior images. We compared a non-optimized baseline prompt with an enhanced multimodal pipeline (&#x201C;Steadi&#x201D;). Text and image outputs were compared based on clinical usefulness, safety, image fidelity/plausibility, and preference between baseline prompt output and our enhanced multimodal pipeline output. Detailed methods and output can be found in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendices 3</xref> and <xref ref-type="supplementary-material" rid="app4">4</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Enhanced model architecture for advice generation and multimodal interaction. CDC: Centers for Disease Control and Prevention; LLM: large language model; STEADI: Stopping Elderly Accidents, Deaths, &#x0026; Injuries.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="aging_v9i1e77591_fig01.png"/></fig></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Initial Advice Generation, Multimodal Communication, and Modification Visualization</title><p>The model takes an uploaded image or video and provides specific, actionable advice supported by evidence-based resources. The proposed architecture successfully applies both additive and subtractive modifications to images, providing users with a concrete visual representation of a safer environment.</p></sec><sec id="s3-2"><title>Model Comparisons</title><p>We find that overall, our raters preferred the &#x201C;Steadi&#x201D; system output of both image and text (40/54 times (74.1%), <xref ref-type="table" rid="table1">Table 1</xref>). We demonstrate that contemporary LLMs produce relatively safe recommendations, regardless of the prompting system, with only one set of recommendations rated as unsafe in the baseline prompting system and none in the enhanced. We find that when given specific issues to visualize, image editing LLMs produce edits with good visualization fidelity (46/54 times, 85.2% for baseline and 43/54 times, 79.6% for enhanced; <xref ref-type="table" rid="table1">Table 1</xref>), and low rates of implausible/hazard-producing edits (6/54 times, 11.1% for both systems; <xref ref-type="table" rid="table1">Table 1</xref>). Text recommendations and visualized outputs ranged from generally &#x201C;somewhat actionable&#x201D; for the baseline system to highly actionable for our enhanced system.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Results of rating study.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Outcome</td><td align="left" valign="bottom">Baseline</td><td align="left" valign="bottom">Enhanced</td><td align="left" valign="bottom">Comparison<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td></tr></thead><tbody><tr><td align="left" valign="top">Q1 Overall clinical usefulness (preference)</td><td align="left" valign="top">Preferred: 10/54 (18.5%)</td><td align="left" valign="top">Preferred: 40/54 (74.1%)</td><td align="left" valign="top">Win rate: 40/50 (80.0%, 95% CI 67.0&#x2010;88.8); sign test <italic>P</italic>&#x2264;.001; ties=4</td></tr><tr><td align="left" valign="top">Q2 Unsafe/inappropriate recommendation rate (Yes)</td><td align="left" valign="top">1/54<break/>(1.9%, 95% CI 0.3&#x2010;9.8)</td><td align="left" valign="top">0/54<break/>(0.0%, 95% CI 0.0&#x2010;6.6)</td><td align="left" valign="top">Risk difference (Enhanced-Baseline): &#x2212;1.9 pp</td></tr><tr><td align="left" valign="top">Q3 Visualization fidelity rate (Yes)</td><td align="left" valign="top">46/54<break/>(85.2%, 95% CI 73.4&#x2010;92.3)</td><td align="left" valign="top">43/54<break/>(79.6%, 95% CI 67.1&#x2010;88.2)</td><td align="left" valign="top">Risk difference (Enhanced-Baseline): &#x2212;5.6 pp</td></tr><tr><td align="left" valign="top">Q4 Hazard-introducing/implausible edit rate (Yes)</td><td align="left" valign="top">6/54<break/>(11.1%, 95% CI 5.2&#x2010;22.2)</td><td align="left" valign="top">6/54<break/>(11.1%, 95% CI 5.2&#x2010;22.2)</td><td align="left" valign="top">Risk difference (Enhanced-Baseline):+0.0 pp</td></tr><tr><td align="left" valign="top">Q5 Actionability (1&#x2010;5 Likert)</td><td align="left" valign="top">median 3.0<break/>(IQR 3.0&#x2010;4.0)</td><td align="left" valign="top">median 5.0<break/>(IQR 4.0&#x2010;5.0)</td><td align="left" valign="top">&#x0394;median (Enhanced-Baseline):+1.0 (IQR 0.0&#x2010;2.0)</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>n=rater case evaluations; outcomes are descriptive; Q1 sign test is exploratory.</p></fn></table-wrap-foot></table-wrap></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study introduces a novel application of multimodal LLMs, leveraging their image-generation capabilities for visualizing personalized home safety recommendations. We demonstrate that enhanced frameworks, such as structured prompting and grounding using trusted resources, produce safe, clinically useful, and actionable outputs that categorically rate better than outputs from baseline LLMs. The inherent flexibility of LLMs supports diverse interaction methods, uniquely enabling users to interact with their &#x201C;consultant&#x201D; in their preferred mode. LLMs may mitigate delays caused by insurance authorizations and restore autonomy to users.</p><p>The visual output capability is also key: generating suggestions directly onto uploaded images offers more intuitive, actionable guidance than abstract text instructions alone. The drive to protect the familiarity of their home from change was identified to be a major motive for older adults rejecting modification advice from OT [<xref ref-type="bibr" rid="ref5">5</xref>]; direct visualization of user-fed images may help overcome this hurdle and increase acceptance. There are still limitations to the technology, namely outputs may be illogical such as the recommended soap placement, and movement of furniture and door in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. However, overall, this study demonstrates that LLMs generally produce visual outputs with high fidelity, low hazard introduction rates, and high actionability.</p><p>To adhere to HIPAA (Health Insurance Portability and Accountability Act) compliance, future work should consider working with LLM providers to sign a HIPAA Business Associate Amendment or other HIPAA-compliant program. Ethical considerations, such as disclosure of privacy and data protection, should be implemented in accordance with WHO guidance on AI in health [<xref ref-type="bibr" rid="ref9">9</xref>].</p></sec><sec id="s4-2"><title>Limitations</title><p>Further testing must be conducted against the current standard for in-home assessments to discover if the proposed model provides comparable advice to professionals. Implementation trials will be needed to mitigate concerns such as the digital divide and ensure accessibility among varying cognitive/visual functions. Implementations must comply with FDA digital health guidance, and characterization and limitation of unsafe output generation must be explored. This model is designed as a supplemental service to be integrated with OT rather than a replacement.</p></sec><sec id="s4-3"><title>Conclusions</title><p>Multimodal LLMs that integrate image generation offer a novel, innovative approach to increasing end users&#x2019; accessibility to personalized home environment recommendations for fall prevention. This capability represents a potential supplement to current care services that may enhance patient understanding, motivation, and adherence, serving as a valuable resource to patients who defer or cannot access in-home safety assessments. Rigorous validation of clinical efficacy and user acceptance is essential to translate this technological potential into improved patient outcomes.</p></sec></sec></body><back><ack><p>We would like to extend our gratitude to Robert Pugliese and MaryEllen Daley for all of their generous support throughout this project. We would also like to thank Dr. Bracken Babula, MD, Dr. Zhe Chen, MD, Dr. Ryan Tomlinson, PhD, Dr. Deanna Gray-Miceli, PhD, CRNP, Dr. Christine Hsieh, MD, and Dr. Brooke Salzman, MD, for their feedback and guidance on this project. No generative AI was used in any portion of the manuscript text generation. We used the generative AI Tool &#x201C;Gemini 1.5 Pro&#x201D; made by Google to draft the system prompt found in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>, with review and editing from the study group. Image portions of <xref ref-type="supplementary-material" rid="app1">Multimedia Appendices 1</xref>, <xref ref-type="supplementary-material" rid="app4">4</xref> and <xref ref-type="supplementary-material" rid="app5">5</xref> were generated using &#x201C;Gemini 2.0 Flash Image Preview&#x201D; as described in the manuscript text as part of the model design. Image portions of <xref ref-type="supplementary-material" rid="app1">Multimedia Appendices 1</xref>, <xref ref-type="supplementary-material" rid="app4">4</xref> and <xref ref-type="supplementary-material" rid="app5">5</xref> were generated using &#x201C;Gemini 2.5 Flash Image&#x201D; as described in the manuscript text as part of the model design.</p></ack><notes><sec><title>Funding</title><p>No external financial support or grants were received from any public, commercial, or not-for-profit entities for the research, authorship, or publication of this article.</p></sec></notes><fn-group><fn fn-type="con"><p>Conceptualization: JD, LZ, BC, JC. Methodology: JD, VS. Resources: JD, LZ, BC, JC.Supervision: RP. Writing &#x2013; Original draft: JD. Writing &#x2013; Revising and editing: JD, VS, LZ, BC, JC.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">CDC</term><def><p>Centers for Disease Control and Prevention</p></def></def-item><def-item><term id="abb3">HIPAA</term><def><p>Health Insurance Portability and Accountability Act</p></def></def-item><def-item><term id="abb4">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb5">MMMU</term><def><p>Massive Multi-discipline Multimodal Understanding and Reasoning</p></def></def-item><def-item><term id="abb6">NICE</term><def><p>National Institute for Health and Care Excellence</p></def></def-item><def-item><term id="abb7">OT</term><def><p>occupational therapy</p></def></def-item><def-item><term id="abb8">STEADI</term><def><p>Stopping Elderly Accidents, Deaths, &#x0026; Injuries</p></def></def-item><def-item><term id="abb9">XML</term><def><p>Extensible Markup Language</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Niedermann</surname><given-names>K</given-names> </name><name name-style="western"><surname>Meichtry</surname><given-names>A</given-names> </name><name name-style="western"><surname>Zindel</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Effectiveness and cost-effectiveness of a single home-based fall prevention program: a prospective observational study based on questionnaires and claims data</article-title><source>BMC Geriatr</source><year>2024</year><month>12</month><day>28</day><volume>24</volume><issue>1</issue><fpage>1044</fpage><pub-id pub-id-type="doi">10.1186/s12877-024-05586-x</pub-id><pub-id pub-id-type="medline">39732675</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Azizan</surname><given-names>A</given-names> </name><name name-style="western"><surname>Justine</surname><given-names>M</given-names> </name></person-group><article-title>Elders&#x2019; exercise and behavioral program: effects on balance and fear of falls</article-title><source>Phys Occup Ther Geriatr</source><year>2015</year><month>10</month><day>2</day><volume>33</volume><issue>4</issue><fpage>346</fpage><lpage>362</lpage><pub-id pub-id-type="doi">10.3109/02703181.2015.1093060</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gillespie</surname><given-names>LD</given-names> </name><name name-style="western"><surname>Robertson</surname><given-names>MC</given-names> </name><name name-style="western"><surname>Gillespie</surname><given-names>WJ</given-names> </name><etal/></person-group><article-title>Interventions for preventing falls in older people living in the community</article-title><source>Cochrane Database Syst Rev</source><year>2012</year><month>09</month><day>12</day><volume>2012</volume><issue>9</issue><fpage>CD007146</fpage><pub-id pub-id-type="doi">10.1002/14651858.CD007146.pub3</pub-id><pub-id pub-id-type="medline">22972103</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lektip</surname><given-names>C</given-names> </name><name name-style="western"><surname>Chaovalit</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wattanapisit</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lapmanee</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nawarat</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yaemrattanakul</surname><given-names>W</given-names> </name></person-group><article-title>Home hazard modification programs for reducing falls in older adults: a systematic review and meta-analysis</article-title><source>PeerJ</source><year>2023</year><volume>11</volume><fpage>e15699</fpage><pub-id pub-id-type="doi">10.7717/peerj.15699</pub-id><pub-id pub-id-type="medline">37489124</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>JJ</given-names> </name><name name-style="western"><surname>Patel</surname><given-names>D</given-names> </name><name name-style="western"><surname>Gadgil</surname><given-names>M</given-names> </name><name name-style="western"><surname>Langness</surname><given-names>S</given-names> </name><name name-style="western"><surname>von Hippel</surname><given-names>CD</given-names> </name><name name-style="western"><surname>Sammann</surname><given-names>A</given-names> </name></person-group><article-title>Understanding barriers to home safety assessment adoption in older adults: qualitative human-centered design study</article-title><source>JMIR Hum Factors</source><year>2025</year><month>06</month><day>24</day><volume>12</volume><fpage>e66854</fpage><pub-id pub-id-type="doi">10.2196/66854</pub-id><pub-id pub-id-type="medline">40554799</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Azizan</surname><given-names>A</given-names> </name></person-group><article-title>Exercise and frailty in later life: a systematic review and bibliometric analysis of research themes and scientific collaborations</article-title><source>IJPS</source><year>2024</year><volume>11</volume><issue>1</issue><fpage>1</fpage><pub-id pub-id-type="doi">10.36922/ijps.3282</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Romero</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Simic</surname><given-names>I</given-names> </name><name name-style="western"><surname>Levy</surname><given-names>C</given-names> </name><name name-style="western"><surname>Sanford</surname><given-names>J</given-names> </name></person-group><article-title>Development and validation of a remote home safety protocol</article-title><source>Disabil Rehabil Assist Technol</source><year>2018</year><month>02</month><volume>13</volume><issue>2</issue><fpage>166</fpage><lpage>172</lpage><pub-id pub-id-type="doi">10.1080/17483107.2017.1300345</pub-id><pub-id pub-id-type="medline">28326967</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Yue</surname><given-names>X</given-names> </name><name name-style="western"><surname>Ni</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>T</given-names> </name><etal/></person-group><article-title>MMMU: a massive multi-discipline multimodal understanding and reasoning benchmark for expert AGI</article-title><source>arXiv</source><comment>Preprint posted online on  Jun 13, 2024</comment><pub-id pub-id-type="doi">10.48550/arXiv.2311.16502</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="report"><article-title>Regulatory considerations on artificial intelligence for health</article-title><year>2023</year><access-date>2025-11-01</access-date><publisher-name>World Health Organization</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.who.int/publications/i/item/9789240078871">https://www.who.int/publications/i/item/9789240078871</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Two-shot image modification architecture.</p><media xlink:href="aging_v9i1e77591_app1.png" xlink:title="PNG File, 941 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>XML Prompt and model parameters.</p><media xlink:href="aging_v9i1e77591_app2.docx" xlink:title="DOCX File, 10 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>Blinded comparison methods.</p><media xlink:href="aging_v9i1e77591_app3.pdf" xlink:title="PDF File, 94 KB"/></supplementary-material><supplementary-material id="app4"><label>Multimedia Appendix 4</label><p>Rating packet.</p><media xlink:href="aging_v9i1e77591_app4.pdf" xlink:title="PDF File, 15079 KB"/></supplementary-material><supplementary-material id="app5"><label>Multimedia Appendix 5</label><p>Usage of model for recommendations for post-stroke patients.</p><media xlink:href="aging_v9i1e77591_app5.png" xlink:title="PNG File, 1063 KB"/></supplementary-material></app-group></back></article>