<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Aging</journal-id><journal-id journal-id-type="publisher-id">aging</journal-id><journal-id journal-id-type="index">31</journal-id><journal-title>JMIR Aging</journal-title><abbrev-journal-title>JMIR Aging</abbrev-journal-title><issn pub-type="epub">2561-7605</issn></journal-meta><article-meta><article-id pub-id-type="publisher-id">53019</article-id><article-id pub-id-type="doi">10.2196/53019</article-id><title-group><article-title>Assessing the Quality of ChatGPT Responses to Dementia Caregivers&#x2019; Questions: Qualitative Analysis</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>Aguirre</surname><given-names>Alyssa</given-names></name><degrees>MSW</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Hilsabeck</surname><given-names>Robin</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Smith</surname><given-names>Tawny</given-names></name><degrees>PharmD</degrees><xref ref-type="aff" rid="aff4">4</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Xie</surname><given-names>Bo</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff5">5</xref><xref ref-type="aff" rid="aff6">6</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>He</surname><given-names>Daqing</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff7">7</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Wang</surname><given-names>Zhendong</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff7">7</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Zou</surname><given-names>Ning</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff7">7</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Neurology, The University of Texas at Austin</institution>, <addr-line>Austin</addr-line><addr-line>TX</addr-line>, <country>United States</country></aff><aff id="aff2"><institution>Steve Hicks School of Social Work, The University of Texas at Austin</institution>, <addr-line>Austin</addr-line><addr-line>TX</addr-line>, <country>United States</country></aff><aff id="aff3"><institution>Glenn Biggs Institute for Alzheimer's &#x0026; Neurodegenerative Diseases, Department of Neurology, University of Texas Health Science Center at San Antonio</institution>, <addr-line>San Antonio</addr-line><addr-line>TX</addr-line>, <country>United States</country></aff><aff id="aff4"><institution>Department of Psychiatry and Behavioral Sciences, The University of Texas at Austin</institution>, <addr-line>Austin</addr-line><addr-line>TX</addr-line>, <country>United States</country></aff><aff id="aff5"><institution>School of Information, The University of Texas at Austin</institution>, <addr-line>Austin</addr-line><addr-line>TX</addr-line>, <country>United States</country></aff><aff id="aff6"><institution>School of Nursing, The University of Texas at Austin</institution>, <addr-line>Austin</addr-line><addr-line>TX</addr-line>, <country>United States</country></aff><aff id="aff7"><institution>School of Computing and Information, University of Pittsburgh</institution>, <addr-line>Pittsburgh</addr-line><addr-line>PA</addr-line>, <country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Du</surname><given-names>Yan</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Hird</surname><given-names>Nick</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Guo</surname><given-names>Ziqiu</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Alyssa Aguirre, MSW<email>alyssa.aguirre@austin.utexas.edu</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>all authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2024</year></pub-date><pub-date pub-type="epub"><day>6</day><month>5</month><year>2024</year></pub-date><volume>7</volume><elocation-id>e53019</elocation-id><history><date date-type="received"><day>22</day><month>09</month><year>2023</year></date><date date-type="rev-recd"><day>15</day><month>02</month><year>2024</year></date><date date-type="accepted"><day>09</day><month>03</month><year>2024</year></date></history><copyright-statement>&#x00A9; Alyssa Aguirre, Robin Hilsabeck, Tawny Smith, Bo Xie, Daqing He, Zhendong Wang, Ning Zou. Originally published in JMIR Aging (<ext-link ext-link-type="uri" xlink:href="https://aging.jmir.org">https://aging.jmir.org</ext-link>), 6.5.2024. </copyright-statement><copyright-year>2024</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Aging, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://aging.jmir.org">https://aging.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://aging.jmir.org/2024/1/e53019"/><abstract><sec><title>Background</title><p>Artificial intelligence (AI) such as ChatGPT by OpenAI holds great promise to improve the quality of life of patients with dementia and their caregivers by providing high-quality responses to their questions about typical dementia behaviors. So far, however, evidence on the quality of such ChatGPT responses is limited. A few recent publications have investigated the quality of ChatGPT responses in other health conditions. Our study is the first to assess ChatGPT using real-world questions asked by dementia caregivers themselves.</p></sec><sec><title>Objectives</title><p>This pilot study examines the potential of ChatGPT-3.5 to provide high-quality information that may enhance dementia care and patient-caregiver education.</p></sec><sec sec-type="methods"><title>Methods</title><p>Our interprofessional team used a formal rating scale (scoring range: 0-5; the higher the score, the better the quality) to evaluate ChatGPT responses to real-world questions posed by dementia caregivers. We selected 60 posts by dementia caregivers from Reddit, a popular social media platform. These posts were verified by 3 interdisciplinary dementia clinicians as representing dementia caregivers&#x2019; desire for information in the areas of memory loss and confusion, aggression, and driving. Word count for posts in the memory loss and confusion category ranged from 71 to 531 (mean 218; median 188), aggression posts ranged from 58 to 602 words (mean 254; median 200), and driving posts ranged from 93 to 550 words (mean 272; median 276).</p></sec><sec sec-type="results"><title>Results</title><p>ChatGPT&#x2019;s response quality scores ranged from 3 to 5. Of the 60 responses, 26 (43%) received 5 points, 21 (35%) received 4 points, and 13 (22%) received 3 points, suggesting high quality. ChatGPT obtained consistently high scores in synthesizing information to provide follow-up recommendations (n=58, 96%), with the lowest scores in the area of comprehensiveness (n=38, 63%).</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>ChatGPT provided high-quality responses to complex questions posted by dementia caregivers, but it did have limitations. ChatGPT was unable to anticipate future problems that a human professional might recognize and address in a clinical encounter. At other times, ChatGPT recommended a strategy that the caregiver had already explicitly tried. This pilot study indicates the potential of AI to provide high-quality information to enhance dementia care and patient-caregiver education in tandem with information provided by licensed health care professionals. Evaluating the quality of responses is necessary to ensure that caregivers can make informed decisions. ChatGPT has the potential to transform health care practice by shaping how caregivers receive health information.</p></sec></abstract><kwd-group><kwd>Alzheimer&#x2019;s disease</kwd><kwd>information technology</kwd><kwd>social media</kwd><kwd>neurology</kwd><kwd>dementia</kwd><kwd>Alzheimer disease</kwd><kwd>caregiver</kwd><kwd>ChatGPT</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Older adults have responded to the COVID-19 pandemic by increasing their internet-enabled behaviors, which include expanding their medical care to the use of web-based platforms [<xref ref-type="bibr" rid="ref1">1</xref>]. Indeed the internet has become the most common source of information among dementia caregivers [<xref ref-type="bibr" rid="ref2">2</xref>], and with recent advances in artificial intelligence (AI), caregivers will increasingly use AI to obtain information about health [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. ChatGPT by OpenAI [<xref ref-type="bibr" rid="ref5">5</xref>], an innovative, dialogue-based large language model that responds to complex natural language inquiries, holds great promise to improve the quality of life of patients with dementia and their caregivers by providing high-quality responses to meet their needs for information [<xref ref-type="bibr" rid="ref4">4</xref>]. On the other hand, several studies have highlighted the limitations of generative AI models in health care, citing the lack of trust and reliability as some of the primary challenges [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. Although there have been studies on the quality of ChatGPT responses to common questions about heart disease [<xref ref-type="bibr" rid="ref8">8</xref>], cirrhosis [<xref ref-type="bibr" rid="ref9">9</xref>], and bariatric surgery [<xref ref-type="bibr" rid="ref10">10</xref>], to our knowledge, no studies have examined the quality of ChatGPT responses to real-world questions posed by dementia caregivers. We have addressed this gap by examining the quality of ChatGPT-3.5 responses to complex questions posted by dementia caregivers on social media.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Overview</title><p>From January to May 2023, a total of 60 social media posts representing dementia caregivers&#x2019; needs for information in 3 areas (memory loss and confusion, aggression, and driving; 20 posts per area) were selected from Reddit, a popular social media platform. These topics were chosen because they are common clinical themes that are often complex and difficult to navigate with potential safety implications. Four seed posts were used in each area to discover the additional 16 posts. Posts were excluded if the poster&#x2019;s main question did not fall into the 3 aforementioned areas as verified by dementia clinicians or if the poster declared they were &#x201C;venting&#x201D; and/or no specific question was asked. Posts that were unclear on whether the person had a dementia diagnosis were excluded to avoid assessing posts that were not clearly dementia related. Word count for posts in the memory loss and confusion category ranged from 71 to 531 (mean 218; median 188), aggression posts ranged from 58 to 602 (mean 254; median 200), and driving posts ranged from 93 to 550 (mean 272; median 276). Of the 60 posts, the caregiver described the person with dementia as their parent (n=34, 56%), grandparent (n=22, 36%), uncle (n=2, 3%), or spouse (n=1, 1.6%). One post did not report relationship. The gender of the person with dementia was described as female in 57% (n=34) of posts and as male in 42% (n=25) of posts. One post did not report gender.</p><p>Three clinicians, each having more than 15 years of experience with patients with dementia and their caregivers, but from diverse disciplines (pharmacy, neuropsychology, and social work), assessed ChatGPT responses to the 60 posts using an adapted rating scale based on Hurtz et al&#x2019;s [<xref ref-type="bibr" rid="ref11">11</xref>] levels of cognitive complexity pertaining to clinical decision-making (<xref ref-type="table" rid="table1">Table 1</xref>). Responses received 1 point for each of the following characteristics: <italic>factuality</italic>, <italic>interpretation</italic>, <italic>application</italic>, <italic>synthesis</italic>, and <italic>comprehensiveness</italic>, with a scoring range of 0-5 for each response, where higher scores indicate higher quality. Table S1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> presents examples of posts for each topic area, ChatGPT responses, and clinician ratings for each response category.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Description of rating scale categories used to measure the quality of ChatGPT responses.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Characteristic</td><td align="left" valign="bottom">Description</td></tr></thead><tbody><tr><td align="left" valign="top">Factuality</td><td align="left" valign="top">Response did not contain inaccurate or false information.</td></tr><tr><td align="left" valign="top">Interpretation</td><td align="left" valign="top">Response adequately interpreted the poster&#x2019;s main need, correctly disregarded nonpriority details, and did not recommend strategies that the poster had already tried.</td></tr><tr><td align="left" valign="top">Application</td><td align="left" valign="top">Response suggested tangible actions (eg, educational information, a change the caregiver could make, and communication strategies such as validation and redirection).</td></tr><tr><td align="left" valign="top">Synthesis</td><td align="left" valign="top">Response contained follow-up recommendations as needed (referrals to help beyond the caregiver-patient dyad, such as support groups, health care professionals, or other community resources).</td></tr><tr><td align="left" valign="top">Comprehensiveness</td><td align="left" valign="top">Response had strong depth, breadth; response was thorough and complete.</td></tr></tbody></table></table-wrap><p>Although the results reported in this paper were based on raters&#x2019; consensus scores, we acknowledge the potential benefits of expanding on ChatGPT responses that originally received different scores. Initially, 1 rater gave a point for comprehensiveness when the majority of suggestions they would provide clinically were conveyed in ChatGPT&#x2019;s response, but another rater did not give the point if they felt it was missing anything at all. It was agreed upon during consensus that if the majority of recommendations were provided, ChatGPT responses would receive full credit for <italic>comprehensiveness</italic>.</p></sec><sec id="s2-2"><title>Ethical Considerations</title><p>This study was approved by the institutional review boards of The University of Texas at Austin (STUDY00003358) and the University of Pittsburgh (STUDY20020007).</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><p>ChatGPT responses in the memory loss and confusion category ranged from 89 to 276 words (mean 170; median 165), 91 to 372 words in the aggression category (mean 221; median 234), and 65 to 359 words in the driving category (mean 175; median 130). At least 2 clinicians agreed on the ratings for all ChatGPT responses, with any disagreements resolved by discussion. ChatGPT scores ranged from 3 to 5. Overall, of the 60 responses, 26 (43%) received 5 points, 21 (35%) received 4 points, and 13 (21.7%) received 3 points (<xref ref-type="table" rid="table2">Table 2</xref>), suggesting high quality. There were no responses that scored a 0, 1, or 2; there were no fabricated responses; and no responses were considered harmful to posters. ChatGPT received the lowest ratings in <italic>comprehensiveness</italic>, followed by <italic>interpretation</italic>, and the highest ratings in <italic>synthesis</italic>, with only 2 out of 60 posts failing to receive the point (<xref ref-type="table" rid="table3">Table 3</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Rating scale results by topic.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Score</td><td align="left" valign="bottom">Memory loss and confusion (n<italic>=</italic>20), n (%)</td><td align="left" valign="bottom">Aggression (n<italic>=</italic>20), n (%)</td><td align="left" valign="bottom">Driving (n<italic>=</italic>20), n (%)</td><td align="left" valign="bottom">Total (N=60), n (%)</td></tr></thead><tbody><tr><td align="char" char="." valign="top">3</td><td align="left" valign="top">6 (30)</td><td align="left" valign="top">3 (15)</td><td align="left" valign="top">4 (20)</td><td align="left" valign="top">13 (22)</td></tr><tr><td align="char" char="." valign="top">4</td><td align="left" valign="top">7 (35)</td><td align="left" valign="top">6 (30)</td><td align="left" valign="top">8 (40)</td><td align="left" valign="top">21 (35)</td></tr><tr><td align="char" char="." valign="top">5</td><td align="left" valign="top">7 (35)</td><td align="left" valign="top">11 (55)</td><td align="left" valign="top">8 (40)</td><td align="left" valign="top">26 (43)</td></tr></tbody></table></table-wrap><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Number of ChatGPT points for each topic.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Characteristic</td><td align="left" valign="bottom">Memory loss and confusion (n=20), n (%)</td><td align="left" valign="bottom">Aggression (n=20), n (%)</td><td align="left" valign="bottom">Driving (n=20), n (%)</td><td align="left" valign="bottom">Total (N=60), n (%)</td></tr></thead><tbody><tr><td align="left" valign="top">Factuality</td><td align="left" valign="top">17 (85)</td><td align="left" valign="top">19 (95)</td><td align="left" valign="top">20 (100)</td><td align="left" valign="top">56 (93)</td></tr><tr><td align="left" valign="top">Interpretation</td><td align="left" valign="top">17 (85)</td><td align="left" valign="top">17 (85)</td><td align="left" valign="top">13 (65)</td><td align="left" valign="top">47 (78)</td></tr><tr><td align="left" valign="top">Application</td><td align="left" valign="top">20 (100)</td><td align="left" valign="top">17 (85)</td><td align="left" valign="top">17 (85)</td><td align="left" valign="top">54 (90)</td></tr><tr><td align="left" valign="top">Synthesis</td><td align="left" valign="top">18 (90)</td><td align="left" valign="top">20 (100)</td><td align="left" valign="top">20 (100)</td><td align="left" valign="top">58 (96)</td></tr><tr><td align="left" valign="top">Comprehensiveness</td><td align="left" valign="top">9 (45)</td><td align="left" valign="top">15 (75)</td><td align="left" valign="top">14 (70)</td><td align="left" valign="top">38 (63)</td></tr></tbody></table></table-wrap></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>In this study, ChatGPT responses to complex, real-world questions posted by dementia caregivers were assessed by dementia clinicians using a clinical decision-making rating scale. ChatGPT was found to produce high-quality responses, suggesting the potential of online chatbots to be a useful source of health information for dementia caregivers. The majority of responses contained factual information (n=56, 93%), with 78% (n=47) of responses correctly interpreting the poster&#x2019;s main need. The majority (n=54, 90%) of ChatGPT responses contained tangible actions the caregiver could apply to their situation. In only 2 instances, follow-up referrals were not suggested when reviewers felt recommendations were needed.</p><p>ChatGPT also had limitations, primarily in the areas of <italic>interpretation</italic> and <italic>comprehensiveness</italic>. In 22% (n=13) of posts, ChatGPT recommended strategies that posters had already explicitly tried, or missed subtleties that affected the accuracy of recommendations, such as failing to recognize that a person placed in a &#x201C;home&#x201D; meant a nursing home facility and not a traditional home. In another instance, ChatGPT recommended considering short-term hospitalization, but the poster already disclosed the person with dementia was currently hospitalized. In 37% (n=22) of posts, ChatGPT&#x2019;s response did not include information that dementia clinicians felt was important or was unable to anticipate future problems that a human clinician might choose to address in response to the same post. For example, if ChatGPT recommended a driving test, it did not suggest what to do if the patient in question refused to take the driving test. The data suggest that ChatGPT has strengths in providing objectively correct information (<italic>factuality</italic>, <italic>application</italic>, and <italic>synthesis</italic>) but is less successful in contextualizing the information it provides (<italic>interpretation</italic> and <italic>comprehensiveness</italic>).</p></sec><sec id="s4-2"><title>Limitations</title><p>Study limitations included potential sample bias and small sample size. Very few posters in this study identified as a spousal caregiver (n=1, 1.6%) even though national studies report that 60% of dementia caregivers are a spouse or partner [<xref ref-type="bibr" rid="ref12">12</xref>]. In selecting social media posts for inclusion, we included only those in which it was clear that the individual had a diagnosis of dementia. Historically, racial and ethnic minority groups are less likely to seek or receive a dementia diagnosis; thus, our sample may have been skewed for race and ethnicity. Posts were from one specific platform, which risked including caregivers with a certain level of technology access and literacy. This study did not evaluate differences in ChatGPT responses at multiple time points, so no conclusions can be made regarding reproducibility. Raters were aware that responses were generated by ChatGPT, which could have influenced stricter grading. Although our 5-point scale graded specific aspects of ChatGPT responses, it might have had a ceiling effect.</p></sec><sec id="s4-3"><title>Conclusions</title><p>This study contributes to the currently small but rapidly growing literature on AI&#x2019;s potential to assist patient-caregiver education by providing high-quality information. Our study illustrates that ChatGPT-3.5 can provide high-quality responses to most questions in the areas of memory loss and confusion, aggression, and driving. Future research should examine family caregivers&#x2019; receptiveness to using ChatGPT, as well as the usefulness of the responses from the perspective of family caregivers. Validated rating scales to assess the quality of ChatGPT responses are still in progress; the field would benefit from a reliable, validated method to evaluate the quality of AI responses to health care questions. We encourage future studies to expand on our findings and investigate how ChatGPT might be used in tandem with information provided by licensed health care professionals.</p></sec></sec></body><back><ack><p>This work was supported in part by the National Institute on Aging of the National Institutes of Health (award number R56AG075770). The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. Editorial support with manuscript development was provided by Dr John Bellquist at the Cain Center for Nursing Research at The University of Texas at Austin School of Nursing.</p></ack><fn-group><fn fn-type="conflict"><p>TS was employed by The University of Texas at Austin during the submission of this work but has since changed positions and is employed by Otsuka America Pharmaceutical, Inc.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Benge</surname><given-names>JF</given-names></name><name name-style="western"><surname>Aguirre</surname><given-names>A</given-names></name><name name-style="western"><surname>Scullin</surname><given-names>MK</given-names></name><etal/></person-group><article-title>Internet-enabled behaviors in older adults during the pandemic: patterns of use, psychosocial impacts, and plans for continued utilization</article-title><source>Work Aging Retire</source><year>2024</year><month>01</month><volume>10</volume><issue>1</issue><fpage>6</fpage><lpage>13</lpage><pub-id pub-id-type="doi">10.1093/workar/waac026</pub-id><pub-id pub-id-type="medline">38196827</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Allen</surname><given-names>F</given-names></name><name name-style="western"><surname>Cain</surname><given-names>R</given-names></name><name name-style="western"><surname>Meyer</surname><given-names>C</given-names></name></person-group><article-title>Seeking relational information sources in the digital age: a study into information source preferences amongst family and friends of those with dementia</article-title><source>Dementia (London)</source><year>2020</year><month>04</month><volume>19</volume><issue>3</issue><fpage>766</fpage><lpage>785</lpage><pub-id pub-id-type="doi">10.1177/1471301218786568</pub-id><pub-id pub-id-type="medline">29999410</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xie</surname><given-names>B</given-names></name><name name-style="western"><surname>Tao</surname><given-names>C</given-names></name><name name-style="western"><surname>Li</surname><given-names>J</given-names></name><name name-style="western"><surname>Hilsabeck</surname><given-names>RC</given-names></name><name name-style="western"><surname>Aguirre</surname><given-names>A</given-names></name></person-group><article-title>Artificial intelligence for caregivers of persons with Alzheimer&#x2019;s disease and related dementias: systematic literature review</article-title><source>JMIR Med Inform</source><year>2020</year><month>08</month><day>20</day><volume>8</volume><issue>8</issue><fpage>e18189</fpage><pub-id pub-id-type="doi">10.2196/18189</pub-id><pub-id pub-id-type="medline">32663146</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lindeman</surname><given-names>DA</given-names></name><name name-style="western"><surname>Kim</surname><given-names>KK</given-names></name><name name-style="western"><surname>Gladstone</surname><given-names>C</given-names></name><name name-style="western"><surname>Apesoa-Varano</surname><given-names>EC</given-names></name></person-group><article-title>Technology and caregiving: emerging interventions and directions for research</article-title><source>Gerontologist</source><year>2020</year><month>02</month><day>14</day><volume>60</volume><issue>Suppl 1</issue><fpage>S41</fpage><lpage>S49</lpage><pub-id pub-id-type="doi">10.1093/geront/gnz178</pub-id><pub-id pub-id-type="medline">32057082</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="web"><article-title>ChatGPT</article-title><source>OpenAI</source><year>2023</year><access-date>2023-09-21</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/chatgpt">https://openai.com/chatgpt</ext-link></comment></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>P</given-names></name><name name-style="western"><surname>Kamel Boulos</surname><given-names>MN</given-names></name></person-group><article-title>Generative AI in medicine and healthcare: promises, opportunities and challenges</article-title><source>Future Internet</source><year>2023</year><month>08</month><volume>15</volume><issue>9</issue><fpage>286</fpage><pub-id pub-id-type="doi">10.3390/fi15090286</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>De Angelis</surname><given-names>L</given-names></name><name name-style="western"><surname>Baglivo</surname><given-names>F</given-names></name><name name-style="western"><surname>Arzilli</surname><given-names>G</given-names></name><etal/></person-group><article-title>ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health</article-title><source>Front Public Health</source><year>2023</year><volume>11</volume><fpage>1166120</fpage><pub-id pub-id-type="doi">10.3389/fpubh.2023.1166120</pub-id><pub-id pub-id-type="medline">37181697</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sarraju</surname><given-names>A</given-names></name><name name-style="western"><surname>Bruemmer</surname><given-names>D</given-names></name><name name-style="western"><surname>Van Iterson</surname><given-names>E</given-names></name><name name-style="western"><surname>Cho</surname><given-names>L</given-names></name><name name-style="western"><surname>Rodriguez</surname><given-names>F</given-names></name><name name-style="western"><surname>Laffin</surname><given-names>L</given-names></name></person-group><article-title>Appropriateness of cardiovascular disease prevention recommendations obtained from a popular online chat-based artificial intelligence model</article-title><source>JAMA</source><year>2023</year><month>03</month><day>14</day><volume>329</volume><issue>10</issue><fpage>842</fpage><lpage>844</lpage><pub-id pub-id-type="doi">10.1001/jama.2023.1044</pub-id><pub-id pub-id-type="medline">36735264</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yeo</surname><given-names>YH</given-names></name><name name-style="western"><surname>Samaan</surname><given-names>JS</given-names></name><name name-style="western"><surname>Ng</surname><given-names>WH</given-names></name><etal/></person-group><article-title>Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma</article-title><source>Clin Mol Hepatol</source><year>2023</year><month>07</month><volume>29</volume><issue>3</issue><fpage>721</fpage><lpage>732</lpage><pub-id pub-id-type="doi">10.3350/cmh.2023.0089</pub-id><pub-id pub-id-type="medline">36946005</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Samaan</surname><given-names>JS</given-names></name><name name-style="western"><surname>Yeo</surname><given-names>YH</given-names></name><name name-style="western"><surname>Rajeev</surname><given-names>N</given-names></name><etal/></person-group><article-title>Assessing the accuracy of responses by the language model ChatGPT to questions regarding bariatric surgery</article-title><source>Obes Surg</source><year>2023</year><month>06</month><volume>33</volume><issue>6</issue><fpage>1790</fpage><lpage>1796</lpage><pub-id pub-id-type="doi">10.1007/s11695-023-06603-5</pub-id><pub-id pub-id-type="medline">37106269</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hurtz</surname><given-names>GM</given-names></name><name name-style="western"><surname>Chinn</surname><given-names>RN</given-names></name><name name-style="western"><surname>Barnhill</surname><given-names>GC</given-names></name><name name-style="western"><surname>Hertz</surname><given-names>NR</given-names></name></person-group><article-title>Measuring clinical decision making: do key features problems measure higher level cognitive processes?</article-title><source>Eval Health Prof</source><year>2012</year><month>12</month><volume>35</volume><issue>4</issue><fpage>396</fpage><lpage>415</lpage><pub-id pub-id-type="doi">10.1177/0163278712446639</pub-id><pub-id pub-id-type="medline">22605792</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><article-title>2023 Alzheimer&#x2019;s disease facts and figures</article-title><source>Alzheimers Dement</source><year>2023</year><month>04</month><volume>19</volume><issue>4</issue><fpage>1598</fpage><lpage>1695</lpage><pub-id pub-id-type="doi">10.1002/alz.13016</pub-id><pub-id pub-id-type="medline">36918389</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Scoring of responses generated from ChatGPT.</p><media xlink:href="aging_v7i1e53019_app1.docx" xlink:title="DOCX File, 20 KB"/></supplementary-material></app-group></back></article>