COMPREHENSIVENESS AND READABILITY OF CHATGPT’S ANSWERS TO DEMENTIA FAMILY CAREGIVERS’ QUESTIONS
2024

Evaluating ChatGPT's Answers for Dementia Caregivers

Sample size: 24 publication Evidence: moderate

Author Information

Author(s): Han Soojeong, Aryoyudanta Bayu, Choi Yong Kyung

Primary Institution: Columbia University School of Nursing

Hypothesis

Can ChatGPT provide comprehensive and readable answers to dementia family caregivers' questions about transitioning to hospice and palliative care?

Conclusion

ChatGPT's answers were found to be only partially comprehensive and very difficult to read.

Supporting Evidence

  • 41.67% of ChatGPT's answers were scored as comprehensive.
  • The median comprehensiveness score was 2.00.
  • The mean FRE score of original answers was 25.37, indicating 'very difficult to read'.
  • The mean FRE score of revised answers was 46.42, indicating 'difficult to read'.
  • The t-test revealed significant differences between the two answers on the FRE scale.

Takeaway

This study looked at how well ChatGPT answers questions from family caregivers of dementia patients, and it found that the answers were hard to understand.

Methodology

Data was extracted from a public online forum, and 24 questions were evaluated using ChatGPT for comprehensiveness and readability.

Limitations

The study used convenience sampling and focused on a limited number of questions.

Participant Demographics

Family caregivers of dementia patients.

Statistical Information

P-Value

0.0001

Statistical Significance

p<0.0001

Digital Object Identifier (DOI)

10.1093/geroni/igae098.3943

Want to read the original?

Access the complete publication on the publisher's website

View Original Publication