AI integration in nephrology: evaluating ChatGPT for accurate ICD-10 documentation and coding
Issued Date
2024-01-01
Resource Type
eISSN
26248212
Scopus ID
2-s2.0-85204040702
Journal Title
Frontiers in Artificial Intelligence
Volume
7
Rights Holder(s)
SCOPUS
Bibliographic Citation
Frontiers in Artificial Intelligence Vol.7 (2024)
Suggested Citation
Abdelgadir Y., Thongprayoon C., Miao J., Suppadungsuk S., Pham J.H., Mao M.A., Craici I.M., Cheungpasitporn W. AI integration in nephrology: evaluating ChatGPT for accurate ICD-10 documentation and coding. Frontiers in Artificial Intelligence Vol.7 (2024). doi:10.3389/frai.2024.1457586 Retrieved from: https://repository.li.mahidol.ac.th/handle/123456789/101321
Title
AI integration in nephrology: evaluating ChatGPT for accurate ICD-10 documentation and coding
Corresponding Author(s)
Other Contributor(s)
Abstract
Background: Accurate ICD-10 coding is crucial for healthcare reimbursement, patient care, and research. AI implementation, like ChatGPT, could improve coding accuracy and reduce physician burden. This study assessed ChatGPT’s performance in identifying ICD-10 codes for nephrology conditions through case scenarios for pre-visit testing. Methods: Two nephrologists created 100 simulated nephrology cases. ChatGPT versions 3.5 and 4.0 were evaluated by comparing AI-generated ICD-10 codes against predetermined correct codes. Assessments were conducted in two rounds, 2 weeks apart, in April 2024. Results: In the first round, the accuracy of ChatGPT for assigning correct diagnosis codes was 91 and 99% for version 3.5 and 4.0, respectively. In the second round, the accuracy of ChatGPT for assigning the correct diagnosis code was 87% for version 3.5 and 99% for version 4.0. ChatGPT 4.0 had higher accuracy than ChatGPT 3.5 (p = 0.02 and 0.002 for the first and second round respectively). The accuracy did not significantly differ between the two rounds (p > 0.05). Conclusion: ChatGPT 4.0 can significantly improve ICD-10 coding accuracy in nephrology through case scenarios for pre-visit testing, potentially reducing healthcare professionals’ workload. However, the small error percentage underscores the need for ongoing review and improvement of AI systems to ensure accurate reimbursement, optimal patient care, and reliable research data.