Navigating the Regulatory Pitfalls of AI Charting
On February 4, Epic rolled out its AI Charting application, which offers a built-in set of AI capabilities for drafting clinician notes and recommending orders. Seven healthcare organizations, including the nonprofit health plan Group Health Cooperative of South Central Wisconsin, adopted the tool upon its launch.
AI scribes listen to conversations between patients and providers and generate clinical documentation in real-time, while ambient AI works in the background to suggest information for clinicians to follow up on. The term AI charting encompasses the entire process of using AI to create, update, and manage patient charges and clinical documentation, including health alerts within the EHR (Electronic Health Record).
Epic originally announced AI Charting as part of its AI tool for clinicians, Art, at its annual Users Group Meeting last August. Alongside AI Charting, Epic introduced Penny, an AI revenue cycle tool, and Emmie, an AI tool that assists patients with tasks such as scheduling appointments and making payments.
According to Epic, AI Charting allows clinicians to stay focused on patient care while the AI tool builds a complete record of a patient visit. Similar tools in the market include Abridge, Suki, and ModMed, which provide various functionalities to enhance clinical documentation workflows.
Understanding the Regulatory Pitfalls Associated with AI Charting
Although AI tools decrease administrative burdens, healthcare systems must ensure that the data sent to CMS (Centers for Medicare & Medicaid Services) or private payers is accurate. The use of AI tools for billing and documentation necessitates proper oversight to avoid upcoding, according to experts.
Additionally, tools like Epic’s AI Charting are subject to HIPAA compliance, raising privacy and security concerns. Organizations must also ensure compliance with Systems and Organization Controls 2 (SOC 2), a data privacy standard that guides the management of client data.
Healthcare systems face challenges with evolving FDA guidelines, with some apps being classified as medical devices requiring FDA approval, while others are not. Although AI Charting is likely not classified as a medical device, if it were, it would be subject to heightened regulatory scrutiny.
Best Practices for Investing in AI Charting
When considering investment in AI clinical documentation and ambient listening tools, organizations should follow these strategies:
- Update Patient Consent Forms: Organizations should update consent forms to inform patients that AI tools are being utilized, ensuring compliance with guidelines like HIPAA.
- Keep a Human in the Loop: It is crucial for healthcare organizations to involve clinicians in the selection and implementation of AI tools, ensuring that human oversight remains a critical component.
- Understand State and Federal Requirements: Organizations should study state guidelines regarding AI tools in conjunction with federal regulations like HIPAA, as compliance is not uniform across jurisdictions.
- Monitor AI Bias: Organizations must be vigilant about potential biases in data collection practices and ensure that AI systems are designed to avoid swaying toward particular ethnicities.
- Create a Heat Map: A heat map can help evaluate the level of regulatory compliance in various areas of the organization, indicating where improvements are needed.
- Understand Vendor Data Usage: It is essential to vet vendors and their AI software to protect patient data and minimize risks related to inappropriate disclosures.
By adhering to these best practices, healthcare organizations can effectively navigate the regulatory landscape surrounding AI charting while optimizing their clinical documentation processes.