The use case for artificial intelligence (AI) and machine learning in healthcare is becoming compelling as the world grapples with the global coronavirus pandemic. But AI has myriad applications in routine healthcare as well and will soon be used regularly throughout the medical field.
As the use of technology and AI proliferate across the healthcare industry, the inevitability of mistakes being made becomes a reality. The whole issue ultimately begs the question of who is liable when AI makes a mistake that is harmful to a patient.
There are no easy answers to this question, particularly in an industry as litigious as healthcare. A recent STAT article breaks the issue down to the AI algorithms used in a particular scenario. If an algorithm is developed directly by a medical facility, that facility would be responsible for any AI mistakes through the legal definition of enterprise liability.
Basically, if a healthcare facility uses AI and removes humans from the decision making process, they will be liable for any mistakes made. However, if the AI algorithm was purchased commercially, the issue becomes more complex. The concept of preemption could be invoked in this case, which means a vendor would lose some of its risk.
Preemption happens if state and federal laws conflict. In the case of the healthcare industry, if a product or technology isn’t mandated by state law, federal law would prevail. And that means the AI vendor would not be required to comply with state laws in this case, and the healthcare facility would be responsible for any problems related to an AI-based decision or diagnosis.
AI, from a litigation standpoint, is somewhat in limbo right now. That’s because AI algorithms are not considered static products, and vendors – let alone the FDA, cannot predict how they will perform in the future. AI algorithms are not currently classified as drugs or devices, and that distinction will influence how the FDA and the court systems handle their successes and failures moving forward.
To further complicate matters, many AI algorithms are being expedited for approval right now. The coronavirus pandemic is only speeding the way for anything that could facilitate healthcare decision making and treatment processes, including developing technologies like AI and machine learning.
The takeaway is that it’s still largely unclear who will be responsible for mistakes made by AI algorithms and software. Healthcare providers should absolutely make use of new technologies to augment and help them, particularly during times of crisis. But they should also be vigilant about AI outcomes and ensure humans are safe checking important decisions in the event AI makes a mistake.
To learn more about how AI technologies are transforming the healthcare industry and beyond, TMC is hosting the Future of Work Expo from June 22-25, 2021 at the Miami Beach Convention Center. The event will explore how AI and machine learning can improve healthcare and business applications, communications, collaboration, contact center and customer service, and marketing and sales experiences and initiatives.
Future of Work Contributor
Intapp, a provider of AI-powered software for professionals in specialized industries, announced new features and a refreshed brand focused on its "In…
Akto recently launched its GenAI Security Testing solution to offer proactive security testing specifically designed for GenAI models and their APIs.
Quantive's new StrategyAI solution hones organizations' efforts for developing, executing, and evaluating smarter, "always-on" business strategies.
The Gen AI fabric platform aims to simplify and accelerate the development and implementation of GenAI applications, which lets businesses harness the…
Eric Vaughan delivered an in-depth keynote presentation on the latest technology revolution, catalyzed by artificial intelligence (AI).