Dr. Scott Gottlieb is a CNBC contributor and is a member of the boards of Pfizer, genetic testing startup Tempus, health-care tech company Aetion Inc. and biotech company Illumina. He is also a partner at the venture capital firm New Enterprise Associates.
Researchers at Harvard presented a study demonstrating an achievement that would challenge any medical student. ChatGPT, a large language model, passed the U.S. Medical Licensing Exam, outperforming about 10 percent of medical students who fail the test annually.
The inevitable question isn’t so much if but when these artificial intelligence devices can step into the shoes of doctors. For some tasks, this medical future is sooner than we think.
To grasp the potential of these tools to revolutionize the practice of medicine, it pays to start with a taxonomy of the different technologies and how they’re being used in medical care.
The AI tools being applied to health care can generally be divided into two main categories. The first is machine learning, which uses algorithms to enable computers to learn patterns from data and make predictions. These algorithms can be trained on a variety of data types, including images.
The second category encompasses natural language processing, which is designed to understand and generate human language. These tools enable a computer to transform human language and unstructured text into machine-readable, organized data. They learn from a multitude of human trial-and-error decisions and emulate a person’s responses.
A key difference between the two approaches resides in their functionality. While machine learning models can be trained to perform specific tasks, large language models can understand and generate text, making them especially useful for replicating interactions with providers.
In medicine, the use of these technologies is generally following one of four different paths. The first encompass large language models that are applied to administrative functions such as processing medical claims or creating and analyzing medical records. Amazon’s HealthScribe is a programmable interface that transcribes conversations between doctors and patients and can extract medical information, allowing providers to create structured records of encounters.
The second bucket involves the use of supervised machine learning to enhance the interpretation of clinical data. Specialties such as radiology, pathology and cardiology are already using AI for image analysis, to read MRIs, evaluate pathology slides or interpret electrocardiograms. In fact, up to 30% of radiology practices have already adopted AI tools. So have other specialties. Google Brain AI has developed software that analyzes images from the back of the eye to diagnose diabetic macular edema and diabetic retinopathy, two common causes of blindness.
Since these tools offer diagnoses and can directly affect patient care, the FDA often categorizes them as medical devices, subjecting them to regulation to verify their accuracy. However, the fact that these tools are trained on closed data sets, where the findings in data or imaging have been rigorously confirmed, gives the FDA increased confidence when assessing these devices’ integrity.
The third broad category comprises AI tools that rely on large language models that extract clinical information from patient-specific data, interpreting it to prompt providers with diagnoses or treatments to consider. Generally known as clinical decision support software, it evokes a picture of an brainy assistant designed to aid, not to supplant, a doctor’s judgment. IBM’s