An agency-wide LLM called Elsa was released weeks ahead of schedule:
Under the Trump administration, the Food and Drug Administration is eagerly embracing artificial intelligence tools that staff members are reportedly calling rushed, buggy, overhyped, and inaccurate.
On Monday, the FDA publicly announced the agency-wide rollout of a large language model (LLM) called Elsa, which is intended to help FDA employees—"from scientific reviewers to investigators." The FDA said the generative AI is already being used to "accelerate clinical protocol reviews, shorten the time needed for scientific evaluations, and identify high-priority inspection targets."
"It can summarize adverse events to support safety profile assessments, perform faster label comparisons, and generate code to help develop databases for nonclinical applications," the announcement promised.
In a statement, FDA Chief AI Officer Jeremy Walsh trumpeted the rollout, saying: "Today marks the dawn of the AI era at the FDA[. W]ith the release of Elsa, AI is no longer a distant promise but a dynamic force enhancing and optimizing the performance and potential of every employee."
Meanwhile, FDA Commissioner Marty Makary highlighted the speed with which the tool was rolled out. "I set an aggressive timeline to scale AI agency-wide by June 30," Makary said. "Today's rollout of Elsa is ahead of schedule and under budget, thanks to the collaboration of our in-house experts across the centers."
However, according to a report from NBC News, Elsa could have used some more time in development. FDA staff tested Elsa on Monday with questions about FDA-approved products or other public information, only to find that it provided summaries that were either completely or partially wrong.
FDA staffers who spoke with Stat news, meanwhile, called the tool "rushed" and said its capabilities were overinflated by officials, including Makary and those at the Department of Government Efficiency (DOGE), which was headed by controversial billionaire Elon Musk. In its current form, it should only be used for administrative tasks, not scientific ones, the staffers said.
"Makary and DOGE think AI can replace staff and cut review times, but it decidedly cannot," one employee said. The staffer also said that the FDA has failed to set up guardrails for the tool's use. "I'm not sure in their rush to get it out that anyone is thinking through policy and use," the FDA employee said.
According to Stat, Elsa is based on Anthropic's Claude LLM and is being developed by consulting firm Deloitte. Since 2020, Deloitte has been paid $13.8 million to develop the original database of FDA documents that Elsa's training data is derived from. In April, the firm was awarded a $14.7 million contract to scale the tech across the agency. The FDA said that Elsa was built within a high-security GovCloud environment and offers a "secure platform for FDA employees to access internal documents while ensuring all information remains within the agency."
(Score: 4, Insightful) by PiMuNu on Tuesday June 10, @02:42PM (1 child)
I worked with colleagues to develop some multivariate pattern recognition to distinguish pions from other subatomic particles in our detectors using one of these fancy fitting routines in about 2005. I recall asking a graduate student what "Machine Learning" meant in about 2015, when the hype was building, to find out it meant any algorithm that has state. At that point I decided it was BS and I haven't seen much evidence to change my mind.
The new class have developed a fancy cut n paste engine, which is nice for those jobs that are all about cut n paste, and can indeed save time.
(Score: 2) by corey on Wednesday June 11, @02:05AM
Yeah, I'm sure your algo was much more efficient and used a fraction of the energy and computing infrastructure an equivalent performing 'AI' system would. This is the main problem I have with it all, it's another example of hardware getting faster, smaller and more energy efficient, but software going the other way at 1.1x the rate, so we go backwards yet require so much more resources to achieve the same thing. The other problem I have is that it's "the next big thing" and "will change everything", so the management want to switch to using 'AI' systems, so they can report they achieved something or "progressed" forward and marketing love it. Oh, the third problem I have is that you were a domain expert in the subatomic particle patterns, so you knew how the algo worked. If someone throws an 'AI' system at it (let's say install a box from OpenAI running some black-box software), nobody knows how the thing works or recognises the patterns, except the people in California who wrote the 'AI' software itself.