S6E8:Youngsoo Choi: Building Reliable Foundation Models
Release Date: 10/15/2025
Science in Parallel
Nearly a decade ago, the U.S Department of Veterans Affairs and the Department of Energy launched the MVP-CHAMPION initiative, not for sports, but as a data-driven strategy for improving healthcare outcomes for veterans and others. Silvia Crivelli of Lawrence Berkeley National Laboratory turned her skills in computational biology toward this new field, especially the problem of identifying veterans at high risk for suicide. As she and her colleagues worked on this challenge, large language models and the notion of foundation models emerged. Now her team is focused on a more comprehensive...
info_outlineScience in Parallel
Foundation models-- LLMs or LLM-like tools-- are a compelling idea for advancing scientific discovery and democratizing computational science. But there's a big gap between these lofty ideas and the trustworthiness of current models. Youngsoo Choi of Lawrence Livermore National Laboratory and his colleagues are thinking about to how to close this chasm. They're engaging with questions such as: What are the essential characteristics that define a foundation model? And how do we make sure that scientists can rely on their results? In this conversation we discuss a position paper that Youngsoo...
info_outlineScience in Parallel
Computational scientists can take on the role of utility players in research, and Steven Wilson is one example. At Arizona State University he's built instruments, carried out experiments and dove deep into computational work. As a postdoc, he's working on a new challenge: building a quantum chemistry startup company. In this episode, he discusses his career that started with 10 years in the United States Navy Nuclear Program, how that military experience shaped his academic studies and the role of the (DOE CSGF) in shaping his research to make chemical reactions more efficient. ...
info_outlineScience in Parallel
While we take a short summer break, we’re posting one of our favorite past episodes and a great follow-up to our last episode with Amanda Randles of Duke University. In 2023, we talked with Joe Insley of Argonne Leadership Computing Facility and Northern Illinois University about data visualization, from the practical process of helping researchers understand their results to showstopping images and animations that make the work accessible to broad audiences. Joe discusses his career path, how he and his team approach visualization projects, his work with students and his advice for...
info_outlineScience in Parallel
Duke University associate professor Amanda Randles' work to simulate and understand human blood flow and its implications demonstrates how high-performance computing paired with scientific principles can help improve human health. In this conversation, she talks about how she brought together early interests in physics, coding, biomedicine and even political science and policy and followed her enthusiasm for the Human Genome Project. She discusses how supercomputers are pushing the boundaries of what researchers can learn about the circulatory system noninvasively and how that knowledge,...
info_outlineScience in Parallel
Understanding how the brain works remains a grand scientific challenge, and it's yet another area where researchers are examining whether foundation models could help them find patterns in complex data. Joel Ye of Carnegie Mellon University talks about his work on foundation models, their potential and limitations and how others can get involved in applying these AI tools. is a Ph.D. student in the program in neural computation at Carnegie Mellon University in Pittsburgh, where he studies ways to understand brain data and brain-computer interfaces. He's a third-year
info_outlineScience in Parallel
Chemists and chemical engineers have modeled molecules for decades, but artificial intelligence and foundation models offer the prospect that researchers could train models with predictive abilities in one area of chemistry that could be fine-tuned for another. Trustworthy chemistry foundation models could help streamline the experimental time and resources needed to discover new medicines or design new batteries. Massachusetts Institute of Technology Ph.D. student Jackson Burns is working on these questions. He describes the promise and challenges of building foundation models in...
info_outlineScience in Parallel
In the second episode in our series on foundation models for science, we discuss Oak Ridge National Laboratory's work and hear about lessons learned from the recent 1000 Scientists AI Jam, a recent event that brought together researchers from several Department of Energy national laboratories, OpenAI and Anthropic. My guest is Prasanna Balaprakash, ORNL's director of AI programs. We talk about how foundation models could help with climate forecasts and his team's 2024 Gordon Bell finalist research and futuristic work that applies principles of swarm intelligence for managing distributed...
info_outlineScience in Parallel
Large language models aren't just powering chatbots like ChatGPT. This type of computational model is an example of a particular flavor of artificial intelligence known as foundation models, which are trained on vast amounts of data to make inferences in new areas. Although text is one rich data source, science offers many more from biology, chemistry, physics and more. Such models open up a tantalizing new set of research questions. How effective are foundation models for science? How could they be improved? Could they help researchers work on challenging questions? And what might they mean...
info_outlineScience in Parallel
Wrapping up our discussion of the 2024 Nobel Prizes in Physics and Chemistry, computer scientist Mansi Sakarvadia and computational structural biologist Josh Vermaas talk about the recent prizes and what they mean for science. You'll hear about how the prizes both break down research barriers and introduce concerns about misinformation and public trust. The research honored with the chemistry prize has already changed how researchers study questions that involve understanding proteins' structures. For more on the 2024 Nobel Prizes, check out . You'll meet: is a Ph.D. student in the and...
info_outlineFoundation models-- LLMs or LLM-like tools-- are a compelling idea for advancing scientific discovery and democratizing computational science. But there's a big gap between these lofty ideas and the trustworthiness of current models.
Youngsoo Choi of Lawrence Livermore National Laboratory and his colleagues are thinking about to how to close this chasm. They're engaging with questions such as: What are the essential characteristics that define a foundation model? And how do we make sure that scientists can rely on their results?
In this conversation we discuss a position paper that Youngsoo and his colleagues wrote to outline these questions and propose starting points for consensus-based answers and the challenges in building foundation models that are robust, reliable and generalizable. That paper also describes the Data-Driven Finite Element Method, or DD-FEM, a tool that they've developed for combining the power of AI and large datasets with physics-based simulation.
You’ll meet:
Youngsoo Choi is a staff scientist at Lawrence Livermore National Laboratory (LLNL) and a member of the lab's Center for Applied Scientific Computing (CASC), which focuses on computational science research for national security problems. Youngsoo completed his Ph.D. in computational and mathematical engineering at Stanford University and carried out postdoctoral research at Stanford and Sandia National Laboratories before joining Livermore in 2017.