Tech

Interactive virtual companion to accelerate discoveries at scientific user facilities

Share
Share
Interactive virtual companion to accelerate discoveries at scientific user facilities
VISION aims to lead the natural-language-controlled scientific expedition with joint human-AI force for accelerated scientific discovery at user facilities. Credit: Brookhaven National Laboratory

A team of scientists at the U.S. Department of Energy’s (DOE) Brookhaven National Laboratory have dreamed up, developed, and tested a novel voice-controlled artificial intelligence (AI) assistant designed to break down everyday barriers for busy scientists.

Known as the Virtual Scientific Companion, or VISION, the generative AI tool—developed by researchers at the Lab’s Center for Functional Nanomaterials (CFN) with support from experts at the National Synchrotron Light Source II (NSLS-II)—offers an opportunity to bridge knowledge gaps at complex instruments, carry out more efficient experiments, save scientists’ time, and overall, accelerate scientific discovery.

The idea is that a user simply has to tell VISION in plain language what they’d like to do at an instrument and the AI companion, tailored to that instrument, will take on the task—whether it’s running an experiment, launching data analysis, or visualizing results. The Brookhaven team recently shared details about VISION in a paper published in Machine Learning: Science and Technology.

“I’m really excited about how AI can impact science and it’s something we as the scientific community should definitely explore,” said Esther Tsai, a scientist in the AI-Accelerated Nanoscience group at CFN. “What we can’t deny is that brilliant scientists spend a lot of time on routine work. VISION acts as an assistant that scientists and users can talk to for answers to basic questions about the instrument capability and operation.”

VISION highlights the close partnership between CFN and NSLS-II, two DOE Office of Science user facilities at Brookhaven Lab. Together they collaborate with facility users on the setup, scientific planning, and analysis of data from experiments at three NSLS-II beamlines, highly specialized measurement tools that enable researchers to explore the structure of materials using beams of X-rays.

Tsai, inspired to alleviate bottlenecks that come with using NSLS-II’s in-demand beamlines, received a DOE Early Career Award in 2023 to develop this new concept. Tsai now leads the CFN team behind VISION, which has collaborated with NSLS-II beamline scientists to launch and test the system at the Complex Materials Scattering (CMS) beamline at NSLS-II, demonstrating the first voice-controlled experiment at an X-ray scattering beamline and marking progress towards the world of AI-augmented discovery.

“At Brookhaven Lab, we’re not only leading in researching this frontier scientific virtual companion concept, we’re also being hands-on, deploying this AI technique on the experimental floor at NSLS-II and exploring how it can be useful to users,” Tsai said.

Talking to AI for flexible workflows

VISION leverages the growing capabilities of large language models (LLMs), the technology at the heart of popular AI assistants such as ChatGPT.

An LLM is an expansive program that creates text modeled on natural human language. VISION exploits this concept, not just to generate text for answering questions but also to generate decisions about what to do and computer code to drive an instrument. Internally, VISION is organized into multiple “cognitive blocks,” or cogs, each comprising an LLM that handles a specific task. Multiple cogs can be put together to form a capable assistant, with the cogs carrying out work transparently for the scientist.

“A user can just go to the beamline and say, ‘I want to select certain detectors’ or ‘I want to take a measurement every minute for five seconds’ or ‘I want to increase the temperature’ and VISION will translate that command into code,” Tsai said.

Those examples of natural language inputs, whether speech, text, or both, are first fed to VISION’s “classifier” cog, which decides what type of task the user is asking about. The classifier routes to the right cog for the task, such as an “operator” cog for instrument control or “analyst” cog for data analysis.

Then, in just a few seconds, the system translates the input into code that’s passed back to the beamline workstation, which the user can review before executing. On the back end, everything is being run on “HAL,” a CFN server optimized for running AI workloads on graphics processing units.

VISION’s use of natural language—that is, how people normally speak—is its key advantage. Since the system is tailored to the instrument the researcher is using, they are liberated from spending time setting up software parameters and can instead focus on the science they are pursuing.

“VISION acts as a bridge between users and the instrumentation, where users can just talk to the system and the system takes care of driving experiments,” said Noah van der Vleuten, a co-author who helped develop VISION’s code generation capability and testing framework. “We can imagine this making experiments more efficient and giving people a lot more time to focus on the science, rather than becoming experts in each instrument’s software.”

The ability to speak to VISION, not only type a prompt, could make workflows even faster, team members noted.

In the fast-paced, ever-evolving world of AI, VISION’s creators also set out to build a scientific tool that can keep up with improving technology, incorporate new instrument capabilities, and scale up as needed to seamlessly navigate multiple tasks.

“A key guiding principle is that we wanted to be modular and adaptable, so we can quickly swap out or replace with new AI models as they become more powerful,” said Shray Mathur, first author on the paper who worked on VISION’s audio-understanding capabilities and overall architecture. “As underlying models become better, VISION becomes better. It’s very exciting because we get to work on some of the most recent technology, and deploy it immediately. We’re building systems that can really benefit users in their research.”

This work builds on a history of AI and machine learning (ML) tools developed by CFN and NSLS-II to aid facility scientists, including for autonomous experiments, data analytics, and robotics. Future versions of VISION could act as a natural interface to these advanced AI/ML tools.

A VISION for an AI Lab Partner
The team behind VISION at the CMS beamline. From left to right, co-authors and Center for Functional Nanomaterials scientists Shray Mathur, Noah van der Vleuten, Kevin Yager, and Esther Tsai. Credit: Kevin Coughlin/Brookhaven National Laboratory

A VISION for human-AI partnership

Now that the architecture for VISION is developed and at a stage where it has an active demonstration at the CMS beamline, the team aims to test it further with beamline scientists and users and eventually bring the virtual AI companion to additional beamlines.

This way, the team can have real discussions with users about what’s truly helpful to them, Tsai said.

“The CFN/NSLS-II collaboration is really unique in the sense that we are working together on this frontier AI development with language models on the experimental floor, at the front-line supporting users,” Tsai said. “We’re getting feedback to better understand what users need and how we can best support them.”

Tsai offered a huge thanks to CMS lead beamline scientist Ruipeng Li for his support and openness to ideas for VISION.

The CMS beamline has already been a testing ground for AI/ML capabilities, including autonomous experiments. When the idea of bringing VISION to the beamline came up, Li saw it as an exciting and fun opportunity.

“We’ve been close collaborators and partners since the beamline was built more than eight years ago,” Li noted. “These concepts enable us to build on our beamline’s potential and continue to push the boundaries of AI/ML applications for science. We want to see how we can learn from this process because we are riding the AI wave now.”

In the bigger picture of AI-augmented scientific research, the development of VISION is a step towards realizing other AI concepts across the DOE complex, including a science exocortex.

Kevin Yager, the AI-Accelerated Nanoscience Group leader at CFN and VISION co-author, envisions the exocortex as an extension of the human brain that researchers can interact with through conversation to generate inspiration and imagination for scientific discovery.

“When I imagine the future of science, I see an ecosystem of AI agents working in coordination to help me advance my research,” Yager said. “The VISION system is an early example of this future—AI assistant that helps you operate an instrument. We want to build more AI assistants, and connect them together into a really powerful network.”

More information:
Shray Mathur et al, VISION: a modular AI assistant for natural human-instrument interaction at scientific user facilities, Machine Learning: Science and Technology (2025). DOI: 10.1088/2632-2153/add9e4

Provided by
Brookhaven National Laboratory


Citation:
Interactive virtual companion to accelerate discoveries at scientific user facilities (2025, June 26)
retrieved 26 June 2025
from

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
Most organisations are at risk thanks to immature supply chain security
Tech

Most organisations are at risk thanks to immature supply chain security

Supply chain attacks are becoming more frequent and more dangerous Many security...

North Korean hackers are hijacking Zoom calls to steal your crypto with scripts buried 10,000 lines deep
Tech

North Korean hackers are hijacking Zoom calls to steal your crypto with scripts buried 10,000 lines deep

Fake Zoom scripts launch malware hidden beneath thousands of lines of code...

I sure hope my PS5 doesn’t die before GTA 6 – reports suggest hardware issues could damage consoles over time
Tech

I sure hope my PS5 doesn’t die before GTA 6 – reports suggest hardware issues could damage consoles over time

Alderon Games founder Matthew Cassells reports supposed liquid metal issues on PS5...