“AI Agents for HEP/NP: From Automation to Knowledge-Driven User Support..” ?
Paul
From:
Torre Wenaus <wenaus AT gmail.com>
Date: Tuesday, April 15, 2025 at 19:38
To: Paul Nilsson <Paul.Nilsson AT cern.ch>
Cc: mkirby AT bnl.gov <mkirby AT bnl.gov>, NPPS leadership team <Phys-npps-mgmt-l AT lists.bnl.gov>, tmaeno AT bnl.gov <tmaeno AT bnl.gov>, Michel Hernandez Villanueva <mhernande1 AT bnl.gov>
Subject: Re: [[Phys-npps-mgmt-l] ] Fwd: "Shovel ready AI" proposals
I really agree with you Paul...
> I’m more interested in log analysis than chatbots (we are going to get that for free so to speak)
The problem I've had with working directly on chatbots is that you do get that for free, and it is moving extremely rapidly, so it seems doubly pointless to work on the chatbot proper. A few months ago RAG
was the way to extend an LLM to specialized knowledge, now comes MCP which obsoletes much of that. But MCP is working with your own system to integrate it with that fast moving AI ecosystem, it will apply to and leverage the 'LLM of the day'. It's a safe worthwhile
investment. And if we get funding, we can apply effort to not just the MCP agent itself (Claude is capable of writing the MCP interface automatically given an API), but to what the MCP is serving up, ie what you're already doing or planning to do.
If you can think of a better title let me know. It is more than 'automation', it serves users as well with high quality information (ops, shifters, analyzers) in an AI context like an LLM, but how to say that
in 2 words.
I really appreciate the input! Some quick comments...
- has to be new and eyecatching, LLMs to improve doc doesn't cut it
- Kirby: yes! as Paul mentioned he's started some of this already, MCP provides a powerful context to do this and more
- Michel: great, will include Belle II ops/shift
- not supported in all models at present, not available in free commercial models, takes more GPU power than we have... all true but not blocking issues, if they give us money for a proposal, all are solved
:-)
Dmitri wants a slide now, for a 5pm meeting with the bosses. Draft for your quick contributions and comments...
Hi,
A few ideas; In principle, the “Ask PanDA”-tool I’m working on could be turned into an “intelligent agent” using MCP. The new version is anyway plug-in
based, so switching to a technology that is essentially doing just that for you is preferable. However, as far as I understand this doesn’t support all models yet and some are via third-party (or even wrappers), as seems to be the case for Llama which might
still be a bit experimental. A limiting factor is also that we don’t have access to that many GPUs, unless we are going down the commercial road. I’m just getting set up at SLAC to start testing Llama, using their GPUs (two I believe, ie the same situation
as NERSC and BNL). As for CERN, they seem to be planning an LLM as a service, but it seems that is at a very early planning stage.
I think it’s good to get some experience with MCP technology and could be developed into something more than a chatbot – actually, personally, I’m
more interested in log analysis than chatbots (we are going to get that for free so to speak), and we also have some ideas about the panda monitor, related to error messages, what they mean and so on. A worrying factor is, as I said, the limited number of
available GPUs. Two GPUs only gives a handful of tokens per second at SLAC e.g., so hardly for production. But we need to start somewhere.
Cheers,
Paul
Hi Torre,
The idea of having something like MCP cooked into services definitely seems like it would have significant advantages by making getting a “data source”
connected via an easy plugin. If I understand correctly, MCP acts as a standardization interface for getting real-time data into an AI client? So for PanDA, you could take MCP to slurp in the current data on running jobs, queues, and then let a user ask the
question “how soon will my jobs be finished?”, and Claude will tell them if they should get a coffee or go to dinner or go on vacation? That’s maybe oversimplified, but it’s absolutely a useful type of thing that would have a huge impact on user experience
and/or operational decisions.
Kirby
Please have a look at the following, tell me if you think it is crazy. Hong wants to have 1-2 pages describing how we could use existing effort in
this FY in order to attract DOE AI money for this FY that they are looking to disperse quickly. I think that in an MCP directed plan as below we could identify time fractions of people in several areas of activity/expertise to get experimental MCP services
in place and client(s) to use them, and deliver something useful in a short time. You can tell me whether I'm crazy, and tell me what we should be proposing to do for this FY25 scenario instead :-)
I include at the bottom the referenced material in the budget briefing.
These are some musings after a weekend trying out pieces of the emerging 'agentic ecosystem'.
A big domain that is about to open up (as big as LLM tech itself, I think) is active agents representing services that can be interrogated and used
by software, LLMs and other human interfaces to create intelligent assistants/actors with powerful capability to act in real time on completely current information, info direct from and curated by the services themselves (no hallucinations from garbage-in
data, no blindness to the time since training). Anthropic established an open source protocol, Model Context Protocol (MCP) a few months ago and it is taking off fast. In my opinion it is worth directly getting into now, much more so than playing around with
LLM chatbots and extensions like RAG. It is the sort of tech that can be the real basis for "end to end integrated AI from collider to detector to analysis" for EIC, for greatly increasing intelligent automation and user-facing capability in the systems and
services we develop like PanDA, etc. This is facility/ops, it is autonomous control of accelerators/experiments, it fits to where the funding apparently is. It is shovel-ready, I think we could start on this instantly in contexts like PanDA, EIC streaming
readout, and the ops environments of running experiments if they were willing to explore the possibilities. It is work that should have significant short-term return from moderate time investment so we could put existing people on it without killing their
useful productivity.
I'll be talking to people about this but I haven't yet. If you want to include something about this in what you're preparing this week, let me know
what you'd like and I'll fast-track the discussing, see whether others agree :-)
Thank you for forwarding!
Hi Torre,
If you have ideas about shovel ready AI projects, please let us know, or work with the other group leaders.
This can potentially recover some of the lost research funding.
Best,
Hong.
From:
Denisov, Dmitri <denisovd AT bnl.gov>
Date: Friday, April 11, 2025 at 4:25 PM
To: Rajagopalan, Srini <srinir AT bnl.gov>, Begel, Michael <begel AT bnl.gov>, Kettell, Steven <kettell AT bnl.gov>,
sallydawsonbnl AT gmail.com <sallydawsonbnl AT gmail.com>, Slosar, Anze <anze AT bnl.gov>, Jaffe,
David <djaffe AT bnl.gov>
Cc: Kotcher, Jonathan <kotcher AT bnl.gov>, Ma, Hong <hma AT bnl.gov>, Deshpande, Abhay <abhay AT bnl.gov>
Subject: "Shovel ready AI" proposals
Folks,
As you did hear OHEP is interested in “shovel ready” AI proposals to at least partly compensate FY25 reductions. You can read slides 11-12 from Alan’s
recent summary (attached) – this is all we know about this topic for now. In order to get organized and as we are meeting with JoAnne on this topic next Tuesday, I suggest all of you consider what activities in your groups could be “recolored” as AI or become
AI with little changes, so the funding can start in FY25.
We will need by next Tuesday morning a few initial examples (1-3 from each area):
1. Title and Abstract (a few sentences, not long).
2. Availability to start activities in FY25.
3. Approximate duration and funding required (for the full duration of the activities and in FY25). I recommend the full time is 1-2 years,
not more.
If you send me your drafts by Tuesday, April 15 morning, I’ll combine them into presentable document to discuss with the lab.
Many thanks, Dmitri.
wAre there opportunities for additional funding in FY 2025?
•Yes, with successful proposals to Hardware-Aware AI and Early Career
Research
•Maybe.
In AI. We will be receptive to shovel ready, well-coordinated AI projects
oIncluding Facilities/Operations
wCosmologists and particle physicists are
early adopters and developers of AI
•An epoch of advanced AI is an epoch of discoveries in fundamental
physics and cosmology
wHEP is built on statistical analyses of PB-scale data to test theory
•HEP requires a deep understanding of probabilities in the interpretation
of data
•This requirement is driving development of AI that can handle probabilistic
rigor
oProbabilistically rigorous AI would be a game changer for many fields
of science, in addition to being invaluable for applications beyond science
•PCAST
Report on Supercharging Research: Harnessing Artificial Intelligence to Meet Global Challenges
wHEP uniquely pushes development of AI at the edge and real time applications
•Autonomous control of particle accelerators and detectors
•AI for applications in remote power constrained environments
oLow-power high performance AI is most likely to be advanced by HEP
wHEP trains an AI ready workforce with experience in real-world applications
using AI on PB scale data
Michael Kirby (he/him/his)
Brookhaven National Laboratory

--
-- Torre Wenaus, BNL NPPS Group Leader, ATLAS and ePIC experiments
-- BNL 510A 1-222 | 631-681-7892
--
-- Torre Wenaus, BNL NPPS Group Leader, ATLAS and ePIC experiments
-- BNL 510A 1-222 | 631-681-7892
|