Why is it so hard to elaborate what AI algorithm / technique they integrate? Would have made this article much better
dcanelhas 4 hours ago [-]
I'm half expecting to see "AI model" appearing as stand-in for "linear regression" at this point in the cycle.
ninjagoo 4 hours ago [-]
> I'm half expecting to see "AI model" appearing as stand-in for "linear regression" at this point in the cycle.
Already the case with consulting companies, have seen it myself
idiotsecant 51 minutes ago [-]
Some career do-nothing-but-make-noise in my organization hired a firm to 'Do AI' on some shitty data and the outcome was basically linear regression. It turns out that you can impressive executives with linear regression if you deliver it enthusiastically enough.
tasuki 3 minutes ago [-]
Tbh, often enough, linear regression is exactly what is needed.
blitzar 3 hours ago [-]
I'm half expecting to see "AI model" appearing as stand-in for "if > 0" at this point in the cycle.
Foobar8568 2 hours ago [-]
This is why I am programming now in Ocaml, files themselves are AI ( ml ).
srean 1 hours ago [-]
I am sure you did not forget that pattern matching.
Vetch 46 minutes ago [-]
This is essentially what any relu based neural network approximately looks like (smoother variants have replaced the original ramp function). AI, even LLMs, essentially reduce to a bunch of code like
let v0 = 0
let v1 = 0.40978399*(0.616*u + 0.291*v)
let v2 = if 0 > v1 then 0 else v1
let v3 = 0
let v4 = 0.377928*(0.261*u + 0.468*v)
let v5 = if 0 > v4 then 0 else v4...
samrus 35 minutes ago [-]
Thats a bit far. Relu does check x>0 but thats just one non-linearity in the linear/non-linear sandwich that makes up universal function approximator theorem. Its more conplex than just x>0
phire 4 hours ago [-]
I'm sure I've seen basic hill climbing (and other optimisation algorithms) described as AI, and then used evidence of AI solving real-world science/engineering problems.
LiamPowell 4 hours ago [-]
Historically this was very much in the field of AI, which is such a massive field that saying something uses AI is about as useful as saying it uses mathematics. Since the term was first coined it's been constantly misused to refer to much more specific things.
From around when the term was first coined: "artificial intelligence research is concerned with constructing machines (usually programs for general-purpose computers) which exhibit behavior such that, if it were observed in human activity, we would deign to label the behavior 'intelligent.'" [1]
That definition moves the goalposts almost by definition, people only stopped thinking that chess demonstrated intelligence when computers started doing it.
Eufrat 3 hours ago [-]
The term artificial intelligence has always been just a buzzword designed to sell whatever it needed to. IMHO, it has no meaningful value outside of a good marketing term. John McCarthy is usually the person who is given credit for coming up with the name and he has admitted in interviews that it was just to get eyeballs for funding.
coherentpony 60 minutes ago [-]
I am somewhat cynically waiting for the AI community to rediscover the last half a century of linear algebra and optimisation techniques.
At some point someone will realise that backpropagation and adjoint solves are the same thing.
yread 3 hours ago [-]
And why not, when linear regression works, it works so well it's basically magic, better than intelligence, artificial or otherwise
plasino 2 hours ago [-]
Having work with people who do that, I can guarantee that’s not the case.
See https://ssummers.web.cern.ch/conifer/ and HSL4ML, these run BDT and CNN
Staross 1 hours ago [-]
That works well to get around patents btw :)
etrautmann 3 hours ago [-]
It seems like most of the implementation is FPGA, which I wouldn’t call “physically burned into silicon.” That’s quite a stretch of language
vultour 4 hours ago [-]
Because if it’s not an LLM it’s not good for the current hype cycle. Calling everything AI makes the line go up.
danielbln 2 hours ago [-]
LLMs also make the cynicism go up among the HN crowd.
fnord77 15 minutes ago [-]
Thanks for tracking this down. I too am annoyed when so-called technical articles omit the actual techniques.
jgalt212 20 minutes ago [-]
Because it does not align with LLM Uber Alles.
jurschreuder 1 hours ago [-]
I've got news for you, everybody with a modern cpu uses this, which use a perceptron for branch prediction.
amelius 8 minutes ago [-]
At this point AI basically means "we didn't know how to solve the problem so we just threw a black box at it".
Glib, but it wont be cost effective at that small scale
quijoteuniv 5 hours ago [-]
A bit of hype in the AI wording here. This could be called a chip with hardcoded logic obtained with machine learning
FartyMcFarter 5 hours ago [-]
AI is not a new thing, and machine learned logic definitely counts as AI.
monkeydust 4 hours ago [-]
For those that have experience with ML, yes. For those that have recently become acquainted with it (more on business side) they seem to really struggle with this in my experience. '
volemo 4 hours ago [-]
Yeah, and don’t forget Eliza!
killingtime74 5 hours ago [-]
Is a LLM logic in weights derived from machine learning?
shlewis 5 hours ago [-]
Well, yes. That's literally what it is.
dmd 5 hours ago [-]
What what is? The article has nothing to do with LLMs. It even explicitly says they don’t use LLMs.
shlewis 2 hours ago [-]
> Is a LLM logic in weights derived from machine learning?
I was just answering this question. LLM logic in weights is fundamentally from machine learning, so yes. Wasn't really saying anything about the article.
quijoteuniv 5 hours ago [-]
Good one… but Is a DB query filter AI? I forgot to say though is sounds like a really cool thing to do
stingraycharles 5 hours ago [-]
Strictly speaking, expert systems are AI as well, as in, an expert comes up with a bunch of if/else rules. So yes technically speaking even if they didn’t acquire the weights using ML and hand-coded them, it could still be called AI.
phire 4 hours ago [-]
It is 100% valid to label an algorithm that plays tic-tac-toe as "AI"
Much of the early AI research was spent on developing various algorithms that could play board games.
Didn't even need computers, one early AI was MENACE [1], a set of 304 matchboxes which could learn how to play noughts and crosses.
Yup this is exactly my point, in the 80s there were plenty of “AI” companies and “fuzzy logic” was the buzzword of the day.
TORcicada 2 hours ago [-]
Thanks for the thoughtful comments and links really appreciated the high-signal feedback.
We've updated the article to better reflect the actual VAE-based AXOL1TL architecture (variational autoencoder for anomaly detection). Added the arXiv paper and Thea Aarrestad's talks to the Primary Sources.
Surac 4 hours ago [-]
Very important! This is not a LLM like the ones so often called AI these days. Its a neural network in a FPGA.
duskdozer 3 hours ago [-]
I guess shows the LLM-companies' marketing worked very well because that's what I immediately thought of.
IshKebab 4 hours ago [-]
> FPGA
So they aren't "burned into silicon" then? The article mentions FPGAs and ASICs but it's a bit vague. I would be surprised if ASICs actually made sense here.
fecal_henge 1 hours ago [-]
They make sense when you consider that 'on detector' electronics has all sorts of constraints that FPGAs cant compete on: Power, Density, Radiation hardness, Material budget.
armcat 3 hours ago [-]
Not on the same extreme level, but I know that some coffee machines use a tiny CNN based model locally/embedded. There is a small super cheap camera integrated in the coffee machine, and the model does three things: (1) classifies the container type in order to select type of coffee, (2) image segmentation - to determine where the cup/hole is placed, (3) regression - to determine the volume and regulate how much coffee to pour.
quantum_state 2 hours ago [-]
CERN has been doing HEP experiments for decades. What did it use before the current incarnation of AI? The AI label seems to be more marketing and superficial than substantial. It’s a bit sad that a place like CERN feels the need to make it public that it is on the bandwagon.
eqvinox 2 hours ago [-]
It doesn't say LLM anywhere.
quantum_state 2 hours ago [-]
Good catch. Corrected. Thanks!
WhyNotHugo 5 hours ago [-]
Intuitively, I’ve always had an impression that using an analogue circuit would be feasible for neural networks (they just matrix multiplication!). These should provide instantaneous output.
Isn’t this kind of approach feasible for something so purpose-built?
> CERN is using extremely small, custom large language models physically burned into silicon chips to perform real-time filtering of the enormous data generated by the Large Hadron Collider (LHC).
sh3rl0ck 6 hours ago [-]
There's no mention of SLMs or LLMs, though.
> This work represents a compelling real-world demonstration of “tiny AI” — highly specialised, minimal-footprint neural networks
FPGAs for Neural Networks have been s thing since before the LLM era.
5 hours ago [-]
100721 5 hours ago [-]
Huh? The first paragraph literally says they are using LLMs
> [ GENEVA, SWITZERLAND — March 28, 2026 ] — CERN is using extremely small, custom large language models physically burned into silicon chips to perform real-time filtering of the enormous data generated by the Large Hadron Collider (LHC).
SiempreViernes 5 hours ago [-]
the site might have fixed it, to me it says "artificial intelligence" instead of LLM, still bad but not" steaming pile of poo on you bank statement" bad
Are they some ancient small-scale integration VLSI design? Do they broadcast on a low-frequency VHF band? Face it: Oxymorons like those are part of the technical world. "VLSI" was a current term back when whole CPUs were made out of fewer transistors than we use for register files now, and "VHF" is low frequency even by commercial broadcasting standards.
rakel_rakel 5 hours ago [-]
haha, yea they are part of it for sure, and I'm not dunking on the use of them, but I rather smile a bit when I stumble upon them.
Like (~9K) Jumbo Frames!
Janicc 4 hours ago [-]
I think chips having a single LLM directly on them will be very common once LLMs have matured/reached a ceiling.
v9v 4 hours ago [-]
Do they actually have ASICs or just FPGAs? The article seems a bit unclear.
seydor 5 hours ago [-]
cern has been using neural networks for decades
mentalgear 5 hours ago [-]
That's what Groq did as well: burning the Transformer right onto a chip (I have to say I was impressed by the simplicity, but afterwards less so by their controversial Kushner/Saudi investment) .
NitpickLawyer 4 hours ago [-]
> That's what Groq did as well: burning the Transformer right onto a chip
Are you perhaps confusing Groq with the Etched approach? IIUC Etched is the company that "burned the transformer onto a chip". Groq uses LPUs that are more generalist (they can run many transformers and some other architectures) and their speed comes from using SRAM.
nerolawa 4 hours ago [-]
the fact that 99% of LHC data is just gone forever is insane
johngossman 17 minutes ago [-]
Not really. Think of the experiment as a very, very high speed camera. They can't store every frame, so they try to capture just the "interesting" ones. They also store some random ones that can be used later as controls or in case they realize they've missed something. That's the whole job of these various layers of algorithms: recognizing interesting frames. Sometimes a new experiment basically just changes the definition of "interesting"
randomNumber7 5 hours ago [-]
Does string theory finally make sense when we ad AI hallucinations?
quantum_state 2 hours ago [-]
This is a good one
logicallee 48 minutes ago [-]
I hope they have good results and keep all the data they need, and identify all the interesting data they're looking for. I do have a cautionary tale about mini neural networks in new experiments. We recently spent a large amount of time training a mini neural network (200k parameters) to make new predictions in a very difficult domain (predicting specific trails for further round collisions in a hash function than anyone did before.) We put up a spiffy internal dashboard[1] where we could tune parameters and see how well the neural network learns the existing results. We got to r^2 of 0.85 (that is very good correlation) on the data that already existed, from other people's records and from the data we solved for previously. It showed such a nicely dropping loss function as it trained, brings tears to the eye, we were pumped to see how it performs on data it didn't see before, data that was too far out to solve for. So many parameters to tune! We thought we could beat the world record by 1 round with it (40 instead of 39 rounds), and then let the community play with it to see if they can train it even better, to predict the inputs that let us brute force 42 round collisions, or even more. We could put up a leaderboard. The possiblities were endless, all it had to do was do extrapolate some input values by one round. We'd take the rest from there with the rest of our solving instrastructure.
After training it fully, we moved on to the inference stage, trying it on the round counts we didn't have data for! It turned out ... to have zero predictive ability on data it didn't see before. This is on well-structured, sensible extrapolations for what worked at lower round counts, and what could be selected based on real algabraic correlations. This mini neural network isn't part of our pipeline now.
When is the price of fabbing silicon coming down, so every SMB can do it?
IshKebab 3 hours ago [-]
My guess would be never. The closest you can get is "multi project wafers" where you get bundled with a load of other projects. As I understand it they're on the order of $100k which is cheap, but if you actually want to design and verify a chip you're looking at at least several million in salaries and software costs. Probably more like $10m, especially if you're paying US salaries. And of course that would be for a low performance design.
I think a better question would be "when are FPGAs going to stop being so ridiculously overpriced". That feels more possible to me (but still unlikely).
fc417fc802 1 hours ago [-]
Doesn't this vary wildly depending on the process node though? The cutting edge stuff keeps getting increasingly ridiculous meanwhile I thought you could get something like 50 nm for cheap. I also remember seeing years ago that some university had a ~micron (IIRC) process that you could order from.
100721 5 hours ago [-]
Does anyone know why they are using language models instead of a more purpose-built statistical model? My intuition is that a language model would either be overfit, or its training data would have a lot of noise unrelated to the application and significantly drive up costs.
5 years ago we would've called it a Machine Learning algorithm. 5 years before that, a Big Data algorithm.
IanCal 5 hours ago [-]
We’ve been calling neural nets AI for decades.
> 5 years before that, a Big Data algorithm.
The DNN part? Absolutely not.
I don’t know why people feel the need for such revisionism but AI has been a field encompassing things far more basic than this for longer than most commenters have been alive.
magicalhippo 5 hours ago [-]
> AI has been a field encompassing things far more basic than this for longer than most commenters have been alive.
When I was 13, having just started programming, I picked up a book from a "junk bin" at a book store on Artificial Intelligence. It must have been from the mid-80s if not older.
It had an entire chapter on syllogism[1] and how to implement a program to spit them out based on user input. As I recall it basically amounted to some string exteaction assuming user followed a template and string concatenation to generate the result. I distinctly recall not being impressed about such a trivial thing being part of a book on AI.
https://arxiv.org/html/2411.19506v1
Why is it so hard to elaborate what AI algorithm / technique they integrate? Would have made this article much better
Already the case with consulting companies, have seen it myself
From around when the term was first coined: "artificial intelligence research is concerned with constructing machines (usually programs for general-purpose computers) which exhibit behavior such that, if it were observed in human activity, we would deign to label the behavior 'intelligent.'" [1]
[1]: https://doi.org/10.1109/TIT.1963.1057864
At some point someone will realise that backpropagation and adjoint solves are the same thing.
https://www.youtube.com/watch?v=8IZwhbsjhvE (From Zettabytes to a Few Precious Events: Nanosecond AI at the Large Hadron Collider by Thea Aarrestad)
Page: https://www.scylladb.com/tech-talk/from-zettabytes-to-a-few-...
(Probably not for this here though.)
I was just answering this question. LLM logic in weights is fundamentally from machine learning, so yes. Wasn't really saying anything about the article.
Much of the early AI research was spent on developing various algorithms that could play board games.
Didn't even need computers, one early AI was MENACE [1], a set of 304 matchboxes which could learn how to play noughts and crosses.
[1] https://en.wikipedia.org/wiki/Matchbox_Educable_Noughts_and_...
So they aren't "burned into silicon" then? The article mentions FPGAs and ASICs but it's a bit vague. I would be surprised if ASICs actually made sense here.
Isn’t this kind of approach feasible for something so purpose-built?
> CERN is using extremely small, custom large language models physically burned into silicon chips to perform real-time filtering of the enormous data generated by the Large Hadron Collider (LHC).
> This work represents a compelling real-world demonstration of “tiny AI” — highly specialised, minimal-footprint neural networks
FPGAs for Neural Networks have been s thing since before the LLM era.
> [ GENEVA, SWITZERLAND — March 28, 2026 ] — CERN is using extremely small, custom large language models physically burned into silicon chips to perform real-time filtering of the enormous data generated by the Large Hadron Collider (LHC).
Like (~9K) Jumbo Frames!
Are you perhaps confusing Groq with the Etched approach? IIUC Etched is the company that "burned the transformer onto a chip". Groq uses LPUs that are more generalist (they can run many transformers and some other architectures) and their speed comes from using SRAM.
After training it fully, we moved on to the inference stage, trying it on the round counts we didn't have data for! It turned out ... to have zero predictive ability on data it didn't see before. This is on well-structured, sensible extrapolations for what worked at lower round counts, and what could be selected based on real algabraic correlations. This mini neural network isn't part of our pipeline now.
[1] screenshot: https://taonexus.com/publicfiles/mar2026/neural-network.png
I think a better question would be "when are FPGAs going to stop being so ridiculously overpriced". That feels more possible to me (but still unlikely).
5 years ago we would've called it a Machine Learning algorithm. 5 years before that, a Big Data algorithm.
> 5 years before that, a Big Data algorithm.
The DNN part? Absolutely not.
I don’t know why people feel the need for such revisionism but AI has been a field encompassing things far more basic than this for longer than most commenters have been alive.
When I was 13, having just started programming, I picked up a book from a "junk bin" at a book store on Artificial Intelligence. It must have been from the mid-80s if not older.
It had an entire chapter on syllogism[1] and how to implement a program to spit them out based on user input. As I recall it basically amounted to some string exteaction assuming user followed a template and string concatenation to generate the result. I distinctly recall not being impressed about such a trivial thing being part of a book on AI.
[1]: https://en.wikipedia.org/wiki/Syllogism
In the 1990s I remember taking my friend's IRC chat history and running it through a Markov model to generate drivel, which was really entertaining.
> The AXOL1TL V5 architecture comprises a VICReg-trained feature extractor stacked on top of a VAE.