3/29/2025 at 2:08:11 PM
I work in the ICU monitoring field, on the R&D team of a company with live systems at dozens of hospitals and multiple FDA approvals. We use extended Kalman filters (i.e. non-blackbox "ML") to estimate certain lab values of patients that are highly indicative of them crashing, based on live data from whatever set of monitors they're hooked up to - and it's highly robust.What the authors of this paper are doing is throwing stuff at the wall to see if it works, and publishing results. That's not necessarily a bad thing at all, but I say this to underline that their results are not at all reflective of SOTA capabilities, and they're not doing much exploration of prior art.
by wswope
3/29/2025 at 2:41:43 PM
Calling EKFs "ML" is certainly a choice.by AlotOfReading
3/29/2025 at 7:00:03 PM
It is a reasonable choice, and especially with the quotes around it, completely understandable.The distinction between statistical inference and machine learning is too blurry to police Kalman filters onto one side.
by getnormality
3/29/2025 at 5:00:52 PM
It's machine learning until you understand how it works, then it's just control theory and filters again.by idiotsecant
3/29/2025 at 5:33:01 PM
Diffusion models are a happy middle ground. :-)by whatshisface
3/29/2025 at 2:58:21 PM
Is it less ML than linear regression?by Epa095
3/29/2025 at 3:27:12 PM
If you want to draw the line between ML and not ML, I think you’ll have to put Kalman filters and linear regression on the non-ML side. You can put support vector machines and neural networks on the ML side.In some sense the exact place you draw the distinction is arbitrary. You could try to characterize where the distinction is by saying that models with fewer parameters and lower complexity tend to be called “not ML”, and models with more parameters and higher complexity tend to be called “ML”.
by klodolph
3/29/2025 at 4:47:27 PM
Linear regression is literally the second lecture of the Stanford ML class. https://cs229.stanford.edu/If you want to say "not neural networks" or not dnn or not llm, sure. But it's obviously machine learning
by wrsh07
3/29/2025 at 5:18:14 PM
When you say it’s “obviously machine learning”, how could that statement possibly be correct? There’s not even broad consensus here… so you don’t get to say that your definition is obviously correct.There are pedagogical reasons why you’d include linear regression in a machine learning course. This is pretty clear to me—they have properties which are extremely important to the field of machine learning field, such as differentiability.
by klodolph
3/29/2025 at 7:14:10 PM
It is obviously machine learning. It is a machine learning algorithm. It is taught in machine learning classes. It is described as a machine learning algorithm literally everywhere. Here's wiki:> Linear regression is also a type of machine learning algorithm, more specifically a supervised algorithm, that learns from the labelled datasets and maps the data points to the most optimized linear functions that can be used for prediction on new datasets
You can pretend it's not because it's not a sophisticated machine learning algorithm, but you are wrong.
by wrsh07
3/29/2025 at 9:46:58 PM
After spending over a decade in both statistics and machine learning I'd say the only reason there isn't a "broad consensus" is because statisticians like to gate-keep, whether that's linear regression, Monte Carlo methods, or Kalman Filters.Linear regression appears in pretty much every ML textbook. Can you confidently say, "this model that appears in every ML textbook is the only model in the ML textbook that isn't an ML model"?
Kalman Filters are like a continuous-state HMM. So why are HMMs considered ML and Kalman Filters not considered ML?
IMO it's an ego thing. They spent decades rigorously analyzing everything about linear models and here come these CS cowboys producing amazing results without any of the careful rigor that statisticians normally apply. It's difficult to argue against real results so the inflexible, hard-nosed statisticians just hang on to whatever they can.
by windsignaling
3/31/2025 at 4:13:17 AM
That’s unfairly harsh to statisticians, IMO. You have two fields of study, statistics and ML. There’s a massive overlap. Gatekeeping? Practitioners from these two fields have different jargon and view things differently from each other.“X is taught in books about subject Y” is a pretty weak argument. I could use it to argue that group theory is quantum mechanics. Scientists and mathematics aren’t fighting over who gets to own group theory—the scientists get to put group theory in their toolboxes and the mathematicians get to study it for itself. Same with ML and statistics. When you do ML, you need certain statistical techniques in your toolbox, so they get taught in your ML books.
by klodolph
3/29/2025 at 7:00:47 PM
If you want to draw a line you could say anything that’s not full AGI isn’t machine learning because of philosophical reasons.But, other than that there’s the only clear line is when the programmer isn’t hard coding results which puts Linear regression over the ML line. I guess you could argue about supervised vs unsupervised algorithms, but that’s going to exclude a lot of what is generally described as ML.
by Retric
3/29/2025 at 5:34:08 PM
Linear regression is ML. You are off base.by computerex
3/29/2025 at 6:15:39 PM
That's cool. Can you explain what machines they were using in the 1800s to do learning on?by genewitch
3/29/2025 at 7:03:09 PM
Does this mean if I simulate a neural network on pen and paper that stops being machine learning?All of these are machine learning techniques. Doing it by hand doesn't change anything. Today we use machines so it's machine learning
by wrsh07
3/30/2025 at 6:48:16 PM
I've worked out a 3x3 neural network with two pieces of labeled data by hand. Give me my statistics degree!by zeroCalories
3/29/2025 at 9:50:21 PM
You're using the word to define the concept rather than the concept to define the word. Wrong order.See: "Program".
by windsignaling
3/29/2025 at 11:32:46 PM
AI professor here.Anything that can separate data apoints can rightly been seen as a "supervised machine learning classifier".
Todemystify the area, I literally introduce my intro to ML lecture by drawing a line on the board, give its equation y = 0.5 x on the backboard, reminding students that they already know this, and then explain how to use it as a spam filter by interpreting the points on either side of the line as good emails versus spam ones.
by jll29
3/30/2025 at 5:33:29 AM
Linear regression is machine learning. At their core neural networks are just repeated linear regression + a non-linearity arranged in interesting ways. The key is that they can be trained to fit data using some optimization protocol (e.g. gradient descent). Just because linear regression has a closed form solution and is conceptually simple doesn't mean anything here.by bigmadshoe
3/29/2025 at 7:49:40 PM
EKFs work by 'learning' the covariance matrix on the fly, so I don't see why not?by CamperBob2
3/29/2025 at 6:54:22 PM
Hence the quotes ;).by wswope
3/29/2025 at 4:27:47 PM
As an intuition on why many people see this as different.PAC Learning is about compression, KF/EKF is more like Taylor expansion.
The specific types of PAC Learning that this paper covers has problems with a simplicity bias, and fairly low sensitivity.
While based on UHATs, this paper may provide some insights.
https://arxiv.org/abs/2502.02393
Obviously LLM and LRMs are the most studied, but even the recent posts on here from anthropic show that without a few high probability entries in the k-top results, confabulations are difficult for transformers.
Obviously there are PAC Learning methods that target anomaly detection, but they are very different than even EKF + Mc
You will note in this paper that even highly weighted features exhibited low sensitivity.
While the industry may find some pathological cases that make the approach usable, autograd and the need for parallelism make the application of this papers methods to tiny variations to multivariate problem ambitious.
They also only trained on medical data. Part of the reason the foundation models do so well is that they encode verifiers from a huge corpus that invalidates the traditional bias variance tradeoffs from the early 90's papers.
But they are still selecting from the needles and don't have access to the hay in the haystack.
The following paper is really not related except it shows how compression exacerbates that problem.
https://arxiv.org/abs/2205.06977
Chaitin's constant encoding the Halting problem, and that it is normal and uncomputable is the extreme top end of computability, but relates to the compression idea.
EKFs have access to the computable reals, and while non-linear, KF and EKFs can be thought of linearization of the approximations as a lens.
If the diagnostic indicators were both ergodic and Markovian, this paper's approach would probably be fairly reliable.
But these efforts are really about finding a many to one reduction that works.
I am skeptical about it in this case for PAC ML, but perhaps they will find a pathological case.
But the tradeoffs between statistical learning and expansive methods are quite different.
Obviously hype cycles drive efforts, I encourage you to look at this years AAAI conference report and see that you are not alone with the frustration on the single minded approach.
IMHO this paper is a net positive, showing that we are moving from a broad exploration to targeted applications.
But that is just my opinion.
by nyrikki
3/29/2025 at 2:40:57 PM
Parameter estimation is ML now?by jvanderbot
3/29/2025 at 3:27:49 PM
I think ML is in quotes for a reason—the reason is because the usage is not typical.by klodolph
3/29/2025 at 3:20:12 PM
Why not? LLMs, vision models, and kalman filters all learn parameters based on data.by thenobsta
3/29/2025 at 3:25:11 PM
A linear regression model can be written and trained as a neural net, has a loss function, all of that. Most if not all ML problems can be formulated as modelling a probability distributionby PaulHoule
3/29/2025 at 3:46:28 PM
That’s too reductive—ML models are statistical models. Statistical models have parameters, and in general cases, you choose the parameters with some kind of optimization algorithm.If you play fast and loose with your definition of “ML”, you’ll end up defining it so that any statistical model is an ML model… in which case, why even bother using two different terms?
ML models are, broadly speaking, the more complicated ones with more parameters, where the behavior of the models is not really known without training. I’m sure you could nitpick to death any definition I give, but that’s fine.
by klodolph
3/29/2025 at 3:58:14 PM
I am sure there are people teach data science classes who look at it in that "reductive" way.From the viewpoint of engineering, scikit-learn provides the same interface to linear regression that it supplies to many other models. Huggingface provides an interface to models that is similar in a lot of ways but I think a 'regression' in that it doesn't provide the bare minimum of model selection facilities needed to reliably make calibrated models. There are many problems where you could use either linear regression or a much more complex model. When it comes to "not known without training" I'm not sure how much of that is the limit of what we know right now and how much is fundamental as in the problem of "we can't really know what a computer program with free loops with do" (Halting problem) or "we can't predict what side of the side Pluto is going to be on in 30 million years" (Deterministic chaos)
(The first industrial model trainer I built was a simultaneously over and under engineered mess like most things in this industry... I didn't appreciate scikit-learn's model selection facilities and even though the data sci's I worked with had a book understanding of them, they didn't really put them to work.)
by PaulHoule
3/29/2025 at 5:25:57 PM
There’s a pedagogical reason to teach things with a kind of reductive definition. It makes a lot of sense.I remember getting cornered by somebody in a statistics class and interrogated about whether I thought neural networks were statistical techniques. In that situation I’ll only answer yes, they are statistical techniques. As far as I can tell, a big chunk of what we do with machine learning is create complicated models with a large number of parameters. We’re not creating something other than statistical models. We just draw a kind of cultural line between traditional statistics and machine learning techniques.
Now that I think about it, maybe if you asked me about the line between traditional statistics and machine learning, I would say that in traditional statistics, you can understand the parameters.
by klodolph
3/29/2025 at 6:30:24 PM
> in traditional statistics, you can understand the parameters.I also think that this is the key differentiator between ML and stats.
Statistical models can be understood formally, which means that not only we know how each parameter affects the predictions, we also know what their estimation uncertainties are, under which assumptions, and how to check that these assumptions are satisfied. Usually, we value these models not only because they're predictive but also because they're interpretable.
In ML there is neither the luxury nor the interest in doing this, all we want is something that predicts as well as possible.
So the difference is not the model itself but what you want to get out of it.
by blackbear_
3/30/2025 at 7:01:52 PM
I've heard professors call ML a form of applied statistics, and I think it's fair to call ML a subfield of statistics that deals with automatically generating statistical models with computers.by zeroCalories
3/29/2025 at 11:39:36 PM
You've equated neural networks with ML.I'm questioning the association of a single maximum likelihood parameter estimation via analytical optimization as ML, because it does not invole any methods beyond calculus and no models other than the system itself, whose parameters we are estimating.
Perhaps I'm wrong, but the power of NN is in an unknown intermediate representation between the data (measurements in estimation) and the prediction. EKF has no such black box.
Which now that I've written it agrees with top level comment so I rescind my question.
by jvanderbot
3/29/2025 at 9:51:47 PM
Neural networks are not ML now?by windsignaling
3/29/2025 at 11:35:45 PM
EKF is a neural network!?by jvanderbot
3/30/2025 at 10:10:08 AM
I think you missed the point of that comment. I was responding to the comment saying "Parameter estimation is ML now?"Neural networks are trained commonly using maximum likelihood estimation, a common parameter estimation technique.
by windsignaling