Optimization or Architecture: What Matters in Non-Linear Filtering?

Ido Greenberg, Netanel Yannay, Shie Mannor

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review


In non-linear filtering, it is traditional to compare
non-linear architectures such as neural networks
to the standard linear Kalman Filter (KF). We observe that this methodology mixes the evaluation
of two separate components: the non-linear architecture, and the numeric optimization method.
In particular, the non-linear model is often optimized, whereas the reference KF model is not.
We argue that both should be optimized similarly.
We suggest the Optimized KF (OKF), which adjusts numeric optimization to the positive-definite
KF parameters. We demonstrate how a significant
advantage of a neural network over the KF may
entirely vanish once the KF is optimized using
OKF. This implies that experimental conclusions
of certain previous studies were derived from a
flawed process. The benefits of OKF over the nonoptimized KF are further studied theoretically and
empirically, where OKF consistently improves the
accuracy in a variety of problems. Experiments
are available on Github, and the OKF on PyPI.
Original languageUndefined/Unknown
Title of host publication1st Workshop on the Synergy of Scientific and Machine Learning Modeling @ ICML2023
StatePublished - 2023
EventSynergy of Scientific and Machine Learning Modeling - Hawaii, United States
Duration: 28 Jul 202328 Jul 2023


ConferenceSynergy of Scientific and Machine Learning Modeling
Country/TerritoryUnited States
Internet address

Cite this