We introduce Optimistic Online Newton (OON), a second-order online learning algorithm that attains regret on strongly curved convex losses while remaining a one-pass, streaming method with memory. The key idea is to maintain a running Hessian proxy from the residual between observed gradients and an optimistic guess; when the guess is accurate, we recover the regret of the best Newton step in hindsight; when it is inaccurate, we degrade gracefully to the standard -regret floor. We give matching lower bounds in the strongly-convex regime and demonstrate empirical gains over AdaGrad and ONS on synthetic and real-world prediction tasks.
Why this matters
Second-order methods deliver fast convergence in offline optimization, but their online counterparts have historically required either tight curvature assumptions or heavy per-round computation. OON sidesteps both: by maintaining a running Hessian proxy seeded with an optimistic prediction , the algorithm recovers logarithmic regret on strongly-curved losses without ever computing a true Hessian.
Main result
For a sequence of -strongly-convex losses with bounded gradients, OON attains
The bound is tight up to a logarithmic factor, matching a lower bound we prove in Section 4 of the paper.
Code & reproducibility
Reference implementations in PyTorch and JAX, alongside reproduction scripts for every figure, are available in the github.com/ooaarg/oon repository. All experiments run on a single GPU in under 30 minutes.