Isaac  Liao

Isaac Liao

Scholar Title

Undergraduate Research and Innovation Scholar

Research Title

Learning to Optimize Quasi-Newton Methods

Cohort

2022–2023

Department

Electrical Engineering and Computer Science

Research Areas
  • Artificial Intelligence and Machine Learning
Supervisor

Marin Soljacic

Abstract

We introduce a novel machine learning optimizer called LODO, which online meta-learns an implicit inverse Hessian of the loss as a subroutine of quasi-Newton optimization. Our optimizer merges Learning to Optimize (L2O) techniques with quasi-Newton methods to learn neural representations of symmetric matrix vector products, which are more flexible than those in other quasi-Newton methods. Unlike other L2O methods, ours does not require any meta-training on a training task distribution, and instead learns to optimize on the fly while optimizing on the test task, adapting to the local characteristics of the loss landscape while traversing it. Theoretically, we show that our optimizer approximates the inverse Hessian in noisy loss landscapes and is capable of representing a wide range of inverse Hessians. We experimentally verify our algorithm’s performance in the presence of noise, and show that simpler alternatives for representing the inverse Hessians worsen performance. Lastly, we use our optimizer to train a semi-realistic deep neural network with 95k parameters, and obtain competitive results against standard neural network optimizers.

Quote

I am interested in the flexibility and generality of techniques in machine learning, and would love to learn and help improve them by studying the foundational ideas which underlie large general classes of machine learning problems. Such ideas include but are not limited to learned optimization, compressed data representation, and generative modeling.

Back to Scholars