User profiles for Jeongyeol Kwon
Jeongyeol KwonUniversity of Wisconsin-Madison Verified email at wisc.edu Cited by 478 |
A fully first-order method for stochastic bilevel optimization
We consider stochastic unconstrained bilevel optimization problems when only the first-order
gradient oracles are available. While numerous optimization methods have been proposed …
gradient oracles are available. While numerous optimization methods have been proposed …
Feed two birds with one scone: Exploiting wild data for both out-of-distribution generalization and detection
Modern machine learning models deployed in the wild can encounter both covariate and
semantic shifts, giving rise to the problems of out-of-distribution (OOD) generalization and OOD …
semantic shifts, giving rise to the problems of out-of-distribution (OOD) generalization and OOD …
Rl for latent mdps: Regret guarantees and a lower bound
In this work, we consider the regret minimization problem for reinforcement learning in latent
Markov Decision Processes (LMDP). In an LMDP, an MDP is randomly drawn from a set of …
Markov Decision Processes (LMDP). In an LMDP, an MDP is randomly drawn from a set of …
On penalty methods for nonconvex bilevel optimization and first-order stochastic approximation
In this work, we study first-order algorithms for solving Bilevel Optimization (BO) where the
objective functions are smooth but possibly nonconvex in both levels and the variables are …
objective functions are smooth but possibly nonconvex in both levels and the variables are …
Global convergence of the EM algorithm for mixtures of two component linear regression
J Kwon, W Qian, C Caramanis… - … on Learning Theory, 2019 - proceedings.mlr.press
The Expectation-Maximization algorithm is perhaps the most broadly used algorithm for
inference of latent variable problems. A theoretical understanding of its performance, however, …
inference of latent variable problems. A theoretical understanding of its performance, however, …
On the minimax optimality of the EM algorithm for learning two-component mixed linear regression
We study the convergence rates of the EM algorithm for learning two-component mixed
linear regression under all regimes of signal-to-noise ratio (SNR). We resolve a long-standing …
linear regression under all regimes of signal-to-noise ratio (SNR). We resolve a long-standing …
On the computational and statistical complexity of over-parameterized matrix sensing
We consider solving the low-rank matrix sensing problem with the Factorized Gradient Descent
(FGD) method when the specified rank is larger than the true rank. We refer to this as over…
(FGD) method when the specified rank is larger than the true rank. We refer to this as over…
EM converges for a mixture of many linear regressions
J Kwon, C Caramanis - International Conference on Artificial …, 2020 - proceedings.mlr.press
We study the convergence of the Expectation-Maximization (EM) algorithm for mixtures of
linear regressions with an arbitrary number $ k $ of components. We show that as long as …
linear regressions with an arbitrary number $ k $ of components. We show that as long as …
Reinforcement learning in reward-mixing mdps
Learning a near optimal policy in a partially observable system remains an elusive challenge
in contemporary reinforcement learning. In this work, we consider episodic reinforcement …
in contemporary reinforcement learning. In this work, we consider episodic reinforcement …
The EM algorithm gives sample-optimality for learning mixtures of well-separated gaussians
J Kwon, C Caramanis - Conference on Learning Theory, 2020 - proceedings.mlr.press
We consider the problem of spherical Gaussian Mixture models with $ k\geq 3$ components
when the components are well separated. A fundamental previous result established that …
when the components are well separated. A fundamental previous result established that …