Department of Mathematics,
University of California San Diego
****************************
PhD Defense
Gaojin He
UC San Diego
Complexity Bounds for Approximately Solving Markov Decision Processes and Properties of Turnpike Functions.
Abstract:
Markov Decision Processes are the major model of controlled stochastic processes in discrete time. Value iteration (VI) is one of the major methods for finding optimal policies. For each discount factor, starting from a finite number of iterations, which is called the turnpike integer, value iteration algorithms always generate decision rules which are deterministic optimal policies for the infinite-horizon problems. This fact justifies the rolling horizon approach for computing infinite-horizon optimal policies by conducting a finite number of value iterations. In this talk, we will first discuss the complexity of using VI to approximately solve MDPs, and then introduce properties of turnpike integers and provide their upper bounds.
-
APM 6402
APM 6402
****************************
Department of Mathematics,
University of California San Diego
****************************
Math 248: Real Analysis Seminar
Professor Shiferaw Berhanu
University of Maryland, College Park
On Sets of Removability Singularities and Propagation of Zeros for Vector Fields
Abstract:
We will present recent results obtained with J. Hounie on the removability sets of singularities for bounded solutions and the propagation of zeros across rough (nondifferentiable) boundaries for solutions of systems of complex vector fields including CR vector fields.
-
APM 6402
APM 6402
****************************
Department of Mathematics,
University of California San Diego
****************************
Postdoc Seminar
Dr. Rishabh Dixit
University of California San Diego
Accelerated gradient methods for nonconvex optimization : asymptotic dynamics and saddle escape
Abstract:
This talk focuses on the problem of understanding the behavior of a general class of accelerated gradient methods on smooth nonconvex functions. Motivated by some recent works that have proposed effective algorithms, based on Polyak’s heavy ball method and the Nesterov accelerated gradient method, to achieve convergence to a local minimum of nonconvex functions, we describe a broad class of Nesterov-type accelerated methods and put forth a rigorous study of these methods encompassing the escape from saddle points and convergence to local minima through an asymptotic analysis. In the asymptotic regime, we first answer an open question of whether Nesterov’s accelerated gradient method (NAG) with variable momentum parameters avoids strict saddle points almost surely with respect to a random initialization and a random step-size. We then develop two metrics of asymptotic rates of convergence and divergence, and evaluate these two metrics for several popular standard accelerated methods such as the NAG and Nesterov’s accelerated gradient with constant momentum (NCM) near strict saddle points. Theoretical results on the asymptotic behavior are demonstrated on the phase retrieval problem.
-
APM 6402
APM 6402
****************************