Grant M. Rotskoff
Assistant Professor of Chemistry
Bio
Grant Rotskoff studies the nonequilibrium dynamics of living matter with a particular focus on selforganization from the molecular to the cellular scale. His work involves developing theoretical and computational tools that can probe and predict the properties of physical systems driven away from equilibrium. Recently, he has focused on characterizing and designing physically accurate machine learning techniques for biophysical modeling. Prior to his current position, Grant was a James S. McDonnell Fellow working at the Courant Institute of Mathematical Sciences at New York University. He completed his Ph.D. at the University of California, Berkeley in the Biophysics graduate group supported by an NSF Graduate Research Fellowship. His thesis, which was advised by Phillip Geissler and Gavin Crooks, developed theoretical tools for understanding nonequilibrium control of the small, fluctuating systems, such as those encountered in molecular biophysics. He also worked on coarsegrained models of the hydrophobic effect and selfassembly. Grant received an S.B. in Mathematics from the University of Chicago, where he became interested in biophysics as an undergraduate while working on free energy methods for largescale molecular dynamics simulations.
Research Summary
My research focuses on theoretical and computational approaches to "mesoscale" biophysics. Many of the cellular phenomena that we consider the hallmarks of living systems occur at the scale of hundreds or thousands of proteins. Processes like the selfassembly of organellesized structures, the dynamics of cell division, and the transduction of signals from the environment to the machinery of the cell are not macroscopic phenomena—they are the result of a fluctuating, nonequilibrium dynamics. Experimentally probing mesoscale systems remains extremely difficult, though it is continuing to benefit from advances in cryoelectron microscopy and superresolution imaging, among many other techniques. Predictive and explanatory models that resolve the essential physics at these intermediate scales have the power to both aid and enrich the understanding we are presently deriving from these experimental developments.
Major parts of my research include:
1. Dynamics of mesoscale biophysical assembly and response.— Biophysical processes involve chemical gradients and timedependent external signals. These inherently nonequilibrium stimuli drive supermolecular organization within the cell. We develop models of active assembly processes and proteinmembrane interactions as a foundation for the broad goal of characterizing the properties of nonequilibrium biomaterials.
2. Machine learning and dimensionality reduction for physical models.— Machine learning techniques are rapidly becoming a central statistical tool in all domains of scientific research. We apply machine learning techniques to sampling problems that arise in computational chemistry and develop approaches for systematically coarsegraining physical models. Recently, we have also been exploring reinforcement learning in the context of nonequilibrium control problems.
3. Methods for nonequilibrium simulation, optimization, and control.— We lack wellestablished theoretical frameworks for describing nonequilibrium states, even seemingly simple situations in which there are chemical or thermal gradients. Additionally, there are limited tools for predicting the response of nonequilibrium systems to external perturbations, even when the perturbations are small. Both of these problems pose key technical challenges for a theory of active biomaterials. We work on optimal control, nonequilibrium statistical mechanics, and simulation methodology, with a particular interest in developing techniques for importance sampling configurations from nonequilibrium ensembles.
Academic Appointments
Honors & Awards

Early Career Research Program Award, Department of Energy (20222027)

Research Scholar Award, Google (2022)

Terman Faculty Fellow, Stanford University (20202022)
202223 Courses
 Advanced Physical Chemistry
CHEM 273 (Win)  Exploring Chemical Research at Stanford
CHEM 91 (Win)  Physical Chemistry III
CHEM 175 (Win) 
Independent Studies (5)
 Advanced Undergraduate Research
CHEM 190 (Aut, Win, Spr)  Directed Instruction/Reading
CHEM 90 (Aut, Spr)  Ph.D. Research
CME 400 (Win)  Research and Special Advanced Work
CHEM 200 (Aut, Win, Spr, Sum)  Research in Chemistry
CHEM 301 (Aut, Win, Spr, Sum)
 Advanced Undergraduate Research

Prior Year Courses
202122 Courses
 Advanced Physical Chemistry
CHEM 273 (Win)  Machine Learning for Chemical and Dynamical Data
CHEM 263 (Aut)  Physical Chemistry III
CHEM 175 (Win)
202021 Courses
 Physical Chemistry III
CHEM 175 (Win)
 Advanced Physical Chemistry
Stanford Advisees

Doctoral Dissertation Reader (AC)
Michael Chen, Ethan Curtis 
Postdoctoral Faculty Sponsor
Clay Batton, Sreekanth Kizhakkumpurath Manikandan 
Doctoral Dissertation Advisor (AC)
Shriram Chennakesavalu, Sherry Li, Andy Mitchell, Emmit Pert
All Publications

Physicsinformed graph neural networks enhance scalability of variational nonequilibrium optimal control
JOURNAL OF CHEMICAL PHYSICS
2022; 157 (7): 074101
Abstract
When a physical system is driven away from equilibrium, the statistical distribution of its dynamical trajectories informs many of its physical properties. Characterizing the nature of the distribution of dynamical observables, such as a current or entropy production rate, has become a central problem in nonequilibrium statistical mechanics. Asymptotically, for a broad class of observables, the distribution of a given observable satisfies a large deviation principle when the dynamics is Markovian, meaning that fluctuations can be characterized in the longtime limit by computing a scaled cumulant generating function. Calculating this function is not tractable analytically (nor often numerically) for complex, interacting systems, so the development of robust numerical techniques to carry out this computation is needed to probe the properties of nonequilibrium materials. Here, we describe an algorithm that recasts this task as an optimal control problem that can be solved variationally. We solve for optimal control forces using neural network ansatz that are tailored to the physical systems to which the forces are applied. We demonstrate that this approach leads to transferable and accurate solutions in two systems featuring large numbers of interacting particles.
View details for DOI 10.1063/5.0095593
View details for Web of Science ID 000840971900002
View details for PubMedID 35987599

Adaptive Monte Carlo augmented with normalizing flows.
Proceedings of the National Academy of Sciences of the United States of America
2022; 119 (10): e2109420119
Abstract
SignificanceMonte Carlo methods, tools for sampling data from probability distributions, are widely used in the physical sciences, applied mathematics, and Bayesian statistics. Nevertheless, there are many situations in which it is computationally prohibitive to use Monte Carlo due to slow "mixing" between modes of a distribution unless handtuned algorithms are used to accelerate the scheme. Machine learning techniques based on generative models offer a compelling alternative to the challenge of designing efficient schemes for a specific system. Here, we formalize Monte Carlo augmented with normalizing flows and show that, with limited prior data and a physically inspired algorithm, we can substantially accelerate sampling with generative models.
View details for DOI 10.1073/pnas.2109420119
View details for PubMedID 35235453

Learning nonequilibrium control forces to characterize dynamical phase transitions
PHYSICAL REVIEW E
2022; 105 (2)
View details for DOI 10.1103/PhysRevE.105.024115
View details for Web of Science ID 000754645400008

Learning nonequilibrium control forces to characterize dynamical phase transitions.
Physical review. E
2022; 105 (21): 024115
Abstract
Sampling the collective, dynamical fluctuations that lead to nonequilibrium pattern formation requires probing rare regions of trajectory space. Recent approaches to this problem, based on importance sampling, cloning, and spectral approximations, have yielded significant insight into nonequilibrium systems but tend to scale poorly with the size of the system, especially near dynamical phase transitions. Here we propose a machine learning algorithm that samples rare trajectories and estimates the associated large deviation functions using a manybody control force by leveraging the flexible function representation provided by deep neural networks, importance sampling in trajectory space, and stochastic optimal control theory. We show that this approach scales to hundreds of interacting particles and remains robust at dynamical phase transitions.
View details for DOI 10.1103/PhysRevE.105.024115
View details for PubMedID 35291069

Probing the theoretical and computational limits of dissipative design.
The Journal of chemical physics
2021; 155 (19): 194114
Abstract
Selfassembly, the process by which interacting components form welldefined and often intricate structures, is typically thought of as a spontaneous process arising from equilibrium dynamics. When a system is driven by external nonequilibrium forces, states statistically inaccessible to the equilibrium dynamics can arise, a process sometimes termed direct selfassembly. However, if we fix a given target state and a set of external control variables, it is not wellunderstood (i) how to designa protocol to drive the system toward the desired state nor (ii) the cost of persistently perturbing the stationary distribution. In this work, we derive a bound that relates the proximity to the chosen target with the dissipation associated with the external drive, showing that highdimensional external control can guide systems toward target distribution but with an inevitable cost. Remarkably, the bound holds arbitrarily far from equilibrium. Second, we investigate the performance of deep reinforcement learning algorithms and provide evidence for the realizability of complex protocols that stabilize otherwise inaccessible states of matter.
View details for DOI 10.1063/5.0067695
View details for PubMedID 34800948

A Dynamical Central Limit Theorem for Shallow Neural Networks
NEURAL INFORMATION PROCESSING SYSTEMS (NIPS). 2020
View details for Web of Science ID 000627697000073