Publications
Publications and Recent Pre-Prints
A complete list of publications (Nov 2022) can be found in my CV or in my Google scholar.
-
Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Aviral Kumar, Rishabh Agarwal, Aaron Courville, Tengyu Ma, George Tucker, Sergey Levine [Paper] [Talk]
Oral Presentation at the Deep Reinforcement Learning Workshop at NeurIPS 2021 (6/140 papers)
Oral Presentation at the Overparameterization Workshop at ICML 2021 (6/45 papers)
Oral Presentation at RL for Real Life Workshop at ICML 2021 (12/90 papers)
International Conference on Learning Representations (ICLR), 2022 (Spotlight Presentation) -
A Workflow for Offline Model-Free Robotic Reinforcement Learning\ Aviral Kumar^, Anikait Singh^, Stephen Tian, Chelsea Finn, Sergey Levine [Paper] [Talk]
Conference on Robot Learning (CoRL), 2021 (Oral Presentation) -
Offline Q-Learning on Diverse Multi-Task Data Scales and Generalizes
Aviral Kumar, Rishabh Agarwal, Xinyang Geng, George Tucker, Sergey Levine [arXiv]
Oral Presentation at the Deep Reinforcement Learning Workshop at NeurIPS 2022 (6/196 papers) -
Data-Driven Offline Optimization for Architecting Hardware Accelerators [Paper] [Talk] [Blog]
Aviral Kumar^, Amir Yazdanbakhsh^, Milad Hashemi, Kevin Swersky, Sergey Levine
International Conference on Learning Representations (ICLR), 2022 -
Pre-Training for Robots: How Offline RL Enables Learning New Tasks from a Handful of Trials
Aviral Kumar^, Anikait Singh^, Frederik Ebert^*, Yanlai Yang, Chelsea Finn, Sergey Levine [arXiv] -
How to Leverage Unlabeled Data in Offline Reinforcement Learning [Paper] [Talk]
Tianhe Yu^, Aviral Kumar^, Yevgen Chebotar, Chelsea Finn, Karol Hausman, Sergey Levine
International Conference on Machine Learning (ICML), 2022 -
Should I Run Offline Reinforcement Learning or Behavioral Cloning?
Aviral Kumar^, Joey Hong^, Anikait Singh, Sergey Levine [Paper] [Blog]
International Conference on Learning Representations (ICLR), 2022 -
Conservative Data-Sharing for Multi-Task Offline Reinforcement Learning [Paper]
Tianhe Yu^, Aviral Kumar^, Yevgen Chebotar, Karol Hausman, Sergey Levine, Chelsea Finn\ Neural Information Processing Systems (NeurIPS), 2021. -
Conservative Objective Models for Effective Offline Model-Based Optimization [Paper] [Blog]
Brandon Trabucco^, Aviral Kumar^, Xinyang Geng, Sergey Levine
International Conference on Machine Learning (ICML), 2021. -
Implicit Under-Parameterization Inhibts Data-Efficient Deep Reinforcement Learning
Aviral Kumar^, Rishabh Agarwal^, Dibya Ghosh, Sergey Levine [arXiv]
International Conference on Learning Representations (ICLR), 2021 -
OPAL: Offline Primitive Discovery For Accelerating Reinforcement Learning
Anurag Ajay, Aviral Kumar, Pulkit Agarwal, Sergey Levine, Ofir Nachum [arXiv]
International Conference on Learning Representations (ICLR), 2021 -
Conservative Safety Crtics for Exploration
Homanaga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkruti, Animesh Garg [arXiv]
International Conference on Learning Representations (ICLR), 2021 -
COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Avi Singh, Albert Yu, Jonathan Yang, Jesse Zhang, Aviral Kumar, Sergey Levine
[arXiv] [website] (October 2020)
4th Conference on Robot Learning (CoRL), 2020\ -
Conservative Q-Learning for Offline Reinforcement Learning
Aviral Kumar, Aurick Zhou, George Tucker, Sergey Levine
[arXiv] [website] [Blog]
Neural Information Processing Systems (NeurIPS), 2020\ -
DisCor: Corrective Feedback in Reinforcement Learning via Distribution Correction
Aviral Kumar, Abhishek Gupta, Sergey Levine [arXiv] [BAIR Blog] (March 2020)
Neural Information Processing Systems (NeurIPS) (Spotlight Presentation, 3% acceptance rate), 2020\ -
Model Inversion Networks for Model-Based Optimization
Aviral Kumar, Sergey Levine [arXiv] (December 2019)
Neural Information Processing Systems (NeurIPS), 2020\ -
One Solution is Not All You Need: Few-Shot Extrapolation Via Structured MaxEnt RL
Saurabh Kumar, Aviral Kumar, Sergey Levine, Chelsea Finn [arXiv] (October 2020)
Neural Information Processing Systems (NeurIPS), 2020\ -
Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction
Aviral Kumar^, Justin Fu^, George Tucker, Sergey Levine
Neural Information Processing Systems (NeurIPS), 2019
[Paper] [Project Page] [BAIR Blog] \ -
Diagnosing Bottlenecks in Deep Q-Learning Algorithms
Justin Fu^, Aviral Kumar^, Matthew Soh, Sergey Levine
International Conference on Machine Learning (ICML) 2019 [Paper] (* Equal Contribution)\ -
Graph Normalizing Flows
Jenny Liu*, Aviral Kumar*, Jimmy Ba, Jamie Kiros, Kevin Swersky
Neural Information Processing Systems (NeurIPS), 2019
[Paper] [Project Page] (* Equal Contribution) -
Trainable Calibration Measures for Neural Networks from Kernel Mean Embeddings
Aviral Kumar, Sunita Sarawagi, Ujjwal Jain
35th International Conference on Machine Learning (ICML) 2018 [Main Paper] -
GRevnet: Improving Graph Neural Networks with Reversible Computation
Aviral Kumar, Jimmy Ba, Jamie Kiros, Kevin Swersky
NeuRIPS 2018 Relational Representation Learning Workshop [Paper] -
Feudal Learning for Large Discrete Action Spaces with Recursive Substructure
Aviral Kumar, Kevin Swersky, Geoffrey Hinton
Hierarchical Reinforcement Learning Workshop, NIPS 2017
[Main Paper]
Tutorials
- Offline Reinforcement Learning: Tutorial, Review and Perspectives on Open Problems
Sergey Levine, Aviral Kumar, George Tucker, Justin Fu
[arXiv] (May 2020)\
Older Pre-Prints
-
Datasets for Data-Driven Deep Reinforcement Learning
Justin Fu, Aviral Kumar, Ofir Nachum, George Tucker, Sergey Levine
[arXiv] [Benchmark] (April 2020)
At RL for Real Life Virtual Conference, 2020.\ -
Reward-Condtioned Policies
Aviral Kumar, Xue Bin Peng, Sergey Levine [arXiv] (December 2019)\ -
Advantage-Weighted Regression: Simple and Scalable Off-Policy Reinforcement Learning
Xue Bin Peng, Aviral Kumar, Grace Zhang, Sergey Levine [Project Page] (October 2019)