Shibhansh Dohare
I am a final year Ph.D. student at the University of Alberta,
advised by Dr. Richard Sutton and Dr. Rupam Mahmood.
I completed my B.Tech. at IIT Kanpur in Computer Science and Engineering.
My long-term research goal is to understand the workings of our minds.
Specifically, to help find the computational principles that give rise to the mind.
In pursuit of this goal, I'm working on various aspects of continual learning, deep learning, and reinforcement learning.
During my Ph.D., I have contributed to exposing a fundamental problem with deep learning systems,
where these systems can lose the ability to learn new things.
I also developed the continual backpropagation algorithm to overcome this problem.
My Ph.D. research has been published in Nature and featured in some popular media outlets,
such as New Scientist.
If you prefer podcasts, I have also discussed my work on the Nature Podcast and AMII's Approximately Correct Podcast.
I am on the job market for Spring/Summer 2025. Feel free to reach out if you think I’d be a good fit for your research position.
Email  / 
CV  / 
Google Scholar  / 
Twitter  / 
Github
|
|
|
Loss of Plasticity in Deep Continual Learning
Shibhansh Dohare,
J. Fernando Hernandez-Garcia,
Qingfeng Lan,
Parash Rahman,
A. Rupam Mahmood,
Richard S. Sutton
Nature 2024
Paper
|
Code
|
Nature Podcast
|
News
We provide first direct demonstrations of plasticity loss in deep continual learning.
We propose a new algorithm, continual backpropagation, that fully maintains plasticity.
Continual backpropagation re-initializes a small fraction of less-used units alongside gradient descent at each update.
|
|
Overcoming Policy Collapse in Deep Reinforcement Learning
Shibhansh Dohare,
Qingfeng Lan,
A. Rupam Mahmood
EWRL 2023
Paper
We show that popular deep RL algorithms, like PPO, do not scale with experience. Their performance gets worse over time. We look deeper into this problem and provide simple solutions to reduce performance degradation.
|
|
Automatic Noise Filtering with Dynamic Sparse Training in Deep Reinforcement Learning
Bram Grooten,
Ghada Sokar,
Shibhansh Dohare,
Elena Mocanu,
Matthew E. Taylor,
Mykola Pechenizkiy,
Decebal Constantin Mocanu
AAMAS 2023
Paper
We show that standard Deep RL algorithms fail when the input contains noisy features.
Dynamic sparse training successfully filters through the noisy features and performs well.
|
|
Gamma-Nets: Generalizing Value Estimation over Timescale
Craig Sherstan,
Shibhansh Dohare,
James MacGlashan,
Johannes Günther,
Patrick M. Pilarski,
AAAI, Oral Presentation, 2020
Paper
We present Gamma-nets, a method for generalizing value function estimation over timescale.
|
|
Unsupervised Semantic Abstractive Summarization
Shibhansh Dohare,
Vivek Gupta,
Harish Karnick,
ACL, Student Research Workshop, 2018
Paper
A novel algorithm for abstractive text summarization based on Abstract Meaning Representation.
|
|