α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Razvan Pascanu
Razvan Pascanu
1
Affiliations
Affiliations
Google DeepMind
20
papers
966
total citations
papers (20)
Understanding the Role of Training Regimes in Continual Learning
NEURIPS 2020
arXiv
265
citations
Continual World: A Robotic Benchmark For Continual Reinforcement Learning
NEURIPS 2021
arXiv
118
citations
Top-KAST: Top-K Always Sparse Training
NEURIPS 2020
arXiv
108
citations
Pointer Graph Networks
NEURIPS 2020
arXiv
69
citations
Deep Reinforcement Learning with Plasticity Injection
NEURIPS 2023
arXiv
64
citations
Why do LLMs attend to the first token?
COLM 2025
arXiv
63
citations
Powerpropagation: A sparsity inducing weight reparameterisation
NEURIPS 2021
arXiv
58
citations
Improving fine-grained understanding in image-text pre-training
ICML 2024
arXiv
46
citations
Universality of Linear Recurrences Followed by Non-linear Projections: Finite-Width Guarantees and Benefits of Complex Eigenvalues
ICML 2024
arXiv
35
citations
The Tunnel Effect: Building Data Representations in Deep Neural Networks
NEURIPS 2023
arXiv
33
citations
Learning to Modulate pre-trained Models in RL
NEURIPS 2023
arXiv
26
citations
Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem
ICML 2024
arXiv
26
citations
How do language models learn facts? Dynamics, curricula and hallucinations
COLM 2025
21
citations
A Large Recurrent Action Model: xLSTM enables Fast Inference for Robotics Tasks
ICML 2025
arXiv
11
citations
On the Role of Optimization in Double Descent: A Least Squares Study
NEURIPS 2021
arXiv
11
citations
Attention as a Hypernetwork
ICLR 2025
arXiv
10
citations
MS-SSM: A Multi-Scale State Space Model for Efficient Sequence Modeling
COLM 2025
arXiv
2
citations
Disentangling Transfer in Continual Reinforcement Learning
NEURIPS 2022
0
citations
RAT: Bridging RNN Efficiency and Attention Accuracy via Chunk-based Sequence Modeling
NEURIPS 2025
arXiv
0
citations
Meta-learning how to Share Credit among Macro-Actions
NEURIPS 2025
arXiv
0
citations