α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Rishabh Joshi
Rishabh Joshi
2
papers
58
total citations
papers (2)
RRM: Robust Reward Model Training Mitigates Reward Hacking
ICLR 2025
arXiv
50
citations
Learning from negative feedback, or positive feedback or both
ICLR 2025
arXiv
8
citations