Donghu Kim

logo_uni_tue          logo_uni_tue

Hi there, welcome!

My name is Donghu Kim. I am on a Master's Degree program in KAIST (advised by Jaegul Choo), studying reinforcement learning and embodied AI with these splendid researchers: Byungkun Lee, Hojoon Lee, Dongyoon Hwang, Hyunseung Kim, Kyungmin Lee, and Youngdo Lee.

My main interest at the moment is inclined towards building an architecture that can maintain plasticity under severe non-stationarity. The fact that training neural networks ironically makes them un-trainable will never not be interesting to me.

I still have a long long way to go; if you want to discuss anything research related, I'd be more than happy to be engaged!

Email  /  CV  /  Google Scholar  /  Github

profile photo

Publications

simbav2
Reinforcement Learning
SimBaV2: Hyperspherical Normalization for Scalable Deep Reinforcement Learning
Hojoon Lee, Youngdo Lee, Takuma Seno, Donghu Kim, Peter Stone, Jaegul Choo.
Under review.
arXiv / project page / github

We further regularize SimBa architecture by projecting both parameters and features onto a hypersphere, leading to better scaling properties in model size and compute.

simba
Reinforcement Learning
SimBa: Simplicity Bias for Scaling Up Parameters in Deep Reinforcement Learning
Hojoon Lee, Dongyoon Hwang, Donghu Kim, Hyunseung Kim, Jun Jet Tai, Kaushik Subramanian, Peter R. Wurman, Jaegul Choo, Peter Stone, Takuma Seno.
ICLR'25, Spotlight.
arXiv / project page / github

We propose a well-regularized architecture that avoids overfitting, allowing parameter and compute scale up in RL.

preprint2024dodont
Reinforcement Learning Skill Discovery
Do’s and Don’ts: Learning Desirable Skills with Instruction Videos
Hyunseung Kim, Byungkun Lee, Hojoon Lee, Dongyoon Hwang, Donghu Kim, Jaegul Choo
NeurIPS'24.
arXiv / project page

We present DoDont, a skill discovery algorithm that learns diverse behaviors while following the behaviors in "do" videos while avoiding the behaviors in "don't" videos.

icml2024atari-pb
Reinforcement Learning Pre-training
ATARI-PB: Investigating Pre-Training Objectives for Generalization in Pixel-Based RL
Donghu Kim*, Hojoon Lee*, Kyungmin Lee*, Dongyoon Hwang, Jaegul Choo.
ICML'24.
arXiv / project page / github / poster

We investigate which pre-training objectives are beneficial for in-distribution, near-out-of-distribution, and far-out-of-distribution generalization in visual reinforcement learning.

icml2024hnt
Reinforcement Learning Plasticity
Slow and Steady Wins the Race: Maintaining Plasticity with Hare and Tortoise Networks
Hojoon Lee, Hyeonseo Cho, Hyunseung Kim, Donghu Kim, Dugki Min, Jaegul Choo, Clare Lyle.
ICML'24.
arXiv / github

To allow the network to continually adapt and generalize, we introduce Hare and Tortoise architecture, inspired by the complementary learning system of the human brain.


Study Materials

Note: These slides are made for studying purposes only, and likely have got something wrong here and there. If you happen to find some, feel free to make fun of me via e-mail :).

  • 25.01.03: Warm Start RL [slides]
  • 24.09.20: Catastrophic Interference in RL [slides]
  • 24.09.06: Understanding Self-Predictive RL [slides]
  • 24.06.21: Automatic Environment Shaping [slides]
  • 24.04.05: Stop Regressing (HL Gauss) [slides]
  • 24.03.08: TD7 [slides]
  • 24.02.23: Introduction to RL (CS285 Lecture2) [slides]
  • 23.11.17: MAE Survey [slides]
  • 23.09.01: ACRO (Multi-step IDM) [slides]

Awards

  • Academic Excellence Award, Korea University 2019, 2022.
  • 2nd Place in NC AI Fellowship ($2000), NCSoft, 2019.
  • Presidential Science Scholarship (Total $40000), Korea Student Aid Foundation, 2018.

Template based on Hojoon Lee's website.