Skip to content
View dyth's full-sized avatar

Block or report dyth

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
dyth/README.md

David Yu-Tung Hui, 許宇同

I am interested in Deep Reinforcement Learning and its application to continuous-control tasks. My research improved the optimization stability of off-policy gradient-based $Q$-learning algorithms over a range of tasks and hyperparameters.

I've written two works along this research direction:

  1. Stabilizing Q-Learning for Continuous Control
    David Yu-Tung Hui
    MSc Thesis, University of Montreal, 2022
    I first investigated the duality between maximizing entropy and maximizing likelihood in the context of RL. I then showed that LayerNorm reduced divergence in $Q$-learning, especially in high-dimensional continuous control tasks.
    [.pdf] [Errata]

  2. Double Gumbel Q-Learning
    David Yu-Tung Hui, Aaron Courville, Pierre-Luc Bacon
    Spotlight at NeurIPS 2023
    We showed that function approximation in $Q$-learning induces two heteroscedastic Gumbel noise sources. An algorithm modeling these noise sources attained almost $2\times$ the aggregate performance of SAC at 1M timesteps over 33 continuous control tasks.
    [.pdf] [Reviews] [Poster (.png)] [5-min talk] [1-hour seminar] [Code (GitHub)] [Errata]

Pinned Loading

  1. doublegum doublegum Public

    NeurIPS 2023 Spotlight

    Python 10 4