[NeurIPS 2022] Exploiting Reward Shifting in Value-Based DRL [Paper] [Code]
Hao Sun, Lei Han, Rui Yang, Xiaoteng Ma, Jian Guo, Bolei Zhou
- A positive reward shifting leads to conservative exploitation, while a negative reward shifting leads to curiosity-driven exploration.
[ICLR 2022] Rethinking Goal-conditioned Supervised Learning and Its Connection to Offline RL [Paper] [Code]
Rui Yang, Yiming Lu, Wenzhe Li, Hao Sun, Meng Fang, Yali Du, Xiu Li, Lei Han, Chongjie Zhang
- We optimize the GCSL with a lower bound of the goal-reaching objective and link the success of GCSL from perspective of offline RL.
[IJCAI 2021] Hierarchical Multi-Scale Gaussian Transformer for Stock Movement Prediction [Paper]
Qianggang Ding, Sifan Wu, Hao Sun, Jiadong Guo, Jian Guo
- We adapt transformer to stock movement predictions.
[AAAI 2021] Adaptive Regularization of Labels [Paper]
Qianggang Ding, Sifan Wu, Hao Sun, Jiadong Guo, Shu-Tao Xia
- We study the correlations between lables to improve model performance.
[NeurIPS 2019 (Spotlight)] Policy Continuation with Hindsight Inverse Dynamics [Paper] [Code] [Homepage]
Hao Sun, Zhizhong Li, Xiaotong Liu, Dahua Lin, Bolei Zhou
- Supervised Learning can be used to solve goal-conditioned RL tasks.
[ICML 2022 DFUQ] DAUX: a Density-based Approach for Uncertainty eXplanations [Paper] [Code]
Hao Sun^, Boris van Breugel^, Jonathan Crabbe, Nabeel Seedat, Mihaela van der Schaar
- We propose a density-based approach to classify uncertain examples.