Hao Sun
I am a Researcher at Google DeepMind. I did my PhD at the University of Cambridge, supervised by Prof. Mihaela van der Schaar. I obtained my M.Phil. at MMLab@CUHK, advised by Prof. Dahua Lin and Prof. Bolei Zhou. I studied Physics in my undergrad at Peking University, and my thesis was supervised by Prof. Zhouchen Lin.
News
π (2025.05) Call for community action! OpenReview Should be Protected and Leveraged as a Community Asset for Research in the Era of Large Language Models
π¦πΉ (2025.07) Attended ACL 2025 Run the ACL 2025 Tutorial (T1): Inverse RL Meets LLM Alignment at Vienna. Slide. We have a write-up for interested readers to find references.
πΊπΈ (2025.07) Invited talk on RL in the Era of LLMs at Intuit AI research.
π¬π§ (2025.06) Invited talk on RL in the Era of LLMs at Jump Trading London.
π (2025.05) Multi-Objective and Personalized Alignment with PCA is accepted by ACL findings! Two Papers on Prompt Optimization are accepted by ACL as main and findings.
π (2025.05) Active Reward Modeling and A Spotlight π Paper on LLM for RL are accepted by ICML!
πΊπΈ (2025.03) Guest lecture on Inverse RL Meets LLMs at the UCLA Reinforcement Learning course.
πΊπΈ (2025.02) Attending AAAI 2025 to run the Tutorial: Inverse RL Meets LLMs. Thanks for joining us in Philadelphia! Slide.
π (2025.02) Our Reward Model Paper Part IV: Multi-Objective and Personalized Alignment with PCA is online.
π (2025.02) Our Reward Model Paper Part III: Infrastructure for Reproducible Reward Model Research is online.
π (2025.02) Our Reward Model Paper Part II: Active Reward Modeling is online.
π (2025.01) Our Reward Model Paper Part I: Foundation, Theory, and Alternatives is accepted by ICLR as an Oral π. It is an amazing experience to work with Yunyi and Jef.