πŸ™‹ About Me


πŸš€ Hi there! I am Hao Sun, a final-year Ph.D. student at the University of Cambridge, supervised by Prof. Mihaela van der Schaar, working at the intersection of reinforcement learning (RL) and large language models (LLMs). During my M.Phil. study at MMLab@CUHK, I was advised by Prof. Dahua Lin and Prof. Bolei Zhou. I hold a B.Sc. in Physics from the Yuanpei College at Peking University, and a B.Sc. from the Guanghua School of Management. My undergraduate thesis was supervised by Prof. Zhouchen Lin.

I am seeking full-time research positions starting in 2025


πŸ“š Research


Research interests and motivations: My research focuses on RL and LLM Alignment (also referred to as post-training). RL is the key toward super-human intelligence, and more powerful LLMs β€” optimized by RL β€” enable humans to learn from machine intelligence through natural language.

I am particularly proud of the following research works:

🧠 Large Language Model Alignment (Since 2023)

πŸ€– Reinforcement Learning (Since 2018)


πŸ“° News!


πŸ‡ΈπŸ‡¬ (2025.04) I’ll attend ICLR 2025 in person.

πŸ‡ΊπŸ‡Έ (2025.03) Guest lecture on Inverse RL Meets LLMs at the UCLA Reinforcement Learning course.

πŸ‡ΊπŸ‡Έ (2025.02) Attending AAAI 2025 to run the Tutorial: Inverse RL Meets LLMs. Thanks for joining us in Philadelphia! Slide.

πŸ“„ (2025.02) Our Reward Model Paper Part IV: Multi-Objective and Personalized Alignment with PCA is online.

πŸ“„ (2025.02) Our Reward Model Paper Part III: Infrastructure for Reproducible Reward Model Research is online.

πŸ“„ (2025.02) Our Reward Model Paper Part II: Active Reward Modeling is online.

πŸ“„ (2025.01) Our Reward Model Paper Part I: Foundation, Theory, and Alternatives is accepted by ICLR as an Oral πŸŽ‰. It is an amazing experience to co-lead this paper with Yunyi and advised by Jef.

πŸ‡¦πŸ‡Ή (2024.12) We will run the Tutorial: Inverse RL Meets LLMs at ACL-2025, see you at Vienna!

πŸ‡¬πŸ‡§ (2024.10) New talk on Inverse RL Meets LLMs at the vdsLab2024 OpenHouse and UCLA Zhou Lab. Slide is online

πŸ“„ (2024.09) Our Data Centric Reward Modeling paper is accepted by the Journal of Data-Centric Machine Learning Research (DMLR).

πŸ‡ΊπŸ‡Έ (2024.08) InverseRLignment is presented at the RL beyond reward workshop (accepted with score 9) at the 1-st RLConference, it builds reward models from SFT data..

πŸ“„ (2024.05) Our RLHF with Dense Reward paper is accepted by ICML 2024.

πŸ‡¬πŸ‡§ (2024.03) Prompt-OIRL and RATP are featured at the Inspiration Exchange, recording is online .

πŸ‡¦πŸ‡Ή (2024.01) 1 RL + LLM Reasoning paper is accepted by ICLR 2024! Prompt-OIRL uses Inverse RL to evaluate and optimize prompts for Math Reasoning.

πŸ‡ΊπŸ‡Έ (2024.01) Invited talk on RLHF at the Intuit AI Research Forum. slide

πŸ‡¨πŸ‡³ (2023.12) Invited talk on RLHF at the Likelihood Lab slide

πŸ‡¨πŸ‡³ (2023.11) Invited talk on RLHF at the CoAI group, THU.. slide

πŸ“„ (2023.10) Prompt-OIRL is selected as an oral presentation πŸŽ‰ at the NeurIPS 2023 ENLSP workshop!

πŸ“„ (2023.10) I wrote an article to share my thoughts as an RL researcher in the Era of LLMs.

πŸ“„ (2023.09) 2 papers on Interpretable Offline RL and Interpretable Uncertainty Quantification are accepted by NeurIPS 2023.