Jiachen Hu
Algorithm Developer at ByteDance. PhD of Peking University.
My name is Jiachen Hu (胡家琛), now working as an algorithm developer at ByteDance. I graduated from Peking University and received the PhD degree in 2025, where I am fortunate to be advised by Professor Liwei Wang, and spend wonderful times working with Chi Jin and Lihong Li remotely for past years. Before becoming a PhD candidate, I obtained my B.S. from Turing Class, Peking University.
I have broad interests in sample efficient reinforcement learning and online learning, especially the application-driven problems. In the past few years, my researches focused on statistically efficient bandits (e.g., multi-armed bandits, linear bandits), online exploration in structured MDPs/POMDPs, and AI methods for mathematics. Please feel free to contact me if you are interested in my researches or having a chat with me!
Contact: nickh at pku.edu.cn
news
| Jul 2025 | I will join ByteDance as an algorithm developer! |
|---|---|
| May 2025 | One paper is accepted at ICML 2025! |
| May 2024 | One paper is accepted at ICML 2024! |
| May 2024 | One paper is accepted at TQC 2024! |
| Jun 2023 | I will visit Princeton University for the next 6 months! |
| Jan 2023 | One paper is accepted at ICLR 2023! |
selected publications
- PreprintNew Sphere Packings from the Antipode ConstructionIn arXiv preprint, 2025
Highlights: We construct non-lattice sphere packings in dimensions 19, 20, 21, 23, 44, 45, and 47, demonstrating record densities that surpass all previously documented results in these dimensions. The construction applies the antipode method to suboptimal cross-sections of \(\Lambda_{24}\) and \(P_{48p}\).
- ICLRProvable Sim-to-real Transfer in Continuous Domain with Partial ObservationsIn International Conference on Learning Representations, 2023
Highlights: We study sim-to-real transfer in continuous domains with partial observations, modeled by linear quadratic Gaussian (LQG) systems. We show that a popular robust adversarial training algorithm can learn a policy from simulation that is competitive to the optimal real-world policy, providing the first provable guarantee in this setting.
- ICLRUnderstanding Domain Randomization for Sim-to-real TransferIn International Conference on Learning Representations(Spotlight, top 6%) , 2022
Highlights: We provide a theoretical framework for domain randomization, modeling the simulator as a set of MDPs with tunable parameters. We prove sharp bounds on the sim-to-real gap and show that successful transfer is achievable without any real-world training samples, highlighting the importance of history-dependent policies.
- ICMLNear-Optimal Representation Learning for Linear Bandits and Linear RLIn Proceedings of the 38th International Conference on Machine Learning, 2021
Highlights: We study multi-task representation learning for linear bandits and episodic RL with linear approximation. Our algorithm achieves \(\tilde{O}(M\sqrt{dkT} + d\sqrt{kMT})\) regret, significantly improving over the \(\tilde{O}(Md\sqrt{T})\) baseline, yielding the first theoretical characterization of multi-task representation learning benefits in RL exploration.
- ICLRDistributed Bandit Learning: Near-Optimal Regret with Efficient CommunicationIn International Conference on Learning Representations, 2020
Highlights: We design communication protocols for distributed bandit learning with $M$ agents under central coordination. For multi-armed bandits, we achieve near-optimal regret with only \(O(M\log(MK))\) communication cost — independent of the time horizon $T$ and matching the lower bound up to a log factor.