Adam Can Converge Without Any Modification On Update Rules Y Zhang, C Chen, N Shi, R Sun, ZQ Luo
Advances in Neural Information Processing Systems, 2022, 2022
39 2022 Provable adaptivity in adam B Wang, Y Zhang, H Zhang, Q Meng, ZM Ma, TY Liu, W Chen
arXiv preprint arXiv:2208.09900, 2022
20 2022 HyperDQN: A randomized exploration method for deep reinforcement learning Z Li, Y Li, Y Zhang, T Zhang, ZQ Luo
International Conference on Learning Representations, 2021
13 2021 When Expressivity Meets Trainability: Fewer than Neurons Can Work J Zhang, Y Zhang, M Hong, R Sun, ZQ Luo
Advances in Neural Information Processing Systems, 2021, 2021
8 2021 Remax: A simple, effective, and efficient method for aligning large language models Z Li, T Xu, Y Zhang, Y Yu, R Sun, ZQ Luo
arXiv preprint arXiv:2310.10505, 2023
3 2023 Fast QLB algorithm and hypothesis tests in logistic model for ophthalmologic bilateral correlated data YQ Lin, YS Zhang, GL Tian, CX Ma
Journal of Biopharmaceutical Statistics 31 (1), 91-107, 2021
3 2021 Why Transformers Need Adam: A Hessian Perspective Y Zhang, C Chen, T Ding, Z Li, R Sun, ZQ Luo
arXiv preprint arXiv:2402.16788, 2024
2024 Does Adam Converge and When? Y Zhang, C Chen, ZQ Luo
ICLR 2022 Blog Track, 2022
2022