About

Welcome to my homepage! My full name is Huy Minh Nguyen, and I am a second-year Ph.D student at the Department of Statistics and Data Sciences, University of Texas at Austin where I am fortunate to be advised by Professor Nhat Ho and Professor Alessandro Rinaldo. Before that, I graduated from Ho Chi Minh City University of Science with a Bachelor’s degree in Mathematics (Summa Cum Laude). I am currently working as a research intern at Microsoft AI.

Email: huynm@utexas.edu

Research Interests

My current research focuses on theoretical foundations for Mixture-of-Experts models. In particular, I try to comprehend the effects of various gating functions (namely softmax gate, top-K sparse softmax gate, dense-to-sparse gate, etc) on the convergence of expert estimation under the Mixture-of-Experts models. Based on insights from these results, I aim to design novel gating functions which help improve the performance of Mixture-of-Experts applications, including Large Language Models, Continual Learning and Medical Images. Additionally, I am also interested in Optimal Transport theory.

(*) denotes equal contribution.

Selected Preprints

Sigmoid Gating is More Sample Efficient than Softmax Gating in Mixture of Experts

Huy Nguyen, Nhat Ho, Alessandro Rinaldo
Under review [arXiv]

Statistical Advantages of Perturbing Cosine Router in Sparse Mixture of Experts

Huy Nguyen, Pedram Akbarian*, Trang Pham*, Trang Nguyen*, Shujian Zhang, Nhat Ho
Under review [arXiv]

Mixture of Experts Meets Prompt-Based Continual Learning

Minh Le, An Nguyen*, Huy Nguyen*, Trang Nguyen*, Trang Pham*, Linh Van Ngo, Nhat Ho
Under review [arXiv]

FuseMoE: Mixture-of-Experts Transformers for Fleximodal Fusion

Xing Han, Huy Nguyen*, Carl Harris*, Nhat Ho, Suchi Saria
Under review [arXiv]

CompeteSMoE - Effective Training of Sparse Mixture of Experts via Competition

Quang Pham, Giang Do, Huy Nguyen, TrungTin Nguyen, Chenghao Liu, Mina Sartipi, Binh T. Nguyen, Savitha Ramasamy, Xiaoli Li, Steven Hoi, Nhat Ho
Under review [arXiv]

Selected Publications on Mixture of Experts

On Least Square Estimation in Softmax Gating Mixture of Experts

Huy Nguyen, Nhat Ho, Alessandro Rinaldo
Proceedings of the ICML, 2024. [arXiv]

Is Temperature Sample Efficient for Softmax Gaussian Mixture of Experts?

Huy Nguyen, Pedram Akbarian, Nhat Ho
Proceedings of the ICML, 2024. [arXiv]

Demystifying Softmax Gating Function in Gaussian Mixture of Experts

Huy Nguyen, TrungTin Nguyen, Nhat Ho
Advances in NeurIPS, 2023 (Spotlight) . [arXiv] [NeurIPS]

Statistical Perspective of Top-K Sparse Softmax Gating Mixture of Experts

Huy Nguyen, Pedram Akbarian, Fanqi Yan, Nhat Ho
Proceedings of the ICLR, 2024. [arXiv] [ICLR]

A General Theory for Softmax Gating Multinomial Logistic Mixture of Experts

Huy Nguyen, Pedram Akbarian, TrungTin Nguyen, Nhat Ho
Proceedings of the ICML, 2024. [arXiv]

Towards Convergence Rates for Parameter Estimation in Gaussian-gated Mixture of Experts

Huy Nguyen*, TrungTin Nguyen*, Khai Nguyen, Nhat Ho
In AISTATS, 2024. [arXiv] [AISTATS]

On Parameter Estimation in Deviated Gaussian Mixture of Experts

Huy Nguyen, Khai Nguyen, Nhat Ho
In AISTATS, 2024. [arXiv]

Selected Publications on Optimal Transport

Entropic Gromov-Wasserstein between Gaussian Distributions

Huy Nguyen*, Khang Le*, Dung Le*, Dat Do, Tung Pham, Nhat Ho
Proceedings of the ICML, 2022. [arXiv] [ICML]

On Multimarginal Partial Optimal Transport: Equivalent Forms and Computational Complexity

Huy Nguyen*, Khang Le*, Khai Nguyen, Tung Pham, Nhat Ho
In AISTATS, 2022. [arXiv] [AISTATS]

On Robust Optimal Transport: Computational Complexity and Barycenter Computation

Huy Nguyen*, Khang Le*, Quang Minh Nguyen, Tung Pham, Hung Bui, Nhat Ho
Advances in NeurIPS, 2021. [arXiv] [NeurIPS]

Recent News

Professional Services

  • Conference Reviewer: ICML (2022,2024), NeurIPS (2022-2024), AISTATS (2022-2024) and ICLR (2024).
  • Journal Reviewer: Electronic Journal of Statistics
  • Workshop Reviewer: Frontier4LCD (ICML 2023).