Team Events Publications Contact
Journal publications
- J. Arbel, S. Girard, H. D. Nguyen, A. Usseglio-Carleve, Multivariate expectile-based
distribution: properties, Bayesian inference, and applications, Journal of Statistical Planning and
Inference, 2023.
Conference publications
- TrungTin Nguyen, Dung Ngoc Nguyen, Hien Duy Nguyen, Faicel Chamroukhi (2023). A non-asymptotic risk bound for model selection in high-dimensional mixture of experts via joint rank and variable selection. hal-03984011. Accepted at AJCAI 2023 as a long oral presentation.
- Huy Nguyen, TrungTin Nguyen, Nhat Ho (2023). Demystifying Softmax Gating Function in Gaussian Mixture of Experts. Thirty-seventh Conference on Neural Information Processing Systems as a spotlight.
- G. Oudoumanessah, C. Lartizien, M. Dojat, F. Forbes. Towards frugal unsupervised detection of subtle abnormalities in medical imaging, International Conference on Medical Image Computing and Computer Assisted Intervention MICCAI, Vancouver, Canada, October 2023, (pdf)
- J. Iollo, C. Heinkele, P. Alliez, F. Forbes. Gradient-based approach for sequential Bayesian optimal design, Colloque Francophone du Traitement du Signal et des Images, GRETSI, Grenoble, August 2023, (pdf)
- S.-K.A. Ng, R. Tawiah, H. Nguyen, F. Forbes. Mixture of linear mixed models for clustering weighted random graphs. 25th international conference on computational statistiscs, Compstat 2023, London UK, August, 2023. Link
Working papers
- T. Nguyen, F. Forbes, J. Arbel, H.D. Nguyen. Bayesian nonparametric mixture of experts for high-dimensional inverse problems, (pdf)
- H. D. Nguyen, T. T. Nguyen, J. Arbel, F. Forbes. Concentration results for approximate Bayesian computation without identifiability. Preprint and supplementary material (pdf)
- Huy Nguyen, Pedram Akbarian, TrungTin Nguyen, Nhat Ho (2023). A General Theory for Softmax Gating Multinomial Logistic Mixture of Experts. arXiv:2310.14188
- Huy Nguyen, TrungTin Nguyen, Khai Nguyen, Nhat Ho (2023). Towards Convergence Rates for Parameter Estimation in Gaussian-gated Mixture of Experts. arXiv:2305.07572.