Hey there! I’m Gaoxiang Luo and I’m a 2nd-year Ph.D. candidate at the University of Minnesota, Twin Cities, advised by Prof. Ju Sun and Prof. Aryan Deshwal. I’m broadly interested in generative models. My pursuit of a doctorate is funded by Zscaler and DSI-MnDRIVE award.
Selecting an optimal set of exemplars is critical for good performance of in-context learning. However, prior exemplar search methods narrowly optimize for predictive accuracy, critically neglecting model calibration—a key determinant of trustworthiness and safe deployment. In this paper, we formulate exemplar selection as a multi-objective optimization problem, explicitly targeting both the maximization of predictive accuracy and the minimization of expected calibration error. We solve this problem with a sample-efficient Combinatorial Bayesian Optimization algorithm (COM-BOM) to find the Pareto-front that optimally trade-offs the two objectives of accuracy and calibration. We evaluate COM-BOM on multiple tasks from un-saturated MMLU-pro benchmark and find that COM-BOM beats or matches the baselines in jointly optimizing the two objectives, while requiring a minimal number of LLM API calls.
@inproceedings{luo-deshwal-2025-com,title={{COM}-{BOM}: {B}ayesian Exemplar Search for Efficiently Exploring the Accuracy-Calibration {P}areto Frontier},author={Luo, Gaoxiang and Deshwal, Aryan},editor={Christodoulopoulos, Christos and Chakraborty, Tanmoy and Rose, Carolyn and Peng, Violet},booktitle={Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing},month=nov,year={2025},address={Suzhou, China},publisher={Association for Computational Linguistics},url={https://aclanthology.org/2025.emnlp-main.1027/},pages={20350--20363},isbn={979-8-89176-332-6},}
@inproceedings{10.1145/3636534.3649369,author={Lai, Haowen and Luo, Gaoxiang and Liu, Yifei and Zhao, Mingmin},title={Enabling Visual Recognition at Radio Frequency},year={2024},isbn={9798400704895},publisher={Association for Computing Machinery},address={New York, NY, USA},url={https://doi.org/10.1145/3636534.3649369},doi={10.1145/3636534.3649369},booktitle={Proceedings of the 30th Annual International Conference on Mobile Computing and Networking},pages={388–403},numpages={16},keywords={RF sensing, mmWave radar, egomotion estimation, 3D imaging, robust perception, machine learning},location={Washington D.C., DC, USA},series={ACM MobiCom '24},media={https://www.bbc.com/news/articles/cm2l1y73mz1o},}
npj Digital Medicine
An in-depth evaluation of federated learning on biomedical natural language processing for information extraction
Language models (LMs) such as BERT and GPT have revolutionized natural language processing (NLP). However, the medical field faces challenges in training LMs due to limited data access and privacy constraints imposed by regulations like the Health Insurance Portability and Accountability Act (HIPPA) and the General Data Protection Regulation (GDPR). Federated learning (FL) offers a decentralized solution that enables collaborative learning while ensuring data privacy. In this study, we evaluated FL on 2 biomedical NLP tasks encompassing 8 corpora using 6 LMs. Our results show that: (1) FL models consistently outperformed models trained on individual clients’ data and sometimes performed comparably with models trained with polled data; (2) with the fixed number of total data, FL models training with more clients produced inferior performance but pre-trained transformer-based models exhibited great resilience. (3) FL models significantly outperformed pre-trained LLMs with few-shot prompting.
@article{Peng2024,author={Peng, Le and Luo, Gaoxiang and Zhou, Sicheng and Chen, Jiandong and Xu, Ziyue and Sun, Ju and Zhang, Rui},title={An in-depth evaluation of federated learning on biomedical natural language processing for information extraction},journal={npj Digital Medicine},year={2024},month=may,day={15},volume={7},number={1},pages={127},issn={2398-6352},doi={10.1038/s41746-024-01126-4},url={https://doi.org/10.1038/s41746-024-01126-4},}
BMVC’23
Rethinking Transfer Learning for Medical Image Classification
Transfer learning (TL) from pretrained deep models is a standard practice in modern medical image classification (MIC). However, what levels of features to be reused are problem-dependent, and uniformly finetuning all layers of pretrained models may be suboptimal. This insight has partly motivated the recent differential TL strategies, such as TransFusion (TF) and layer-wise finetuning (LWFT), which treat the layers in the pretrained models differentially. In this paper, we add one more strategy into this family, called TruncatedTL, which reuses and finetunes appropriate bottom layers and directly discards the remaining layers. This yields not only superior MIC performance but also compact models for efficient inference, compared to other differential TL methods. Our code is available at: https://github.com/sun-umn/TTL
@inproceedings{peng2023rethinking,title={Rethinking Transfer Learning for Medical Image Classification},author={Peng, Le and Liang, Hengyue and Luo, Gaoxiang and Li, Taihui and Sun, Ju},year={2023},publisher={British Machine Vision Association},booktitle={Proceedings of the 34th British Machine Vision Conference,},series={BMVC '23},location={Aberdeen, UK},url={https://papers.bmvc2023.org/0881.pdf},}
JAMIA
Evaluation of federated learning variations for COVID-19 diagnosis using chest radiographs from 42 US and European hospitals
Federated learning (FL) allows multiple distributed data holders to collaboratively learn a shared model without data sharing. However, individual health system data are heterogeneous. “Personalized” FL variations have been developed to counter data heterogeneity, but few have been evaluated using real-world healthcare data. The purpose of this study is to investigate the performance of a single-site versus a 3-client federated model using a previously described Coronavirus Disease 19 (COVID-19) diagnostic model. Additionally, to investigate the effect of system heterogeneity, we evaluate the performance of 4 FL variations.We leverage a FL healthcare collaborative including data from 5 international healthcare systems (US and Europe) encompassing 42 hospitals. We implemented a COVID-19 computer vision diagnosis system using the Federated Averaging (FedAvg) algorithm implemented on Clara Train SDK 4.0. To study the effect of data heterogeneity, training data was pooled from 3 systems locally and federation was simulated. We compared a centralized/pooled model, versus FedAvg, and 3 personalized FL variations (FedProx, FedBN, and FedAMP).We observed comparable model performance with respect to internal validation (local model: AUROC 0.94 vs FedAvg: 0.95, P = .5) and improved model generalizability with the FedAvg model (P \< .05). When investigating the effects of model heterogeneity, we observed poor performance with FedAvg on internal validation as compared to personalized FL algorithms. FedAvg did have improved generalizability compared to personalized FL algorithms. On average, FedBN had the best rank performance on internal and external validation.FedAvg can significantly improve the generalization of the model compared to other personalization FL algorithms; however, at the cost of poor internal validity. Personalized FL may offer an opportunity to develop both internal and externally validated algorithms.
@article{10.1093/jamia/ocac188,author={Peng, Le and Luo, Gaoxiang and Walker, Andrew and Zaiman, Zachary and Jones, Emma K and Gupta, Hemant and Kersten, Kristopher and Burns, John L and Harle, Christopher A and Magoc, Tanja and Shickel, Benjamin and Steenburg, Scott D and Loftus, Tyler and Melton, Genevieve B and Gichoya, Judy Wawira and Sun, Ju and Tignanelli, Christopher J},title={{Evaluation of federated learning variations for COVID-19 diagnosis using chest radiographs from 42 US and European hospitals}},journal={Journal of the American Medical Informatics Association},volume={30},number={1},pages={54-63},year={2022},month=oct,issn={1527-974X},doi={10.1093/jamia/ocac188},url={https://doi.org/10.1093/jamia/ocac188},eprint={https://academic.oup.com/jamia/article-pdf/30/1/54/47829262/ocac188.pdf},}
KDD’22
SAMCNet: Towards a Spatially Explainable AI Approach for Classifying MxIF Oncology Data
The goal of spatially explainable artificial intelligence (AI) classification approach is to build a classifier to distinguish two classes (e.g., responder, non-responder) based on the their spatial arrangements (e.g., spatial interactions between different point categories) given multi-category point data from two classes. This problem is important for generating hypotheses towards discovering new immunotherapies for cancer treatment as well as for other applications in biomedical research and microbial ecology. This problem is challenging due to an exponential number of category subsets which may vary in the strength of their spatial interactions. Most prior efforts on using human selected spatial association measures may not be sufficient for capturing the relevant spatial interactions (e.g., surrounded by) which may be of biological significance. In addition, the related deep neural networks are limited to category pairs and do not explore larger subsets of point categories. To overcome these limitations, we propose a Spatial-interaction Aware Multi-Category deep neural Network (SAMCNet) architecture and contribute novel local reference frame characterization and point pair prioritization layers for spatially explainable classification. Experimental results on multiple cancer datasets (e.g., MxIF) show that the proposed architecture provides higher prediction accuracy over baseline methods. A real-world case study demonstrates that the proposed work discovers patterns that are missed by the existing methods and has the potential to inspire new scientific discoveries.
@inproceedings{10.1145/3534678.3539168,author={Farhadloo, Majid and Molnar, Carl and Luo, Gaoxiang and Li, Yan and Shekhar, Shashi and Maus, Rachel L. and Markovic, Svetomir and Leontovich, Alexey and Moore, Raymond},title={SAMCNet: Towards a Spatially Explainable AI Approach for Classifying MxIF Oncology Data},year={2022},isbn={9781450393850},publisher={Association for Computing Machinery},address={New York, NY, USA},url={https://doi.org/10.1145/3534678.3539168},doi={10.1145/3534678.3539168},booktitle={Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining},pages={2860–2870},numpages={11},keywords={mxif, oncology, spatial interactions, spatially explainable classifier},location={Washington DC, USA},series={KDD '22},}
In my free time, I enjoy playing with my dog 🐶 and recording guitar covers 🎸.