National Science Foundation (NSF)

CAREER: Automated Multimodal Learning for Healthcare

Multimodal learning is one of the central tasks of artificial intelligence (AI), which aims to effectively fuse and model multimodal data to gain a better understanding of the world around us. Many multimodal fusion strategies have been proposed, ranging from manually designed policies to advanced automated machine learning (AutoML)-based approaches. Although AutoML-based solutions outperform handcrafted ones, they are still far from optimal due to their lack of generalizability in model design and failure to account for the unique characteristics of multimodal data. This project takes the multimodal healthcare predictive modeling task as a representative example, aiming to discover and identify the optimal way to fuse multimodal data via a new learning paradigm, i.e., automated multimodal learning, with minimal human interventions. The success of this project will yield new fundamental knowledge in various fields, including automated machine learning, multimodal deep learning, and healthcare predictive modeling. The new automated multimodal learning paradigm will revolutionize multimodal data mining by automatically searching for new and complex yet optimal fusion strategies from the data, potentially motivating researchers and domain experts to understand the multimodal data better. In addition, recognizing unique research challenges posed by the unique nature of multimodal data in the healthcare domain and providing customized solutions will advance the research of healthcare predictive modeling significantly.
To meet these goals, the investigator proposes to equip automated multimodal learning with the ability to model the unique challenges of multimodal health data, including data size variety, noise, and missing modalities. The investigator also proposes to validate the proposed research for different multimodal fusion tasks in healthcare informatics and beyond and gather feedback from experts to refine the proposed research. The results of this project will provide a needed paradigm shift toward automated multimodal data fusion, impacting a broad range of research fields, including machine learning, data mining, and healthcare informatics. The proposed research will also make an enduring contribution to multimodal predictive modeling in clinical practice and other domains. The generated data, source codes, and software tools will be made available to researchers worldwide. The open platform will expedite research, enhance global collaborations in this field, and provide longstanding value for academia, healthcare organizations, and health industries. The proposed education plan will help to ensure that graduates are well equipped to design and evaluate machine learning solutions and cultivate K-12 students' interest in computer science and informatics. It will also lead to a more diverse population of undergraduate research assistants and enhance collaboration and networking among graduate students.

Students

  • Aofei Chang (Penn State)
  • Xiaochen Wang (Penn State)
  • Yuan Zhong (Penn State)

  • Publications

  • Asymmetrical Reciprocity-based Federated Learning for Resolving Disparities in Medical Diagnosis
    Jiaqi Wang*, Ziyi Yin*, Quanzeng You, Lingjuan Lyu, and Fenglong Ma
    Proceedings of the 31st SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 2025), August 3 - 7, 2025, Toronto, ON, Canada. (* equal contribution, August cycle research track acceptance rate: 19%)
  • Multimodal Artificial Intelligence in Healthcare
    Jiaqi Wang, Xiaochen Wang, Yuan Zhong, Ziyi Yin, Aofei Chang, Cao Xiao, and Fenglong Ma
    Conference Tutorial at the Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI 2025), February 25 - March 4, 2025, Philadelphia, Pennsylvania, USA.
  • FedMeKI: A Benchmark for Scaling Medical Foundation Models via Federated Knowledge Injection
    Jiaqi Wang*, Xiaochen Wang*, Lingjuan Lyu, Jinghui Chen and Fenglong Ma
    Proceedings of the Thirty-eighth Annual Conference on Neural Information Processing Systems (NeurPIS 2024), December 9-15, 2024, Vancouver, Canada. (Spotlight, * equal contribution, datasets and benchmarks track acceptance rate: 25.3%)
  • Asymmetric Mutual Learning for Decentralized Federated Medical Imaging
    Jiaqi Wang, Houping Xiao and Fenglong Ma
    Proceedings of the 15th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (ACM-BCB 2024), November 22-25, 2024, Shenzhen, China. (Oral acceptance rate: 17.6%)
  • FEDKIM: Adaptive Federated Knowledge Injection into Medical Foundation Models
    Xiaochen Wang*, Jiaqi Wang*, Houping Xiao, Jinghui Chen and Fenglong Ma
    Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP 2024 Main), November 12 –16, Miami, Florida. (* equal contribution)
  • BIPEFT: Budget-Guided Iterative Search for Parameter Efficient Fine-Tuning of Large Pretrained Language Models
    Aofei Chang, Jiaqi Wang, Han Liu, Parminder Bhatia, Cao Xiao, Ting Wang and Fenglong Ma
    Findings of the 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP 2024 Findings), November 12 –16, Miami, Florida.
  • Synthesizing Multimodal Electronic Health Records via Predictive Diffusion Models
    Yuan Zhong, Xiaochen Wang, Jiaqi Wang, Xiaokun Zhang, Yaqing Wang, Mengdi Huai, Cao Xiao and Fenglong Ma
    Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 2024), Aug. 25-29, 2024, Barcelona, Spain.
  • Unity in Diversity: Collaborative Pre-training Across Multimodal Medical Sources
    Xiaochen Wang, Junyu Luo, Jiaqi Wang, Yuan Zhong, Xiaokun Zhang, Yaqing Wang, Parminder Bhatia, Cao Xiao and Fenglong Ma
    Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024), August 11-16, 2024, Bangkok, Thailand.
  • Recent Advances in Predictive Modeling with Electronic Health Records[Paper][Slides]
    Jiaqi Wang, Junyu Luo, Muchao Ye, Xiaochen Wang, Yuan Zhong, Aofei Chang, Guanjie Huang, Ziyi Yin, Cao Xiao, Jimeng Sun and Fenglong Ma
    Proceedings of the 33rd International Joint Conference on Artificial Intelligence (IJCAI 2024), Aug. 3-9, 2024, Jeju, South Korea. (Survey Track, acceptance rate: 20.7%)
  • Automated Fusion of Multimodal Electronic Health Records for Better Medical Predictions
    Suhan Cui, Jiaqi Wang, Yuan Zhong, Han Liu, Ting Wang and Fenglong Ma
    Proceedings of the 24th SIAM International Conference on Data Mining (SDM 2024), accepted. (acceptance rate: 29.2%)
  • MedDiffusion: Boosting Health Risk Prediction via Diffusion-based Data Augmentation
    Yuan Zhong, Suhan Cui, Jiaqi Wang, Ziyi Yin, Yaqing Wang, Houping Xiao, Mengdi Huai, Ting Wang and Fenglong Ma
    Proceedings of the 24th SIAM International Conference on Data Mining (SDM 2024), accepted. (acceptance rate: 29.2%)
  • Hierarchical Pretraining on Multimodal Electronic Health Records
    Xiaochen Wang, Junyu Luo, Jiaqi Wang, Ziyi Yin, Suhan Cui, Yuan Zhong, Yaqing Wang and Fenglong Ma
    Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP 2023), December 6-10, 2023, Singapore, accepted.
  • ClinicalRisk: A New Therapy-related Clinical Trial Dataset for Predicting Risk Levels and Risk Factors
    Junyu Luo, Zhi Qiao, Lucas Glass, Cao Xiao and Fenglong Ma
    Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM 2023), Oct. 21-25, 2023, Birmingham, UK, accepted.
  • pADR: Towards Personalized Adverse Drug Reaction Prediction
    Junyu Luo, Cheng Qian, Xiaochen Wang, Lucas Glass and Fenglong Ma
    Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM 2023), Oct. 21-25, 2023, Birmingham, UK, accepted.

  • Dissemination

  • AAAI 2025 Tutorial [Link]
  • IJCAI 2024 AI4TS Workshop Keynote [Link]
  • ICPR 2024 PRHA Workshop Keynote [Link]

  • Softwares

  • AutoFM (SDM 2024)
  • MedDiffusion (SDM 2024)
  • MedNMP (EMNLP 2023)