Bio. I am an assistant professor in the School of Science and Engineering at the Chinese University of Hong Kong, Shenzhen (CUHK-Shenzhen). Before that, I received the Ph.D. degree from Texas A&M University (2019-2024), and conducted postdoctoral research at the California Institute of Technology (2024-2025). Please refer to my CV for more information.
I am starting the AI for Life Sciences Laboratory to develop machine learning algorithms for structural data mining (e.g. graphs, point clouds, and fields), and leveraging them to build simulators of living organisms at multiple scales (e.g. virtual tissues). [join us]
News.
2025/03. Receive the Distinguished Graduate Student Award for Excellence in Research from the Association of Former Students @ TAMU. [news]
2025/01. Excited to release the Cellular Interaction Foundation Model (CIFM), an AI foundation model capable of simulating living systems like DeepSeek/ChatGPT – it inputs cellular microenvironments (like language contexts) and query locations, and outputs gene expressions (like next word tokens) for those locations (2025/03: Accepted @ MLGenX Workshop, ICLR’25). [huggingface] 🎉
2024/12. Co-organize AI Bootcamp VIII on Graph Machine Learning @ Caltech.
2024/10. “Correlational Lagrangian Schrödinger Bridge: Learning Dynamics with Population-Level Regularization” (biology-inspired diffusion models under correlation conservation) is accepted @ AIDrugX Workshop, NeurIPS’24. [poster] 🎉
2024/06. Pass the Ph.D. final defense titled “Generalizable Graph AI for Biomedicine: Data-Driven Self-Supervision and Principled Regularization”, and become a Ph.D. 🎓🎉
2024/04. Participate in the community effort of CAGI6 Rare Genomes Project with the outcome accepted @ Human Genomics’24.
2024/03. “Multi-Modal Contrastive Learning for Proteins by Combining Domain-Informed Views” (multi-modal protein representation learning) is accepted @ MLGenX Workshop, ICLR’24. [poster]
2024/01. “Latent 3D Graph Diffusion” (latent diffusion models for 3D graphs) is accepted @ ICLR’24. [poster]
2023/04. Receive the Quality Graduate Student Award from ECEN @ TAMU.
2023/01. “Graph Domain Adaptation via Theory-Grounded Spectral Regularization” (model-based risk bound analysis of GDA) is accepted @ ICLR’23. [poster] 🎉
2022/10. “Does Inter-Protein Contact Prediction Benefit from Multi-Modal Data and Auxiliary Tasks?” (multi-modal/task protein-protein interface prediction) is accepted @ MLSB Workshop, NeurIPS’22. [poster]
2022/09. “Augmentations in Hypergraph Contrastive Learning: Fabricated and Generative” (contrastive learning on hypergraphs) is accepted @ NeurIPS’22. [poster]
2022/06. “Cross-Modality and Self-Supervised Protein Embedding for Compound-Protein Affinity and Contact Prediction” (multi-modal self-supervision in CPAC) is accepted @ Bioinformatics’22 (MoML’22, ECCB’22). [poster]
2022/01. “Bayesian Modeling and Uncertainty Quantification for Learning to Optimize: What, Why, and How” (Bayesian learning to optimize) is accepted @ ICLR’22. [poster]
2021/12. Receive the NSF Student Travel Awards from WSDM’22.
2021/10. “Bringing Your Own View: Graph Contrastive Learning without Prefabricated Data Augmentations” (generative augmentations in GraphCL) is accepted @ WSDM’22. [poster]
2021/09. Receive the Chevron Scholarship from ECEN @ TAMU.
2021/07. Serve as the session chair of Semisupervised and Unsupervised Learning @ ICML’21 and talk.
2021/05. “Graph Contrastive Learning Automated” (long presentation, automatic augmentation selection in GraphCL) is accepted @ ICML’21. [video] 🎉
2021/03. “Probabilistic Constructive Interference Precoding for Imperfect CSIT” (robust CI precoding in wireless communication) is accepted @ TVT’21. 🎉
2020/09. “Cross-Modality Protein Embedding for Compound-Protein Affinity and Contact Prediction” (cross-modality learning in CPAC) is accepted @ MLSB Workshop, NeurIPS’20. [poster]
2020/09. “Graph Contrastive Learning with Augmentations” (contrastive learning in GNN pre-training) is accepted @ NeurIPS’20. [poster]
2020/06. “When Does Self-Supervision Help Graph Convolutional Networks?” (self-supervision in GCNs) is accepted @ ICML’20.
2020/02. “L2-GCN: Layer-Wise and Learned Efficient Training of Graph Convolutional Networks” (efficient GCN training) is accepted @ CVPR’20.
2019/08 – 2024/08. Attend Texas A&M University for the Ph.D. Degree in Electrical Engineering, advised by Prof. Yang Shen.
2019/02. Receive the Electrical and Computer Engineering PhD Merit Fellowship from ECEN @ TAMU.