Bio


Jinxi Xiang (Derek) is a Postdoctoral Researcher at Stanford University School of Medicine, working with Prof. Ruijiang Li to pioneer AI-driven innovations for precision oncology. He earned his Ph.D. in Instrumentation Engineering from Tsinghua University in 2021. Prior to Stanford, Dr. Xiang spent two years as a Senior Researcher at Tencent AI Lab, leading projects in computational pathology, image/video compression, and generative video models for gaming applications.

With expertise bridging computer vision, medical image analysis, and AI, his interdisciplinary approach emphasizes translating technical advances into clinical practice, particularly through computational pathology and multimodal AI. Dr. Xiang’s current work focuses on developing scalable AI tools to optimize cancer diagnosis and treatment, fostering collaborations between engineers, clinicians, and industry partners.

Professional Education


  • Bachelor of Engineering, Wuhan University (2016)
  • Doctor of Philosophy, Tsinghua University (2021)
  • Visiting PhD, University of Edinburgh, UK, Medical Imaging (2020)

Stanford Advisors


Patents


  • Jinxi Xiang, S Yang, J Zhang, D Jiang, Y Hou, X Han. "United States Patent US App. 18/378,405 Image detection method and apparatus", Tencent
  • Z Yang, S Yang, Jinxi Xiang, J Zhang, X Han. "United States Patent US App. 18/626,165 Method and apparatus for training image recognition model, device, and medium", Tencent
  • Z Yang, S Yang, Jinxi Xiang, J Zhang, X Han. "United States Patent US App. 18/641,184 Method for determining lesion region, and model training method and apparatus", Tencent
  • S Yang, Jinxi Xiang, J Zhang, X Han. "United States Patent US App. 18/642,802 Image encoder training method and apparatus, device, and medium", Tencent
  • G Yonghang, K Tian, Jinxi Xiang, J Zhang. "United States Patent US App. 18/816,556 Video compression method and apparatus, video decompression method and apparatus, computer device, and storage medium", Tencent
  • F Luo, Jinxi Xiang, K Tian, J Zhang. "United States Patent US App. 18/931,813 Video compression method, video decoding method, and related apparatuses", Tencent
  • Lv Yue, Jinxi Xiang, J Zhang, X Han. "United States Patent US App. 19/089,142 Image compression method and apparatus, electronic device, computer program product, and storage medium", Tencent
  • K Tian, J Zhang, Jinxi Xiang, Y Guan. "United States Patent US App. 19/217,091 Data encoding method and apparatus, data decoding method and apparatus, computer device, and storage medium", Tencent
  • Jinxi Xiang, F Luo, J Zhang. "United States Patent US App. 19/226,621 Image processing method and apparatus, computer device, and computer-readable storage medium", Tencent
  • K Tian, J Zhang, Jinxi Xiang. "United States Patent US App. 19/228,298 Video encoding and decoding processing method and apparatus, computer device, and storage medium", Tencent
  • F Luo, Jinxi Xiang, J Zhang. "United States Patent US App. 19/328,727 Image enhancement method and apparatus, electronic device, computer-readable storage medium, and computer program product", Tencent
  • S Yang, Jinxi Xiang, J Zhang, X Han. "United States Patent US Patent 12,499,150 Image encoder training method and apparatus, device, and medium", Tencent

Current Research and Scholarly Interests


I develop machine leanring methods to autonomate the digital pathology.

All Publications


  • AI-enabled virtual spatial proteomics from histopathology for interpretable biomarker discovery in lung cancer. Nature medicine Li, Z., Li, Y., Xiang, J., Wang, X., Yang, S., Zhang, X., Eweje, F., Chen, Y., Luo, X., Li, Y., Mulholland, J., Bergstrom, C., Kim, T., Olguin, F. M., Willens, S., Nirschl, J. J., West, R., Neal, J., Diehn, M., Li, R. 2026

    Abstract

    Spatial proteomics enables high-resolution mapping of protein expression and can transform our understanding of biology and disease. However, major challenges remain for clinical translation, including cost, complexity and scalability. Here we present H&E to protein expression (HEX), an AI model designed to computationally generate spatial proteomics profiles from standard histopathology slides. Trained and validated on 819,000 histopathology image tiles with matched protein expression from 382 tumor samples, HEX accurately predicts the expression of 40 biomarkers encompassing immune, structural and functional programs. HEX demonstrates substantial performance gains over alternative methods for protein expression prediction from H&E images. We develop a multimodal data integration approach that combines the original H&E image and AI-derived virtual spatial proteomics to enhance outcome prediction. Applied to six independent non-small-cell lung cancer cohorts totaling 2,298 patients, HEX-enabled multimodal integration improved prognostic accuracy by 22% and immunotherapy response prediction by 24-39% compared with conventional clinicopathological and molecular biomarkers. Biological interpretation revealed spatially organized tumor-immune niches predictive of therapeutic response, including the co-localization of T helper cells and cytotoxic T cells in responders, and immunosuppressive tumor-associated macrophage and neutrophil aggregates in non-responders. HEX provides a low-cost and scalable approach to study spatial biology and enables the discovery and clinical translation of interpretable biomarkers for precision medicine.

    View details for DOI 10.1038/s41591-025-04060-4

    View details for PubMedID 41491099

  • Pancancer outcome prediction via a unified weakly supervised deep learning model. Signal transduction and targeted therapy Yuan, W., Chen, Y., Zhu, B., Yang, S., Zhang, J., Mao, N., Xiang, J., Li, Y., Ji, Y., Luo, X., Zhang, K., Xing, X., Kang, S., Xiao, D., Wang, F., Wu, J., Zhang, H., Tang, H., Maurya, H., Corredor, G., Barrera, C., Zhou, Y., Pandav, K., Zhao, J., Jain, P., Delasos, L., Huang, J., Yang, K., Teknos, T. N., Lewis, J., Koyfman, S., Pennell, N. A., Yu, K. H., Han, X., Zhang, J., Wang, X., Madabhushi, A. 2025; 10 (1): 285

    Abstract

    Accurate prognosis prediction is essential for guiding cancer treatment and improving patient outcomes. While recent studies have demonstrated the potential of histopathological images in survival analysis, existing models are typically developed in a cancer-specific manner, lack extensive external validation, and often rely on molecular data that are not routinely available in clinical practice. To address these limitations, we present PROGPATH, a unified model capable of integrating histopathological image features with routinely collected clinical variables to achieve pancancer prognosis prediction. PROGPATH employs a weakly supervised deep learning architecture built upon the foundation model for image encoding. Morphological features are aggregated through an attention-guided multiple instance learning module and fused with clinical information via a cross-attention transformer. A router-based classification strategy further refines the prediction performance. PROGPATH was trained on 7999 whole-slide images (WSIs) from 6,670 patients across 15 cancer types, and extensively validated on 17 external cohorts with a total of 7374 WSIs from 4441 patients, covering 12 cancer types from 8 consortia and institutions across three continents. PROGPATH achieved consistently superior performance compared with state-of-the-art multimodal prognosis prediction models. It demonstrated strong generalizability across cancer types and robustness in stratified subgroups, including early- and advanced-stage patients, treatment cohorts (radiotherapy and pharmaceutical therapy), and biomarker-defined subsets. We further provide model interpretability by identifying pathological patterns critical to PROGPATH's risk predictions, such as the degree of cell differentiation and extent of necrosis. Together, these results highlight the potential of PROGPATH to support pancancer outcome prediction and inform personalized cancer management strategies.

    View details for DOI 10.1038/s41392-025-02374-w

    View details for PubMedID 40897689

    View details for PubMedCentralID PMC12405520

  • Foundation Model for Predicting Prognosis and Adjuvant Therapy Benefit From Digital Pathology in GI Cancers. Journal of clinical oncology : official journal of the American Society of Clinical Oncology Wang, X., Jiang, Y., Yang, S., Wang, F., Zhang, X., Wang, W., Chen, Y., Wu, X., Xiang, J., Li, Y., Jiang, X., Yuan, W., Zhang, J., Yu, K. H., Ward, R. L., Hawkins, N., Jonnagaddala, J., Li, G., Li, R. 2025: JCO2401501

    Abstract

    Artificial intelligence (AI) holds significant promise for improving cancer diagnosis and treatment. Here, we present a foundation AI model for prognosis prediction on the basis of standard hematoxylin and eosin-stained histopathology slides.In this multinational cohort study, we developed AI models to predict prognosis from histopathology images of patients with GI cancers. First, we trained a foundation model using over 130 million patches from 104,876 whole-slide images on the basis of self-supervised learning. Second, we fine-tuned deep learning models for predicting survival outcomes and validated them across seven cohorts, including 1,619 patients with gastric and esophageal cancers and 2,594 patients with colorectal cancer. We further assessed the model for predicting survival benefit from adjuvant chemotherapy.The AI models predicted disease-free survival and disease-specific survival with a concordance index of 0.726-0.797 for gastric cancer and 0.714-0.757 for colorectal cancer in the validation cohorts. The models stratified patients into high-risk and low-risk groups, with 5-year survival rates of 49%-52% versus 76%-92% in gastric cancer and 43%-72% versus 81%-98% in colorectal cancer. In multivariable analysis, the AI risk scores remained an independent prognostic factor after adjusting for clinicopathologic variables. Compared with stage alone, an integrated model consisting of stage and image information improved prognosis prediction across all validation cohorts. Finally, adjuvant chemotherapy was associated with improved survival in the high-risk group but not in the low-risk group (treatment-model interaction P = .01 and .006) for stage II/III gastric and colorectal cancer, respectively.The pathology foundation model can accurately predict survival outcomes and complement clinicopathologic factors in GI cancers. Pending prospective validation, it may be used to improve risk stratification and inform personalized adjuvant therapy.

    View details for DOI 10.1200/JCO-24-01501

    View details for PubMedID 40168636

  • Deep Learning-Enabled Integration of Histology and Transcriptomics for Tissue Spatial Profile Analysis. Research (Washington, D.C.) Ge, Y., Leng, J., Tang, Z., Wang, K., U, K., Zhang, S. M., Han, S., Zhang, Y., Xiang, J., Yang, S., Liu, X., Song, Y., Wang, X., Li, Y., Zhao, J. 2025; 8: 0568

    Abstract

    Spatially resolved transcriptomics enable comprehensive measurement of gene expression at subcellular resolution while preserving the spatial context of the tissue microenvironment. While deep learning has shown promise in analyzing SCST datasets, most efforts have focused on sequence data and spatial localization, with limited emphasis on leveraging rich histopathological insights from staining images. We introduce GIST, a deep learning-enabled gene expression and histology integration for spatial cellular profiling. GIST employs histopathology foundation models pretrained on millions of histology images to enhance feature extraction and a hybrid graph transformer model to integrate them with transcriptome features. Validated with datasets from human lung, breast, and colorectal cancers, GIST effectively reveals spatial domains and substantially improves the accuracy of segmenting the microenvironment after denoising transcriptomics data. This enhancement enables more accurate gene expression analysis and aids in identifying prognostic marker genes, outperforming state-of-the-art deep learning methods with a total improvement of up to 49.72%. GIST provides a generalizable framework for integrating histology with spatial transcriptome analysis, revealing novel insights into spatial organization and functional dynamics.

    View details for DOI 10.34133/research.0568

    View details for PubMedID 39830364

    View details for PubMedCentralID PMC11739434

  • A vision-language foundation model for precision oncology. Nature Xiang, J., Wang, X., Zhang, X., Xi, Y., Eweje, F., Chen, Y., Li, Y., Bergstrom, C., Gopaulchan, M., Kim, T., Yu, K. H., Willens, S., Olguin, F. M., Nirschl, J. J., Neal, J., Diehn, M., Yang, S., Li, R. 2025

    Abstract

    Clinical decision-making is driven by multimodal data, including clinical notes and pathological characteristics. Artificial intelligence approaches that can effectively integrate multimodal data hold significant promise in advancing clinical care1,2. However, the scarcity of well-annotated multimodal datasets in clinical settings has hindered the development of useful models. In this study, we developed the Multimodal transformer with Unified maSKed modeling (MUSK), a vision-language foundation model designed to leverage large-scale, unlabelled, unpaired image and text data. MUSK was pretrained on 50 million pathology images from 11,577 patients and one billion pathology-related text tokens using unified masked modelling. It was further pretrained on one million pathology image-text pairs to efficiently align the vision and language features. With minimal or no further training, MUSK was tested in a wide range of applications and demonstrated superior performance across 23 patch-level and slide-level benchmarks, including image-to-text and text-to-image retrieval, visual question answering, image classification and molecular biomarker prediction. Furthermore, MUSK showed strong performance in outcome prediction, including melanoma relapse prediction, pan-cancer prognosis prediction and immunotherapy response prediction in lung and gastro-oesophageal cancers. MUSK effectively combined complementary information from pathology images and clinical reports and could potentially improve diagnosis and precision in cancer therapy.

    View details for DOI 10.1038/s41586-024-08378-w

    View details for PubMedID 39779851

    View details for PubMedCentralID 9586871

  • Deep learning-based diagnosis and survival prediction of patients with renal cell carcinoma from primary whole slide images. Pathology Chen, S., Wang, X., Zhang, J., Jiang, L., Gao, F., Xiang, J., Yang, S., Yang, W., Zheng, J., Han, X. 2024

    Abstract

    There is an urgent clinical demand to explore novel diagnostic and prognostic biomarkers for renal cell carcinoma (RCC). We proposed deep learning-based artificial intelligence strategies. The study included 1752 whole slide images from multiple centres. Based on the pixel-level of RCC segmentation, the diagnosis diagnostic model achieved an area under the receiver operating characteristic curve (AUC) of 0.977 (95% CI 0.969-0.984) in the external validation cohort. In addition, our diagnostic model exhibited excellent performance in the differential diagnosis of RCC from renal oncocytoma, which achieved an AUC of 0.951 (0.922-0.972). The graderisk for the recognition of high-grade tumour achieved AUCs of 0.840 (0.805-0.871) in the Cancer Genome Atlas (TCGA) cohort, 0.857 (0.813-0.894) in the Shanghai General Hospital (General) cohort, and 0.894 (0.842-0.933) in the Clinical Proteomic Tumor Analysis Consortium (CPTAC) cohort, for the recognition of high-grade tumour. The OSrisk for predicting 5-year survival status achieved an AUC of 0.784 (0.746-0.819) in the TCGA cohort, which was further verified in the independent general cohort and the CPTAC cohort, with AUCs of 0.774 (0.723-0.820) and 0.702 (0.632-0.765), respectively. Moreover, the competing-risk nomogram (CRN) showed its potential to be a prognostic indicator, with a hazard ratio (HR) of 5.664 (3.893-8.239, p<0.0001), outperforming other traditional clinical prognostic indicators. Kaplan-Meier survival analysis further illustrated that our CRN could significantly distinguish patients with high survival risk. Deep learning-based artificial intelligence could be a useful tool for clinicians to diagnose and predict the prognosis of RCC patients, thus improving the process of individualised treatment.

    View details for DOI 10.1016/j.pathol.2024.05.012

    View details for PubMedID 39168777

  • Towards Real-Time Neural Video Codec for Cross-Platform Application Using Calibration Information Tian, K., Guan, Y., Xiang, J., Zhang, J., Han, X., Yang, W., ACM ASSOC COMPUTING MACHINERY. 2023: 7961-7970