<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom">
 
  <title>Smart Lab News</title>
  <link href="https://hkustsmartlab.github.io/"/>
  <link type="application/atom+xml" rel="self" href="https://hkustsmartlab.github.io//news.xml"/>
  <updated>2026-04-09T22:59:06+08:00</updated>
  <id>https://hkustsmartlab.github.io/</id>
  <author>
    <name>Smart Lab</name>
    <!-- <email>you@example.com</email> -->
  </author>

  
  <entry>
    <id>https://hkustsmartlab.github.io//2026/04/01/dgr</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2026/04/01/dgr/"/>
    <title>[Nature Communications] Generative AI for misalignment-resistant virtual staining to accelerate histopathology workflows</title>
    <published>2026-04-01T00:00:00+08:00</published>
    <updated>2026-04-01T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2026/04/01/dgr/</uri>
    </author>
    <content type="html">&lt;p&gt;Recently, Prof. Hao Chen’s team at HKUST, in collaboration with The Chinese University of Hong Kong (CUHK) and Nanfang Hospital of Southern Medical University, published a new paper in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;Nature Communications&lt;/font&gt;&lt;/strong&gt; (IF = 15.7): &lt;strong&gt;“Generative AI for misalignment-resistant virtual staining to accelerate histopathology workflows”&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;The study presents &lt;strong&gt;DGR (Decoupled Generation and Registration)&lt;/strong&gt;, a misalignment-resistant virtual staining framework that addresses spatial distortions caused by tissue deformation during chemical staining, and provides a scalable solution for practical pathology workflows.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig1.png&quot; alt=&quot;DGR Fig 1&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;Histopathology is a cornerstone of clinical diagnosis, relying on chemically stained tissue slides to reveal disease-related morphology and molecular signals. In practice, however, conventional staining is labor-intensive, time-consuming, tissue-consuming, and environmentally costly.&lt;/p&gt;

&lt;p&gt;Virtual staining has emerged as a promising alternative by translating one imaging modality into another with deep learning. Yet most existing methods require perfectly aligned paired data for pixel-level supervision. This assumption is difficult to satisfy in real-world workflows, because tissue deformation during processing introduces unavoidable spatial misalignment, and repeated staining on the same section often damages tissue integrity.&lt;/p&gt;

&lt;p&gt;To address this bottleneck, we propose DGR, a robust framework with a cascaded registration strategy that decouples image generation from alignment correction. This design enables high-fidelity virtual staining under imperfectly paired data, reducing data curation burden and improving practicality for clinical deployment.&lt;/p&gt;

&lt;p&gt;Across five datasets and four staining tasks, DGR demonstrates clear improvements: average gains of &lt;strong&gt;3.2%&lt;/strong&gt; on internal datasets and &lt;strong&gt;10.1%&lt;/strong&gt; on external datasets. Under severe misalignment, DGR improves PSNR by up to &lt;strong&gt;3.4 dB&lt;/strong&gt; (&lt;strong&gt;23.8%&lt;/strong&gt;) over baseline methods. In blinded pathologist evaluation, experienced pathologists achieved only around &lt;strong&gt;52%&lt;/strong&gt; accuracy in distinguishing virtual from chemical staining, indicating no statistically significant diagnostic difference.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fig. 1&lt;/strong&gt; summarizes the end-to-end virtual staining workflow and key evaluations:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;(a) pipeline from tissue sampling to virtual stain generation,&lt;/li&gt;
  &lt;li&gt;(b) dataset composition,&lt;/li&gt;
  &lt;li&gt;(c–e) quantitative comparison on PSNR/SSIM/LPIPS,&lt;/li&gt;
  &lt;li&gt;(f) robustness under severe synthetic misalignment,&lt;/li&gt;
  &lt;li&gt;(g–h) blinded pathology evaluation on H&amp;amp;E and PAS-AB.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;

&lt;p&gt;DGR contains two key modules:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;&lt;strong&gt;Registration for Noise Reduction (R1):&lt;/strong&gt; aligns generated images with roughly paired ground-truth targets, reducing the impact of noisy misregistration in reconstruction supervision.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Position-Consistency Generation (R2):&lt;/strong&gt; enforces spatial consistency between generated outputs and input images through adversarial training, ensuring the generator focuses on stain translation rather than geometric warping.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Unlike prior coupled approaches (e.g., RegGAN), DGR explicitly separates generation and registration. This avoids a common failure mode where generators hide structural misalignment and rely on downstream registration to compensate. Importantly, DGR can be integrated into existing virtual staining pipelines without modifying their backbone architectures.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig2.png&quot; alt=&quot;DGR Fig 2&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;In short, DGR imposes dual constraints: one for reliable reconstruction supervision under noisy pairing (R1), and one for preserving spatial consistency between input and generated images (R2). This enables robust learning from roughly paired data while maintaining anatomical faithfulness.&lt;/p&gt;

&lt;h2 id=&quot;results&quot;&gt;Results&lt;/h2&gt;

&lt;p&gt;We evaluated DGR across five datasets and four staining translation tasks:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Autofluorescence to H&amp;amp;E:&lt;/strong&gt; DGR achieved the best quantitative and perceptual performance, including PSNR 22.914 dB (+4.4%) and SSIM 0.766 (+4.8%), while reducing LPIPS and FID by 15.9% and 12.0% compared with the next-best method.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;H&amp;amp;E to PAS-AB:&lt;/strong&gt; On internal testing, DGR improved PSNR and SSIM by 1.2% and 1.0%; on external testing, gains reached 10.1% (PSNR) and 8.0% (SSIM), showing strong generalization.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;H&amp;amp;E to mIHC:&lt;/strong&gt; DGR delivered the best image quality and improved downstream classification performance (UniToPatho +2.4%, GCHTID +1.2%), while achieving the highest nucleus segmentation Dice score (0.422).&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;H&amp;amp;E stain normalization:&lt;/strong&gt; DGR outperformed all baselines in both fidelity and distribution-level metrics (PSNR 23.823, SSIM 0.734, FID 10.253, KID 0.007).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In a blinded pathologist study, experts distinguished virtual from chemical staining at near-chance level (about 52% accuracy), with no statistically significant difference. Under severe simulated misalignment, DGR also maintained clear robustness, improving PSNR by up to 3.4 dB over baseline methods.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig3.png&quot; alt=&quot;DGR Fig 3&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Task 1: Autofluorescence → H&amp;amp;E&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Best PSNR: &lt;strong&gt;22.914 dB&lt;/strong&gt; (+4.4%)&lt;/li&gt;
  &lt;li&gt;Best SSIM: &lt;strong&gt;0.766&lt;/strong&gt; (+4.8%)&lt;/li&gt;
  &lt;li&gt;Best LPIPS/FID: LPIPS 0.159 and FID 20.264 (15.9% and 12.0% lower than the next best method)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig4.png&quot; alt=&quot;DGR Fig 4&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Task 2: H&amp;amp;E → PAS-AB&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Internal test: PSNR/SSIM gains of +1.2%/+1.0%&lt;/li&gt;
  &lt;li&gt;External test: stronger gains of +10.1%/+8.0%&lt;/li&gt;
  &lt;li&gt;LPIPS reduction on both internal and external sets&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig5.png&quot; alt=&quot;DGR Fig 5&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Task 3: H&amp;amp;E → mIHC&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Best image-level metrics across compared methods&lt;/li&gt;
  &lt;li&gt;Best downstream task gains (+2.4% on UniToPatho, +1.2% on GCHTID)&lt;/li&gt;
  &lt;li&gt;Highest nucleus segmentation Dice (0.422), indicating superior structure consistency&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Task 4: H&amp;amp;E stain normalization&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Best fidelity and distribution scores simultaneously: PSNR 23.823, SSIM 0.734, FID 10.253, KID 0.007&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig6.png&quot; alt=&quot;DGR Fig 6&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Blinded human evaluation&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;250 virtual + 250 chemical H&amp;amp;E images, and 250 virtual + 250 chemical PAS-AB images&lt;/li&gt;
  &lt;li&gt;Accuracy around 52.4% in both settings, with non-significant statistical difference&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Misalignment robustness&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Built 11,918 well-aligned H&amp;amp;E–PAS-AB pairs and injected five levels of synthetic misalignment (rotation/translation/scaling)&lt;/li&gt;
  &lt;li&gt;DGR consistently outperformed all baselines across all misalignment levels, with up to &lt;strong&gt;+3.4 dB PSNR&lt;/strong&gt; under severe misalignment&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_dgr_fig7.png&quot; alt=&quot;DGR Fig 7&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;DGR reframes misalignment from a nuisance to remove into an inherent property of histopathology workflows that models should tolerate. By decoupling generation and registration with dual constraints, DGR improves robustness to imperfect pairing while preserving anatomical consistency and staining realism.&lt;/p&gt;

&lt;p&gt;The framework lowers data acquisition barriers, scales to practical clinical settings, and offers a general strategy for building resilient virtual staining systems. Future work will extend DGR to more staining modalities, tissue types, and end-to-end pathology applications such as tumor detection and grading.&lt;/p&gt;

&lt;p&gt;More broadly, this work shifts the focus from unrealistic data perfection to algorithmic resilience, which is a critical step toward replacing resource-intensive chemical workflows in routine practice.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;resources&quot;&gt;Resources&lt;/h2&gt;

&lt;p&gt;For more details, please see our paper &lt;a href=&quot;https://doi.org/10.1038/s41467-026-71038-2&quot;&gt;Generative AI for misalignment-resistant virtual staining to accelerate histopathology workflows&lt;/a&gt; via Nature Communications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Code&lt;/strong&gt;: &lt;a href=&quot;https://github.com/birkhoffkiki/DTR&quot;&gt;https://github.com/birkhoffkiki/DTR&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Paper&lt;/strong&gt;: “Generative AI for misalignment-resistant virtual staining to accelerate histopathology workflows” (Nature Communications, 2026)&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Authors&lt;/strong&gt;: Jiabo Ma, Wenqiang Li (co-first authors)&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Corresponding Author&lt;/strong&gt;: Hao Chen (jhc@cse.ust.hk), Professor, Department of Computer Science and Engineering, HKUST&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Collaborations&lt;/strong&gt;: HKUST, The Chinese University of Hong Kong, Nanfang Hospital of Southern Medical University&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2026/03/18/mrblock</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2026/03/18/mrblock/"/>
    <title>[ICLR 2026] Exploiting Low-Dimensional Manifold of Features for Few-Shot Whole Slide Image Classification</title>
    <published>2026-03-18T00:00:00+08:00</published>
    <updated>2026-03-18T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2026/03/18/mrblock/</uri>
    </author>
    <content type="html">&lt;p&gt;A joint research effort from &lt;strong&gt;The Chinese University of Hong Kong (CUHK)&lt;/strong&gt;, &lt;strong&gt;SmartX Lab&lt;/strong&gt;, and &lt;strong&gt;Nanyang Technological University (NTU)&lt;/strong&gt; has been accepted at &lt;strong&gt;ICLR 2026&lt;/strong&gt;, one of the premier machine learning conferences. The study presents &lt;strong&gt;MR Block (Manifold Residual Block)&lt;/strong&gt;, a plug-and-play, geometry-aware drop-in replacement for standard linear layers in Multiple Instance Learning (MIL) models, specifically targeting the challenge of &lt;strong&gt;few-shot Whole Slide Image (WSI) classification&lt;/strong&gt; in computational pathology.&lt;/p&gt;

&lt;p&gt;MR Block decomposes a linear projection into two parallel paths: a &lt;strong&gt;fixed random geometric anchor&lt;/strong&gt; that preserves the intrinsic manifold structure of pathology foundation model features, and a &lt;strong&gt;trainable low-rank residual path (LRP)&lt;/strong&gt; for task-specific adaptation. This design introduces a structured inductive bias that simplifies learning into a more tractable residual-fitting problem, &lt;strong&gt;achieving state-of-the-art performance with significantly fewer trainable parameters&lt;/strong&gt;.&lt;/p&gt;

&lt;h2 id=&quot;background&quot;&gt;Background&lt;/h2&gt;

&lt;p&gt;Histopathology is the gold standard for disease diagnosis, and computational analysis of Whole Slide Images (WSIs) faces two structural constraints. First, WSIs operate at the gigapixel scale, making Multiple Instance Learning (MIL) the de facto paradigm: each slide is represented as a &lt;strong&gt;bag&lt;/strong&gt; of patch features. Second, expert annotations are costly and scarce, and real-world data often involves few labeled slides with only slide-level labels, making models highly susceptible to overfitting.&lt;/p&gt;

&lt;p&gt;To understand the root cause of overfitting beyond the learning algorithm itself, this work examines the &lt;strong&gt;intrinsic geometric structure of features&lt;/strong&gt;. Based on the manifold hypothesis, the authors analyze feature representations on the Camelyon16 dataset using different feature extractors (CONCH, UNI, ResNet-50), providing multi-angle evidence that these representations lie on a &lt;strong&gt;low-dimensional, nonlinear manifold&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Spectral analysis&lt;/strong&gt; reveals an effective rank of only 29.7 (against CONCH’s 512-dimensional space), confirming low-dimensionality.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;t-SNE visualization&lt;/strong&gt; shows clear clustering topology.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Tangent space analysis&lt;/strong&gt; demonstrates non-flat, distance-dependent geometric drift—quantitative evidence of the manifold’s intrinsic curvature, ruling out a purely linear subspace hypothesis.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_mrblock_fig1.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Based on these observations, the authors argue that a key driver of few-shot overfitting is &lt;strong&gt;geometric&lt;/strong&gt;: while pathology foundation models produce features with a fragile low-dimensional manifold structure, existing MIL models fail to preserve it. The primary source of distortion is the most ubiquitous and indispensable component—the &lt;strong&gt;linear layer&lt;/strong&gt;. Linear layers appear in projections, attention computation, and classification heads, yet they are inherently geometry-agnostic. Tangent space analysis provides direct evidence: &lt;strong&gt;trained linear layers significantly distort the intrinsic geometry of the manifold&lt;/strong&gt;, causing models to learn overly complex mappings in few-shot settings that both violate the low-rank nature of features and discard the geometric priors learned during pre-training.&lt;/p&gt;

&lt;h2 id=&quot;related-work&quot;&gt;Related Work&lt;/h2&gt;

&lt;p&gt;Existing MIL approaches for WSI classification can be broadly grouped by how they handle feature geometry:&lt;/p&gt;

&lt;h3 id=&quot;a-standard-mil-backbones&quot;&gt;A. Standard MIL Backbones&lt;/h3&gt;

&lt;p&gt;Attention-based approaches such as &lt;strong&gt;ABMIL&lt;/strong&gt; and &lt;strong&gt;CATE&lt;/strong&gt; are effective general frameworks but rely on unconstrained linear layers that are inherently geometry-agnostic. In few-shot settings, this leads to manifold distortion and overfitting.&lt;/p&gt;

&lt;h3 id=&quot;b-few-shot-specialized-methods&quot;&gt;B. Few-Shot Specialized Methods&lt;/h3&gt;

&lt;p&gt;Methods such as &lt;strong&gt;ViLaMIL&lt;/strong&gt; and &lt;strong&gt;FOCUS&lt;/strong&gt; are specifically designed for few-shot WSI classification and represent the current state of the art. However, they do not explicitly account for the geometric structure of pre-trained features, leaving the manifold distortion problem unaddressed.&lt;/p&gt;

&lt;h3 id=&quot;c-manifold-residual-block-mr-block&quot;&gt;C. Manifold Residual Block (MR Block)&lt;/h3&gt;

&lt;p&gt;MR Block takes a fundamentally different approach by directly addressing the geometric distortion caused by linear layers. Rather than designing a new MIL backbone, it provides a &lt;strong&gt;plug-and-play replacement&lt;/strong&gt; for standard linear layers, explicitly preserving and leveraging the low-dimensional manifold geometry of foundation model features.&lt;/p&gt;

&lt;h2 id=&quot;methodology&quot;&gt;Methodology&lt;/h2&gt;

&lt;h3 id=&quot;preliminaries&quot;&gt;Preliminaries&lt;/h3&gt;

&lt;p&gt;In the bag-level MIL setting, a WSI is divided into non-overlapping patches, encoded by a pre-trained feature extractor into patch features, forming a bag. An MIL aggregator (typically attention pooling) produces a slide-level feature, which is then fed into a classifier. Modern MIL models use attention pooling that simultaneously yields patch-level importance scores.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Spectral analysis&lt;/strong&gt; estimates the intrinsic dimensionality of representations via the eigenvalue distribution of the Gram matrix, computing the Von Neumann entropy and effective rank. &lt;strong&gt;Tangent space analysis&lt;/strong&gt; probes the non-linear structure beyond low-dimensionality by constructing a neighborhood graph on normalized features and estimating local tangent spaces via PCA at each point—quantifying how much local geometry varies with position.&lt;/p&gt;

&lt;h3 id=&quot;mr-block-architecture&quot;&gt;MR Block Architecture&lt;/h3&gt;

&lt;p&gt;To mitigate the geometric degradation introduced by standard linear layers, the authors propose &lt;strong&gt;MR Block&lt;/strong&gt;, a parameter-efficient, plug-and-play, geometry-aware alternative. As illustrated below, MR Block decomposes the linear mapping into two parallel paths:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_mrblock_fig2.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Given input &lt;strong&gt;x&lt;/strong&gt; ∈ ℝ^{d_in}, MR Block is defined as:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;MR(x) = W_anchor · x + Up(GELU(Down(x)))&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;where:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Geometric Anchor Path&lt;/strong&gt;: W_anchor ∈ ℝ^{d_out × d_in} is a &lt;strong&gt;fixed random matrix&lt;/strong&gt; (Kaiming uniform initialized, never updated). It serves as a geometric anchor that approximately preserves the original feature topology while acting as a &lt;strong&gt;spectral sharpener&lt;/strong&gt; to enhance spectral discriminability.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Low-rank Residual Path (LRP)&lt;/strong&gt;: Down ∈ ℝ^{r × d_in} and Up ∈ ℝ^{d_out × r} are &lt;strong&gt;trainable matrices&lt;/strong&gt; with a bottleneck rank parameter r ≪ d_in. This structural bottleneck explicitly aligns with the low effective rank of features, modeling only task-relevant residuals.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Initialization&lt;/strong&gt;: Up is initialized to all zeros, so MR Block initially behaves as W_anchor alone, contributing zero residual at the start of training. The LRP activates only when it can improve the training objective, counteracting the geometric distortion that LRP alone would introduce.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Parameter efficiency&lt;/strong&gt;: When r ≪ d_in, LRP has r(d_in + d_out) parameters—strictly fewer than a standard linear layer’s d_in × d_out, theoretically reducing overfitting risk.&lt;/p&gt;

&lt;h2 id=&quot;experimental-results&quot;&gt;Experimental Results&lt;/h2&gt;

&lt;h3 id=&quot;comparison-with-state-of-the-art-methods&quot;&gt;Comparison with State-of-the-Art Methods&lt;/h3&gt;

&lt;p&gt;Table 1 summarizes results across multiple datasets. Three key conclusions emerge:&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_mrblock_table1.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Consistent Improvement&lt;/strong&gt;: Whether on large cohorts with artificially constructed few-shot settings or on naturally few-shot treatment response datasets, MR-augmented models consistently outperform their respective baselines across different datasets and shot numbers. On Camelyon16, TCGA-NSCLC, and TCGA-RCC, MR versions match or surpass current state-of-the-art methods (ViLaMIL and FOCUS) &lt;strong&gt;with significantly fewer trainable parameters&lt;/strong&gt;.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Parameter Efficiency&lt;/strong&gt;: Replacing standard linear layers with MR Block produces a smaller model that performs better, and this pattern holds across multiple MIL backbones. This confirms that MR’s gains come from a beneficial low-rank inductive bias, not from increased model capacity.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Stability&lt;/strong&gt;: As k increases, all methods improve steadily; MR versions show the largest gains at moderate shot counts and remain competitive at higher shot counts. MR also exhibits lower variance across runs in several settings, suggesting improved training stability.&lt;/p&gt;
  &lt;/li&gt;
&lt;/ol&gt;

&lt;h2 id=&quot;ablation-studies&quot;&gt;Ablation Studies&lt;/h2&gt;

&lt;h3 id=&quot;component-ablation&quot;&gt;Component Ablation&lt;/h3&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_mrblock_table2.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Table 2 ablation validates the decoupled design. Removing the LRP consistently degrades performance, confirming LRP’s necessity for task adaptation. Removing the geometric anchor—whether retaining or discarding the residual connection—significantly harms performance, highlighting its dual role as both a &lt;strong&gt;geometric anchor&lt;/strong&gt; and a &lt;strong&gt;spectral sharpener&lt;/strong&gt;. Most critically, &lt;strong&gt;making the anchor trainable causes catastrophic performance collapse&lt;/strong&gt;, providing direct empirical evidence that unconstrained linear layers tend to disrupt feature manifolds, while the MR design better preserves them.&lt;/p&gt;

&lt;h3 id=&quot;capacity-matched-analysis&quot;&gt;Capacity-Matched Analysis&lt;/h3&gt;

&lt;p&gt;To disentangle “fewer parameters” from “geometric inductive bias,” the authors construct a capacity-matched MR-ABMIL where MR Block’s trainable parameter count exactly matches the original gated attention layer. On Camelyon16 and RCC, capacity-matched MR-ABMIL significantly outperforms ABMIL across all shot settings; results on NSCLC are broadly comparable. Since these gains are achieved with identical parameter counts, they directly demonstrate that &lt;strong&gt;MR’s geometry-aware structure—not parameter reduction alone—plays the central role in few-shot performance gains&lt;/strong&gt;.&lt;/p&gt;

&lt;h3 id=&quot;rank-sensitivity-analysis&quot;&gt;Rank Sensitivity Analysis&lt;/h3&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_mrblock_fig3.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The residual rank r controls information flow in the low-rank path. Sensitivity analysis shows performance saturates at approximately &lt;strong&gt;r ≈ 32&lt;/strong&gt; across evaluated datasets and shot settings. This saturation point closely matches the theoretically predicted effective rank of the features (~29.7), confirming that a low-rank path suffices to capture the principal task information encoded in the manifold structure. The simpler MR-ABMIL shows a pronounced peak at r = 32, while the more expressive MR-CATE shows only marginal gains beyond this—attributed to CATE’s additional modeling capacity capturing finer feature interactions beyond the main manifold.&lt;/p&gt;

&lt;h2 id=&quot;interpretability-in-extreme-resource-constrained-settings&quot;&gt;Interpretability in Extreme Resource-Constrained Settings&lt;/h2&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2026_mrblock_fig6.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Figure 6 shows heatmaps generated by MR-CATE in the extreme 2-shot setting. Standard CATE under 2-shot supervision typically fails to produce meaningful attention maps, while MR-CATE exhibits significantly stronger robustness. In the original images (top), blue curves mark approximate tumor boundaries; corresponding heatmaps are shown below. Notably, MR-CATE also captures finer-grained boundaries beyond those in the original annotations—even within the blue tumor boundaries, the model accurately distinguishes between different morphological patterns, demonstrating sensitivity to heterogeneity within tumors and surrounding normal tissue.&lt;/p&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;This work re-examines few-shot WSI classification overfitting from a geometric perspective. The authors provide quantitative and visual evidence that pathology foundation model features exhibit a fragile low-dimensional manifold geometry, and identify a common failure mode in MIL models: their indispensable linear layers systematically destroy this manifold structure due to a lack of geometric awareness.&lt;/p&gt;

&lt;p&gt;MR Block addresses this by combining a &lt;strong&gt;fixed random geometric anchor&lt;/strong&gt; that preserves manifold structure with a &lt;strong&gt;low-rank residual path&lt;/strong&gt; for parameter-efficient task-specific adaptation. Extensive experiments not only demonstrate state-of-the-art performance but also empirically support the geometric diagnosis, offering a new geometry-aware paradigm for building more robust models—with implications extending beyond computational pathology.&lt;/p&gt;

&lt;h3 id=&quot;future-directions&quot;&gt;Future Directions&lt;/h3&gt;

&lt;ul&gt;
  &lt;li&gt;Extend MR Block to other MIL tasks beyond WSI classification.&lt;/li&gt;
  &lt;li&gt;Explore more sophisticated spectral shaping strategies for the geometric anchor.&lt;/li&gt;
  &lt;li&gt;Apply geometry-aware inductive biases to other domains where pre-trained features exhibit manifold structure (e.g., natural image few-shot learning, medical image segmentation).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The code is open-sourced at &lt;a href=&quot;https://github.com/BearCleverProud/MR-Block&quot;&gt;https://github.com/BearCleverProud/MR-Block&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;For more details, check out the full paper:&lt;br /&gt;
&lt;strong&gt;Conghao Xiong, Zhengrui Guo, Zhe Xu, Yifei Zhang, Raymond Kay-yu Tong, Si Yong Yeo, Hao Chen, Joseph J. Y. Sung, and Irwin King. “Exploiting Low-Dimensional Manifold of Features for Few-Shot Whole Slide Image Classification.” arXiv preprint arXiv:2505.15504, 2026.&lt;/strong&gt;&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/12/29/seminar</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/12/29/seminar/"/>
    <title>[Seminar] Publishing in Nature Biomedical Engineering</title>
    <published>2025-12-29T00:00:00+08:00</published>
    <updated>2025-12-29T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/12/29/seminar/</uri>
    </author>
    <content type="html">&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_cse_seminar_Jennifer.png&quot; alt=&quot;2025 cse seminar Jennifer&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Publishing in top-tier journals such as &lt;em&gt;Nature Biomedical Engineering&lt;/em&gt; can be highly competitive and challenging. This seminar will provide an insider’s perspective on the editorial and review processes at the journal, and offer practical advice for researchers aiming to publish their work there.&lt;/p&gt;

&lt;p&gt;Dr. Jennifer Haskell is an Associate Editor at &lt;em&gt;Nature Biomedical Engineering&lt;/em&gt;, part of the Nature Portfolio. She joined the journal in September 2024 after conducting research at the University of Exeter, where she worked on applications of Raman spectroscopy for cancer detection. At the journal, she handles manuscripts in the areas of biomedical imaging, cellular immunotherapy, and genetic engineering, giving her a broad view of current trends in biomedical engineering research.&lt;/p&gt;

&lt;p&gt;In this talk, Dr. Haskell will discuss what the editors look for in submissions, how manuscripts are evaluated, and common reasons for rejection. She will also share tips on framing your research story, preparing a strong manuscript, and navigating peer review. The seminar will be particularly valuable for early-career researchers and students who plan to submit their work to high-impact journals.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Title:&lt;/strong&gt; Publishing in &lt;em&gt;Nature Biomedical Engineering&lt;/em&gt;&lt;br /&gt;
&lt;strong&gt;Speaker:&lt;/strong&gt; Dr. Jennifer Haskell (Associate Editor, &lt;em&gt;Nature Biomedical Engineering&lt;/em&gt;)&lt;br /&gt;
&lt;strong&gt;Time:&lt;/strong&gt; 8 Jan 2026 (Thu), 10:00 am – 11:00 am&lt;br /&gt;
&lt;strong&gt;Venue:&lt;/strong&gt; LTB, Academic Building, HKUST&lt;br /&gt;
&lt;strong&gt;Host:&lt;/strong&gt; Dr. Hao Chen&lt;/p&gt;

&lt;p&gt;Everyone is welcome to attend!&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/12/11/freetumor</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/12/11/freetumor/"/>
    <title>[Nature Communications] Large-scale generative tumor synthesis in computed tomography images for improving tumor recognition</title>
    <published>2025-12-11T00:00:00+08:00</published>
    <updated>2025-12-11T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/12/11/freetumor/</uri>
    </author>
    <content type="html">&lt;p&gt;Recently, the SmartX Lab team has completed a groundbreaking project on a Generative AI (GAI) model for tumor synthesis. Published in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;Nature Communications&lt;/font&gt;&lt;/strong&gt;, this work presents a tumor synthesis model (FreeTumor) in CT images for improvint tumor recognition.&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;AI-driven tumor recognition unlocks new possibilities for precise tumor screening and diagnosis. However, the progress is heavily hampered by the scarcity of annotated datasets, demanding extensive efforts by radiologists. To this end, we introduce FreeTumor, a Generative AI framework to enable large-scale tumor synthesis for mitigating data scarcity. Specifically, FreeTumor effectively leverages limited labeled data and large-scale unlabeled data for training. Unleashing the power of large-scale data, FreeTumor is capable of synthesizing a large number of realistic tumors for augmenting training datasets. We curate a large-scale dataset comprising 161,310 Computed Tomography (CT) volumes for tumor synthesis and recognition, with only 2.3% containing annotated tumors. 13 board-certified radiologists are engaged to discern between synthetic and real tumors, rigorously validating the quality of synthetic tumors. Through high-quality tumor synthesis, FreeTumor showcases a notable superiority over state-of-the-art tumor recognition methods, indicating promising prospects in clinical applications.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_freetumor1.png&quot; alt=&quot;freetumor&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_freetumor2.png&quot; alt=&quot;freetumor&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The framework of FreeTumor, including two stages: (1) Large-Scale Generative Tumor Synthesis Training. We first leverage labeled data to train a baseline segmentation model as the discriminator of the tumor synthesis model. Second, we leverage both labeled and unlabeled data to train the tumor synthesis model. Specifically, we train the generator to synthesize tumors on health organs while the discriminator is utilized to discriminate the reality of synthetic tumors. (2) Tumor Synthesis for Large-Scale Segmentation Training. We employ the generator to synthesize tumors on healthy organs for augmenting segmentation training datasets, while the discriminator is employed for quality control of synthetic tumors.&lt;/p&gt;

&lt;h2 id=&quot;results&quot;&gt;Results&lt;/h2&gt;

&lt;p&gt;To validate the fidelity of synthetic tumors, we engaged 13 board-certified radiologists in a Visual Turing Test to discern between synthetic and real tumors. Rigorous clinician evaluation validates the high quality of our synthetic tumors, as they achieved only 51.1% sensitivity and 60.8% accuracy in distinguishing our synthetic tumors from real ones.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_freetumor3.png&quot; alt=&quot;freetumor&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Through high-quality tumor synthesis, FreeTumor scales up the recognition training datasets by over 40 times, showcasing a notable superiority over state-of-the-art AI methods including various synthesis methods and foundation models.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_freetumor4.png&quot; alt=&quot;freetumor&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;AI-driven tumor recognition has received increasing attention in recent years, yet the progress is heavily hampered by the scarcity of annotated datasets. Early attempts mainly focus on advancing network architectures to improve tumor recognition. Although encouraging results have been demonstrated, the scarcity of annotated datasets still heavily hampered further development. 
To this end, numerous medical foundation models have been introduced to tackle the challenges of data scarcity. Although these foundation models can leverage unlabeled data in self-supervised pre-training, they still fail to utilize unlabeled data during segmentation training and remain constrained by the limited scale of annotated datasets.&lt;/p&gt;

&lt;p&gt;Thus, tumor synthesis emerges as a promising solution to mitigate the scarcity of annotated tumor datasets, which can synthesize a large number of tumors on images for augmenting training datasets. Early attempts investigated image processing and generative models for tumor synthesis. However, these methods fail to integrate large-scale data into synthesis training, thus hindering the improvements of downstream tumor recognition. In addition, these methods largely ignore the importance of quality control in synthesizing tumors, while low-quality synthetic tumors will pose a negative impact on downstream training.&lt;/p&gt;

&lt;p&gt;To this end, we introduce FreeTumor to address the aforementioned challenges. First, FreeTumor adopts an effective adversarial-based synthesis training framework to leverage both labeled and unlabeled data, facilitating the integration of large-scale unlabeled data in synthesis training. Second, FreeTumor further employs an adversarial-based discriminator to discard low-quality synthetic tumors, enabling automatic quality control of large-scale synthetic tumors in the subsequent segmentation training. In this way, FreeTumor facilitates the utilization of large-scale data in both synthesis and segmentation training, demonstrating superior performances compared with previous methods.&lt;/p&gt;

&lt;p&gt;Although FreeTumor has demonstrated promising results in tumor recognition, there are still numerous areas for growth and improvement. In our work, we collected 12 annotated datasets from public resources for training and validation, which are commonly used in existing research for the five types of tumors/lesions we studied. With more annotated tumor datasets for training, the performance of FreeTumor could be further improved. In the future, we will consistently collect more annotated datasets to advance our model.
Moving forward, we will extend the application of FreeTumor to encompass other tumor types. Furthermore, generative models, including GAN and diffusion models, have also demonstrated promising results in the applications of other medical imaging modalities, e.g., X-ray and pathology images. In the future, we will explore adapting FreeTumor to other medical imaging modalities, which require further dataset curation and more evaluation.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;resources&quot;&gt;Resources&lt;/h2&gt;

&lt;p&gt;For more details, please see our paper &lt;a href=&quot;https://www.nature.com/articles/s41467-025-66071-6&quot;&gt;Large-scale generative tumor synthesis in computed tomography images for improving tumor recognition&lt;/a&gt; via Nature Communications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Citation&lt;/strong&gt;:&lt;br /&gt;
L. Wu, et al, “Large-scale generative tumor synthesis in computed tomography images for improving tumor recognition,” in Nature Communications, doi: https://doi.org/10.1038/s41467-025-66071-6.&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/12/03/voco</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/12/03/voco/"/>
    <title>[TPAMI] Large-Scale 3D Medical Image Pre-training with Geometric Context Priors</title>
    <published>2025-12-03T00:00:00+08:00</published>
    <updated>2025-12-03T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/12/03/voco/</uri>
    </author>
    <content type="html">&lt;p&gt;Recently, the SmartX Lab team has completed a groundbreaking project on the CT foundation model. Published in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)&lt;/font&gt;&lt;/strong&gt;, this work presents a new foundation model (VoCo) for 3D medical images with a comprehensive evaluation benchmark.&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;The scarcity of annotations poses a significant challenge in medical image analysis, which demands extensive efforts from radiologists, especially for high-dimension 3D medical images. Large-scale pre-training has emerged as a promising label-efficient solution, owing to the utilization of large-scale data, large models, and advanced pre-training techniques. However, its development in medical images remains underexplored. The primary challenge lies in harnessing large-scale unlabeled data and learning high-level semantics without annotations. We observe that 3D medical images exhibit consistent geometric context, i.e., consistent geometric relations between different organs, which leads to a promising way for learning consistent representations. Motivated by this, we introduce a simple-yet-effective Volume Contrast (VoCo) framework to leverage geometric context priors for self-supervision. Given an input volume, we extract base crops from different regions to construct positive and negative pairs for contrastive learning. Then we predict the contextual position of a random crop by contrasting its similarity to the base crops. In this way, VoCo implicitly encodes the inherent geometric context into model representations, facilitating high-level semantic learning without annotations. Extensive experiments highlight the superiority of VoCo, showcasing promising transferability to unseen modalities and datasets. VoCo notably enhances performance on datasets with limited labeled cases and significantly expedites fine-tuning convergence.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_voco1.png&quot; alt=&quot;VoCo&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;

&lt;div style=&quot;text-align: center;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_voco2.png&quot; alt=&quot;VoCo&quot; style=&quot;width: 80%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;The pivotal procedure is to generate position labels for self-supervision. We propose to leverage the inherent geometric context priors in 3D medical images. Given an input volume V, we first randomly crop a sub-volume k, with the objective of constructing positive and negative pairs with k for contrastive learning. Specifically, we propose to employ position encoding to generate n non-overlap base crops q-i, where each base crop represents a distinct region of the input volume.&lt;/p&gt;

&lt;p&gt;Within human body anatomy, various organs are situated in distinct regions, leading to a potential way for us to form positive and negative pairs. As shown in Fig., the random crop k and the positive base crops q-pos exhibit overlap areas, whereas the negative base crops q-neg, lacking such overlaps, more likely encompass different organs (not absolutely). For example, k and q-pos both contain stomach, pancreas, vein, aorta, and cava, while k and q-neg exhibit different organ information. Thus, we can employ the position encoding to construct positive and negative pairs for contrastive learning.&lt;/p&gt;

&lt;p&gt;Previous contrastive learning methods mainly employ InfoNCE loss to maximize the mutual information of positive pairs. In this paper, we propose to generate labels with specific values to supervise the correlation extent of positive pairs, i.e., with labels to reflect how similar between k and q-pos. It can be observed that the correlation between k and q-pos is associated with their overlap proportions. Intuitively, if a positive base crop q-pos shares more overlap areas with k, this q-pos would be more similar to k. Thus, we propose to assign the overlap proportions as the values of position labels y, enabling us to measure the similarity between k and q-pos. In contrast, the position labels y of q-neg are assigned to 0.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_voco3.png&quot; alt=&quot;VoCo&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Overall framework of VoCo. (a) First, we generate base crops q with corresponding position labels y. Then we input the random crop k and base crops q for contextual position prediction. Specifically, we employ a student-teacher module to project k and q separately, where the teacher projector is frozen and updated from the student projector with Exponential Moving Average (EMA). Finally, we conduct volume contrast between k and q to predict similarity s, where s is supervised by position labels y. (b) We use the position labels to supervise the intra-volume contrast on k, q-pos, and q-neg, where k, q-pos, and q-neg are from the same volume. (c) We extract random crop k-A and base crops q-B from different volumes V-A and V-B for inter-volume contrast.&lt;/p&gt;

&lt;h2 id=&quot;experiments&quot;&gt;Experiments&lt;/h2&gt;

&lt;p&gt;We build the largest benchmark in this field, where we open-source the implementation of more than 50 downstream tasks, including segmentation, classification, registration, and Visual Language Processing (VLP). Extensive experiments demonstrate the superiority of VoCo. Consistent and significant improvements across 51 tasks are highlighted, &lt;em&gt;i.e.&lt;/em&gt;, average +3.62% over baseline and +2.19% above the second-best model.&lt;/p&gt;

&lt;div style=&quot;text-align: center;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_voco4.png&quot; alt=&quot;VoCo&quot; style=&quot;width: 80%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;
&lt;p&gt;In this paper, we proposed a simple-yet-effective Volume Contrast (VoCo) framework for large-scale 3D medical image pre-training. Inspired by the consistent geometric relation between different organs, we propose to leverage the geometric context priors to learn consistent semantic representations for SSL. VoCo can also be seamlessly integrated into a semi-supervised learning framework for omni-supervised pre-training. To facilitate the study of large-scale 3D medical image pre-training, we curated the existing largest medical image pre-training dataset PreCT-160K, which encompasses 160K CT volumes covering diverse anatomical structures. We further delve into the scaling law of model capacity and propose guidelines for tailoring different model sizes to various medical tasks. To evaluate the effectiveness of pre-training, we establish a comprehensive evaluation benchmark encompassing 51 downstream datasets across various tasks. Extensive experiments highlighted the superior performance of VoCo compared with previous methods.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;resources&quot;&gt;Resources&lt;/h2&gt;

&lt;p&gt;For more details, please see our paper &lt;a href=&quot;https://ieeexplore.ieee.org/document/11274411&quot;&gt;Large-Scale 3D Medical Image Pre-training with Geometric Context Priors&lt;/a&gt; via TPAMI.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Citation&lt;/strong&gt;:&lt;br /&gt;
L. Wu, J. Zhuang and H. Chen, “Large-Scale 3D Medical Image Pre-Training With Geometric Context Priors,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, doi: 10.1109/TPAMI.2025.3639593.&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/10/22/smartpath</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/10/22/smartpath/"/>
    <title>[Press Release] SmartLab Introduces SmartPath: An AI Pathology Platform Transforming End-to-End Cancer Care</title>
    <published>2025-10-22T00:00:00+08:00</published>
    <updated>2025-10-22T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/10/22/smartpath/</uri>
    </author>
    <content type="html">&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartpath_cover.png&quot; alt=&quot;SmartPath AI Pathology Platform Banner&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;SmartLab&lt;/strong&gt; at The Hong Kong University of Science and Technology (HKUST) today launched &lt;strong&gt;SmartPath&lt;/strong&gt;, a comprehensive artificial intelligence (AI) system designed to transform the entire pathology workflow for cancer care. Led by Assistant Professor &lt;strong&gt;Hao Chen&lt;/strong&gt;, Director of the Collaboration Center for Medical and Engineering Innovation, SmartPath provides integrated support for clinical diagnosis, subtyping, biomarker quantification, treatment response assessment, and prognostic follow-up across a wide spectrum of cancers—accelerating turnaround times and enhancing personalized treatment.&lt;/p&gt;

&lt;h3 id=&quot;groundbreaking-features-for-end-to-end-clinical-support&quot;&gt;Groundbreaking Features for End-to-End Clinical Support&lt;/h3&gt;

&lt;p&gt;Developed from one of the largest and most diverse pathology datasets—over &lt;strong&gt;500,000 whole-slide images&lt;/strong&gt; spanning &lt;strong&gt;34 major tissue sites&lt;/strong&gt;—SmartPath assists healthcare professionals with &lt;strong&gt;100+ clinical tasks&lt;/strong&gt;, including cancer classification, subtyping, treatment response evaluation, survival prediction, and automated pathology report generation.&lt;/p&gt;

&lt;p&gt;SmartPath is powered by two integrated large AI models:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Generalizable Pathology Foundation Model (GPFM)&lt;/strong&gt;&lt;br /&gt;
A unified framework for accurate tumor identification, subtyping, and biomarker quantification across diverse tissues. GPFM supports survival outcome prediction and treatment response assessment, forming a data-driven foundation for personalized therapy.&lt;br /&gt;
Sources: arXiv (Towards a generalizable pathology foundation model via unified knowledge distillation, 2024), SmartX Lab publications.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Multimodal Knowledge-enhanced Whole-slide Pathology Foundation Model (mSTAR)&lt;/strong&gt;&lt;br /&gt;
Fuses whole-slide pathology images with contextual data (pathology reports and transcriptomics) to enable minute-level automated report generation and powerful visual question-answering on slide regions.&lt;br /&gt;
Sources: arXiv (A multimodal knowledge-enhanced whole-slide pathology foundation model, 2024), SmartX Lab technical page.&lt;/p&gt;
  &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Engineered for seamless clinical integration, SmartPath streamlines the cancer care cycle—from rapid slide analysis and proactive risk alerts to AI-assisted reporting—reducing diagnostic bottlenecks and enabling pathologists to focus on complex decisions.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartpath_features.png&quot; alt=&quot;SmartPath Features and Workflow&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;proven-performance-in-rigorous-clinical-trials&quot;&gt;Proven Performance in Rigorous Clinical Trials&lt;/h3&gt;

&lt;p&gt;SmartPath is undergoing &lt;strong&gt;multi-center prospective validation&lt;/strong&gt; with top-tier hospitals in Hong Kong and the Chinese Mainland. In comprehensive benchmarking, it significantly outperformed existing models. A &lt;strong&gt;recent prospective study at Nanfang Hospital&lt;/strong&gt; reported &lt;strong&gt;accuracy &amp;gt;95%&lt;/strong&gt; across multiple cancers (including lung, breast, and colorectal), confirming SmartPath’s ability to enhance diagnostic accuracy, reliably predict patient survival, and rapidly generate detailed pathology reports.&lt;br /&gt;
Source: HKUST official news release; Medical Xpress; The Standard.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartpath_clinical.png&quot; alt=&quot;Clinical Prospective Validation at Partner Hospitals&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;leadership-voices&quot;&gt;Leadership Voices&lt;/h3&gt;

&lt;p&gt;Prof. &lt;strong&gt;Hao Chen&lt;/strong&gt;, Director of the Collaboration Center for Medical and Engineering Innovation and Assistant Professor in the Departments of Computer Science and Engineering and Chemical and Biological Engineering, said:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“SmartPath has been built and validated with a strong network of clinical partners. Across a wide array of real-world clinical tasks, the system consistently ranks first in benchmarking—especially in malignancy identification and treatment response prediction. With continuous real-world feedback, SmartPath keeps learning and improving, setting a new standard for intelligent, personalized medicine.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Prof. &lt;strong&gt;Liang Li&lt;/strong&gt;, Director of the Department of Pathology at &lt;strong&gt;Nanfang Hospital&lt;/strong&gt; and Professor at &lt;strong&gt;Southern Medical University&lt;/strong&gt;, commented:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“Preliminary results from our prospective trials are highly encouraging. SmartPath improves malignancy identification, provides reliable prognostic predictions, and significantly shortens diagnostic turnaround time through rapid generation of preliminary reports—crucial for time-sensitive cancer cases. This is the future of pathology, where AI augments precision and empowers data-driven clinical decisions.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartpath_leadership.png&quot; alt=&quot;Prof. Hao Chen and Prof. Liang Li at SmartPath Launch&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;real-world-impact-and-collaboration&quot;&gt;Real-World Impact and Collaboration&lt;/h3&gt;

&lt;p&gt;SmartPath is being deployed with &lt;strong&gt;over a dozen leading hospitals&lt;/strong&gt; across Hong Kong and the Chinese Mainland, enabling robust validation across diverse patient populations and clinical tasks. The benchmark and framework established by HKUST and its partners set a new standard for computational pathology in precision oncology and smart healthcare worldwide. Ongoing research is expanding SmartPath to additional cancer types—including &lt;strong&gt;rare and genetically complex malignancies&lt;/strong&gt;—to further enhance predictive models and patient stratification.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartpath_team.png&quot; alt=&quot;HKUST SmartPath Research Team and Clinical Partners&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;hr /&gt;

&lt;h3 id=&quot;about-smartx-lab&quot;&gt;About SmartX Lab&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;SmartX Lab&lt;/strong&gt;, led by Prof. Hao Chen at HKUST, is dedicated to advancing trustworthy AI technologies for healthcare and science. Our research spans large-scale models for healthcare, computer-assisted intervention, AI for science, and bioinformatics. Our mission is to drive a transformative revolution in medical practice and scientific discovery, shaping a healthier future.&lt;/p&gt;

&lt;h3 id=&quot;about-the-hong-kong-university-of-science-and-technology&quot;&gt;About The Hong Kong University of Science and Technology&lt;/h3&gt;

&lt;p&gt;&lt;a href=&quot;https://hkust.edu.hk/news/hkust-ai-system-delivers-comprehensive-cancer-diagnosis&quot;&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/a&gt; is a world-class university that excels in innovative education, research excellence, and impactful knowledge transfer. With a holistic and interdisciplinary approach, HKUST was ranked 3rd in THE’s Young University Rankings 2024, and 19th worldwide and No.1 in Hong Kong in THE Impact Rankings 2025. Thirteen HKUST subjects were ranked among the world’s top 50 in the QS World University Rankings by Subject 2025, with Data Science and Artificial Intelligence ranked 17th globally and first in Hong Kong. Over 80% of HKUST research was rated “internationally excellent” or “world leading” in Hong Kong’s latest Research Assessment Exercise. As of July 2025, HKUST members have founded over 1,900 active start-ups, including 10 unicorns and 17 exits (IPO or M&amp;amp;A).&lt;/p&gt;

&lt;h3 id=&quot;sources-and-further-reading&quot;&gt;Sources and Further Reading&lt;/h3&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;a href=&quot;https://hkust.edu.hk/news/hkust-ai-system-delivers-comprehensive-cancer-diagnosis&quot;&gt;HKUST Official News&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href=&quot;https://medicalxpress.com/news/2025-10-ai-comprehensive-cancer-diagnosis.html&quot;&gt;Medical Xpress coverage&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href=&quot;https://www.chinadailyhk.com/hk/article/622161&quot;&gt;China Daily (HK) report&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href=&quot;https://english.dotdotnews.com/a/202510/21/AP68f76420e4b08d29053ae77d.html&quot;&gt;Dot News (HK) report&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href=&quot;https://www.thestandard.com.hk/hong-kong-news/article/314639/HKUST-develops-AI-pathology-analysis-system-with-over-95-accuracy&quot;&gt;The Standard (HK) report&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href=&quot;https://www.nature.com/articles/s41551-025-01488-4&quot;&gt;GPFM (Nature Biomedical Engineering)&lt;/a&gt;&lt;/li&gt;
  &lt;li&gt;&lt;a href=&quot;https://arxiv.org/abs/2407.15362&quot;&gt;mSTAR (arXiv)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/10/10/ecmtnbc</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/10/10/ecmtnbc/"/>
    <title>[EClinicalMedicine] Multi-task Deep Learning System achieves accurate identification and prognosis prediction of triple-negative breast cancer</title>
    <published>2025-10-10T00:00:00+08:00</published>
    <updated>2025-10-10T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/10/10/ecmtnbc/</uri>
    </author>
    <content type="html">&lt;p&gt;Recently, a joint research team from the &lt;strong&gt;The First Affiliated Hospital, Zhejiang University School of Medicine&lt;/strong&gt; and the &lt;strong&gt;Hong Kong University of Science and Technology (HKUST)&lt;/strong&gt;, together with partner hospitals in China, announced a major breakthrough in AI-assisted triple negative breast cancer (TNBC) identification and prognosis prediction. Their work introduces the &lt;strong&gt;TRiple-negative breast cancer Identification and Prognosis prediction (TRIP) System&lt;/strong&gt; designed for simultaneously identifying triple-negative breast cancer and predicting its prognosis based on H&amp;amp;E slides. Published in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;EClinicalMedicine&lt;/font&gt;&lt;/strong&gt; (IF = 10.0, CAS Medicine Q1), this study demonstrates substantial accuracy gains and time savings for pathologists in clinical scenarios.&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;In this study, we developed and rigorously validated the TRIP system, leveraging &lt;strong&gt;4,898 breast cancer patient samples&lt;/strong&gt; (including &lt;strong&gt;over 1,000 TNBC cases&lt;/strong&gt;) from five hospitals in China and the public TCGA dataset. The system accurately identifies TNBC subtype and predicts patient’s disease-free survival and overall survival using H&amp;amp;E slides, providing an innovative paradigm for integrated AI-assisted pathology image analysis of TNBC.&lt;/p&gt;

&lt;p&gt;The development of TRIP not only provides a new diagnostic and prognostic tool for TNBC, the most aggressive and heterogeneous breast cancer subtype, but also injects new momentum into intelligent pathology and precision oncology.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmtnbc_fig1.png&quot; alt=&quot;framework&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;clinical-challenges&quot;&gt;Clinical Challenges&lt;/h2&gt;

&lt;p&gt;Triple-negative breast cancer, the most aggressive subtype of breast cancer, accounts for approximately 15% of all breast cancer cases. Its cancer cells do not express estrogen receptors (ER), progesterone receptors (PR), and human epidermal growth factor receptor 2 (HER2), lacking clear therapeutic targets. The five-year survival rate for patients is only approximately 75%, significantly lower than that of other breast cancer subtypes (over 90%). Current clinical diagnosis relies primarily on immunohistochemistry (IHC), which is costly, time-consuming, and requires stringent tissue specimens. Furthermore, due to tumor heterogeneity, even patients with the same TNM stage have varying prognoses. Existing prognostic stratification methods based on clinicopathological features are limited in effectiveness, necessitating an urgent need for more efficient and accurate diagnostic and prognostic tools. Previous AI researches on triple-negative breast cancer have been plagued by small sample sizes (less than 600 TNBC patients), limited validation cohorts, and single clinical tasks, making it difficult to meet clinical needs. The TRIP system is developed by integrating multi-center large-sample data with innovative algorithms, and achieves the “identification + prognosis” multi-task integration for the first time, filling the technological gap in this field.&lt;/p&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;

&lt;p&gt;​​The exceptional performance of the TRIP system stems from multiple innovations in its underlying technical architecture, establishing a fully intelligent solution from pathology image analysis to clinical prediction:​​&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;​​Unified Multiple-Instance Learning Framework:​​ Unlike traditional single-task AI models, the TRIP system employs the same AI network structure simultaneously supporting two key modules: “Triple-Negative Breast Cancer (TNBC) Identification” and “Disease-Free Survival (DFS)/Overall Survival (OS) Prediction,” enabling efficient analysis of WSIs without manual intervention.&lt;/li&gt;
  &lt;li&gt;Effective and Efficient Long-sequence Modeling:​​ The system utilizes a state-of-the-art pathology foundation model GPFM to extract patch-level features, which can automatically adapts to staining variations across different hospitals, eliminating the need for additional staining normalization. Then, a bidirectional Mamba encoder effectively captures long-range dependencies between patches across the entire slide, addressing the limitation of traditional Mamba models that early information in long sequences is easy to be forgot, thereby enhancing the accuracy of pathology image analysis.&lt;/li&gt;
  &lt;li&gt;Dynamic Adaptation for Enhanced Generalizability:​​ To tackle challenges like staining and scanner differences across hospitals, the system integrates a Test-Time Adaptation (TTA) strategy. By fine-tuning only the parameters of normalization layer, it significantly improves prediction robustness on data from external sources, laying a solid foundation for real-world clinical deployment.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmtnbc_fig2.png&quot; alt=&quot;network structure&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;validation-and-results&quot;&gt;Validation and Results&lt;/h2&gt;
&lt;p&gt;To validate the effectiveness of the TRIP system, we constructed a large-scale dataset covering five tertiary hospitals in China and the public TCGA dataset. A total of 4,898 breast cancer patients (including more than 1,000 TNBC patients) were included, making this the &lt;strong&gt;world’s largest AI system validation study for triple-negative breast cancer&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TNBC identification performance&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Internal cohort.&lt;/strong&gt; AUC: 0.980 (95% CI: 0.958-0.996), Sensitivity: 0.963, Specificity: 0.857, Accuracy: 0.934.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;External cohort.&lt;/strong&gt; AUCs: 0.916 (95% CI: 0.848-0.959), 0.936 (95% CI: 0.907-0.962), 0.860 (95% CI: 0.779-0.930), and 0.890 (95% CI: 0.841-0.929) in SDPH, SRRS, WHCH, and TCGA, respectively, which were significantly better than conventional AI models (MaxMIL, AttMIL).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmtnbc_fig3.png&quot; alt=&quot;TNBC identification performance&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Survival analysis performance&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Disease-free survival.&lt;/strong&gt; Internal C-index: 0.747±0.070 (95% CI: 0.617-0.852), external C-index: 0.731±0.047 and 0.732±0.043&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Overall survival.&lt;/strong&gt; Internal C-index: 0.744±0.075 (95% CI: 0.602-0.865), external C-index: 0.720±0.034 and 0.721±0.030&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Kaplan-Meier analysis.&lt;/strong&gt; The TRIP system can accurately divide patients into high-risk and low-risk groups. The difference between the two groups is statistically significant (P-values ​​&amp;lt; 0.0033), providing a clear risk stratification for clinical decision-making.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmtnbc_fig4.png&quot; alt=&quot;Survival analysis performance&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;interpretability&quot;&gt;Interpretability&lt;/h2&gt;

&lt;p&gt;To address the “black box” problem of AI models, the TRIP system incorporates pathology heatmap visualization technology to visually mark tissue regions that are critical for diagnosis and prognosis. Pathologists found that the areas the system focuses on closely align with typical pathological features of triple-negative breast cancer. For example, in highly malignant cases, the heatmap highlights areas of nuclear atypia, tumor necrosis, and an immunosuppressive microenvironment; areas with abundant lymphoplasmacytic infiltration correlate with a better prognosis, which is generally consistent with clinical pathology.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmtnbc_fig5.png&quot; alt=&quot;Interpretability&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;To further validate the system’s reliability, the research team conducted the multi-omics analysis. Transcriptome data from 211 patients revealed 116 differentially expressed genes between high- and low-risk groups stratified by the TRIP system, and identified three molecular subtypes (C1-C3) with distinct immune and tumor-promoting signaling profiles. Among them, the C2 subtype showed significantly better prognosis than the C1 and C3 subtypes (P-values ​​&amp;lt; 0.05), due to its high expression of immune-related pathways (such as the interferon-γ response pathway). This result is highly consistent with the conclusions of internationally recognized TNBC molecular classification studies, confirming the scientific validity of the TRIP system’s prognostic assessment at the molecular level.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmtnbc_fig6.png&quot; alt=&quot;Multi-omics analysis&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;clinical-benefits-and-limitations&quot;&gt;Clinical Benefits and Limitations&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Clinical Benefits&lt;/strong&gt;:&lt;br /&gt;
The successful development of the TRIP system brings multiple clinical benefits to the diagnosis and treatment of triple-negative breast cancer:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Simplified diagnostic workflow: TNBC can be rapidly identified using only H&amp;amp;E slides, eliminating the need for IHC testing. This significantly reduces testing costs, shortens reporting times, and reduces the workload of pathologists, making it particularly suitable for resource-limited settings.&lt;/li&gt;
  &lt;li&gt;Optimized treatment decisions: Accurate prognostic stratification helps clinicians identify high-risk patients and chosse tailored treatment plans. It also enables treatment de-escalation for low-risk patients, reducing over-medication.&lt;/li&gt;
  &lt;li&gt;Improved equity in diagnosis and treatment: The system can be deployed on a workstation equipped with a single GPU with 12GB GPU memory, offering a low hardware threshold and facilitating its adoption in hospitals at all levels, helping to narrow the gap in diagnosis and treatment across different regions.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Limitations&lt;/strong&gt;:&lt;br /&gt;
The current research data mainly comes from postoperative tissue samples, while clinical variables (such as age and TNM stage) are not yet included, limiting its direct application in preoperative scenarios. Moreover, further prospective studies can be conducted to verify the effectiveness of the system in core needle biopsy (CNB) samples; at the same time, clinical, imaging and genomic data can be integrated to build a multimodal AI model to continuously improve system performance.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;resources&quot;&gt;Resources&lt;/h2&gt;

&lt;p&gt;For more details, see the paper: &lt;a href=&quot;https://doi.org/10.1016/j.eclinm.2025.103557&quot;&gt;Development and validation of an artificial intelligence system for triple-negative breast cancer identification and prognosis prediction: a multicentre retrospective study&lt;/a&gt; via EClinicalMedicine.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Citation&lt;/strong&gt;:&lt;br /&gt;
X.M. Zhang, H.J. Zhou, Q. Chen, et al. Development and validation of an artificial intelligence system for triple-negative breast cancer identification and prognosis prediction: a multicentre retrospective study. EClinicalMedicine (2025). https://doi.org/10.1016/j.eclinm.2025.103557&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/09/29/ecmknee</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/09/29/ecmknee/"/>
    <title>[EClinicalMedicine] Multi-task Deep Learning System Enhances Integrated Non-invasive MRI Diagnosis of Nine Knee Abnormalities</title>
    <published>2025-09-29T00:00:00+08:00</published>
    <updated>2025-09-29T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/09/29/ecmknee/</uri>
    </author>
    <content type="html">&lt;p&gt;Recently, a joint research team from the &lt;strong&gt;Southern Medical University Third Affiliated Hospital&lt;/strong&gt; and the &lt;strong&gt;Hong Kong University of Science and Technology (HKUST)&lt;/strong&gt;, together with partner hospitals in southern China, announced a major breakthrough in AI-assisted knee MRI interpretation. Their work introduces a &lt;strong&gt;multi-task deep learning system (DLS)&lt;/strong&gt; designed for integrated diagnosis of nine common knee abnormalities. Published in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;EClinicalMedicine&lt;/font&gt;&lt;/strong&gt; (IF = 10.0, CAS Medicine Q1), this study demonstrates substantial accuracy gains and time savings for radiologists in real-world settings.&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;Knee MRI diagnosis poses unique challenges due to complex soft tissue anatomy, multiple imaging sequences, and nuanced structural details. Conventional AI solutions tend to focus on single abnormalities, limiting clinical utility. The new &lt;strong&gt;DLS&lt;/strong&gt; tackles this gap by providing a “generalist” approach—covering **meniscal tears, cartilage defects, ACL/PCL/MCL/LCL injuries, infrapatellar fat pad (IFP) injury, synovial plica, and cysts—within a unified diagnostic workflow.&lt;/p&gt;

&lt;p&gt;The large-scale, multicentre study involved &lt;strong&gt;13,419 patients&lt;/strong&gt;, &lt;strong&gt;14,962 MRI exams&lt;/strong&gt;, and over &lt;strong&gt;1 million individual images&lt;/strong&gt;, and adopted a rigorous &lt;strong&gt;stepwise validation strategy&lt;/strong&gt; incorporating multi-reader, multi-case experiments and randomized controlled trials.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmknee_fig1.png&quot; alt=&quot;DLS Knee MRI Study&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;clinical-challenges&quot;&gt;Clinical Challenges&lt;/h2&gt;

&lt;p&gt;Radiologists often struggle with knee MRI interpretation due to:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Diagnostic complexity&lt;/strong&gt;: Multiple coexisting pathologies with subtle boundaries.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Current AI limitations&lt;/strong&gt;: Most systems are narrow-focus, lacking panoramic assessment.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Workflow inefficiency&lt;/strong&gt;: A single case may require numerous sequence reviews, often taking 5–8 minutes, increasing cognitive load.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmknee_fig2.png&quot; alt=&quot;DLS Knee MRI Performance&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The DLS employs a &lt;strong&gt;coarse-to-fine, multi-plane attention framework&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Coarse localization&lt;/strong&gt;: A 3D U-Net identifies the meniscal region and crops 256×256 ROIs to reduce irrelevant background.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Multi-plane feature extraction&lt;/strong&gt;: Sagittal, coronal, and axial PD fat-suppressed sequences are processed separately and then fused for classification.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Attention-guided defect focus&lt;/strong&gt;: The Attention Object Localization Module (AOLM) directs focus to discriminative areas, improving small-lesion detection.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Explainability&lt;/strong&gt;: Grad-CAM heatmaps visualize AI-attention overlap with expert-identified pathology regions, fostering clinical trust.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The system was implemented in PyTorch, and the training/testing datasets were strictly separated into internal and two external test sets.&lt;/p&gt;

&lt;h2 id=&quot;validation-and-results&quot;&gt;Validation and Results&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Multi-centre model performance&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Major lesions&lt;/strong&gt; (meniscal tear, cartilage defect, ACL tear) AUC: Internal 0.898, External I 0.852, External II 0.812&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Minor lesions&lt;/strong&gt; (PCL, MCL, LCL, IFP injury, plica, cysts) AUC: Internal 0.815, External I 0.744, External II 0.774&lt;/li&gt;
  &lt;li&gt;Accuracy range: Internal 73.1%–95.6%; External I 63.3%–89.3%; External II 65.5%–83.5%&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmknee_fig3.png&quot; alt=&quot;DLS Knee MRI Performance&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Reader studies&lt;/strong&gt;:&lt;/p&gt;
&lt;ol&gt;
  &lt;li&gt;&lt;strong&gt;Step 1&lt;/strong&gt; – DLS vs low/high seniority radiologists: Comparable accuracy in major lesions for both groups; notable gains in cartilage defects and cyst detection for low-seniority readers.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Step 2&lt;/strong&gt; – Multi-reader re-read (External I, washout period): Accuracy, sensitivity, and specificity improved across all readers; mean reading times reduced by ~30.5s (junior) and ~26.0s (senior).&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Step 3&lt;/strong&gt; – Randomized controlled trial (External II): DLS-assisted group showed overall accuracy gains of 4.2%–8.8% (statistically significant), with further reading time reductions (~35.5s junior, ~30.7s senior).&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmknee_fig4.png&quot; alt=&quot;DLS Knee MRI Performance&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;clinical-insights&quot;&gt;Clinical Insights&lt;/h2&gt;

&lt;p&gt;The DLS proved helpful in addressing nine high-difficulty scenarios, such as early-stage cartilage changes, partial ACL tears, chronic PCL “normal appearances,” and distinguishing cysts from synovial recess fluid. Attention-guided multi-plane processing enabled more stable lesion identification in these “error-prone” cases.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_ecmknee_fig5.png&quot; alt=&quot;Grad-CAM Examples&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;impact-and-outlook&quot;&gt;Impact and Outlook&lt;/h2&gt;

&lt;p&gt;From a radiology perspective, the &lt;strong&gt;integrated approach&lt;/strong&gt; reflects real-world needs for simultaneous interpretation of diverse knee pathologies. As a &lt;strong&gt;second reader&lt;/strong&gt;, DLS can mitigate both under-diagnosis and over-diagnosis risks while improving efficiency.&lt;/p&gt;

&lt;p&gt;From an AI perspective, the &lt;strong&gt;coarse-to-fine attention paradigm&lt;/strong&gt; demonstrates quantifiable benefits, with attention-guided lesion localization showing clear performance boosts in ablation studies. The explainable heatmaps align closely with human expert strategies, increasing clinical adoption potential.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Limitations&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Retrospective, regional multicentre scope; broader geographic, prospective studies are needed.&lt;/li&gt;
  &lt;li&gt;Reference standards based on MRI + report; limited arthroscopic gold-standard confirmation.&lt;/li&gt;
  &lt;li&gt;Current reliance on PD fat-suppressed sequences; broader protocol adaptability needed.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Future Directions&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Large-scale prospective evaluations to measure long-term outcome and efficiency gains.&lt;/li&gt;
  &lt;li&gt;Extend “localization–recognition” AI workflow to shoulder, hip, and ankle joints for a generalizable musculoskeletal AI platform.&lt;/li&gt;
&lt;/ul&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;resources&quot;&gt;Resources&lt;/h2&gt;

&lt;p&gt;For more details, see the paper: &lt;a href=&quot;https://doi.org/10.1016/j.eclinm.2025.103534&quot;&gt;Development of a multi-task deep learning system for classification of nine common knee abnormalities on MRI: a large-scale, multicentre, stepwise validation study&lt;/a&gt; via EClinicalMedicine.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Citation&lt;/strong&gt;:&lt;br /&gt;
Xie, Z., Qiu, Z., Li, Y., et al. Development of a multi-task deep learning system for classification of nine common knee abnormalities on MRI: a large-scale, multicentre, stepwise validation study. EClinicalMedicine (2025). https://doi.org/10.1016/j.eclinm.2025.103534&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/09/10/csig2025</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/09/10/csig2025/"/>
    <title>[Call for Participants] CSIG Youth Scientist Conference 2025 Forum</title>
    <published>2025-09-10T00:00:00+08:00</published>
    <updated>2025-09-10T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/09/10/csig2025/</uri>
    </author>
    <content type="html">&lt;p&gt;The 2025 CSIG Youth Scientist Conference, initiated by the Youth Working Committee of the China Society of Image and Graphics, will be held from September 18-21, 2025 at the Wyndham Grand Qingdao Yingsha Beach. The conference is hosted by the China Society of Image and Graphics and co-organized by the Ocean University of China, Shandong University, and the Youth Working Committee of the China Society of Image and Graphics. Academician Yaonan Wang, Professor Houjie Wang, Professor Junyu Dong, and Professor Huimin Ma will serve as the conference chairs.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig1.png&quot; alt=&quot;2025 csig2025 cover&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;overview&quot;&gt;Overview&lt;/h2&gt;

&lt;p&gt;CSIG Youth Scientist Conference 2025 will be a gathering of elites and distinguished guests. 4 keynote speakers will bring us the latest insights on image and graphics, 30 thematic forums and 4 workshops will showcase the diversity of opinions in the field, and more than 200 high-level academic reports will surely spark ideas. Youth talent forums such as the Youth Talent Support Program Forum, Excellent Doctoral Dissertation Forum, and Academic Rising Star Forum will also provide a platform for young talents to speak freely, exchange ideas, and grow. At that time, 2000+ scholars and doctoral students from academia and industry in the field of image and graphics and interdisciplinary subjects will gather together to celebrate this academic event. The biomedical image analysis sub-forum, is one of the thematic forums of the conference.&lt;/p&gt;

&lt;h2 id=&quot;agenda&quot;&gt;Agenda&lt;/h2&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig2.png&quot; alt=&quot;2025 csig2025 agenda&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;chairs-and-keynote-speakers&quot;&gt;Chairs and Keynote Speakers&lt;/h2&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig3.png&quot; alt=&quot;2025 csig2025 chairs and keynote speakers&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;chairs-information&quot;&gt;Chairs Information&lt;/h3&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig4.png&quot; alt=&quot;Prof. Qing Cai&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Qing Cai&lt;/strong&gt; - &lt;em&gt;Ocean University of China&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Qing Cai is an Associate Professor and Doctoral Supervisor at the Ocean University of China. He has successively won the titles of “Taishan Scholar” Young Expert in Shandong Province, Shandong Provincial Outstanding Youth Fund, Shandong Provincial Artificial Intelligence Science and Technology Award - Outstanding Youth Award, and Ocean University of China “Youth Talent Project”. He is a Senior Member of CCF, a member of the CCF-CV Committee, a member of the CCF-MM Special Committee, a member of the CCF-AI Special Committee, a member of the CSIG Youth Working Committee, a director of the Shandong Artificial Intelligence Society, and a member of YOCSEF Qingdao. His main research direction is artificial intelligence and medical-engineering interdisciplinary direction, including medical image processing, disease-assisted diagnosis, and 3D reconstruction. Related work has been published in important academic conferences and journals at home and abroad as the first or corresponding author: CVPR, AAAI(x4), IJCAI, ACM MM, IEEE TIP(x4), IEEE TNNLS(x2), IEEE TCSVT, PR(x3), etc. He serves as a reviewer for CVPR, ICCV, ECCV, NeurIPS, ICLR, ICML, AAAI, IJCV, IEEE TIP, IEEE TNNLS, IEEE TCYB, etc. He presides over projects such as the National Natural Science Foundation of China (General &amp;amp; Youth), the Shandong Provincial Natural Science Foundation (Outstanding Youth &amp;amp; Youth), the China Postdoctoral Science Foundation, and the Ocean University of China Youth Talent Project.&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig5.png&quot; alt=&quot;Prof. Yinghuan Shi&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Yinghuan Shi&lt;/strong&gt; - &lt;em&gt;Nanjing University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Yinghuan Shi, PhD, is a Professor and Doctoral Supervisor in the School of Computer Science and Technology at Nanjing University, Assistant Dean, and also the main person in charge of the Medical Artificial Intelligence Platform of the National Research Institute of Health and Medical Big Data at Nanjing University. She received her Bachelor’s and Doctoral degrees from the Department of Computer Science and Technology (now the School of Computer Science) at Nanjing University in 2007 and 2013, respectively. Her research interests include machine learning, pattern recognition, and their interdisciplinary research in medical image processing, AI for Science, etc. In recent years, she has presided over the National Natural Science Foundation for Excellent Young Scientists, the National Natural Science Foundation Key Project, the National Key R&amp;amp;D Program Digital Diagnosis Key Special Project, the National Science and Technology Innovation 2030 - New Generation Artificial Intelligence Major Project, and the Jiangsu Provincial Frontier Technology R&amp;amp;D Program Project. She has published more than 80 papers in CCF-A conferences and IEEE/ACM journals. She has published a popular science book “Artificial Intelligence in Your Pocket - AI and Health Care”. She has received honors such as the first Outstanding Graduate Moral Education Tutor of Nanjing University, the Wu Wenjun Artificial Intelligence Outstanding Youth Award, the China Association for Science and Technology Youth Talent Support Project, the second prize of the Jiangsu Provincial Natural Science Award (second finisher), and the Chinese People’s Liberation Army Military Medical Achievement Award (third finisher).&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig6.png&quot; alt=&quot;Prof. Xiao Jia&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Xiao Jia&lt;/strong&gt; - &lt;em&gt;Shandong University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Xiao Jia is a Professor and Doctoral Supervisor at the School of Control Science and Engineering, Shandong University. He was selected into the National High-level Youth Talent Program, Distinguished Young and Middle-aged Scholars of Shandong University (First Level), Taishan Scholar Youth Expert of Shandong Province, and Shandong Provincial Excellent Youth Science Fund Project (Overseas), and undertakes a number of national and provincial natural science funds. He received his bachelor’s degree from the Department of Automation, Shandong University, his doctorate degree from the Department of Electronic Engineering, The Chinese University of Hong Kong, and was a postdoctoral fellow at Stanford University. Relying on the Key Laboratory of Machine Intelligence and System Control of the Ministry of Education and the Institute of Artificial Intelligence and System Control, he conducts research work. His main research directions include machine learning, multimodal intelligent perception, vision-language large models, and intelligent medical systems. He has published more than 40 papers in international academic journals and conferences such as PIEEE, EU, TASE, ICRA, IROS, and MICCAI.&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig7.png&quot; alt=&quot;Prof. Miaojing Shi&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Miaojing Shi&lt;/strong&gt; - &lt;em&gt;Tongji University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Miaojing Shi is a Professor at the School of Electronics and Information Engineering, Tongji University, Vice Dean of the Scientific and Technological Research Institute, and a Visiting Professor at King’s College London. She graduated from Peking University with a Ph.D. and has served as a researcher at the French National Institute for Research in Computer Science and Automation, and as an Assistant Professor and Associate Professor in the Department of Informatics at King’s College London. Her main research interests include computer vision and medical image processing. She has published more than 90 high-level journal and conference papers. She has presided over more than 10 projects including the National Natural Science Foundation of China, the UK Engineering and Physical Sciences Research Council project, and the European Research Council project. Recently, she won the France-China Committee Personal Science and Technology Innovation Award, the King’s College London Annual Contribution Award, and the Tongji University May Fourth Youth Medal. He is a Young Thousand Talents, a Senior Member of IEEE, and a Fellow of the UK Higher Education Academy.&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig8.png&quot; alt=&quot;Prof. Daoqiang Zhang&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Daoqiang Zhang&lt;/strong&gt; - &lt;em&gt;Nanjing University of Aeronautics and Astronautics&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Daoqiang Zhang is a Professor and Dean of the School of Artificial Intelligence, Nanjing University of Aeronautics and Astronautics, and Director of the Key Laboratory of Brain-Machine Intelligence Technology of the Ministry of Education. He was selected as a national leading talent and a Fellow of the International Association for Pattern Recognition (IAPR Fellow), and was funded by the National Natural Science Foundation for Excellent Youth and Key Projects. He serves as an editor of journals such as IEEE Trans. Medical Imaging, Pattern Recognition, Machine Intelligence Research, and Intelligent Medicine, as well as the deputy editor of the journal “Data Acquisition and Processing”. He serves as a supervisor of the China Society of Image and Graphics, a director of the Alzheimer’s Disease Prevention and Treatment Association, deputy director of the Graphics Big Data Special Committee of the Chinese Society of Graphics, a standing member of the Machine Learning Special Committee of the Chinese Artificial Intelligence Society, a standing member of the Medical Information and Control Branch of the Chinese Society of Biomedical Engineering, and director of the Medical Image Processing Special Committee of the Jiangsu Artificial Intelligence Society. His main research directions are artificial intelligence, machine learning, medical image analysis, brain-computer interface, etc. He has published more than 200 academic papers, which have been cited more than 20,000 times. He has won 1 second prize of the National Natural Science Award, and 1 first prize and 1 second prize of the Ministry of Education Natural Science Award. Doctoral students/postdoctoral fellows he has supervised have won the MICCAI Young Scientist Award, an important international conference in the field of medical imaging, twice. He has been selected as an Elsevier Highly Cited Chinese Scholar for 10 consecutive years from 2014-2023.&lt;/p&gt;

&lt;h3 id=&quot;keynote-speakers-information&quot;&gt;Keynote Speakers Information&lt;/h3&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig9.png&quot; alt=&quot;Prof. Yang Chen&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
  &lt;strong&gt;Prof. Yang Chen&lt;/strong&gt; - &lt;em&gt;Southeast University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Report Title&lt;/strong&gt;: Intelligent Medical Imaging and Processing&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;: The report will focuses on high-quality imaging technology based on feature learning in clinical task-driven intelligent medical imaging, the embedding of core algorithm research and development of domestic medical imaging equipment, and clinical task-driven medical image processing. It mainly talks about four parts: intelligent medical imaging, imaging algorithm application, intelligent image processing and application, and thinking on medical-engineering interdisciplinary research.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Speaker Profile&lt;/strong&gt;: Professor Chen Yang conducts scientific research on medical imaging algorithms and intelligent image analysis, serving domestic high-end medical equipment. He has published more than 100 papers and is a Highly Cited Chinese Scholar released by Elsevier from 2022-2024. He is currently a professor at the School of Computer Science and Engineering, Southeast University, a winner of the National Science Fund for Distinguished Young Scholars, and the person in charge of the Key R&amp;amp;D Program of the Ministry of Science and Technology.&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig10.png&quot; alt=&quot;Prof. Yong Xia&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
  &lt;strong&gt;Prof. Yong Xia&lt;/strong&gt; - &lt;em&gt;Northwestern Polytechnical University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Report Title&lt;/strong&gt;: Intelligent Computing for Medical Imaging - Challenges and Practices&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;: With the rapid development of deep learning technology, intelligent computing for medical imaging has made significant progress, but it still faces many challenges, especially how to build high-performance and reliable models in the case of scarce labeled data and long-tail distribution of diseases. In order to cope with these challenges, the application of pre-training technology in medical image analysis has gradually attracted attention. Related research is committed to using relevant or even irrelevant medical image data to pre-train the model to improve its ability to analyze various modalities of medical images. On this basis, researchers are working to build large-scale foundation models and develop model fine-tuning techniques to improve the generalization performance of models on different diagnostic tasks. This report will delve into the main challenges faced by pre-training technology in medical image analysis, including insufficient data labeling, data dimensionality issues, limitations of model capabilities, and the construction of foundation models. By sharing the research experience and insights of the research group in these fields, this report will also explore the application opportunities and challenges of pre-training and foundation models in medical image analysis, in order to provide useful references and inspirations for researchers in related fields.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Speaker Profile&lt;/strong&gt;: Yong Xia is a Professor at the School of Computer Science/School of Artificial Intelligence, Northwestern Polytechnical University, and a member of the National Engineering Laboratory for Integrated Space-Air-Ground-Sea Big Data Application Technology. His research direction is intelligent computing for medical imaging. In the past 5 years, he has published more than 100 papers in JAMA Network Open, Radiology, IEEE-TPAMI/TMI/TIP/TNNLS, IJCV, MedIA, NeurIPS, CVPR, ECCV, MICCAI, AAAI, and IJCAI. Google citations are more than 17,000 times, H-index=61, and he has won the top three in more than 10 international subject competitions. He serves as a director of the Chinese Society for Stereology, a standing member of the Digital Medicine Branch of the China Computer Federation, etc.&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig11.png&quot; alt=&quot;Prof. Yong Liu&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
  &lt;strong&gt;Prof. Yong Liu&lt;/strong&gt; - &lt;em&gt;Beijing University of Posts and Telecommunications&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Report Title&lt;/strong&gt;: Brain Connection, Brain Network and Cognitive Ability: A Reflection Taking AD Application as an Example&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;: Multimodal magnetic resonance brain imaging can non-invasively provide information on the structure and functional activity of the human brain. Brain network research has opened up new avenues for understanding brain information processing mechanisms and evaluating new intervention programs. Developing precise and effective brain network computing theories and methods, and on this basis, clarifying the structural and functional organization rules, information processing patterns, and regulatory mechanisms of brain networks on cognitive functions has become a common scientific frontier for information science, brain science, etc. We will discuss with you the origin of the ideas of brain connection and brain network research based on brain imaging, the research progress of some special populations, and the progress of application in Alzheimer’s disease (AD). The focus is on reporting the recent progress of the team in selecting AD, a typical representative of neurodegenerative diseases, as the research object, from individualized brain atlases, individualized precise brain connection patterns, the development of non-invasive transcranial photobiomodulation systems, and AD intervention paradigms.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Speaker Profile&lt;/strong&gt;: Yong Liu, PhD, Professor, Vice Dean of the School of Artificial Intelligence, Beijing University of Posts and Telecommunications. His main research direction is intelligent understanding and clinical application of brain imaging. He has published more than 50 papers as the corresponding (including co-) author, including Science Advances, Alzheimer &amp;amp; Dementia, eClinicalMedicine, Biological Psychiatry, Science Bulletin, etc., and has been granted 7 patents. As the project leader, he has undertaken projects including the National Natural Science Foundation Youth Fund (Category A) project, the Ministry of Science and Technology Science and Technology Innovation 2030 - Major Project, the National Key R&amp;amp;D Program Key Special Project, the National Natural Science Foundation Key Project, and the Beijing Natural Science Foundation for Distinguished Young Scientists (2020). He has won academic awards such as the Wu Wenjun Artificial Intelligence Science Award First Prize (2019, ranked 2nd), and has been continuously selected as an Elsevier Highly Cited Chinese Scholar (2020–).&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig12.png&quot; alt=&quot;Prof. Hao Chen&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
  &lt;strong&gt;Prof. Hao Chen&lt;/strong&gt; - &lt;em&gt;Hong Kong University of Science and Technology&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Report Title&lt;/strong&gt;: Pathology Large Models for Precision Cancer Diagnosis and Treatment: Challenges and Opportunities&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;: With the deep integration of artificial intelligence and digital pathology, pathology large models are becoming a new paradigm in the field of precision cancer diagnosis and treatment. This report will systematically explore the frontier progress of pathology large models in precision cancer diagnosis, molecular subtyping prediction, treatment response evaluation, and prognostic analysis, revealing its breakthrough potential to achieve “micro-meso-macro” association analysis through multimodal data fusion (such as whole-slide images, imaging, genomics, and clinical information). Further explore how pathology large models promote the paradigm shift of precision cancer diagnosis and treatment from “experience-driven” to “calculation-driven”.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Speaker Profile&lt;/strong&gt;: Hao Chen, Assistant Professor, Department of Computer Science and Engineering, Department of Chemical and Biological Engineering, and Department of Life Science, Hong Kong University of Science and Technology, Director of the Medical Engineering Interdisciplinary Joint Innovation Center. His research interests include medical large models, computational pathology, multimodal fusion, medical image analysis, computer-aided minimally invasive diagnosis and treatment, etc. He has published more than 100 papers (Google Scholar citations more than 35,000 times, h-index 79) in top journals and conferences such as Nature Biomedical Engineering, Nature Communications, Lancet Digital Health, Nature Machine Intelligence, Jama, MICCAI, IEEE-TMI, MIA, CVPR, ICCV, etc. He has been continuously selected as a Stanford University global top 2% scientist and a Clarivate Analytics global highly cited scientist. He has won the 2023 Asian Young Scientist Award, the second prize of the National Ministry of Education Excellent Achievement Award, the first prize of the Beijing Science and Technology Progress Award, and the 2019 MICCAI Young Scientist Impact Award, a top conference in artificial intelligence medical imaging. He serves as an editor of journals including IEEE TMI, TNNLS, J-BHI, and CMIG, and serves as an area chair and program committee member of multiple international conferences such as ICLR, CVPR, ACM MM, and MICCAI. He has led the team to win 15 international medical image analysis challenge championships.&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_csig_fig13.png&quot; alt=&quot;Prof. Yitian Zhao&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
  &lt;strong&gt;Prof. Yitian Zhao&lt;/strong&gt; - &lt;em&gt;Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Report Title&lt;/strong&gt;: Beyond The Eye: Exploring Neurodegenerative Diseases with Intelligent Analysis of Ophthalmic Images&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;: Due to the complexity of the brain, the ambiguity of evaluation indicators, and the lack of direct observation methods, early screening for neurodegenerative diseases such as Alzheimer’s disease (AD) is a difficult problem in the medical field. Clinically routine methods such as PET, MRI, DSA, and cognitive tests have problems such as large equipment, high cost, radiation or trauma, and long detection time, which are not suitable for popularization at the grassroots level and large-scale population screening. Therefore, exploring efficient and non-invasive screening methods has become an urgent need and research hotspot for early prevention. The eye and brain have a high degree of developmental homology and functional similarity, and the eye has the potential to become an excellent observation window for studying the brain. This report will introduce the reporter’s research work in multimodal ophthalmic medical image processing in recent years, mainly including image enhancement, structure extraction and segmentation, feature quantification, and disease diagnosis algorithm research. Focusing on the difficulties in large-scale screening of neurodegenerative diseases such as Alzheimer’s disease, this report explores the correlation between brain diseases and various eye image features, reveals the internal connections and patterns between brain diseases and retinal structural changes, and provides a scientific basis and technical support for early accurate assisted diagnosis.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Speaker Profile&lt;/strong&gt;: Yitian Zhao is a researcher and doctoral supervisor at the Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, and the deputy director of the Advanced Diagnosis and Treatment Laboratory. He is mainly engaged in the research of artificial intelligence + ophthalmic medical image analysis. In recent years, he has focused on ophthalmic multimodal images and carried out intelligent diagnosis algorithm research and equipment development for eye, brain, heart and other related diseases. He has successively received support from the National Excellent Youth Fund, the Ministry of Science and Technology Key R&amp;amp;D Program Young Scientists, the Zhejiang Provincial Outstanding Youth Fund, and the China Association for Science and Technology Youth Talent Support Project, and has presided over more than ten projects including the National Natural Science Foundation. He has published more than 100 papers in top international journals and conferences in the industry such as Nature Machine Intelligence, npj Digital Medicine, IEEE PAMI, IJCV, IEEE-TMI, MedIA, CVPR, and MICCAI, with more than 9,200 academic citations and single paper highest citation times over 2300 times. He currently serves as an editor/associate editor for journals such as The Innovation (IF: 32.1) IEEE Trans. Medical Imaging (IF: 10.6) and Medical Physics (IF: 4.5).&lt;/p&gt;

&lt;h2 id=&quot;more-information&quot;&gt;More Information&lt;/h2&gt;

&lt;p&gt;For more information, please visit the &lt;a href=&quot;http://youth.csig.org.cn/CSIG2025&quot;&gt;official website&lt;/a&gt;. We look forward to your participation!&lt;/p&gt;

</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/09/07/icidm2025</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/09/07/icidm2025/"/>
    <title>[Call for Participants] 8th International Symposium on Image Computing and Digital Medicine</title>
    <published>2025-09-07T00:00:00+08:00</published>
    <updated>2025-09-07T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/09/07/icidm2025/</uri>
    </author>
    <content type="html">&lt;p&gt;The 8th International Symposium on Image Computing and Digital Medicine (ISICDM 2025) will be held at Shenzhen University from December 20 to 22, 2025. This year, ISICDM will feature a sub-forum on “Empowering Precision Medicine through Intelligent Digital Pathology,” which will bring together top scholars from around the world to discuss the latest research and applications of intelligent digital pathology in precision medicine.&lt;/p&gt;

&lt;h2 id=&quot;overview&quot;&gt;Overview&lt;/h2&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig1.png&quot; alt=&quot;ICIDM 2025&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;With the rapid advancement of artificial intelligence, digital and intelligent pathology is becoming a core driver of precision medicine. By integrating high-resolution pathological images with multi-omics data, and leveraging cutting-edge technologies such as large AI models and multimodal fusion, intelligent digital pathology enables quantitative analysis of microscopic disease features, significantly improving the accuracy and efficiency of pathological diagnosis. This provides essential support for personalized treatment and prognosis evaluation.&lt;/p&gt;

&lt;p&gt;However, real-world implementation still faces critical challenges — including model interpretability &amp;amp; transparency, patient privacy protection, and ethical compliance.&lt;/p&gt;

&lt;p&gt;This forum will bring together prominent experts from pathology, AI, clinical medicine, and bioinformatics to discuss:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;Breakthroughs in frontier technologies&lt;/li&gt;
  &lt;li&gt;Clinical application scenarios&lt;/li&gt;
  &lt;li&gt;Standardization and regulatory considerations&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Through interdisciplinary dialogue, the forum aims to connect technological innovation with clinical needs, driving the paradigm shift from experience-based medicine to data-driven medicine, and contributing new momentum to the global precision medicine ecosystem.&lt;/p&gt;

&lt;h2 id=&quot;forum-agenda&quot;&gt;Forum Agenda&lt;/h2&gt;
&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig2.png&quot; alt=&quot;ICIDM 2025&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;chairs&quot;&gt;Chairs&lt;/h2&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig3.png&quot; alt=&quot;Prof. Hao Chen&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Hao Chen&lt;/strong&gt; - &lt;em&gt;Hong Kong University of Science and Technology&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Assistant Professor (Research) at the Departments of Computer Science &amp;amp; Engineering, Chemical &amp;amp; Biological Engineering, and Division of Life Science; Director of the Joint Innovation Center for Med-Engineering Crossover.&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Research interests: large medical foundation models, computational pathology, multimodal data fusion, medical image analysis, interpretable deep learning, AI-assisted minimally invasive diagnosis.&lt;/li&gt;
  &lt;li&gt;Published over 200 papers in &lt;em&gt;Nature Biomedical Engineering&lt;/em&gt;, &lt;em&gt;Nature Communications&lt;/em&gt;, &lt;em&gt;The Lancet Digital Health&lt;/em&gt;, &lt;em&gt;Nature Machine Intelligence&lt;/em&gt;, &lt;em&gt;JAMA&lt;/em&gt;, MICCAI, IEEE Transactions on Medical Imaging, &lt;em&gt;Medical Image Analysis&lt;/em&gt;, CVPR, and ICCV, with more than 35,000 Google Scholar citations and a Nature Index of 79.&lt;/li&gt;
  &lt;li&gt;Recognized as Stanford University Top 2% Scientist (multiple years) and Clarivate Highly Cited Researcher.&lt;/li&gt;
  &lt;li&gt;Awards: 2023 Asian Young Scientist Award, Ministry of Education (China) Scientific Research Excellence Award (2nd Prize), Beijing Science &amp;amp; Technology Progress Award (1st Prize), MICCAI 2019 Young Scientist Award.&lt;/li&gt;
  &lt;li&gt;Serves on editorial boards of IEEE RBME, IEEE TMI, IEEE TNNLS, JBHI, and CMIG; Area Chair for ICLR, CVPR, ACM MM, and MICCAI; led teams to over 15 international medical image analysis challenge championships.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig4.png&quot; alt=&quot;Prof. Liansheng Wang&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Liansheng Wang&lt;/strong&gt; - &lt;em&gt;Xiamen University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Professor at the School of Informatics and jointly appointed Professor at the School of Medicine; Vice Director of the Digital Fujian Institute of Big Data for Health; Director of the XMU Medical AI Research Institute; Chair of MICS; Vice Chair of the AI Group of the Radiology Branch of the Fujian Medical Association; PhD from the Chinese University of Hong Kong.&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Research focus: medical image processing, AI-assisted diagnosis.&lt;/li&gt;
  &lt;li&gt;Published over 120 papers in &lt;em&gt;Nature Machine Intelligence&lt;/em&gt;, &lt;em&gt;Nature Communications&lt;/em&gt;, IEEE Transactions on Medical Imaging, &lt;em&gt;Medical Image Analysis&lt;/em&gt;, CVPR, AAAI, etc.&lt;/li&gt;
  &lt;li&gt;PI/Co-PI for NSFC instrumentation projects, China’s “Innovation 2030” Megaprojects, National Key R&amp;amp;D Program, and NSFC general/youth projects.&lt;/li&gt;
  &lt;li&gt;Awards include Tencent Rhino-Bird Research Award, Fujian Provincial Science &amp;amp; Technology Progress Award (2nd Prize), 2023 Tian Zhaowu Interdisciplinary Research Award (1st Prize) at XMU.&lt;/li&gt;
  &lt;li&gt;Led teams to win 11 international medical imaging competitions.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig5.png&quot; alt=&quot;Prof. Jun Xu&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Jun Xu&lt;/strong&gt; - &lt;em&gt;Nanjing University of Information Science &amp;amp; Technology&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;Vice Dean of the School of Artificial Intelligence; Level-II Professor; Executive Director of the Institute of Smart Healthcare.&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;PhD from Zhejiang University; postdoctoral researcher and visiting professor at Rutgers University and Case Western Reserve University.&lt;/li&gt;
  &lt;li&gt;Roles: Deputy Director of the Medical Image Analysis Committee of the Jiangsu AI Society; Member of the Digital Pathology &amp;amp; AI Committee of the Chinese Society of Pathology; rotating Chair of the 4th Youth Symposium on Medical Image Computing.&lt;/li&gt;
  &lt;li&gt;Published in &lt;em&gt;Nature Communications&lt;/em&gt;, &lt;em&gt;Radiology&lt;/em&gt;, IEEE Transactions on Medical Imaging, &lt;em&gt;Medical Image Analysis&lt;/em&gt;; listed as Stanford University Top 2% Scientist.&lt;/li&gt;
  &lt;li&gt;PI for NSFC joint key projects, NSFC general projects, National Key R&amp;amp;D Program key projects, and provincial/ministerial grants.&lt;/li&gt;
  &lt;li&gt;Research focuses: medical image computation, computational pathology, quantitative analysis of imaging and pathological slices for disease classification, risk prediction, diagnosis, treatment response, and prognosis evaluation.&lt;/li&gt;
&lt;/ul&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;keynote-speakers--topics&quot;&gt;Keynote Speakers &amp;amp; Topics&lt;/h2&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig6.png&quot; alt=&quot;Prof. Lin Yang&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Lin Yang&lt;/strong&gt; - &lt;em&gt;Westlake University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Digital Pathology &amp;amp; Artificial Intelligence: Current Status and Future Perspectives&lt;/em&gt;     &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Diagnostic pathology is the foundation and gold standard for cancer identification, but high inter-observer variability significantly impacts diagnostic consistency and efficiency, particularly in areas with shortages of pathologists. Despite rapid developments in computer-aided diagnosis (CAD), whole-slide pathology diagnosis still faces significant challenges in real-world settings. This talk reviews the latest progress in digital pathology, presents AI algorithms and solutions developed by our team to address practical challenges, and discusses open questions hindering wider clinical adoption in China.      &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; From 2009–2011, served as Assistant Professor at Rutgers University (Departments of Pathology, Radiology, and Biomedical Engineering). From 2011–2014, Assistant Professor at the University of Kentucky. From 2014–2019, Preeminence Hire at the University of Florida with tenure as Associate Professor across three departments. Since 2020, Professor at Westlake University focusing on AI, medical imaging, machine learning, computer vision, and medical foundation models. Author of 100+ publications in &lt;em&gt;Nature Machine Intelligence&lt;/em&gt;, &lt;em&gt;Nature Medicine&lt;/em&gt;, CVPR, ECCV with over 10,000 citations. Winner of MICCAI Young Scientist Awards (2015, 2024), ECCV 2024 Best Paper Finalist; listed in Stanford University’s Top 2% Scientists.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig7.png&quot; alt=&quot;Prof. Yongbing Zhang&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Yongbing Zhang&lt;/strong&gt; - &lt;em&gt;Harbin Institute of Technology, Shenzhen&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Whole-Slide Pathology Scanning and Computational Analysis&lt;/em&gt;         &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Histopathology imaging and computational analysis are the clinical gold standards for cancer diagnosis. Current imaging technologies suffer from low speed and precision; manual diagnosis is labor-intensive and prone to subjective bias, potentially resulting in missed or incorrect diagnoses. This talk introduces the evolution of computational pathology, our lab’s work on high-speed whole-slide scanning and AI-assisted computational diagnosis, and future development prospects.             &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Professor and Doctoral Supervisor; recipient of the National Science Fund for Excellent Young Scholars. Has authored 100+ publications in &lt;em&gt;Nature&lt;/em&gt; sub-journals, IEEE Transactions, NeurIPS, CVPR, ICCV; holds over 50 patents. Winner of the China National Science &amp;amp; Technology Progress Award (2nd Prize); recognized as Guangdong Province Science &amp;amp; Technology Innovation Leading Talent.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig8.png&quot; alt=&quot;Prof. Munning Wang&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Munning Wang&lt;/strong&gt; - &lt;em&gt;Fudan University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Multi-Instance Learning with Bag and Instance-Level Optimization for Digital Pathology Diagnosis&lt;/em&gt;      &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Digital pathology diagnosis is typically framed as a multi-instance learning (MIL) classification task supervised only by bag-level labels, often neglecting instance classification. Accurate instance classification is vital for interpretability and biomarker discovery; furthermore, it can improve bag-level classification performance. This talk presents our research on optimizing both bag and instance classification, including works published at NeurIPS and ICCV.       &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Professor at the Digital Medicine Center of Fudan University; Deputy Director of the Shanghai Key Laboratory of Medical Image Computing and Computer-Assisted Intervention. Research areas: AI-based digital pathology, medical imaging AI, and computer-aided drug screening. Author of 80+ papers in IEEE TPAMI, IEEE TMI, BIB, NeurIPS, ICML, CVPR, ICCV; recipient of China National Technology Invention Award (2nd Prize), Shanghai Youth Science &amp;amp; Technology Talent, and Shanghai Outstanding Technical Leader honors.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig9.png&quot; alt=&quot;Prof. Sheng Huang&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Sheng Huang&lt;/strong&gt; - &lt;em&gt;Chongqing University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Key Instance Mining: Identifying Local Diagnostic Cues in Pathological Images&lt;/em&gt;         &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Addresses the visual redundancy of ultra-high-resolution whole-slide images (WSIs) using a multi-instance learning (MIL) framework to study key instance evaluation and selection. Models the weakly supervised relationship between image patches and diagnostic labels, employing attention mechanisms, local pseudo-classifiers, and clustering strategies to identify discriminative local cues. Promotes interpretability by transitioning from global to local evidence in AI-assisted pathology analysis.         &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Professor at the School of Big Data &amp;amp; Software Engineering, Chongqing University; Vice Director of the Big Data Intelligence Research Institute; ACM Chongqing Rising Star Awardee. Research focuses on medical image processing, open-world pattern recognition, and intelligent industrial inspection. Author of 60+ publications in IEEE TIP, TIFS, TNNLS, TMI, ICCV, CVPR, AAAI, IJCAI; serves as Area Chair for IJCAI, PRCV; executive committee member for multiple academic societies.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig10.png&quot; alt=&quot;Prof. Jingang Yu&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Prof. Jingang Yu&lt;/strong&gt; - &lt;em&gt;South China University of Technology&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Instance Segmentation in Pathology Images and Clinical Applications&lt;/em&gt;            &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Accurate instance-level segmentation of pathological structures such as nuclei and glands from digital pathology images is a key task in computational pathology. It underpins many clinical applications including tumor microenvironment analysis and precise immunohistochemistry interpretation, yet faces challenges such as labor-intensive annotation, strong heterogeneity, and structural complexity. This talk is divided into three parts: (1) overview of challenges, solutions, and current research status in pathology instance segmentation; (2) our recent work under weak annotation settings, including few-shot learning, semi-supervised learning, and approaches based on vision-language large models; (3) clinical use cases in related technologies.               &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Professor at the School of Automation, South China University of Technology; “Pearl River Talent Plan” Distinguished Young Scholar of Guangdong Province. Has studied at Xi’an Jiaotong University, Huazhong University of Science and Technology, and University of Nebraska–Lincoln (USA). Focuses on AI algorithms and translational applications for pathology. First/corresponding author of 40+ papers in TMI, MIA, MICCAI, TIP, CVPR; PI of 10+ national/provincial projects; first inventor on 12 patents worth nearly RMB 5 million in completed tech transfers; AI pathology products have been deployed in multiple hospitals. Academic roles include Deputy Secretary-General of the Visual Cognition &amp;amp; Computing Committee of CSIG and MICS Executive Committee Member.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig11.png&quot; alt=&quot;Assoc. Prof. Yushan Zheng&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Assoc. Prof. Yushan Zheng&lt;/strong&gt; - &lt;em&gt;Beihang University&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Multi-Level Representation Learning for Whole-Slide Images and Tumor-Aided Diagnosis&lt;/em&gt;      &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Whole-slide image (WSI) representation learning is central to digital pathology analysis and tumor-aided diagnosis. This talk introduces a multi-level representation learning framework spanning from the microscopic level of cells to the macroscopic level of tissues, as well as integration with cross-case multimodal structured learning. The framework effectively improves tumor screening, molecular status prediction, and prognosis evaluation across multiple auxiliary diagnostic systems.          &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Associate Professor at Beihang University, focusing on digital pathology image processing and tumor-aided diagnosis. PI of over 10 projects including National NSFC, Beijing Natural Science Foundation, and industry-academia collaborations. Author of 50+ papers in IEEE TMI, Medical Image Analysis, AAAI, MICCAI; holder of 10+ patents. Recognitions include the Young Scientist Award from the Chinese Society for Stereology and Biomedical Engineering.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig12.png&quot; alt=&quot;Assoc. Prof. Chu Han&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Assoc. Prof. Chu Han&lt;/strong&gt; - &lt;em&gt;Guangdong Provincial People’s Hospital&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Efficient Computation and Processing of Digital Pathology Images&lt;/em&gt;         &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Computational pathology can assist in precise cancer diagnosis and treatment, but the extremely high resolution of pathology images poses challenges for annotation and computation. To address low annotation efficiency, our team developed multi-task learning, weak supervision, and self-supervised learning strategies that greatly reduce the need for manual expert labeling. To improve computational efficiency, lightweight network models and knowledge distillation were employed to enable large-scale, rapid pathology image analysis. This work provides core technical support for building efficient, automated intelligent pathology diagnostic systems.               &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Guangdong Distinguished Young Scholar; PhD in Computer Science from The Chinese University of Hong Kong; Associate Researcher and PhD Supervisor at Guangdong Provincial People’s Hospital; PI at the Guangdong Key Laboratory for Intelligent Medical Imaging Analysis &amp;amp; Application. Research focuses on AI algorithms for oncology imaging and computational pathology. Author of 70+ papers in TPAMI, TNNLS, TMI, MedIA, ACM TOG, CVPR, MICCAI; filed 20+ patents (6 granted). PI for multiple NSFC and Guangdong grants; awards include Guangdong S&amp;amp;T Progress Award (1st Prize, 2024) and 2023 National Digital Health Application Contest Grand Prize (1st place out of 39 teams).&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig13.png&quot; alt=&quot;Assoc. Prof. Jun Shi&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Assoc. Prof. Jun Shi&lt;/strong&gt; - &lt;em&gt;Hefei University of Technology&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;AI-Based Intelligent Pathology Image Analysis for Cancer-Aided Diagnosis and Treatment&lt;/em&gt;     &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; Pathological diagnosis is the gold standard in cancer treatment. AI technologies have been increasingly applied to digital pathology image analysis, reshaping cancer diagnosis and treatment. This talk focuses on our practical research in AI-assisted cancer care, including advances in histopathology image analysis and multimodal pathology data modeling for tumor classification/grading, biomarker prediction, prognosis evaluation, and therapeutic response prediction, as well as challenges and future directions.        &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Associate Professor at the School of Software, Hefei University of Technology; Member of MICS. Holds a PhD in Pattern Recognition &amp;amp; Intelligent Systems from Beihang University; former researcher at CETC 38th Research Institute. Author of 40+ papers in IEEE TMI, MIA, Journal of Pathology, MICCAI, AAAI, BIBM; inventor on 10+ patents. PI/co-PI of over 10 research projects, including NSFC and Anhui Provincial grants; reviewer for IEEE TMI, MIA, TIP, JBHI, TCSVT, MICCAI, AAAI, BIBM.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig14.png&quot; alt=&quot;Dr. Tian Shen&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Dr. Tian Shen&lt;/strong&gt; - &lt;em&gt;SenseTime Medical&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;AI-Driven Smart Pathology Department Construction&lt;/em&gt;           &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; The development of smart, digital pathology departments is a hot direction globally. Artificial intelligence can significantly improve diagnostic accuracy and workflow efficiency, but real-world issues such as a wide variety of disease types and multimodal long-tail distributions create barriers. This talk shares cutting-edge industry practices in integrating segmentation, classification, and detection into pathology department operations, and highlights how multimodal large models and model production tools can address long-tail challenges in diagnosis.                &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; PhD in Computer Science from Lehigh University (USA). Formerly with Siemens Research and Tencent AI Lab, specializing in microscopic image processing and medical AI. Over a decade of industry experience in algorithm R&amp;amp;D, product innovation, and quality management. Currently COO of SenseTime Medical, overseeing AI pathology and imaging product deployment and usage.&lt;/p&gt;

&lt;p&gt;&lt;br /&gt;&lt;/p&gt;

&lt;div style=&quot;display: flex; align-items: center; margin-bottom: 15px;&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_icidm2025_fig15.png&quot; alt=&quot;Assoc. Prof. Wei Shao&quot; style=&quot;width: 100px; height: 100px; object-fit: cover; border-radius: 8px; margin-right: 20px;&quot; /&gt;
    &lt;strong&gt;Assoc. Prof. Wei Shao&lt;/strong&gt; - &lt;em&gt;Nanjing University of Aeronautics and Astronautics&lt;/em&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Talk:&lt;/strong&gt; &lt;em&gt;Computational Pathology Radiogenomics&lt;/em&gt;           &lt;br /&gt;
&lt;strong&gt;Abstract:&lt;/strong&gt; This talk presents our recent progress in intelligent pathology radiogenomics for tumor diagnosis, including tumor microenvironment component analysis, multimodal fusion, open-set active learning for nucleus detection, zero-shot tissue segmentation, image–gene association analysis, and spatial transcriptomics-based gene prediction, with applications to breast, lung, and liver cancer prognosis and treatment.                 &lt;br /&gt;
&lt;strong&gt;Bio:&lt;/strong&gt; Associate Professor at the College of Artificial Intelligence, NUAA; PhD Supervisor; member of the iBrain team (headed by Prof. Daoqiang Zhang). PI of two NSFC general projects; author of 55 publications in &lt;em&gt;Nature Communications&lt;/em&gt; (2), &lt;em&gt;Cell Reports&lt;/em&gt; (2), IEEE TMI (12), as well as CVPR, ICCV, NeurIPS; over 3,200 Google Scholar citations. Honors include 2× MICCAI Young Scientist Awards (only mainland China recipient), 2025 MOE Natural Science Award (2nd Prize), and Stanford University Top 2% Scientist listing (2024).&lt;/p&gt;

&lt;h2 id=&quot;about-isicdm-2025&quot;&gt;About ISICDM 2025&lt;/h2&gt;

&lt;p&gt;The forum represents a unique opportunity to engage with leading experts in digital pathology and AI, explore cutting-edge research, and discuss practical implementation challenges. Through the diverse expertise of our distinguished speakers, participants will gain comprehensive insights into the latest technological advances and their clinical applications. We warmly welcome researchers, clinicians, and industry professionals to join us in shaping the future of intelligent digital pathology.&lt;/p&gt;

</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/09/02/gpfm</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/09/02/gpfm/"/>
    <title>[Nature Biomedical Engineering] A generalizable pathology foundation model using a unified knowledge distillation pretraining framework</title>
    <published>2025-09-02T00:00:00+08:00</published>
    <updated>2025-09-02T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/09/02/gpfm/</uri>
    </author>
    <content type="html">&lt;p&gt;Recently, the SmartX Lab team, in collaboration with several leading institutions, has completed a groundbreaking project on the pathology foundation model. Published in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;Nature Biomedical Engineering&lt;/font&gt;&lt;/strong&gt;, this work presents a comprehensive evaluation of foundation models in pathology and introduces novel solutions to address current limitations.&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;Pathology, often referred to as the “gold standard” in cancer diagnosis, has seen significant advancements with the development of artificial intelligence (AI). However, existing models are often task-specific and lack the generalizability needed for complex clinical workflows. To address this limitation, a collaborative research team led by Professor Hao Chen at HKUST’s SmartX Lab, along with institutions such as Southern Medical University, Shanghai AI Lab, and others, has introduced the Generalizable Pathology Foundation Model (GPFM).&lt;/p&gt;

&lt;p&gt;GPFM represents a milestone in pathology AI, excelling across 72 clinical tasks spanning six major categories, including diagnosis, prognosis, and quality assurance (QA). The findings have been published in &lt;strong&gt;&lt;font color=&quot;red&quot;&gt;Nature Biomedical Engineering&lt;/font&gt;&lt;/strong&gt;, highlighting the potential for this model to address the limitations of current “specialized” AI systems.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_gpfm.png&quot; alt=&quot;GPFM&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;related-works&quot;&gt;Related Works&lt;/h2&gt;

&lt;p&gt;Existing pathology AI systems often excel in specific tasks but struggle with broader adaptability. These challenges include:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;Task-specificity: Models may achieve high accuracy in tissue classification but fail in survival prediction or report generation.&lt;/li&gt;
  &lt;li&gt;Generalization gap: Without a unified evaluation framework, the adaptability of models across diverse pathology tasks remains untested.&lt;/li&gt;
  &lt;li&gt;Complex clinical applications: The lack of a “comprehensive AI” hinders the deployment of AI into integrated clinical workflows.&lt;/li&gt;
  &lt;li&gt;To systematically evaluate these limitations, the team developed the first framework for generalizable pathology AI, assessing models on six task categories across &lt;strong&gt;72&lt;/strong&gt; clinical benchmarks. Existing models achieved an average rank of only 3.7, with the best model leading in just six tasks (see &lt;strong&gt;Figure 2&lt;/strong&gt;).&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;
&lt;p&gt;The success of GPFM lies in three critical components:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;
    &lt;p&gt;A Unified Pathology Evaluation Framework
The team introduced a comprehensive benchmark set to test the true generalizability of pathology models. This framework covers &lt;strong&gt;six&lt;/strong&gt; task types: whole-slide classification, survival analysis, pathology QA, ROI classification, pathology report generation, and pathology image retrieval.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;Knowledge Distillation
The researchers implemented a novel, dual-engine knowledge distillation framework:&lt;/p&gt;

    &lt;p&gt;&lt;em&gt;Expert Distillation&lt;/em&gt;: Leveraging the strengths of high-performing models such as UNI and Phikon to integrate specialized knowledge.              &lt;br /&gt;
 &lt;em&gt;Self-Distillation&lt;/em&gt;: Facilitating cross-scale alignment of tissue features, enhancing generalization across microscopic and macroscopic levels.                 &lt;br /&gt;
 &lt;em&gt;Large-Scale Pretraining&lt;/em&gt;: GPFM was pretrained on an extensive dataset of &lt;strong&gt;190&lt;/strong&gt; million image-level samples, sourced from over &lt;strong&gt;95,000&lt;/strong&gt; whole-slide images (WSIs) spanning &lt;strong&gt;34&lt;/strong&gt; tissue types. This unprecedented scale ensures the model’s robustness and adaptability to unseen clinical data.&lt;/p&gt;
  &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_gpfm_fig1.png&quot; alt=&quot;GPFM&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;clinical-validation&quot;&gt;Clinical Validation&lt;/h2&gt;
&lt;p&gt;The GPFM was comprehensively validated against state-of-the-art models across a diverse set of clinical scenarios, demonstrating significant improvements in accuracy and generalizability. In &lt;strong&gt;whole-slide image classification&lt;/strong&gt;, a critical diagnostic task, GPFM achieved an average rank of 1.22 across 36 tasks, outperforming the previous leader UNI (average rank: 3.60). The model also recorded a mean AUC of 0.891, surpassing UNI by 1.6% (P &amp;lt; 0.001).&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_gpfm_fig2.png&quot; alt=&quot;WSI Classification Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;For &lt;strong&gt;survival prediction&lt;/strong&gt;, which requires modeling complex prognostic data, GPFM held an average rank of 2.1 across 15 tasks, securing a top-2 position in 13 of them. It achieved a C-Index score of 0.665, representing a 3.4% improvement over UNI (P &amp;lt; 0.001).&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_gpfm_fig3.png&quot; alt=&quot;Survival Analysis Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;In the domain of &lt;strong&gt;ROI classification&lt;/strong&gt;, GPFM achieved the best average rank of 1.88 across 16 tasks, outperforming Prov-Gigapath (rank: 3.09), with the highest mean AUC of 0.946 (+0.2%, P &amp;lt; 0.001). Beyond these core tasks, GPFM demonstrated strong performance in additional areas such as pathology QA, image retrieval, and pathology report generation. These results showcase GPFM’s versatility and robustness, highlighting its potential for broad clinical application. For a detailed breakdown of these tasks, refer to the published study.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_gpfm_fig4.png&quot; alt=&quot;ROI Classification Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;translational-potential&quot;&gt;Translational Potential&lt;/h2&gt;
&lt;p&gt;Building on the capabilities of GPFM, the team developed &lt;strong&gt;SmartPath&lt;/strong&gt;, a next-generation diagnostic tool designed for intraoperative workflows. SmartPath is tailored to support diagnosis in five high-incidence cancers, including lung, breast, and gastrointestinal cancers.&lt;/p&gt;

&lt;p&gt;Currently under active deployment, SmartPath aims to accelerate the adoption of digital pathology by improving diagnostic accuracy and streamlining clinical workflows.&lt;/p&gt;

&lt;video controls=&quot;&quot; style=&quot;width: 100%; height: auto;&quot;&gt;
  &lt;source src=&quot;/static/media/SmartPath.mp4&quot; type=&quot;video/mp4&quot; /&gt;
  Your browser does not support the video tag.
&lt;/video&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;resources&quot;&gt;Resources&lt;/h2&gt;

&lt;p&gt;For more details, please see our paper &lt;a href=&quot;https://www.nature.com/articles/s41551-025-01488-4&quot;&gt;A generalizable pathology foundation model using a unified knowledge distillation pretraining framework&lt;/a&gt; via Nature Biomedical Engineering.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Citation&lt;/strong&gt;:&lt;br /&gt;
Ma, J., Guo, Z., Zhou, F. et al. A generalizable pathology foundation model using a unified knowledge distillation pretraining framework. Nat. Biomed. Eng (2025). https://doi.org/10.1038/s41551-025-01488-4&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/08/09/survey</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/08/09/survey/"/>
    <title>[IJCAI 2025] A Survey of Pathology Foundation Models: Recent Advances and Future Directions</title>
    <published>2025-08-09T00:00:00+08:00</published>
    <updated>2025-08-09T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/08/09/survey/</uri>
    </author>
    <content type="html">&lt;p&gt;A collaborative work by &lt;strong&gt;The Chinese University of Hong Kong&lt;/strong&gt;, &lt;strong&gt;SmartX Lab&lt;/strong&gt;, and &lt;strong&gt;Nanyang Technological University&lt;/strong&gt; titled &lt;strong&gt;“A Survey of Pathology Foundation Models: Recent Advances and Future Directions”&lt;/strong&gt; has been accepted at &lt;strong&gt;IJCAI 2025&lt;/strong&gt;, a premier conference in artificial intelligence. This survey offers a comprehensive and systematic overview of &lt;strong&gt;Pathology Foundation Models (PFMs)&lt;/strong&gt;, a transformative direction in &lt;strong&gt;computational pathology&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;This work delivers the &lt;strong&gt;first hierarchical classification framework&lt;/strong&gt; for pathology foundation models (PFMs), establishes a &lt;strong&gt;systematic evaluation benchmark&lt;/strong&gt;, and delineates &lt;strong&gt;critical technical challenges and future research priorities&lt;/strong&gt; for advancing the field.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;background&quot;&gt;Background&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Computational pathology (CPath)&lt;/strong&gt; enables &lt;strong&gt;AI-powered analysis of whole slide images (WSIs)&lt;/strong&gt; for disease diagnosis and prognosis. WSIs are gigapixel pathology images, and the dominant analysis framework is &lt;strong&gt;Multiple Instance Learning (MIL)&lt;/strong&gt;. In MIL, a WSI is divided into smaller &lt;strong&gt;image patches&lt;/strong&gt;, from which a &lt;strong&gt;feature extractor&lt;/strong&gt; produces embeddings that are subsequently aggregated by an &lt;strong&gt;aggregator&lt;/strong&gt; to yield slide-level predictions.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_survey_fig1.png&quot; alt=&quot;Prompt Comparison&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Historically, models like &lt;strong&gt;ImageNet-pretrained ResNet-50&lt;/strong&gt; served as extractors, but these suffer from &lt;strong&gt;domain mismatch&lt;/strong&gt;, failing to fully capture pathology-specific features such as subtle staining patterns or hierarchical tissue structures. &lt;strong&gt;Pathology Foundation Models (PFMs)&lt;/strong&gt;—large-scale pathology-pretrained networks, often using &lt;strong&gt;self-supervised learning (SSL)&lt;/strong&gt;—address this gap, enabling robust morphological representation and better performance on downstream tasks.&lt;/p&gt;

&lt;p&gt;Yet, despite promising results, PFMs still face &lt;strong&gt;unique challenges&lt;/strong&gt; in development, scalability, and deployment.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;hierarchical-taxonomy-of-pfms&quot;&gt;Hierarchical Taxonomy of PFMs&lt;/h2&gt;

&lt;p&gt;Our proposed taxonomy systematically organizes PFMs along three key dimensions using a &lt;strong&gt;top-down analytical framework&lt;/strong&gt;:&lt;br /&gt;
(1) &lt;strong&gt;Model scope&lt;/strong&gt;, which categorizes PFMs by their functional emphasis—focusing primarily on feature extraction, feature aggregation, or joint optimization of both;&lt;br /&gt;
(2) &lt;strong&gt;Model pretraining&lt;/strong&gt;, which dissects image-centric pretraining methodologies at the slide, patch, and multimodal levels;&lt;br /&gt;
(3) &lt;strong&gt;Model architecture&lt;/strong&gt;, which classifies PFMs by parameter scale and structural complexity.&lt;br /&gt;
This framework enables comprehensive, consistent comparison across PFMs.&lt;/p&gt;

&lt;h3 id=&quot;model-scope&quot;&gt;Model Scope&lt;/h3&gt;

&lt;p&gt;The standard MIL workflow for WSIs involves three stages: patch extraction, feature extraction, and feature aggregation. Patch extraction is already mature; as a result, a model’s overall performance depends critically on the &lt;strong&gt;quality of its extractor and aggregator&lt;/strong&gt;. WSIs themselves have &lt;strong&gt;hierarchical tissue organization&lt;/strong&gt;—the extractor captures local morphological details, while the aggregator models global structural patterns. The synergy between these two components largely determines diagnostic accuracy. Based on functional emphasis, PFMs can be divided into three categories:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Extractor-Oriented PFMs&lt;/strong&gt; form the current research mainstream. Their popularity is driven by two factors:&lt;br /&gt;
(1) The critical role of high-quality feature extraction;&lt;br /&gt;
(2) The urgent need to address domain adaptation challenges when transferring ImageNet-pretrained CNNs to pathology.&lt;br /&gt;
This design philosophy mirrors clinical workflows, where pathologists rely on fine-grained cellular features at the patch level to make diagnoses. For instance, &lt;strong&gt;CTransPath&lt;/strong&gt; pioneered a semantic-based contrastive learning method to train a CNN–Transformer hybrid extractor on 15 million patches. &lt;strong&gt;REMEDIS&lt;/strong&gt; demonstrated cross-domain limitations of ResNet-50 across medical imaging modalities, reinforcing the necessity for pathology-specific extractors—a point further validated by later works such as &lt;strong&gt;Virchow&lt;/strong&gt; and &lt;strong&gt;SINAI&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Aggregator-Oriented PFMs&lt;/strong&gt; focus on the slide-level representation. Since the aggregator is the only MIL component trained with direct supervision from true WSI labels, it plays a pivotal role in slide-level tasks. Yet, research in this direction remains limited. &lt;strong&gt;CHIEF&lt;/strong&gt; was the first to show the value of aggregator pretraining, using anatomical-site supervision to create site-aware aggregation. More recent works like &lt;strong&gt;MADELEINE&lt;/strong&gt;, &lt;strong&gt;TITAN&lt;/strong&gt;, and &lt;strong&gt;THREAD&lt;/strong&gt; employed multimodal data to pretrain aggregators while keeping patch features frozen—improving performance in data-constrained scenarios. This reflects an increasing recognition of the aggregator’s importance, consistent with transfer learning principles: pretraining on large datasets alleviates downstream data scarcity. However, results from &lt;strong&gt;CHIEF&lt;/strong&gt; showed that its pretrained aggregator could sometimes underperform simple linear probes on extractors. Possible reasons include small pretraining model sizes or domain bias conflicting with generalized features—indicating the need for further study on large-scale aggregator pretraining.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Hybrid Optimization PFMs&lt;/strong&gt; pretrain both extractor and aggregator, aiming to maximize synergy between these components. &lt;strong&gt;HIPT&lt;/strong&gt; pioneered this by hierarchically pretraining the first two layers of the extractor (excluding the final layer), yielding significant performance gains. &lt;strong&gt;Prov-GigaPath&lt;/strong&gt; followed a similar strategy, pretraining a ViT extractor alongside a LongNet slide encoder; however, LongNet produced instance-level features rather than a single slide-level embedding, requiring additional pooling (e.g., ABMIL or non-parametric methods) for classification. &lt;strong&gt;TANGLE&lt;/strong&gt; pretrained a ViT extractor and an ABMIL aggregator guided by transcriptomic data. &lt;strong&gt;mSTAR&lt;/strong&gt; introduced a reverse pipeline: pretraining a multimodal aggregator first, then using it to pretrain the extractor, achieving a fully pretrained hybrid design.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_survey_fig2.png&quot; alt=&quot;Prompt Comparison&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;From recent trends, two key insights emerge:&lt;br /&gt;
First, research focus is shifting from extractor pretraining toward aggregator pretraining—a likely consequence of extractor performance plateauing and the growing appreciation of aggregator importance, especially in limited-data settings.&lt;br /&gt;
Second, aggregators are increasingly &lt;strong&gt;hierarchically dependent&lt;/strong&gt;: later models often reuse earlier ones’ capabilities. For example, &lt;strong&gt;TITAN&lt;/strong&gt; builds on features from &lt;strong&gt;CONCHv1.5&lt;/strong&gt;, which itself is based on &lt;strong&gt;UNI&lt;/strong&gt;—forming a dependency chain where TITAN’s performance rests on CONCHv1.5, which in turn depends on UNI.&lt;/p&gt;

&lt;h3 id=&quot;model-pretraining&quot;&gt;Model Pretraining&lt;/h3&gt;

&lt;p&gt;PFM pretraining methods fall into &lt;strong&gt;supervised&lt;/strong&gt; and &lt;strong&gt;self-supervised learning (SSL)&lt;/strong&gt;. SSL dominates due to its ability to extract rich morphological representations without manual labels; in fact, &lt;strong&gt;CHIEF&lt;/strong&gt; is the only surveyed PFM to use supervised aggregator pretraining. SSL approaches can be grouped into two directions:&lt;br /&gt;
&lt;strong&gt;Pure vision methods&lt;/strong&gt; (contrastive learning, masked image modeling, self-distillation) and &lt;strong&gt;cross-modal methods&lt;/strong&gt; (multimodal alignment).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Contrastive Learning&lt;/strong&gt; aims to minimize the distance between positive sample pairs and maximize separation between negative pairs. Key developments include:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;SimCLR&lt;/strong&gt;: established strong data augmentation and large-batch training regimes;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;MoCo v3&lt;/strong&gt;: stabilized ViT performance in SSL;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;CLIP&lt;/strong&gt;: extended contrastive alignment to image–text pairs;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;CoCa&lt;/strong&gt;: unified contrastive pretraining with text generation.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_survey_fig3.png&quot; alt=&quot;Prompt Comparison&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;In pathology, &lt;strong&gt;REMEDIS&lt;/strong&gt; used SimCLR to enhance robustness and data efficiency; &lt;strong&gt;TANGLE&lt;/strong&gt; augmented SimCLR with gene expression reconstruction and intra-slide patch alignment; &lt;strong&gt;Pathoduet&lt;/strong&gt; extended MoCo v3 with cross-scale localization and stain-transfer tasks to tackle stain variability and tissue heterogeneity. CLIP’s multimodal alignment has been widely deployed—not only in extractors (PLIP) but also in aggregators (&lt;strong&gt;Prov-GigaPath&lt;/strong&gt;, &lt;strong&gt;mSTAR&lt;/strong&gt;, &lt;strong&gt;MADELEINE&lt;/strong&gt;, &lt;strong&gt;THREAD&lt;/strong&gt;). &lt;strong&gt;KEEP&lt;/strong&gt; improved CLIP-based extractors by integrating curated knowledge graphs to clean image–text pairs. CoCa-based frameworks have been adopted for both extractors (&lt;strong&gt;CONCH&lt;/strong&gt;, trained on 1.17M image–caption pairs) and aggregators (&lt;strong&gt;PRISM&lt;/strong&gt;, &lt;strong&gt;TITAN&lt;/strong&gt;).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Masked Image Modeling (MIM)&lt;/strong&gt; predicts masked image regions to learn context-aware features. While &lt;strong&gt;SimMIM&lt;/strong&gt; simplified design with random masking and lightweight decoders, &lt;strong&gt;MAE&lt;/strong&gt; employed heavy masking with asymmetric encoder–decoder architectures. Pathology models like &lt;strong&gt;SINAI&lt;/strong&gt; (pretrained a ViT on 3.2B patches) have demonstrated MIM’s scalability; follow-ups like &lt;strong&gt;MUSK&lt;/strong&gt; (BEiT-3) and &lt;strong&gt;BEPH&lt;/strong&gt; (BEiTv2) confirmed its utility. &lt;strong&gt;Prov-GigaPath&lt;/strong&gt; further showed that MIM benefits aggregator pretraining—applying MAE to its LongNet encoder.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Self-Distillation&lt;/strong&gt; enables models to learn from their own predictions, often combining “teacher” and “student” networks. &lt;strong&gt;DINO&lt;/strong&gt; introduced the momentum encoder with multi-crop training; &lt;strong&gt;iBOT&lt;/strong&gt; integrated patch-level masking into self-distillation; &lt;strong&gt;DINOv2&lt;/strong&gt; improved scale stability. In pathology:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Phikon&lt;/strong&gt; applied iBOT to 43M patches from 16 tumor types;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Phikon-v2&lt;/strong&gt; used DINOv2 across 456M patches;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;RudolfV&lt;/strong&gt; merged DINOv2 with pathologist knowledge across 58 tissue types;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Hibou&lt;/strong&gt; scaled DINOv2 to 1.2B patches.&lt;br /&gt;
Self-distillation has also been used for aggregators—&lt;strong&gt;TITAN&lt;/strong&gt; applied iBOT for generic aggregator learning. Domain-specific refinements include &lt;strong&gt;PLUTO&lt;/strong&gt; (combining DINOv2, MAE, and Fourier loss on 195M patches), &lt;strong&gt;GPFM&lt;/strong&gt; (a unified knowledge distillation framework mixing MIM, self-distillation, and expert knowledge), and &lt;strong&gt;Virchow2&lt;/strong&gt; (pathology-specific augmentations and redundancy reduction for DINOv2).&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;model-architecture&quot;&gt;Model Architecture&lt;/h3&gt;

&lt;p&gt;PFM architecture design hinges on &lt;strong&gt;backbone choice, parameter count, and scale&lt;/strong&gt;. Scale is primarily determined by parameter count; the authors introduce a &lt;strong&gt;ViT-based size taxonomy&lt;/strong&gt; to standardize scale comparisons including the following sizes: &lt;strong&gt;XS&lt;/strong&gt; (2.78M params), &lt;strong&gt;S&lt;/strong&gt; (21.7M), &lt;strong&gt;B&lt;/strong&gt; (86.3M), &lt;strong&gt;L&lt;/strong&gt; (307M), &lt;strong&gt;H&lt;/strong&gt; (632M), &lt;strong&gt;g&lt;/strong&gt; (1.13B), and &lt;strong&gt;G&lt;/strong&gt; (1.9B).&lt;/p&gt;

&lt;p&gt;Extractor architectures include CNNs (ResNet) and Transformers (ViT, Swin, BEiT/BEiTv2, FlexiViT, multimodal BEiT-3). Aggregator architectures are dominated by ABMIL variants and Perceiver models.&lt;/p&gt;

&lt;p&gt;Analysis reveals notable patterns:&lt;/p&gt;
&lt;ol&gt;
  &lt;li&gt;&lt;strong&gt;ABMIL&lt;/strong&gt; is the most widely used aggregator, while extractors are predominantly Transformer-based.&lt;/li&gt;
  &lt;li&gt;ViT-L is the most common backbone, with smaller variants (ViT-S, ViT-B) used for efficiency.&lt;/li&gt;
  &lt;li&gt;Extractors often have significantly more parameters than aggregators (e.g., ViT-L extractors vs. ViT-XS aggregators), leading to &lt;strong&gt;data–model scale mismatches&lt;/strong&gt;.&lt;/li&gt;
  &lt;li&gt;Model sizes are steadily increasing—from ViT-B in early PFMs to ViT-L/H/g/G in recent designs—highlighting an ongoing scale-up trend despite computational costs.&lt;/li&gt;
&lt;/ol&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;evaluation-framework&quot;&gt;Evaluation Framework&lt;/h2&gt;

&lt;p&gt;The survey structures &lt;strong&gt;PFM evaluation tasks&lt;/strong&gt; into four broad categories:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;&lt;strong&gt;Slide-Level&lt;/strong&gt;: WSI classification, survival prediction, retrieval, segmentation.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Patch-Level&lt;/strong&gt;: Patch classification, retrieval, segmentation—testing extractors directly.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Multimodal&lt;/strong&gt;: Cross-modal retrieval, report generation, visual question answering.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Biological&lt;/strong&gt;: Gene mutation detection, molecular subtype classification.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Leading models such as &lt;strong&gt;CONCH, UNI, GPFM, TITAN&lt;/strong&gt; provide &lt;strong&gt;comprehensive multi-setting benchmarks&lt;/strong&gt;, enabling evaluation across &lt;strong&gt;zero-shot, few-shot, and fully supervised scenarios&lt;/strong&gt;.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;key-challenges--future-directions&quot;&gt;Key Challenges &amp;amp; Future Directions&lt;/h2&gt;

&lt;p&gt;While PFMs have made substantial strides, several open problems must be addressed to unlock their full research and clinical impact. These challenges span from &lt;strong&gt;domain-specific learning&lt;/strong&gt; to &lt;strong&gt;deployment infrastructure&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Pathology-Specific Pretraining&lt;/strong&gt;&lt;br /&gt;
Current PFMs largely rely on techniques designed for natural image analysis, insufficiently tailored to the unique textures, staining variations, and hierarchical tissue organization in pathology. There is a pressing need for &lt;strong&gt;domain-specialized algorithms&lt;/strong&gt;, designed for both single-modality (WSI) and multi-modality (WSI + reports + omics) contexts, to capture nuanced visual–biological correlations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. End-to-End WSI Learning&lt;/strong&gt;&lt;br /&gt;
The two-stage extractor–aggregator paradigm often leads to misalignment between local patch and global WSI representations. Future systems must achieve &lt;strong&gt;fully integrated gigapixel-scale training&lt;/strong&gt;, simultaneously optimizing both components while efficiently handling the massive computational and memory requirements.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Data–Model Scalability&lt;/strong&gt;&lt;br /&gt;
Scaling trends for PFMs—considering dataset size, patch count, model capacity, and data diversity—are not yet fully understood. Research must determine optimal scaling strategies, develop adaptive architectures, and improve curation pipelines to ensure both &lt;strong&gt;quality and diversity&lt;/strong&gt; of training data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Efficient Federated Learning&lt;/strong&gt;&lt;br /&gt;
Training with multi-institutional datasets without data sharing is essential but computationally challenging, especially in &lt;strong&gt;non-IID&lt;/strong&gt; settings. Innovations in algorithm design are needed to reduce communication overhead, improve training stability, and preserve patient privacy, enabling large-scale collaborative learning at lower cost.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5. Robustness Across Institutions&lt;/strong&gt;&lt;br /&gt;
Performance degradation due to institutional differences—scanner types, staining protocols, imaging resolutions—remains a critical obstacle. Approaches to &lt;strong&gt;domain generalization, adaptive normalization, and style-invariant modeling&lt;/strong&gt; will be necessary to ensure consistent diagnostic performance across diverse clinical environments.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;6. Retrieval-Augmented Pathology Language Models&lt;/strong&gt;&lt;br /&gt;
Integrating &lt;strong&gt;retrieval-augmented generation (RAG)&lt;/strong&gt; with PFMs could enrich decision support by combining visual analysis with retrieved pathology-specific knowledge bases, such as atlases or literature. This multimodal integration could yield &lt;strong&gt;explainable, context-aware AI tools&lt;/strong&gt; for clinical pathology.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;7. Model Adaptation &amp;amp; Maintenance&lt;/strong&gt;&lt;br /&gt;
Given the high cost of full retraining, PFMs must be &lt;strong&gt;maintainable and updatable&lt;/strong&gt; as new diseases emerge, staining protocols evolve, or institutional practices change. Lightweight continual learning and parameter-efficient tuning strategies will be key to maintaining clinical relevance.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;This IJCAI 2025 survey delivers a &lt;strong&gt;structured taxonomy&lt;/strong&gt;, a &lt;strong&gt;comprehensive evaluation framework&lt;/strong&gt;, and a &lt;strong&gt;forward-looking research roadmap&lt;/strong&gt; for pathology foundation models. By addressing the outlined challenges, the authors envision PFMs progressing toward &lt;strong&gt;greater generalization, clinical reliability, and theoretical advancement&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Here are the links to the paper and the GitHub repository:
&lt;a href=&quot;https://arxiv.org/abs/2504.04045&quot;&gt;Read the full paper on arXiv&lt;/a&gt;&lt;br /&gt;
&lt;a href=&quot;https://github.com/BearCleverProud/AwesomeWSI&quot;&gt;Explore resources on GitHub&lt;/a&gt;&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/07/29/smartcare</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/07/29/smartcare/"/>
    <title>[Press Release] SmartLab at HKUST Launches SmartCare AI Medical Platform to Transform Patient-Centered Healthcare</title>
    <published>2025-07-29T00:00:00+08:00</published>
    <updated>2025-07-29T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/07/29/smartcare/</uri>
    </author>
    <content type="html">&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartcare_cover.png&quot; alt=&quot;SmartCare AI Platform Banner&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;SmartX Lab&lt;/strong&gt; at the &lt;strong&gt;Hong Kong University of Science and Technology (HKUST)&lt;/strong&gt;, led by Assistant Professor Hao Chen, has officially launched the &lt;strong&gt;SmartCare AI Medical Platform&lt;/strong&gt; at the HKUST campus clinic. This milestone marks a new era in patient-centered healthcare, harnessing the power of artificial intelligence to transform the entire patient journey—from triage to treatment—while streamlining clinical workflows for healthcare professionals.&lt;/p&gt;

&lt;h3 id=&quot;pilot-program-real-world-impact-at-hkust-clinic&quot;&gt;Pilot Program: Real-World Impact at HKUST Clinic&lt;/h3&gt;

&lt;p&gt;A six-month pilot study of SmartCare is currently underway at the HKUST Clinic, inviting over 15,000 students, faculty, and staff to participate and provide feedback. The pilot epitomizes SmartX Lab’s commitment to translating cutting-edge AI research into practical solutions that directly benefit both patients and clinicians.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartcare_trial.png&quot; alt=&quot;SmartX Lab SmartCare Team at HKUST Clinic&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;core-features-and-innovation&quot;&gt;Core Features and Innovation&lt;/h3&gt;

&lt;p&gt;SmartCare is built on HKUST’s MedDr—one of the world’s largest open-source, generalist multimodal foundation models for medicine. The platform is designed to:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Streamline clinical workflows&lt;/strong&gt;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Enhance patient-provider interactions&lt;/strong&gt;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Support next-generation medical education&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Key features include:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Intelligent Pre-visit Triage:&lt;/strong&gt; Patients scan a QR code to submit medical histories and symptoms via text or voice, even before entering the clinic. The AI-powered triage system helps nurses optimize resource allocation and patient flow, significantly reducing waiting times.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;AI-Powered Consultation Assistant:&lt;/strong&gt; The platform supports real-time, multilingual transcription (Cantonese, Mandarin, English), enabling doctors to focus on patients instead of manual note-taking, and fostering personalized care.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Automated Medical Documentation:&lt;/strong&gt; SmartCare auto-generates over 30 types of medical documents—including referral letters, prescriptions, and medical certificates—and provides AI-powered post-visit follow-up, reducing administrative burden for clinicians.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Virtual Patient Module:&lt;/strong&gt; This patent-pending feature enables simulated consultations with instant AI feedback, paving the way for innovative medical education and training.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;built-on-world-class-team-research-meddr2-and-gsco&quot;&gt;Built on World-Class Team Research: MedDr2 and GSCo&lt;/h3&gt;

&lt;p&gt;SmartCare is &lt;strong&gt;powered by MedDr2&lt;/strong&gt;—the world’s largest open-source, generalist multimodal foundation model for medicine—&lt;strong&gt;developed by the SmartX Lab team based on the innovative &lt;a href=&quot;https://arxiv.org/html/2404.15127v2&quot;&gt;GSCo&lt;/a&gt; framework&lt;/strong&gt;. This foundational work was led by &lt;a href=&quot;https://sunanhe.github.io/&quot;&gt;Sunan He&lt;/a&gt; and &lt;a href=&quot;https://jerrrynie.github.io/&quot;&gt;Yuxiang Nie&lt;/a&gt;, supervised by Prof. Hao Chen, with a team of outstanding collaborators, and sets a new paradigm for trustworthy AI in healthcare.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartcare_meddr.png&quot; alt=&quot;SmartX Lab SmartCare Introduction&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;MedDr2&lt;/strong&gt; is built upon the &lt;strong&gt;GSCo framework&lt;/strong&gt;, which enables seamless integration of both &lt;strong&gt;generalist foundation models (GFMs)&lt;/strong&gt; and &lt;strong&gt;specialist models&lt;/strong&gt;. This collaborative approach allows MedDr2 to deliver robust, precise, and scalable medical image analysis, demonstrating exceptional instruction-following, in-context learning, and generalization across diverse medical domains. GSCo introduces advanced collaborative mechanisms—including &lt;strong&gt;Mixture-of-Expert Diagnosis (MoED)&lt;/strong&gt; and &lt;strong&gt;Retrieval-Augmented Diagnosis (RAD)&lt;/strong&gt;—empowering MedDr2 to incorporate specialist knowledge for accurate diagnosis, even in out-of-domain or complex scenarios. This &lt;strong&gt;team-driven innovation&lt;/strong&gt; is the scientific foundation of SmartCare’s advanced capabilities in clinical workflow optimization, intelligent triage, and automated documentation, ensuring the platform is both cutting-edge and clinically impactful.&lt;/p&gt;

&lt;h3 id=&quot;leadership-voices&quot;&gt;Leadership Voices&lt;/h3&gt;

&lt;p&gt;Prof. Hao Chen, Director of SmartX Lab and Assistant Professor at HKUST, remarked:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“SmartCare is the culmination of SmartX Lab’s mission to develop trustworthy, real-world AI for healthcare. Leveraging our MedDr foundation model, SmartCare excels at complex medical queries, medical image interpretation, and seamless clinical integration. Following our campus pilot, we are exploring collaborations with Gleneagles Hospital Hong Kong and CUHK Medical Centre for further real-world validation and broader impact. HKUST’s AI innovation ecosystem empowers SmartCare to redefine clinical efficiency and patient outcomes.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Dr. Justin Cheng, SmartCare’s CEO, Co-Founder, and Registered Medical Practitioner at HKUST Clinic, added:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“SmartCare fundamentally shifts the clinical paradigm. By automating administrative tasks and documentation, SmartCare allows doctors to focus on the human connection at the heart of medicine, delivering truly personalized and high-quality care. As both a clinician and researcher, I am excited by the transformative impact SmartCare brings to daily practice.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Prof. Samuel Yu Chung-Toi, Director of the Health, Safety, and Environment Office at HKUST, commented:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“The SmartCare pilot at our campus clinic exemplifies HKUST’s commitment to impactful, cross-disciplinary innovation. By providing SmartX Lab with resources and an environment for real-world testing, we ensure our solutions address the needs of both healthcare providers and patients.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_smartcare_speech.png&quot; alt=&quot;SmartCare Launch at HKUST Clinic&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;foundation-and-partnership&quot;&gt;Foundation and Partnership&lt;/h3&gt;

&lt;p&gt;SmartCare’s development is rooted in SmartX Lab’s research excellence and HKUST’s vibrant AI innovation ecosystem. In May 2025, SmartX Lab facilitated a tripartite partnership to integrate SmartCare’s platform with PanopticAI’s camera-based vital sign monitoring—another HKUST-nurtured innovation—at a new ambulatory care center in Admiralty, further enhancing operational efficiency and service quality.&lt;/p&gt;

&lt;h3 id=&quot;leadership-and-research-excellence&quot;&gt;Leadership and Research Excellence&lt;/h3&gt;

&lt;p&gt;The SmartCare project is led by Prof. Hao Chen (Director, SmartX Lab; Assistant Professor, Computer Science &amp;amp; Engineering, HKUST), Dr. Justin Cheng (CEO &amp;amp; Co-Founder, SmartCare; Registered Medical Practitioner, HKUST Clinic), and Prof. Samuel Yu Chung-Toi (Director, Health, Safety, and Environment Office, HKUST). The team is dedicated to advancing AI medical foundation models and next-generation clinical technologies, translating scientific breakthroughs into tangible healthcare benefits.&lt;/p&gt;

&lt;hr /&gt;

&lt;h3 id=&quot;about-smartx-lab&quot;&gt;About SmartX Lab&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;SmartX Lab&lt;/strong&gt;, led by Prof. Hao Chen at HKUST, is dedicated to advancing trustworthy AI technologies for healthcare and science. Our research spans large-scale models for healthcare, computer-assisted intervention, AI for science, and bioinformatics. Our mission is to drive a transformative revolution in medical practice and scientific discovery, shaping a healthier future.&lt;/p&gt;

&lt;h3 id=&quot;about-the-hong-kong-university-of-science-and-technology&quot;&gt;About The Hong Kong University of Science and Technology&lt;/h3&gt;

&lt;p&gt;&lt;a href=&quot;https://hkust.edu.hk/&quot;&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/a&gt; is a world-class university driving innovative education, research excellence, and impactful knowledge transfer. With a holistic and interdisciplinary approach, HKUST was ranked 3rd in the Times Higher Education’s Young University Rankings 2024 and 19th worldwide in the THE Impact Rankings 2025. Thirteen HKUST subjects were ranked among the world’s top 50 in the QS World University Rankings by Subject 2025, with “Data Science and Artificial Intelligence” ranked 17th globally and first in Hong Kong. Over 80% of our research was rated “internationally excellent” or “world leading” in Hong Kong’s latest Research Assessment Exercise. As of May 2025, HKUST members have founded over 1,800 active start-ups, including 10 Unicorns and 17 exits (IPO or M&amp;amp;A).&lt;/p&gt;

&lt;h3 id=&quot;media-links&quot;&gt;Media Links&lt;/h3&gt;

&lt;p&gt;RTHK: &lt;a href=&quot;https://news.rthk.hk/rthk/ch/component/k2/1815646-20250729.htm?spTabChangeable=0&quot;&gt;Link (Traditional Chinese only)&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hong Kong Economic Journal: &lt;a href=&quot;https://www.hkej.com/instantnews/current/article/4149132/%E7%A7%91%E5%A4%A7%E7%A0%94%E7%99%BCSmartCare%E6%99%BA%E8%83%BD%E9%86%AB%E7%99%82%E5%B9%B3%E5%8F%B0+%E6%8F%90%E5%8D%87%E5%B0%B1%E8%A8%BA%E6%95%88%E7%8E%87&quot;&gt;Link (Traditional Chinese only)&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;HKET &lt;a href=&quot;https://news.hket.com/article/3985417/%E4%BA%BA%E5%B7%A5%E6%99%BA%E8%83%BD%E9%86%AB%E7%99%82%E5%B9%B3%E5%8F%B0%EF%BD%9C%E7%A7%91%E5%A4%A7%E6%8E%A8%E4%BA%BA%E5%B7%A5%E6%99%BA%E8%83%BD%E9%86%AB%E7%99%82%E5%B9%B3%E5%8F%B0%20SmartCare%20%E7%B3%BB%E7%B5%B1%E3%80%80%E9%86%AB%E7%94%9F%EF%BC%9A%E8%A8%BA%E6%96%B7%E6%99%82%E9%96%93%E5%8A%A0%E5%BF%AB%E4%BA%86%E4%B8%89%E5%88%86%E4%B8%80?mtc=20023&quot;&gt;Link (Traditional Chinese only)&lt;/a&gt;&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/07/27/colab</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/07/27/colab/"/>
    <title>[Press Release] HKUST and Wuhan Union Hospital Jointly Establish the Medical-Engineering Joint Innovation Center</title>
    <published>2025-07-27T00:00:00+08:00</published>
    <updated>2025-07-27T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/07/27/colab/</uri>
    </author>
    <content type="html">&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_union_medeng_center_cover.png&quot; alt=&quot;HKUST &amp;amp; Wuhan Union Hospital Joint Innovation Center Banner&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;The &lt;strong&gt;Hong Kong University of Science and Technology (HKUST)&lt;/strong&gt; and &lt;strong&gt;Wuhan Union Hospital, Tongji Medical College of Huazhong University of Science and Technology&lt;/strong&gt; have officially inaugurated the &lt;strong&gt;Medical-Engineering Joint Innovation Center&lt;/strong&gt; at HKUST. This milestone marks a new era of interdisciplinary collaboration, combining the strengths of engineering, artificial intelligence (AI), and clinical medicine to accelerate medical technology innovation for global healthcare advancement.&lt;/p&gt;

&lt;h3 id=&quot;launch-ceremony-and-strategic-collaboration&quot;&gt;Launch Ceremony and Strategic Collaboration&lt;/h3&gt;

&lt;p&gt;The unveiling ceremony was attended by Prof. Zhang Yu, Party Secretary of Wuhan Union Hospital, Prof. Wang Hongbo, Deputy Secretary, and their delegation, alongside HKUST representatives including Prof. Kwang-Ting Cheng, Vice-President for Research and Development, and Prof. Chung Chi Shing, Associate Dean of Engineering. Distinguished guests toured the HKUST Neumann Institute and engaged in in-depth discussions with HKUST scholars regarding the center’s development plans, joint innovation funding, and collaborative training models for medical and graduate students.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_union_medeng_center_team.png&quot; alt=&quot;Delegations from HKUST and Wuhan Union Hospital at the inauguration&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;core-focus-and-innovation-goals&quot;&gt;Core Focus and Innovation Goals&lt;/h3&gt;

&lt;p&gt;According to the collaboration agreement, the Joint Innovation Center will focus on interdisciplinary medical-engineering research, prioritizing AI-driven medical imaging, ultrasound, radiology, and pathology. Through academic exchanges, faculty training, joint research projects, and translational initiatives, the center aims to accelerate the development of large multimodal medical AI models, intelligent diagnosis systems, precision medicine, and smart medical devices.&lt;/p&gt;

&lt;p&gt;Prof. Hao Chen, Co-Director of the Center, led guests on a visit to the HKUST Neumann Institute, showcasing the university’s cutting-edge research infrastructure.&lt;/p&gt;

&lt;h3 id=&quot;leadership-voices&quot;&gt;Leadership Voices&lt;/h3&gt;

&lt;p&gt;Prof. Kwang-Ting Cheng, HKUST Vice-President for Research and Development, remarked:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“The rapid advancement of artificial intelligence is a key driver of medical innovation. The deep integration of medicine, engineering, and data science opens new frontiers in disease prevention, diagnosis, monitoring, and treatment. HKUST’s strengths in data fusion, algorithm innovation, and high-performance computing, combined with Wuhan Union Hospital’s extensive clinical resources, will catalyze breakthroughs in AI-powered healthcare. This collaboration highlights the vast potential of cross-disciplinary and cross-regional partnerships, accelerating the translation of intelligent diagnostics and precision medicine to real-world clinical impact.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Prof. Zhang Yu, Party Secretary of Wuhan Union Hospital, commented:&lt;/p&gt;
&lt;blockquote&gt;
  &lt;p&gt;“The accelerating technological revolution and industrial transformation make medical-engineering integration one of the most promising directions for innovation. As a leading national medical institution, Wuhan Union Hospital offers rich clinical experience and comprehensive resources. This strategic partnership leverages complementary strengths and shared resources, advancing medical science and technology. We look forward to deepening collaboration, harnessing AI for medical progress, and contributing to the Healthy China initiative.”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hkust_union_medeng_center_gift.png&quot; alt=&quot;Exchange of commemorative gifts between Prof. Kwang-Ting Cheng and Prof. Zhang Yu&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;h3 id=&quot;foundation-of-the-partnership&quot;&gt;Foundation of the Partnership&lt;/h3&gt;

&lt;p&gt;Wuhan Union Hospital is one of China’s premier medical institutions, serving 6.9 million outpatients, 420,000 inpatients, and performing 156,000 surgeries annually—providing vital clinical support for large-scale, multimodal medical AI model development.&lt;/p&gt;

&lt;p&gt;The collaboration between HKUST and Wuhan Union Hospital began in 2023, with the first official visit by Prof. Zhang Yu. In 2024, both parties signed a formal agreement and established the first base of the Medical-Engineering Joint Innovation Center at Wuhan Union Hospital. Since then, the partnership has achieved breakthrough progress in AI and medical imaging, including the development of the world’s largest ultrasound AI model trained on data from nearly one million patients, achieving diagnostic accuracy for thyroid nodules on par with experienced clinicians.&lt;/p&gt;

&lt;h3 id=&quot;leadership-and-research-excellence&quot;&gt;Leadership and Research Excellence&lt;/h3&gt;

&lt;p&gt;The Center is jointly led by Prof. Hao Chen (Assistant Professor, Computer Science &amp;amp; Engineering, Chemical &amp;amp; Biological Engineering, and Life Science, HKUST) and Prof. Zheng Chuansheng (Director of Radiology, Wuhan Union Hospital), with Prof. Huang Ziwei (Associate Professor, Chemical &amp;amp; Biological Engineering, HKUST) and Prof. Xie Mingxing (Director of Ultrasound, Wuhan Union Hospital) as Deputy Directors.&lt;/p&gt;

&lt;p&gt;Prof. Chen and Prof. Huang’s teams have recently developed four novel AI medical foundation models and an innovative cellular imaging technology capable of accurately detecting residual cancer cells, driving medical innovation and translating scientific breakthroughs into clinical benefit.&lt;/p&gt;

&lt;hr /&gt;

&lt;h3 id=&quot;about-smartx-lab&quot;&gt;About SmartX Lab&lt;/h3&gt;

&lt;p&gt;SmartX Lab, led by Prof. Hao Chen, is committed to pushing the boundaries of trustworthy AI technologies for healthcare and science. Our research directions include large-scale models for healthcare, computer-assisted intervention, AI for science, and bioinformatics, etc. Our ultimate goal is to spearhead a transformative revolution in medical practices and scientific discoveries, paving the way for a brighter and healthier future.&lt;/p&gt;

&lt;h3 id=&quot;about-the-hong-kong-university-of-science-and-technology&quot;&gt;About The Hong Kong University of Science and Technology&lt;/h3&gt;

&lt;p&gt;&lt;a href=&quot;https://hkust.edu.hk/&quot;&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/a&gt; is a world-class university that excels in driving innovative education, research excellence, and impactful knowledge transfer. With a holistic and interdisciplinary pedagogy approach, HKUST was ranked 3rd in the Times Higher Education’s Young University Rankings 2024, while 12 of its subjects were ranked among the world’s top 50 in the QS World University Rankings by Subject 2024, with “Data Science and Artificial Intelligence” being ranked first in Hong Kong and 10th in the world. Our graduates are highly competitive, consistently ranking among the world’s top 30 most sought-after employees. In terms of research and entrepreneurship, over 80% of our work was rated “internationally excellent” or “world leading” in the latest Research Assessment Exercise 2020 of Hong Kong’s University Grants Committee. As of October 2024, HKUST members have founded over 1,800 active start-ups, including 10 Unicorns and 14 exits (IPO or M&amp;amp;A).&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/07/14/lai4bm2025</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/07/14/lai4bm2025/"/>
    <title>[Press Release] The 2025 International Workshop on Large AI for Biomedical Imaging (LAI4BM 2025) Successfully Held at HKUST</title>
    <published>2025-07-14T00:00:00+08:00</published>
    <updated>2025-07-14T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/07/14/lai4bm2025/</uri>
    </author>
    <content type="html">&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/lai4bm2025_cover.png&quot; alt=&quot;LAI4BM 2025 Workshop Banner&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;The &lt;strong&gt;International Workshop on Large AI Models for Biomedicine (LAI4BM 2025)&lt;/strong&gt; successfully concluded on July 12, 2025, at the HKUST Jockey Club Institute for Advanced Study. This groundbreaking workshop brought together leading researchers, practitioners, and innovators from around the world to explore the latest advancements in Large AI Models for Biomedical Applications.&lt;/p&gt;

&lt;h3 id=&quot;workshop-overview&quot;&gt;Workshop Overview&lt;/h3&gt;

&lt;p&gt;The LAI4BM 2025 Workshop focused on cutting-edge AI technologies, machine learning techniques, and their transformative applications in healthcare, medical research, and biomedical data analysis. Through keynote presentations, technical sessions, and interactive panel discussions, attendees gained valuable insights into the latest developments in AI-driven biomedical research.&lt;/p&gt;

&lt;p&gt;The event successfully fostered collaboration between AI researchers and biomedical professionals, accelerating innovation in healthcare technology and medical AI applications. The workshop featured distinguished speakers from prestigious institutions including Harvard University, Stanford University, Imperial College London, and leading healthcare organizations.&lt;/p&gt;

&lt;h3 id=&quot;distinguished-speakers-and-their-contributions&quot;&gt;Distinguished Speakers and Their Contributions&lt;/h3&gt;

&lt;p&gt;The workshop featured an impressive lineup of international experts who shared their cutting-edge research and insights:&lt;/p&gt;

&lt;h3 id=&quot;morning-session-highlights&quot;&gt;Morning Session Highlights&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Prof. Pranav Rajpurkar&lt;/strong&gt; from Harvard University delivered a compelling keynote on “Beyond Assistance: Rethinking AI-Human Integration in Radiology,” exploring how AI can transform medical imaging beyond simple assistance tools.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Ruijiang Li&lt;/strong&gt; from Stanford University presented “Multi-modal Foundation AI for Precision Oncology,” discussing the integration of multiple data modalities for personalized cancer treatment.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mr. Dennis Lee&lt;/strong&gt; from Hong Kong Hospital Authority shared practical insights on “Reimagining Healthcare with AI: A Real-World Application of Clinical Management Systems in Hong Kong Hospital Authority,” providing valuable real-world implementation perspectives.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Dong Liang&lt;/strong&gt; from SIAT, Chinese Academy of Sciences, presented “Magnetic Resonance Live imaging: from Photograph to Videography,” showcasing advances in real-time medical imaging.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Kyongtae Tyler Bae&lt;/strong&gt; from The University of Hong Kong discussed “Clinical Implementation of AI in Radiology,” addressing practical challenges and solutions in clinical deployment.&lt;/p&gt;

&lt;h3 id=&quot;afternoon-session-highlights&quot;&gt;Afternoon Session Highlights&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Prof. Xiaomin Ouyang&lt;/strong&gt; from HKUST presented “Building Foundation Model-Powered Multimodal Sensing Systems for Daily Healthcare,” exploring AI applications in everyday health monitoring.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Cheong Kin Ronald Chan&lt;/strong&gt; from CUHK &amp;amp; Hong Kong Hospital Authority delivered an insightful talk on “What Doctors Think, Want and Fear About the Large AI Model?” providing crucial clinical perspectives.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Zheng Li&lt;/strong&gt; from CUHK showcased “Intelligent Surgical Assistive Robots,” demonstrating the future of AI-powered surgical assistance.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Xin Wang&lt;/strong&gt; from CUHK presented “Deep Learning Models for Cancer Subtype Classification to Advance Precision Oncology,” highlighting AI’s role in personalized cancer treatment.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Dr. Yang Cheng&lt;/strong&gt; from AstraZeneca shared industry insights on “Charting the AI Pathway: Opportunities and Challenges in Pharma Development.”&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prof. Chen Qin&lt;/strong&gt; from Imperial College London concluded with “Artificial Intelligence Meets Medical Imaging: From Signals to Interpretation,” bridging technical innovation with clinical application.&lt;/p&gt;

&lt;h3 id=&quot;panel-discussions-addressing-critical-questions&quot;&gt;Panel Discussions: Addressing Critical Questions&lt;/h3&gt;

&lt;p&gt;The workshop featured two thought-provoking panel discussions moderated by Prof. Hao Chen:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;“Are We There for Deploying Large AI Models in Biomedical Applications?”&lt;/strong&gt; - This morning panel explored the current readiness and challenges in implementing large AI models in clinical settings.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;“How Will Large AI Models Reshape the Future of Biomedicine?”&lt;/strong&gt; - The afternoon panel discussed future directions and transformative potential of AI in healthcare.&lt;/p&gt;
  &lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;These discussions provided valuable insights into both the opportunities and challenges facing the field, fostering meaningful dialogue between researchers, clinicians, and industry professionals.&lt;/p&gt;

&lt;h3 id=&quot;organizing-excellence&quot;&gt;Organizing Excellence&lt;/h3&gt;

&lt;p&gt;The workshop was expertly organized by a distinguished committee of HKUST faculty members:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Prof. Hao Chen&lt;/strong&gt; - Assistant Professor of CSE, CBE and LIFS, HKUST&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Prof. Jiguang Wang&lt;/strong&gt; - Padma Harilela Associate Professor of CBE and LIFS, HKUST&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Prof. Kai Liu&lt;/strong&gt; - Cheng Professor of Life Science, HKUST&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Prof. Yingcong Chen&lt;/strong&gt; - Assistant Professor at AI Thrust, Information Hub of HKUST (GZ)&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Prof. Can Yang&lt;/strong&gt; - Dr Tai-chin Lo Associate Professor of Mathematics, HKUST&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Prof. Xiaofang Zhou&lt;/strong&gt; - Otto Poon Professor of Engineering &amp;amp; Chair Professor of CSE, HKUST&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;supporting-organizations&quot;&gt;Supporting Organizations&lt;/h3&gt;

&lt;p&gt;The workshop was supported by several key organizations:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;HKUST Collaborative Center for Medical and Engineering Innovation&lt;/li&gt;
  &lt;li&gt;State Key Laboratory of Nervous System Disorders&lt;/li&gt;
  &lt;li&gt;Center for Medical Imaging and Analysis&lt;/li&gt;
  &lt;li&gt;Department of Computer Science and Engineering, HKUST&lt;/li&gt;
  &lt;li&gt;Department of Chemical and Biological Engineering, HKUST&lt;/li&gt;
  &lt;li&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;workshop-impact-and-success&quot;&gt;Workshop Impact and Success&lt;/h3&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/lai4bm2025_group_photo_1.jpg&quot; alt=&quot;LAI4BM 2025 Group Photo 1&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Workshop participants and speakers gathered for group photo&lt;/em&gt;&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/lai4bm2025_group_photo_2.jpg&quot; alt=&quot;LAI4BM 2025 Group Photo 2&quot; style=&quot;width: 100%; max-width: 800px;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;The workshop successfully achieved its objectives of:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Bringing together leading experts from academia, industry, and healthcare&lt;/li&gt;
  &lt;li&gt;Facilitating knowledge exchange on the latest AI developments in biomedicine&lt;/li&gt;
  &lt;li&gt;Fostering collaborations between AI researchers and biomedical professionals&lt;/li&gt;
  &lt;li&gt;Addressing practical challenges in deploying AI solutions in healthcare settings&lt;/li&gt;
  &lt;li&gt;Exploring future directions for AI in biomedical applications&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;looking-forward&quot;&gt;Looking Forward&lt;/h3&gt;

&lt;p&gt;The LAI4BM 2025 Workshop has set a strong foundation for continued collaboration and innovation in the field of AI for biomedicine. The insights shared, connections made, and discussions held during this event will undoubtedly contribute to advancing the field and improving healthcare outcomes through AI technology.&lt;/p&gt;

&lt;p&gt;The organizers extend their heartfelt gratitude to all speakers, attendees, and sponsors who made this workshop a tremendous success. The enthusiasm and engagement demonstrated by all participants highlight the bright future ahead for AI applications in biomedicine.&lt;/p&gt;

&lt;hr /&gt;

&lt;p&gt;&lt;em&gt;The LAI4BM 2025 Workshop was organized by SmartX Lab at HKUST and supported by various departments and centers at The Hong Kong University of Science and Technology.&lt;/em&gt;&lt;/p&gt;

&lt;h3 id=&quot;about-smartx-lab&quot;&gt;About SmartX Lab&lt;/h3&gt;

&lt;p&gt;SmartX Lab, led by Prof. Hao Chen, is committed to pushing the boundaries of trustworthy AI technologies for healthcare and science. Our research directions include large-scale models for healthcare, computer-assisted intervention, AI for science, and bioinformatics, etc. Our ultimate goal is to spearhead a transformative revolution in medical practices and scientific discoveries, paving the way for a brighter and healthier future.&lt;/p&gt;

&lt;h3 id=&quot;about-the-hong-kong-university-of-science-and-technology&quot;&gt;About The Hong Kong University of Science and Technology&lt;/h3&gt;

&lt;p&gt;&lt;a href=&quot;https://hkust.edu.hk/&quot;&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/a&gt; is a world-class university that excels in driving innovative education, research excellence, and impactful knowledge transfer. With a holistic and interdisciplinary pedagogy approach, HKUST was ranked 3rd in the Times Higher Education’s Young University Rankings 2024, while 12 of its subjects were ranked among the world’s top 50 in the QS World University Rankings by Subject 2024, with “Data Science and Artificial Intelligence” being ranked first in Hong Kong and 10th in the world. Our graduates are highly competitive, consistently ranking among the world’s top 30 most sought-after employees. In terms of research and entrepreneurship, over 80% of our work was rated “internationally excellent” or “world leading” in the latest Research Assessment Exercise 2020 of Hong Kong’s University Grants Committee. As of October 2024, HKUST members have founded over 1,800 active start-ups, including 10 Unicorns and 14 exits (IPO or M&amp;amp;A).&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/05/21/reg2rg</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/05/21/reg2rg/"/>
    <title>[IEEE-TMI] Reg2RG: Large Language Model with Region-Guided Referring and Grounding for CT Report Generation</title>
    <published>2025-05-21T00:00:00+08:00</published>
    <updated>2025-05-21T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/05/21/reg2rg/</uri>
    </author>
    <content type="html">&lt;p&gt;A recent study from SmartX Lab, led by Zhi-Xuan Chen, has been accepted by &lt;em&gt;IEEE Transactions on Medical Imaging&lt;/em&gt; (Impact Factor: 8.9). The work presents &lt;strong&gt;Reg2RG&lt;/strong&gt;, a novel region-guided large language model (LLM) framework for chest CT report generation. By explicitly linking image regions with diagnostic text, Reg2RG significantly improves the accuracy, interpretability, and clinical value of automated CT reporting.&lt;/p&gt;

&lt;h2 id=&quot;introduction&quot;&gt;Introduction&lt;/h2&gt;

&lt;p&gt;In medical image analysis, automated report generation aims to assist clinicians in interpreting and articulating the visual content of scans. The challenge lies in comprehending complex anatomical structures and expressing findings accurately in natural language. Chest CT report generation is a representative task that requires precise lesion localization, feature interpretation, and the synthesis of clinically relevant textual descriptions. However, most existing methods rely on end-to-end full-image modeling, where the entire image is encoded to train language models. Although these methods achieve initial automation, they commonly exhibit poor localization, diagnostic bias, and limited clinical interpretability.&lt;/p&gt;

&lt;p&gt;To address this, we focus on a central question: Can a clear correspondence be established between specific visual regions in medical images and their respective diagnostic paragraphs in textual reports? Inspired by cross-modal tasks such as image-text matching, referring expression, and visual grounding, Reg2RG introduces a core concept: extracting discriminative local image features via region masks and embedding them as tokens into large language models. This allows the model to generate natural language reports that not only follow clinical conventions but also explicitly refer to the relevant anatomical regions in the CT scan, thereby improving the traceability and interpretability of diagnostic conclusions.&lt;/p&gt;

&lt;p&gt;Moreover, chest CT scans are high-dimensional, with lesions that may be subtle and dispersed across multiple anatomical regions. Thus, simultaneously modeling both local anomalies and global anatomical context is essential for generating clinically valuable reports. To address this challenge, Reg2RG adopts a dual-branch encoding architecture that processes local region features and global semantics independently, then integrates them during the language generation phase. This design enhances both localized diagnostic precision and global contextual consistency, enabling the language model to produce well-structured, accurate, and region-aware diagnostic paragraphs. The overall framework of Reg2RG is illustrated in Figure 1.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_fig1.png&quot; alt=&quot;Reg2RG Framework&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Figure 1: Overview of the Reg2RG framework. The model extracts local region features and global context, then generates region-referred diagnostic paragraphs using a large language model.&lt;/em&gt;&lt;/p&gt;

&lt;h2 id=&quot;related-works&quot;&gt;Related Works&lt;/h2&gt;

&lt;h3 id=&quot;medical-image-report-generation&quot;&gt;Medical Image Report Generation&lt;/h3&gt;
&lt;p&gt;Medical image report generation is a critical task in multimodal understanding, aiming to automatically generate clinically meaningful descriptions from complex medical images. Early approaches typically adopted encoder-decoder frameworks, using convolutional neural networks (CNNs) to extract global image features, followed by recurrent neural networks (RNNs) or Transformer-based models to produce diagnostic text. These methods focused on capturing overall image semantics but often overlooked fine-grained representation and localization of specific lesions, leading to reports that lacked interpretability and diagnostic specificity. To improve spatial awareness, some studies introduced attention mechanisms to allow models to focus on key regions. However, these approaches still relied on global features, which made it difficult to explicitly align textual fragments with specific image regions. In recent years, the advent of multimodal large language models (LLMs) has significantly improved report generation quality, but region-level semantic description and grounding consistency remain substantial challenges.&lt;/p&gt;

&lt;h3 id=&quot;region-level-referring-and-grounding&quot;&gt;Region-Level Referring and Grounding&lt;/h3&gt;
&lt;p&gt;In general vision-language tasks, various studies have explored integrating LLMs with region-level features to enhance fine-grained perception. However, applications in the medical imaging domain remain scarce. Existing methods often fail to capture semantic relationships between different regions or rely solely on coarse positional information when referring to image areas. This results in imprecise descriptions that do not meet the high standards of detail and accuracy required for clinical diagnosis. To address this, we propose a new region-level understanding framework tailored for report generation. Through a decoupled representation strategy, we preserve complete geometric information while enhancing texture details, satisfying the clinical demands for sensitivity to lesion location and morphology. Additionally, we incorporate global context to model inter-regional relationships, improving overall image understanding. A region-text alignment mechanism is also introduced to explicitly link visual regions with report content, enabling efficient and coherent multi-region expression in a single inference step—thus advancing the field of region-level visual perception in medical imaging.&lt;/p&gt;

&lt;h2 id=&quot;method&quot;&gt;Method&lt;/h2&gt;
&lt;h3 id=&quot;overall-framework-overview&quot;&gt;Overall Framework Overview&lt;/h3&gt;
&lt;p&gt;Traditional CT report generation methods primarily rely on global features of the entire scan, which often fail to capture subtle abnormalities located in specific anatomical regions, thus limiting the diagnostic accuracy and value of the generated reports. To overcome this limitation, our approach introduces multiple local regional features extracted using a generic segmentation module and combines them with global contextual information to guide report generation. Each local feature corresponds to a specific anatomical region, providing more targeted, fine-grained semantic information. After fusing local and global features, these representations are fed into a large language model for inference. This allows the model to accurately identify and describe multiple anatomical areas, resulting in diagnostic reports that are more comprehensive and clinically interpretable.&lt;/p&gt;

&lt;h3 id=&quot;decoupled-local-feature-representation-strategy&quot;&gt;Decoupled Local Feature Representation Strategy&lt;/h3&gt;
&lt;p&gt;To efficiently preserve high-resolution details of local regions, we propose a decoupled representation strategy that separates local features into texture and geometric components, which are processed independently and then merged into a unified regional representation.
For texture feature extraction, we first apply region masks to isolate target areas from the original CT images. To reduce redundancy and retain high-quality detail, we crop the extracted region images into smaller, focused patches. These are then encoded using an image encoder and compressed through a feature projection module to produce visual embeddings compatible with the language model.
For geometric features, we retain the uncropped region masks to preserve spatial position and scale within the original image. These masks are processed through a lightweight encoding network and mapped into representations that can also be understood by the language model.
Finally, the texture and geometric features are concatenated to form a complete local representation, which is subsequently used for region-level language generation tasks.&lt;/p&gt;

&lt;h3 id=&quot;global-local-feature-collaboration-mechanism&quot;&gt;Global-Local Feature Collaboration Mechanism&lt;/h3&gt;
&lt;p&gt;Different anatomical regions in medical images often exhibit strong structural and semantic interdependencies. Relying solely on local information is insufficient for comprehensive understanding. To address this, we introduce a global-local feature collaboration mechanism that jointly models both levels of information to enhance contextual consistency and narrative coherence in the generated reports. Specifically, the model receives both independent local regional features and global visual features extracted from the entire CT volume. These features are processed by a unified projection module and embedded into the input prompt of the language model.&lt;/p&gt;

&lt;p&gt;We design a structured prompt in which global and regional features are represented by special tokens that are subsequently replaced with their corresponding embeddings. The language model uses these embeddings during inference, allowing it to reason about relationships between different anatomical regions and produce reports that are consistent and explainable.&lt;/p&gt;

&lt;h3 id=&quot;region-report-alignment-training-strategy&quot;&gt;Region-Report Alignment Training Strategy&lt;/h3&gt;
&lt;p&gt;To enhance the model’s ability to refer accurately to specific regions during report generation, we introduce a region-report alignment training strategy. This approach requires the model to identify the anatomical region associated with each paragraph before generating the corresponding diagnostic content.                            &lt;br /&gt;
During training, we prepend a prefix to each ground-truth regional report indicating the region it describes. For instance, the model is prompted with text such as “Region i corresponds to the heart,” thereby guiding it to perform anatomical identification before generating the report. To prevent the model from memorizing the input order, we randomly shuffle the sequence of local features in each iteration. This forces the model to recognize region identity based solely on the embedded features rather than positional bias.                     &lt;br /&gt;
The regional reports are still trained using standard language modeling loss, and the generation remains in free-form textual format. This strategy significantly improves the model’s regional awareness, allowing each paragraph of the generated report to not only be accurate in content but also traceable to a specific anatomical location, thus enhancing interpretability and clinical reliability.&lt;/p&gt;

&lt;h2 id=&quot;experimental-results&quot;&gt;Experimental Results&lt;/h2&gt;

&lt;h3 id=&quot;datasets--metrics&quot;&gt;Datasets &amp;amp; Metrics&lt;/h3&gt;

&lt;p&gt;Reg2RG was evaluated on two large-scale chest CT datasets:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;RadGenome-ChestCT&lt;/strong&gt;: 25,000+ region-annotated CTs with structured reports.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;CTRG-Chest-548K&lt;/strong&gt;: 1,800 CTs with unstructured reports, auto-segmented for region-level evaluation.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Performance was assessed using:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;NLG metrics&lt;/strong&gt;: BLEU, METEOR, ROUGE-L (textual similarity)&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Clinical efficacy metrics&lt;/strong&gt;: Precision, recall, F1 (diagnostic accuracy, via RadBERT extraction)&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Region recognition accuracy&lt;/strong&gt;: Correctly linking text to anatomical regions&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;experimental-setup-and-baselines&quot;&gt;Experimental Setup and Baselines&lt;/h3&gt;
&lt;p&gt;We use the SAT segmentation model to process CT volumes and resample all images to a spatial resolution of 256×256×64. Local texture features are extracted using a pretrained ViT3D encoder and adapted via a Perceiver-based dimensionality reduction module. The same ViT3D backbone is used for global feature extraction. Geometric features are encoded by a lightweight three-layer ViT3D model and mapped into the language model’s input space through fully connected layers.                     &lt;br /&gt;
For language decoding, we adopt the LLaMA2-7B model with the LoRA fine-tuning strategy for efficiency. The model is optimized using AdamW with a fixed learning rate and a warm-up schedule at the beginning. Training is conducted for 6 epochs on the RadGenome dataset and 10 epochs on the CTRG dataset using two NVIDIA 3090 GPUs with a batch size of 16. To reduce memory usage, we apply the ZeRO optimization strategy along with gradient checkpointing.                      &lt;br /&gt;
We compare Reg2RG against several state-of-the-art 3D report generation baselines including CT2Rep, RadFM, and M3D, as well as two 2D-based models: R2GenGPT and MedVInT. All models use LLaMA2-7B as the language decoder and receive the same preprocessed input to ensure fairness. For 2D models, we convert 3D volumes into multi-channel 2D representations.&lt;/p&gt;

&lt;h3 id=&quot;quantitative-results&quot;&gt;Quantitative Results&lt;/h3&gt;
&lt;p&gt;As shown in Table 1, Reg2RG outperforms all competing methods on NLG metrics across both datasets. On the RadGenome-ChestCT dataset, Reg2RG surpasses the next-best model, MedVInT, on all language metrics. In particular, it achieves over 9% relative improvement in METEOR, indicating superior vocabulary diversity and semantic coherence. On the CTRG-Chest-548K dataset, Reg2RG also leads in BLEU and METEOR, though its ROUGE-L score is slightly lower, primarily due to the more fragmented structure of region-level reports, which affects the longest common subsequence matching.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table1.png&quot; alt=&quot;NLG Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 1: Reg2RG achieves state-of-the-art results on both language and clinical efficacy metrics compared to strong baselines.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;As shown in Table 2, Reg2RG also demonstrates clear superiority in Clinical Efficacy (CE) on the RadGenome dataset. It achieves significant improvements in precision, recall, and F1 score compared to all other models. In particular, the F1 score improves by nearly 20%, validating Reg2RG’s reliability in capturing critical diagnostic information. Notably, Reg2RG maintains a strong balance between precision and recall, whereas many models trade off one for the other, highlighting its robust diagnostic capability.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table2.png&quot; alt=&quot;Clinical Efficacy Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 2: Comparison of Reg2RG with other models on Clinical Efficacy (CE) metrics.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Additionally, we evaluate the model’s performance in region recognition. As reported in Table 3, recognition accuracy exceeds 95% for eight out of ten anatomical regions. Lower performance on the lungs and pleura is attributed to segmentation challenges rather than model limitations. Once accurate masks are available, Reg2RG shows consistently strong recognition performance.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table3.png&quot; alt=&quot;Clinical Efficacy Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 3: Region recognition accuracy of Reg2RG.&lt;/em&gt;&lt;/p&gt;

&lt;h3 id=&quot;ablation-studies&quot;&gt;Ablation Studies&lt;/h3&gt;

&lt;p&gt;Ablation experiments confirm the importance of each module:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Removing &lt;strong&gt;local feature decoupling&lt;/strong&gt; (texture/geometry separation) degrades region identification.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;To further assess the contribution of each module to overall model performance, we conducted a series of ablation experiments.
As shown in Table 4, removing the local feature decoupling (LFD) strategy results in performance degradation across several metrics. This highlights the importance of maintaining high-resolution texture and separate geometric information for accurately identifying abnormal regions.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table4.png&quot; alt=&quot;Ablation Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 4: Effectiveness of Local Feature Decoupling (LFD) strategy.&lt;/em&gt;&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;Using only texture or only geometry reduces performance; combining both with global context yields best results.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In Table 5, we examine the effects of different feature combinations. Using only texture features without spatial location information limits model performance. The inclusion of geometric features improves results significantly, and further addition of global context features enhances the model’s ability to capture semantic relationships across regions.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table5.png&quot; alt=&quot;Ablation Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 5: Effectiveness of texture (TXT), geometric (GEO), and global (GLB) features.&lt;/em&gt;&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;Disabling &lt;strong&gt;region-report alignment&lt;/strong&gt; leads to less accurate and less interpretable reports.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In Table 6, we explore the impact of the Region-Report Alignment (RRA) training strategy. Without region-guided prompts, the model can still generate coherent text, but the accuracy and consistency of region-specific content degrade notably. This confirms that regional prompting is crucial for reinforcing semantic alignment between visual input and generated text.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table6.png&quot; alt=&quot;Ablation Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 6: Effectiveness of Region-Report Alignment (RRA) strategy.&lt;/em&gt;&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;Larger LLMs (LLaMA2-7B) provide significant gains over smaller models.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Furthermore, we evaluate how model performance varies with different language model sizes. As shown in Table 7, replacing LLaMA2-7B with a smaller model such as GPT-2 leads to substantial drops in performance across all metrics. This suggests that large models possess superior capabilities for region-level referring, cross-region reasoning, and complex report generation.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_table7.png&quot; alt=&quot;Ablation Results&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Table 7: Effectiveness of language model size on Reg2RG performance.&lt;/em&gt;&lt;/p&gt;

&lt;h3 id=&quot;report-quality-analysis&quot;&gt;Report Quality Analysis&lt;/h3&gt;
&lt;p&gt;To gain deeper insights into the quality of generated reports, we conducted statistical analysis on report length distributions and compared them with real-world references.                      &lt;br /&gt;
As shown in Figure 2, the length distribution of reports generated by Reg2RG closely mirrors that of real clinical reports, exhibiting less deviation than the MedVInT baseline. This indicates that Reg2RG produces more complete and information-rich outputs, reducing the likelihood of missing findings.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_fig2.png&quot; alt=&quot;Report Length Distribution&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Figure 2: Report length distribution comparison between Reg2RG and other methods.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;A case study in Figure 3 demonstrates that Reg2RG accurately identifies and describes abnormalities across multiple anatomical regions, whereas competing methods frequently miss or misidentify findings. The output shown in Figure 4 reveals clearly delineated report sections, each explicitly referring to a specific region. This structure greatly enhances the clinical interpretability and practical utility of the model output. The region-aware alignment mechanism not only improves the precision of generated report but also provides clinicians with precise spatial references for validation.&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_fig3.png&quot; alt=&quot;Case Study&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Figure 3: Report comparison between Reg2RG and the second-best model.&lt;/em&gt;&lt;/p&gt;

&lt;div align=&quot;center&quot;&gt;
  &lt;p&gt;&lt;img src=&quot;/static/img/news/2025_reg2rg_fig4.png&quot; alt=&quot;Case Study&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;
&lt;/div&gt;

&lt;p&gt;&lt;em&gt;Figure 4: Example of a region-aligned diagnostic report generated by Reg2RG.&lt;/em&gt;&lt;/p&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;This paper presents Reg2RG, a region-guided framework for chest CT report generation, which introduces region-aware mechanisms to enhance diagnostic accuracy, fine-grained expression, and interpretability. Unlike traditional approaches that rely solely on global image features, Reg2RG integrates anatomical region features with global contextual information to produce precise multi-region diagnostic reports.                       &lt;br /&gt;
We propose a decoupled local feature strategy to model both texture and geometric information and further enhance inter-region semantic relationships through a global-local collaboration mechanism. Additionally, a region-report alignment training strategy reinforces the alignment between visual and linguistic modalities, improving the model’s region-specific reasoning and reporting capability.                     &lt;br /&gt;
Experimental results demonstrate that Reg2RG surpasses state-of-the-art methods in both language generation quality and clinical accuracy. The framework shows strong generalizability and practical potential. In future work, we aim to extend the approach to additional imaging modalities and diagnostic tasks, further advancing the practical deployment of automated medical report generation.&lt;/p&gt;

&lt;p&gt;For more details, code and models are available at:&lt;br /&gt;
&lt;a href=&quot;https://arxiv.org/pdf/2411.15539&quot;&gt;Paper (arXiv)&lt;/a&gt; | &lt;a href=&quot;https://github.com/zhi-xuan-chen/Reg2RG&quot;&gt;GitHub&lt;/a&gt; | &lt;a href=&quot;https://huggingface.co/Trusure/Reg2RG&quot;&gt;HuggingFace&lt;/a&gt;&lt;/p&gt;

&lt;h3 id=&quot;references&quot;&gt;References&lt;/h3&gt;

&lt;p&gt;Z. Chen, Y. Bie, H. Jin, and H. Chen, “Large Language Model with Region-guided Referring and Grounding for CT Report Generation,” &lt;em&gt;IEEE Transactions on Medical Imaging&lt;/em&gt;, doi: 10.1109/TMI.2025.3559923.&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/05/19/deepbreath</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/05/19/deepbreath/"/>
    <title>[Call for Papers] Deep-Brea3th 2025: AI and Imaging for Breast Care Workshop to be Held at MICCAI 2025</title>
    <published>2025-05-19T00:00:00+08:00</published>
    <updated>2025-05-19T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/05/19/deepbreath/</uri>
    </author>
    <content type="html">&lt;p&gt;&lt;strong&gt;Daejeon, South Korea – September 23-27, 2025&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The second edition of the Deep-Brea3th Workshop on AI and Imaging for Diagnostic and Treatment Challenges in Breast Care will take place during MICCAI 2025 in Daejeon, South Korea. This international event will spotlight the latest advancements in artificial intelligence (AI) and multimodal imaging for breast cancer diagnosis and treatment, uniting leading researchers, clinicians, and industry experts to drive innovation and cross-disciplinary collaboration in breast health.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_deepbreath.png&quot; alt=&quot;Deep-Brea3th 2025&quot; style=&quot;width: 100%; max-width: 800px; margin: 0 auto; display: block;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;breast-cancer-a-global-challenge-ai-as-a-solution&quot;&gt;Breast Cancer: A Global Challenge, AI as a Solution&lt;/h2&gt;

&lt;p&gt;Breast cancer remains the most common malignant tumor among women worldwide, characterized by significant heterogeneity and complex diagnostic needs. Medical imaging—including mammography, ultrasound, MRI, PET/CT, and pathology—plays a crucial role in detection and characterization. With recent breakthroughs in AI, automated breast image interpretation and clinical decision support are rapidly advancing, offering new hope for improved outcomes and personalized care.&lt;/p&gt;

&lt;h2 id=&quot;three-international-challenges-to-advance-the-field&quot;&gt;Three International Challenges to Advance the Field&lt;/h2&gt;

&lt;p&gt;Deep-Brea3th 2025 will feature three associated challenges, fostering open innovation and benchmarking in key areas:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;ODELIA 2025&lt;/strong&gt;: Breast MRI Challenge&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;MAMA-MIA 2025&lt;/strong&gt;: Tumor Segmentation and Treatment Response Prediction in Breast MRI&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;UUSIC 2025&lt;/strong&gt;: Universal Ultrasound Image Challenge for Multi-Organ Classification and Segmentation&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;broad-scope-focused-impact&quot;&gt;Broad Scope, Focused Impact&lt;/h2&gt;

&lt;p&gt;The workshop welcomes submissions in all areas of AI for breast cancer, including (but not limited to):&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;Breast imaging (mammography, ultrasound, MRI, PET/CT, pathology, etc.)&lt;/li&gt;
  &lt;li&gt;Detection, segmentation, and classification&lt;/li&gt;
  &lt;li&gt;Breast cancer screening and risk prediction&lt;/li&gt;
  &lt;li&gt;Image registration, synthesis, and reconstruction&lt;/li&gt;
  &lt;li&gt;Multimodal imaging fusion&lt;/li&gt;
  &lt;li&gt;Treatment response and drug selection&lt;/li&gt;
  &lt;li&gt;Lymph node status and molecular subtypes&lt;/li&gt;
  &lt;li&gt;Tumor microenvironment and recurrence prediction&lt;/li&gt;
  &lt;li&gt;Radiomics, reader studies, and federated/swarm learning&lt;/li&gt;
  &lt;li&gt;Natural language processing and large language models&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;high-quality-peer-review-and-open-science&quot;&gt;High-Quality Peer Review and Open Science&lt;/h2&gt;

&lt;p&gt;Submissions are invited in two tracks:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Full Papers&lt;/strong&gt;: Up to 8 pages (plus 2 pages for references), double-blind reviewed, to be published in the Springer LNCS MICCAI Satellite Events proceedings. Selected papers may be recommended for high-impact SCI-indexed journals.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Abstracts&lt;/strong&gt;: Up to 2500 characters, including a supporting figure, for both new and recently published work. Accepted abstracts will be presented orally or as posters and made publicly available.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All accepted contributions are eligible for Best Student Paper, Best Workshop Paper, and Best Abstract awards.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Submission portal:&lt;/strong&gt; &lt;a href=&quot;https://cmt3.research.microsoft.com/DeepBreath2025&quot;&gt;CMT Submission System&lt;/a&gt;&lt;/p&gt;

&lt;h2 id=&quot;key-dates&quot;&gt;Key Dates&lt;/h2&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;May 13, 2025:&lt;/strong&gt; Submissions open&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;June 25, 2025:&lt;/strong&gt; Submission deadline&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;July 16, 2025:&lt;/strong&gt; Notification of decisions&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;July 30, 2025:&lt;/strong&gt; Camera-ready deadline&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;August 1, 2025:&lt;/strong&gt; Full program released&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;September 23-27, 2025:&lt;/strong&gt; Workshop at MICCAI 2025&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;international-leadership-and-expertise&quot;&gt;International Leadership and Expertise&lt;/h2&gt;

&lt;p&gt;Deep-Brea3th 2025 is organized by an international committee of experts from the Netherlands, Germany, China, the UK, USA, Spain, Switzerland, Greece, and more, ensuring a high standard of academic rigor and global impact.&lt;/p&gt;

&lt;h2 id=&quot;join-us-shape-the-future-of-breast-health-with-ai&quot;&gt;Join Us: Shape the Future of Breast Health with AI&lt;/h2&gt;

&lt;p&gt;We invite researchers, clinicians, and innovators from around the world to submit their work and participate in this landmark event. For more information, submission guidelines, and updates, please visit our &lt;a href=&quot;https://cmt3.research.microsoft.com/DeepBreath2025&quot;&gt;website&lt;/a&gt; or contact the organizing committee at &lt;strong&gt;miccai.deepbreath@gmail.com&lt;/strong&gt;.&lt;/p&gt;

&lt;hr /&gt;

&lt;p&gt;&lt;strong&gt;Contact:&lt;/strong&gt;&lt;br /&gt;
Deep-Brea3th 2025 Organizing Committee&lt;br /&gt;
Email: miccai.deepbreath@gmail.com&lt;br /&gt;
Submission Portal: https://cmt3.research.microsoft.com/DeepBreath2025&lt;/p&gt;

&lt;hr /&gt;

&lt;p&gt;&lt;em&gt;Let’s use AI to change the future of breast care—together.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/05/06/news</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/05/06/news/"/>
    <title>[Press Release] SmartX Lab Members Shine at 2025 Undergraduate Research Opportunities Program Awards</title>
    <published>2025-05-06T00:00:00+08:00</published>
    <updated>2025-05-06T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/05/06/news/</uri>
    </author>
    <content type="html">&lt;p&gt;&lt;strong&gt;SmartX Lab is proud to announce the outstanding achievements of its students and faculty at the 2025 Kerry Holdings Limited Undergraduate Research Opportunities Program (UROP) Awards.&lt;/strong&gt; The award presentation ceremony, held on April 23, recognized innovative research across multiple disciplines, with HKUST students taking home several top honors.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_press_2.jpg&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;top-honors-for-biomedical-ai-research&quot;&gt;Top Honors for Biomedical AI Research&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;LIU Runsheng&lt;/strong&gt; (COMP, Year 3), supervised by Prof. CHEN Hao from the Department of Computer Science and Engineering, clinched both the &lt;strong&gt;Champion&lt;/strong&gt; and &lt;strong&gt;Best Poster Award&lt;/strong&gt; for his groundbreaking project, “GAInS: Gradient Anomaly-aware Biomedical Instance Segmentation.”&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_press_4.jpg&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;Biomedical instance segmentation is crucial for accurately identifying and outlining structures such as tissues and cells in medical images. Traditional methods often struggle with overlapping or touching cells, treating each scenario separately and missing the connections between them. LIU’s GAInS approach introduces a new way to detect and refine these complex regions by analyzing the local gradient anomalies—essentially, the sharp changes in image intensity that signal different structures. His method uses a novel Gradient Anomaly Mapping Module (GAMM) to create detailed maps of these regions, and an Adaptive Local Refinement Module (ALRM) that fine-tunes the segmentation boundaries. The result is a system that outperforms current state-of-the-art techniques, as demonstrated in extensive experiments across multiple biomedical datasets. The work has also been recognized internationally, having been published at the &lt;a href=&quot;https://www.computer.org/csdl/proceedings-article/bibm/2024/10822187/23oomcXkdfq&quot;&gt;IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2024&lt;/a&gt;.&lt;/p&gt;

&lt;h3 id=&quot;advancing-medical-image-analysis-with-ai&quot;&gt;Advancing Medical Image Analysis with AI&lt;/h3&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_press_3.jpg&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;GU Yi&lt;/strong&gt; (COMP, Graduated), also supervised by Prof. CHEN Hao, was named &lt;strong&gt;Second Runner-Up&lt;/strong&gt; for his project, “Deep learning for medical image analysis.” GU’s research tackles the challenge of detecting medical anomalies—critical for early disease diagnosis—using artificial intelligence. Traditional models are trained only on healthy data and must identify when something unusual appears. However, these models often struggle to confidently distinguish between normal and abnormal cases.&lt;/p&gt;

&lt;p&gt;GU’s solution, D2UE (Diversified Dual-space Uncertainty Estimation), introduces two key innovations. First, a technique called Redundancy-Aware Repulsion (RAR) ensures that AI models learn to agree when analyzing healthy images and disagree when faced with potential anomalies. Second, his Dual-Space Uncertainty (DSU) method examines both the outputs of the models and how those outputs change with slight variations in the input, allowing for more sensitive detection of subtle anomalies. Tested across five medical datasets, D2UE demonstrated superior performance compared to existing methods, marking a significant step forward in AI-powered medical diagnostics.&lt;/p&gt;

&lt;h3 id=&quot;looking-ahead&quot;&gt;Looking Ahead&lt;/h3&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_press_1.jpg&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The UROP Office encourages both faculty and students to participate in the upcoming Summer 2025 program. Project proposals and student applications are now open. For further details, please visit the &lt;a href=&quot;https://urop.hkust.edu.hk/Awards_and_Sponsorships/awardee_2025&quot;&gt;UROP website&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Congratulations to all awardees for their dedication to advancing research and innovation at HKUST!&lt;/p&gt;

&lt;h3 id=&quot;about-smartx-lab&quot;&gt;About SmartX Lab&lt;/h3&gt;

&lt;p&gt;SmartX Lab, led by Prof. Hao Chen, is committed to pushing the boundaries of trustworthy AI technologies for healthcare and science. Our research directions include large-scale models for healthcare, computer-assisted intervention, AI for science, and bioinformatics, etc. Our ultimate goal is to spearhead a transformative revolution in medical practices and scientific discoveries, paving the way for a brighter and healthier future.&lt;/p&gt;

&lt;h3 id=&quot;about-the-hong-kong-university-of-science-and-technology&quot;&gt;About The Hong Kong University of Science and Technology&lt;/h3&gt;

&lt;p&gt;&lt;a href=&quot;https://hkust.edu.hk/&quot;&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/a&gt; is a world-class university that excels in driving innovative education, research excellence, and impactful knowledge transfer. With a holistic and interdisciplinary pedagogy approach, HKUST was ranked 3rd in the Times Higher Education’s Young University Rankings 2024, while 12 of its subjects were ranked among the world’s top 50 in the QS World University Rankings by Subject 2024, with “Data Science and Artificial Intelligence” being ranked first in Hong Kong and 10th in the world. Our graduates are highly competitive, consistently ranking among the world’s top 30 most sought-after employees. In terms of research and entrepreneurship, over 80% of our work was rated “internationally excellent” or “world leading” in the latest Research Assessment Exercise 2020 of Hong Kong’s University Grants Committee. As of October 2024, HKUST members have founded over 1,800 active start-ups, including 10 Unicorns and 14 exits (IPO or M&amp;amp;A).&lt;/p&gt;
</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/04/30/news</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/04/30/news/"/>
    <title>[Press Release] MOME Model: AI Empowers Breast Cancer Diagnosis | Dialogue between SmartX Lab and Shenzhen People’s Hospital</title>
    <published>2025-04-30T00:00:00+08:00</published>
    <updated>2025-04-30T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/04/30/news/</uri>
    </author>
    <content type="html">&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_hetao_talk_cover.png&quot; alt=&quot;HKUST SmartLab&quot; style=&quot;width: 100%;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;On April 30, 2025, the &lt;strong&gt;SmartX Lab&lt;/strong&gt; team, led by &lt;strong&gt;Prof. Hao Chen&lt;/strong&gt;, had a fruitful dialogue with &lt;strong&gt;Shenzhen People’s Hospital&lt;/strong&gt;. The discussion focused on the &lt;a href=&quot;https://smartx.cse.ust.hk/2024/08/28/mome/&quot;&gt;MOME model&lt;/a&gt;, an AI-powered breast cancer diagnosis model developed by SmartX Lab. This model has shown remarkable performance in assisting medical professionals with breast cancer diagnosis.&lt;/p&gt;

&lt;p&gt;The dialogue was attended by &lt;strong&gt;Prof. Chen&lt;/strong&gt; and &lt;strong&gt;Dr. Wu Mingxiang&lt;/strong&gt; from Shenzhen People’s Hospital. The meeting provided an excellent opportunity for both teams to exchange ideas and explore potential collaborations in the field of AI-driven healthcare.&lt;/p&gt;

&lt;p&gt;The MOME model, which stands for &lt;strong&gt;Mixture of Modality Experts&lt;/strong&gt;, is a state-of-the-art AI model that integrates the  multiparametric MRI (mpMRI) data. The model has been trained on the largest dataset and has demonstrated its ability to provide reliable diagnostic support for medical professionals.&lt;/p&gt;

&lt;p&gt;The collaboration between SmartX Lab and Shenzhen People’s Hospital aims to further refine the MOME model and explore its potential applications in clinical settings. The dialogue highlighted the importance of interdisciplinary collaboration in advancing AI technologies for healthcare, ultimately benefiting patients and improving medical outcomes.
The meeting also emphasized the significance of AI in transforming the healthcare landscape, particularly in the early detection and diagnosis of diseases. By leveraging advanced AI technologies, healthcare professionals can enhance their diagnostic capabilities, leading to more accurate and timely interventions for patients.&lt;/p&gt;

&lt;p&gt;Following the meeting, both teams expressed their enthusiasm for future collaborations and the potential impact of AI in healthcare. The SmartX Lab team is committed to continuing its research in AI-driven healthcare solutions, with the goal of improving patient outcomes and revolutionizing medical practices.&lt;/p&gt;

&lt;p&gt;Video of the dialogue is available as follows (in Mandarin Chinese with English subtitles):&lt;/p&gt;
&lt;video controls=&quot;&quot; style=&quot;width: 100%; height: auto;&quot;&gt;
  &lt;source src=&quot;/static/media/Hetao_Talk.mp4&quot; type=&quot;video/mp4&quot; /&gt;
  Your browser does not support the video tag.
&lt;/video&gt;

&lt;h3 id=&quot;about-smartx-lab&quot;&gt;About SmartX Lab&lt;/h3&gt;

&lt;p&gt;SmartX Lab, led by Prof. Hao Chen, is committed to pushing the boundaries of trustworthy AI technologies for healthcare and science. Our research directions include large-scale models for healthcare, computer-assisted intervention, AI for science, and bioinformatics, etc. Our ultimate goal is to spearhead a transformative revolution in medical practices and scientific discoveries, paving the way for a brighter and healthier future.&lt;/p&gt;

&lt;h3 id=&quot;about-the-hong-kong-university-of-science-and-technology&quot;&gt;About The Hong Kong University of Science and Technology&lt;/h3&gt;

&lt;p&gt;&lt;a href=&quot;https://hkust.edu.hk/&quot;&gt;The Hong Kong University of Science and Technology (HKUST)&lt;/a&gt; is a world-class university that excels in driving innovative education, research excellence, and impactful knowledge transfer. With a holistic and interdisciplinary pedagogy approach, HKUST was ranked 3rd in the Times Higher Education’s Young University Rankings 2024, while 12 of its subjects were ranked among the world’s top 50 in the QS World University Rankings by Subject 2024, with “Data Science and Artificial Intelligence” being ranked first in Hong Kong and 10th in the world. Our graduates are highly competitive, consistently ranking among the world’s top 30 most sought-after employees. In terms of research and entrepreneurship, over 80% of our work was rated “internationally excellent” or “world leading” in the latest Research Assessment Exercise 2020 of Hong Kong’s University Grants Committee. As of October 2024, HKUST members have founded over 1,800 active start-ups, including 10 Unicorns and 14 exits (IPO or M&amp;amp;A).&lt;/p&gt;

</content>
  </entry>
  
  <entry>
    <id>https://hkustsmartlab.github.io//2025/02/28/focus</id>
    <link type="text/html" rel="alternate" href="https://hkustsmartlab.github.io//2025/02/28/focus/"/>
    <title>[CVPR 2025] FOCUS: Knowledge-Enhanced Adaptive Visual Compression for Few-Shot Whole Slide Image Classification</title>
    <published>2025-02-28T00:00:00+08:00</published>
    <updated>2025-02-28T00:00:00+08:00</updated>
    <author>
      <name>Cheng Jin</name>
      <uri>/2025/02/28/focus/</uri>
    </author>
    <content type="html">&lt;p&gt;Recetnly, SmartX Lab’s new work &lt;strong&gt;FOCUS: Knowledge-Enhanced Adaptive Visual Compression for Few-Shot Whole Slide Image Classification&lt;/strong&gt;, accepted at &lt;strong&gt;CVPR 2025&lt;/strong&gt;, a top-tier conference in computer vision. This study introduces &lt;strong&gt;FOCUS&lt;/strong&gt;, a novel framework designed to tackle the &lt;strong&gt;few-shot weakly supervised learning (FSWL) problem in whole slide image (WSI) classification&lt;/strong&gt;. By leveraging &lt;strong&gt;adaptive visual compression and knowledge enhancement&lt;/strong&gt;, FOCUS effectively &lt;strong&gt;filters out irrelevant regions and prioritizes diagnostically significant areas&lt;/strong&gt;, addressing key challenges in pathology AI under data-scarce conditions.&lt;/p&gt;

&lt;h2 id=&quot;background&quot;&gt;Background&lt;/h2&gt;

&lt;p&gt;Whole slide image (WSI) classification is a crucial task in &lt;strong&gt;computational pathology (CPath)&lt;/strong&gt;, enabling &lt;strong&gt;automated disease diagnosis&lt;/strong&gt; from high-resolution pathology images. This is particularly important for &lt;strong&gt;cancer detection&lt;/strong&gt;, where accurate classification can significantly impact clinical decision-making. However, WSIs are &lt;strong&gt;extremely high-resolution&lt;/strong&gt; (often exceeding &lt;strong&gt;100,000 × 100,000 pixels&lt;/strong&gt;), and diagnostic information is &lt;strong&gt;sparsely distributed&lt;/strong&gt; across the image. Traditional &lt;strong&gt;fully supervised learning&lt;/strong&gt; approaches require extensive &lt;strong&gt;expert-annotated training data&lt;/strong&gt;, which is costly and limited due to &lt;strong&gt;privacy constraints and expert availability&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;To address this, &lt;strong&gt;few-shot weakly supervised learning (FSWL)&lt;/strong&gt; has emerged as a promising solution, allowing models to learn effectively from &lt;strong&gt;limited labeled samples&lt;/strong&gt; combined with &lt;strong&gt;unlabeled data&lt;/strong&gt;. A widely used approach in FSWL is &lt;strong&gt;multiple instance learning (MIL)&lt;/strong&gt;, where WSIs are divided into smaller &lt;strong&gt;image patches&lt;/strong&gt;, and features are aggregated to form a slide-level representation. However, when labeled samples are extremely scarce, &lt;strong&gt;MIL struggles to distinguish diagnostically relevant patches from irrelevant ones&lt;/strong&gt;. Additionally, existing methods &lt;strong&gt;underutilize pathology foundation models (FMs) and language priors&lt;/strong&gt;, limiting their effectiveness in few-shot scenarios.&lt;/p&gt;

&lt;p&gt;This raises a key question:&lt;br /&gt;
&lt;strong&gt;How can we efficiently compress visual information and focus on diagnostically relevant regions under few-shot conditions?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;FOCUS addresses this by integrating &lt;strong&gt;foundation models (FMs) and language knowledge&lt;/strong&gt; into an &lt;strong&gt;adaptive compression mechanism&lt;/strong&gt;, ensuring that critical diagnostic features are prioritized. Figure 1 illustrates the overall framework.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_focus_fig1.png&quot; alt=&quot;FOCUS Framework&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;related-work&quot;&gt;Related Work&lt;/h2&gt;

&lt;h3 id=&quot;a-multiple-instance-learning-mil&quot;&gt;A. Multiple Instance Learning (MIL)&lt;/h3&gt;

&lt;p&gt;MIL is the dominant approach for WSI analysis, treating each WSI as a &lt;strong&gt;bag&lt;/strong&gt; and its patches as &lt;strong&gt;instances&lt;/strong&gt;. Patch features are aggregated to form a &lt;strong&gt;slide-level representation&lt;/strong&gt;, making MIL well-suited for &lt;strong&gt;weakly supervised learning&lt;/strong&gt;. However, traditional MIL &lt;strong&gt;requires large labeled datasets&lt;/strong&gt;, making it ineffective in few-shot settings. FOCUS enhances MIL by introducing &lt;strong&gt;adaptive visual compression&lt;/strong&gt; to improve feature selection.&lt;/p&gt;

&lt;h3 id=&quot;b-pathology-foundation-models-fms&quot;&gt;B. Pathology Foundation Models (FMs)&lt;/h3&gt;

&lt;p&gt;Recent pathology foundation models (e.g., &lt;strong&gt;CONCH, UNI, GPFM, Virchow&lt;/strong&gt;) have demonstrated &lt;strong&gt;strong feature extraction capabilities&lt;/strong&gt; through large-scale pretraining. However, current methods &lt;strong&gt;limit their use to feature extraction&lt;/strong&gt;, failing to &lt;strong&gt;leverage their semantic understanding&lt;/strong&gt;. FOCUS utilizes FMs for &lt;strong&gt;global redundancy removal and semantic relevance assessment&lt;/strong&gt;, improving feature selection.&lt;/p&gt;

&lt;h3 id=&quot;c-few-shot-weakly-supervised-learning-fswl&quot;&gt;C. Few-Shot Weakly Supervised Learning (FSWL)&lt;/h3&gt;

&lt;p&gt;FSWL aims to perform &lt;strong&gt;WSI classification with minimal labeled samples&lt;/strong&gt;. Existing methods incorporate &lt;strong&gt;language guidance or additional visual samples&lt;/strong&gt;, but they often &lt;strong&gt;struggle with multi-resolution requirements or reliance on extra reference samples&lt;/strong&gt;. FOCUS introduces a &lt;strong&gt;three-stage compression strategy&lt;/strong&gt;, offering a &lt;strong&gt;more generalizable and efficient solution&lt;/strong&gt;.&lt;/p&gt;

&lt;h2 id=&quot;methodology&quot;&gt;Methodology&lt;/h2&gt;

&lt;p&gt;FOCUS employs a &lt;strong&gt;three-stage progressive visual compression strategy&lt;/strong&gt;, consisting of the following steps:&lt;/p&gt;

&lt;h3 id=&quot;a-global-redundancy-removal&quot;&gt;A. &lt;strong&gt;Global Redundancy Removal&lt;/strong&gt;&lt;/h3&gt;

&lt;p&gt;Each WSI is divided into &lt;strong&gt;non-overlapping patches&lt;/strong&gt;, and features are extracted using &lt;strong&gt;pretrained pathology foundation models&lt;/strong&gt;. To remove redundant information, FOCUS applies &lt;strong&gt;cosine similarity-based filtering&lt;/strong&gt; within a &lt;strong&gt;local sliding window&lt;/strong&gt;. A &lt;strong&gt;dynamic threshold&lt;/strong&gt; is computed based on &lt;strong&gt;mean similarity and standard deviation&lt;/strong&gt;, ensuring that highly redundant patches are eliminated.&lt;/p&gt;

\[\hat{\mathbf{b}}_i=\frac{\mathbf{b}_i}{\left\|\mathbf{b}_i\right\|_2}, \quad S_{i j}=\hat{\mathbf{b}}_i \cdot \hat{\mathbf{b}}_j, \quad \tau_g=\mu(S)+\sigma(S)\]

&lt;h3 id=&quot;b-language-guided-visual-token-prioritization&quot;&gt;B. &lt;strong&gt;Language-Guided Visual Token Prioritization&lt;/strong&gt;&lt;/h3&gt;

&lt;p&gt;After redundancy removal, the remaining patch features are &lt;strong&gt;matched with expert-provided textual descriptions&lt;/strong&gt;. A &lt;strong&gt;text encoder&lt;/strong&gt; converts expert descriptions into &lt;strong&gt;text tokens&lt;/strong&gt;, which are then mapped to &lt;strong&gt;visual features&lt;/strong&gt;. A &lt;strong&gt;cross-modal attention mechanism&lt;/strong&gt; calculates similarity scores, ranking visual tokens by &lt;strong&gt;semantic relevance&lt;/strong&gt;. The top-ranked tokens are &lt;strong&gt;selected for further processing&lt;/strong&gt;, ensuring that diagnostically significant regions are prioritized.&lt;/p&gt;

\[\begin{gathered}
\mathbf{A}=\operatorname{softmax}\left(\frac{\left(\mathbf{T} W_q\right)\left(\mathbf{B} W_k\right)^{\top}}{\sqrt{d}}\right), \quad r_i=\frac{1}{t_1+t_2} \sum_{j=1}^{t_1+t_2} \mathbf{A}_{j i}, \\
k=\min \left(M_{\max }, \gamma N^{\prime}\right), \quad \mathbf{B}_s=\left\{\mathbf{b}_i \mid \operatorname{rank}\left(r_i\right) \leq k\right\},
\end{gathered}\]

&lt;h3 id=&quot;c-sequential-visual-token-compression&quot;&gt;C. &lt;strong&gt;Sequential Visual Token Compression&lt;/strong&gt;&lt;/h3&gt;

&lt;p&gt;To further refine the feature set, FOCUS applies &lt;strong&gt;sequential compression&lt;/strong&gt; to remove &lt;strong&gt;local redundancy&lt;/strong&gt; among selected visual tokens. Cosine similarity is computed between &lt;strong&gt;adjacent tokens&lt;/strong&gt;, and a &lt;strong&gt;dynamic thresholding mechanism&lt;/strong&gt; iteratively filters out redundant tokens. This process preserves &lt;strong&gt;spatial coherence&lt;/strong&gt; while eliminating unnecessary information.&lt;/p&gt;

\[\begin{gathered}
s_{j, j+1}=\frac{\mathbf{b}_j^{l \top} \mathbf{b}_{j+1}}{\left\|\mathbf{b}_j\right\|_2 \cdot\left\|\mathbf{b}_{j+1}\right\|_2}, \quad j \in\{1, \ldots, k-1\} \\
\operatorname{mask}_j= \begin{cases}1, &amp;amp; \text { if } \min \left(s_{j-1, j}, s_{j, j+1}\right)&amp;lt;\theta_i \\
0, &amp;amp; \text { otherwise }\end{cases}
\end{gathered}\]

&lt;h3 id=&quot;d-cross-modal-aggregation-and-final-prediction&quot;&gt;D. &lt;strong&gt;Cross-Modal Aggregation and Final Prediction&lt;/strong&gt;&lt;/h3&gt;

&lt;p&gt;The compressed visual tokens are fused with &lt;strong&gt;text embeddings&lt;/strong&gt; using a &lt;strong&gt;multi-head attention mechanism&lt;/strong&gt;. The attention scores are normalized and passed through a &lt;strong&gt;fully connected layer&lt;/strong&gt; to generate the final classification probabilities. The model is trained &lt;strong&gt;end-to-end&lt;/strong&gt; using a &lt;strong&gt;cross-entropy loss function&lt;/strong&gt;, optimizing feature extraction, cross-modal attention, and classification jointly.&lt;/p&gt;

\[\begin{gathered}
\operatorname{Head}_i=\operatorname{softmax}\left(\frac{\mathbf{Q} W_q^i\left(\mathbf{K} W_k^i\right)^{\top}}{\sqrt{d_k}}\right) \mathbf{V} W_v^i, \\
\mathbf{O}=\operatorname{LayerNorm}\left(\operatorname{Concat}\left(\operatorname{Head}_1, \ldots, \operatorname{Head}_h\right) W_o\right), \\
P\left(Y \mid \mathbf{B}_c, \mathbf{T}\right)=\operatorname{softmax}\left(W_c \mathbf{O}+\beta_c\right),
\end{gathered}\]

&lt;h2 id=&quot;experimental-results&quot;&gt;Experimental Results&lt;/h2&gt;

&lt;h3 id=&quot;datasets&quot;&gt;Datasets&lt;/h3&gt;

&lt;p&gt;FOCUS was evaluated on &lt;strong&gt;three benchmark pathology datasets&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;TCGA-NSCLC&lt;/strong&gt;: A dataset from &lt;strong&gt;The Cancer Genome Atlas&lt;/strong&gt;, containing &lt;strong&gt;lung adenocarcinoma (LUAD) and lung squamous cell carcinoma (LUSC)&lt;/strong&gt; WSIs (~1,000 slides).&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;CAMELYON&lt;/strong&gt;: A dataset from the &lt;strong&gt;CAMELYON16 and CAMELYON17 challenges&lt;/strong&gt;, including &lt;strong&gt;normal and metastatic lymph node samples&lt;/strong&gt; (~900 slides).&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;UBC-OCEAN&lt;/strong&gt;: A dataset from the &lt;strong&gt;University of British Columbia&lt;/strong&gt;, containing &lt;strong&gt;five ovarian cancer subtypes&lt;/strong&gt; (~500 slides).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Experiments were conducted under &lt;strong&gt;4-shot, 8-shot, and 16-shot settings&lt;/strong&gt;, where only &lt;strong&gt;4, 8, or 16 labeled WSIs per class&lt;/strong&gt; were used for training.&lt;/p&gt;

&lt;h3 id=&quot;evaluation-metrics&quot;&gt;Evaluation Metrics&lt;/h3&gt;

&lt;p&gt;The model was evaluated using:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Balanced Accuracy (Balanced ACC)&lt;/strong&gt;: Accounts for class imbalance.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Area Under the ROC Curve (AUC)&lt;/strong&gt;: Measures classification performance.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;F1 Score&lt;/strong&gt;: Reflects overall classification quality.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;key-findings&quot;&gt;Key Findings&lt;/h3&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_focus_fig2.png&quot; alt=&quot;FOCUS Performance&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;p&gt;The above Table summarizes FOCUS’s performance:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;TCGA-NSCLC&lt;/strong&gt;: Achieves &lt;strong&gt;81.9% Balanced ACC and 91.5% AUC&lt;/strong&gt; under &lt;strong&gt;4-shot settings&lt;/strong&gt;, outperforming all baselines.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;CAMELYON&lt;/strong&gt;: Maintains high accuracy across &lt;strong&gt;low-shot settings&lt;/strong&gt;, reaching &lt;strong&gt;70.1% ACC in 4-shot experiments&lt;/strong&gt;.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;UBC-OCEAN&lt;/strong&gt;: Demonstrates strong performance in &lt;strong&gt;multi-class classification&lt;/strong&gt;, achieving &lt;strong&gt;70.4%, 77.3%, and 86.4% Balanced ACC&lt;/strong&gt; in &lt;strong&gt;4-shot, 8-shot, and 16-shot settings&lt;/strong&gt;, respectively.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;effect-of-foundation-model-selection&quot;&gt;Effect of Foundation Model Selection&lt;/h3&gt;

&lt;p&gt;We evaluated different &lt;strong&gt;foundation models (FMs)&lt;/strong&gt; for feature extraction, including &lt;strong&gt;CONCH, UNI, GPFM, and Virchow&lt;/strong&gt;. &lt;strong&gt;CONCH consistently outperformed others&lt;/strong&gt;, particularly in &lt;strong&gt;low-shot settings&lt;/strong&gt;, highlighting the importance of &lt;strong&gt;vision-language pretrained models&lt;/strong&gt; for pathology tasks.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_focus_fig3.png&quot; alt=&quot;FM Comparison&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;h3 id=&quot;impact-of-prompt-generation&quot;&gt;Impact of Prompt Generation&lt;/h3&gt;

&lt;p&gt;We also examined the effect of &lt;strong&gt;different LLM-generated prompts&lt;/strong&gt;. Prompts generated by &lt;strong&gt;Claude-3.5-Sonnet&lt;/strong&gt; led to the highest &lt;strong&gt;Balanced ACC (86.4%) in 16-shot settings&lt;/strong&gt;, outperforming &lt;strong&gt;GPT-3.5-Turbo (84.9%)&lt;/strong&gt;, demonstrating the impact of &lt;strong&gt;high-quality textual guidance&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/static/img/news/2025_focus_fig4.png&quot; alt=&quot;Prompt Comparison&quot; style=&quot;width: 100%; height: auto;&quot; /&gt;&lt;/p&gt;

&lt;h2 id=&quot;conclusion&quot;&gt;Conclusion&lt;/h2&gt;

&lt;p&gt;FOCUS introduces a &lt;strong&gt;knowledge-enhanced adaptive visual compression framework&lt;/strong&gt; for &lt;strong&gt;few-shot WSI classification&lt;/strong&gt;, addressing key challenges in &lt;strong&gt;data-scarce pathology AI&lt;/strong&gt;. By integrating &lt;strong&gt;foundation models and language priors&lt;/strong&gt;, FOCUS significantly improves &lt;strong&gt;diagnostic feature selection&lt;/strong&gt; and &lt;strong&gt;classification accuracy&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Future work will explore:&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Expanding FOCUS to more computational pathology tasks&lt;/strong&gt;.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Enhancing multi-resolution feature modeling&lt;/strong&gt;.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Applying FOCUS to broader medical AI applications&lt;/strong&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For more details, check out the full paper:&lt;br /&gt;
&lt;strong&gt;Guo, Zhengrui, et al. “FOCUS: Knowledge-Enhanced Adaptive Visual Compression for Few-Shot Whole Slide Image Classification.” arXiv preprint arXiv:2411.14743 (2024).&lt;/strong&gt;&lt;/p&gt;
</content>
  </entry>
  
 
</feed>
