Multiclass Liver Cirrhosis Segmentation and Severity Staging on T2-Weighted MRI Using 3UResNet Explainability and Per-Class Deep Expert Classifiers

1University of Barishal, 2Bangladesh Open University, 3Patuakhali Science and Technology University, 4AppsCode Inc., 5Jahangirnagar University, 6Shaheed Suhrawardy Medical College and Hospital, 7Military Institute of Science and Technology
Class 1

Class A (Mild)

Class 2

Class B (Moderate)

Class 3

Class C (Severe)

Liver Cirrhosis Classification: Visualizing the Progression Across 3 Critical Severity Levels.

Methodology Diagram

Stepwise procedure of (A) Segmentation and (B) Classification of liver cirrhosis based on chirrMRI600+ dataset.

Abstract

Liver cirrhosis demands accurate, interpretable MRI analysis to support timely diagnosis and staging, yet existing methods often decouple segmentation from classification and lack the explainability required for clinical use. This work introduces 3UResNet, an ensemble of U-Net, TransUNet++, and Attention U-Net with ResNet50 as encoder, designed for robust multiclass cirrhosis segmentation on 2D T2-weighted MRI from the CirrMRI600+ dataset. A total of 5,364 training, 674 validation, and 664 test samples across mild, moderate, and severe classes were preprocessed with augmentation, normalization, and mask binarization. This research also designed three Vision Transformer–based models, namely SwinUnet, CCTUnet, and EaNetUnet for segmentation. On the test set, 3UResNet achieved a Dice score of 0.9516 and a mean Intersection over Union of 0.9077 across all classes, outperforming the transformer baselines. After accurate segmentation, this study used the segmented masks to extract shape features and applied convolutional neural networks (CNN) to extract visual features for classification. DenseNet121 performed best for mild cirrhosis (95%), while CoAnNet achieved higher accuracy for moderate (80%) and severe cirrhosis (94%). Here, a WeightedRandom sampler and focal loss were employed to mitigate class imbalance during training, ensuring results reflect real-world clinical distributions. Model interpretability is strengthened with LIME and Grad-CAM, highlighting clinically meaningful liver regions and aligning attention with cirrhotic anatomy across severity levels. This study demonstrates a practical end-to-end framework uniting high-fidelity segmentation with severity staging and transparent rationale, paving the way for clinical adoption.

Proposed Framework: 3UResNet Pipeline

🔴 The Problem

  • Existing methods often decouple segmentation from classification, leading to loss of diagnostic context.
  • Current AI tools lack the explainability required for clinical trust.
  • Class imbalance in medical datasets hinders accurate staging of severity.

🟡 Why We Solved It

  • To enable automated, timely diagnosis and precise staging of Liver Cirrhosis on T2-Weighted MRI.
  • To outperform single-model Baselines (e.g., SwinUnet) using a robust ensemble approach.
  • To bridge the gap between pixel-level segmentation and patient-level diagnosis.

🟢 How We Solved It (3UResNet)

  • Proposed Model: 3UResNet, an ensemble of U-Net, TransUNet++, and Attention U-Net with a ResNet50 encoder.
  • Pipeline: First, precise segmentation (Dice: 0.9516), followed by extraction of Shape & Visual features.
  • Classification: Utilized DenseNet121 (for Mild) and CoAnNet (for Moderate/Severe) with WeightedRandom Sampler to mitigate imbalance.
Architecture of 3UResNet

Figure: The proposed 3UResNet Ensemble Architecture and Classification Pipeline.

Performance Evaluation

Comprehensive analysis of the proposed 3UResNet model demonstrating superior accuracy and clinical relevance.

Quantitative Results: Segmentation

Performance comparison of 3UResNet against Transformer-based and CNN-based baselines using Dice Score and Mean IoU.

Table 1: Segmentation Performance of Integrated Transformer & ViT Models

Model Architecture Mild Moderate Severe
Encoder Decoder Dice mIoU Dice mIoU Dice mIoU
ResNet50TransUNetPP (3UResNet) 0.95320.91060.95780.91910.93100.8710
ResNet50Attention U-Net 0.96050.92390.95720.91800.93260.8738
ResNet50UNet 0.95560.91500.95340.91090.93300.8745
ResNet50LinTransUnet 0.94160.88960.95590.91560.92820.8661
VGG16U²-Net 0.93150.87180.94110.88880.85630.7487
InceptionV3SynergyNet 0.95750.91840.95460.91310.93390.8759
Swin-Unet 0.93190.87250.94090.88840.91730.8472
CCT-Unet 0.93210.87280.93470.87740.92810.8658
EaNet 0.92870.86690.93900.88500.92420.8592

Classification Performance & Analysis

Evaluation of the proposed Hybrid Fusion Models and comparison with state-of-the-art studies.

Table 2: Classification Performance (SubsetRandomSampler)

Model Mild Moderate Severe
Feature Extractor Prec.Rec.Acc. Prec.Rec.Acc. Prec.Rec.Acc.
DenseNet-121 0.880.650.65 0.230.360.36 0.560.600.60
ResNet-101 0.830.710.71 0.230.380.38 0.510.400.40
EfficientNet-b0 0.840.710.71 0.230.310.31 0.550.590.59
ResNet34 0.900.710.71 0.250.420.42 0.520.480.48
ResNet50+GAP 0.750.770.77 0.270.080.08 0.480.710.71

Table 3: Classification Performance (WeightedRandomSampler)

Model Mild Moderate Severe
Feature Extractor Prec.Rec.Acc. Prec.Rec.Acc. Prec.Rec.Acc.
Hybrid7Net6 (Proposed) 0.970.770.77 0.560.770.77 0.670.750.75
DenseNet-121 0.830.950.95 0.130.050.05 0.640.770.77
ConvNeXt 0.820.950.95 0.200.110.11 0.630.660.63
CoAtNet 0.780.860.86 0.210.140.14 0.510.540.54
Hybrid10Net 0.840.910.91 0.790.260.26 0.580.770.77

Table 4: Training Configurations

Configuration DenseNet121 (Mild) CoAtNet (Moderate) CoAtNet (Severe)
Data Augmentation Resize, HFlip 0.5, ShiftScaleRotate 0.05, BrightnessContrast, GaussNoise Same base + stronger minority aug (HFlip 0.8, VFlip 0.5, more noise) Same as Moderate with focus on Severe minority
Loss Function FocalLoss (alpha=None, gamma=2.0/3.0) FocalLoss with class weights (Moderate weighted 1.5) FocalLoss with class weights (no multiplier)
Sampler WeightedRandomSampler oversampling rare classes WeightedRandomSampler with inverse frequency + class weights Same as Moderate
Optimizer AdamW + ReduceLROnPlateau Same as DenseNet121 Same as DenseNet121

Table 5: Comparison of Studies on Liver Chirrhosis

Study Dataset Segmentation Method Classification Method Performance Explainability
Jha et al. [2025] T2W 2D MedSegDiff N/A Dice: 0.76 Not discussed
Gupta et al. [2025] T2W 2D 2D nnU-Net Masked Attn ResNet-18 Dice: 0.96, Acc: 78% Grad-CAM
This Study T2W 2D 3UResNet50 Hybrid7Net6 Dice: ~0.95 | Acc: Mild 0.95, Mod 0.80, Sev 0.94 LIME + Grad-CAM

Key Findings

Summary of the major breakthroughs achieved in this study.

🎯

Superior Segmentation

The proposed 3UResNet model achieved a Dice Score of ~0.95, significantly outperforming standard ViT and CNN baselines by preserving local anatomical details.

⚖️

Robust Classification

Our Hybrid7Net6 framework successfully handled class imbalance, achieving high accuracy (Mild: 95%, Severe: 94%) using WeightedRandomSampler.

🧠

Explainable AI (XAI)

Integration of LIME and Grad-CAM provided visual validation, confirming that the model focuses on relevant liver regions (e.g., fragmentation) for decision making.

🏥

Clinical Applicability

The end-to-end pipeline demonstrates that automated systems can reliably grade cirrhosis severity from 2D MRI slices without invasive biopsy.

Future Directions

Prospective avenues to further enhance the system's capability and deployment.

  • 3D Volume Analysis: Extending the current 2D slice-based approach to 3D volumetric analysis to capture the complete spatial context of liver fibrosis.
  • Multi-Modal Integration: Incorporating clinical data (blood biomarkers, patient history) alongside MRI images to improve diagnostic precision.
  • Federated Learning: Implementing privacy-preserving federated learning to train models across multiple hospitals without sharing sensitive patient data.
  • Real-Time Deployment: Developing a lightweight version of the model for mobile applications or real-time clinical support systems.

Honors & Presentations

Our research has been recognized and presented at prestigious academic forums.

🏆

GPC Research & Creative Activities Forum

University of Missouri (Fall 2025)

🥈 2nd Place
Research Poster Competition
🏆 People’s Choice Award
Most Recognized Research Poster
🏛️

Stanford AI+Health Conference

Stanford University (2025)

📢 Featured Presentation
Selected for poster presentation among top research works in AI & Healthcare.

*Presented as a research poster to the global AI health community.

BibTeX

@article{setu2025explainable,
  author    = {Setu, Deblina Mazumder and Dey, Samrat Kumar and Islam, Tania and Howlader, Arpita and Biswas, Saurov Chandra and Mazumder, Rashed and Siddiqi, Umme Raihan and Rahman, Md. Mahbubur},
  title     = {Multiclass Liver Cirrhosis Segmentation and Severity Staging on T2-Weighted MRI Using 3UResNet Explainability and Per-Class Deep Expert Classifiers},
  journal   = {},
  year      = {},
}