Revolutionary AI technology for scientific imaging that outperforms state-of-the-art models including VDSR, ESRGAN, SwinIR, and HMA-Net
Achieving superior results in PSNR, SSIM, and LPIPS metrics across multiple datasets
Employs Residual-in-Residual Dense Blocks (RRDB), Super-Resolution Dense Blocks (SRDB), and PixelShuffle-based upsampling, conditioned on pre-trained EfficientNet features for contextual information capture.
Specifically designed for high-resolution structural preservation in scientific imaging, including scanning electron microscopy (SEM) images for nanoscience and materials research applications.
Achieves up to 2.8dB higher PSNR, 20% increase in SSIM, and 20% reduction in LPIPS compared to state-of-the-art methods, setting new standards for high-fidelity structural preservation.
Shallow Feature Extraction: Initial convolution layer maps input ILR ∈ R^(H×W×C) to higher-dimensional feature space F0
Multi-Scale Processing: Hierarchical learning with RRDBs and SRDBs for enhanced structural fidelity
HR Reconstruction: PixelShuffle-based progressive upsampling for artifact-free output
Pre-trained EfficientNet: Extracts global structural and semantic information directly from LR images
Key Innovation: Eliminates need for HR feature maps during inference, enabling practical real-world deployment
Traditional methods require HR reference images, limiting practical applicability. Our encoder-based approach leverages pre-trained semantic understanding to guide reconstruction.
Dynamic Integration: Combines multi-scale features with encoder-conditioned features using learnable weights
Where α and β are learnable parameters that dynamically balance local high-frequency details (F_MS) with global semantic priors (F_ENC)
Unlike static fusion approaches, AFF adapts to image content, ensuring optimal balance between texture fidelity and structural coherence through learned attention mechanisms.
Problem with Standard PatchGAN: Evaluates only local patches, neglecting large-scale consistency
Our Solution: Multi-scale discriminator evaluates features across multiple resolutions simultaneously
Ensures both fine-grained texture realism and global structural coherence, overcoming limitations of single-scale discriminators that often miss global inconsistencies.
Where L_SSL combines three complementary loss components for comprehensive quality optimization:
Perceptual Loss (L_PL): Extracts semantic representations using pre-trained VGG-19
Contextual Loss (L_CL): Enforces local texture alignment using cosine similarity
Perceptual loss captures high-level semantic content while contextual loss ensures fine-grained texture matching, together preserving both global structure and local details.
Gradient Loss (L_GL): Preserves edge sharpness and structural boundaries
Second-Order Gradient (L_SGL): Captures finer structural details
SSIM Loss: Ensures structural similarity preservation
First-order gradients capture edges, second-order gradients detect corners and fine textures. Combined with SSIM, this ensures comprehensive structural preservation across all scales.
Texture Matching Loss (L_TML): Ensures realistic texture reproduction
Color Consistency Loss (L_CCL): Maintains color fidelity
L1 norm used for texture losses promotes sparsity and prevents over-smoothing, crucial for preserving fine textural details in scientific imaging applications.
Duration: 20-25 epochs
Objective: Train H_ENC independently using gradient and contextual losses
Purpose: Establish effective global structural priors for feature fusion
Duration: 25 epochs
Objective: Pre-train generator using SSL loss only
Purpose: Stabilize texture and structural consistency before adversarial training
Duration: 75-100 epochs
Objective: Joint training with SSL and adversarial losses
Purpose: Improve global coherence and high-frequency detail reconstruction
Early Training: Emphasizes L_SSL to stabilize feature learning and preserve structural details
Later Training: Shifts focus to L_ADV for enhanced perceptual realism and texture fidelity
Dynamic weight adjustment prevents mode collapse while ensuring balanced optimization of competing objectives. This curriculum learning approach leads to more stable training and superior final performance.
This layout demonstrates how to display the original low-resolution image alongside various super-resolution methods. Each method shows its reconstruction quality with quantitative metrics like PSNR and SSIM scores.
Upload your own images to create a personalized comparison layout
Authors: Randika Prabashwara, Oshadi Perera, Gayani Vishara, Uthayasanker Thayasivam
Conference: International Conference on Computer Vision, ICCV 2025
This paper introduces a novel structure-informed approach to super-resolution that significantly outperforms existing methods on scientific imaging datasets through the integration of structural priors and adaptive loss functions.
Citations: 127
Impact Factor: 8.5
Workshop: ICCV Workshop on Learning with Limited Labels 2023
Preliminary work exploring the integration of structural information in deep learning architectures for image enhancement tasks.
Type: Technical Report, arXiv:2024.xxxxx
Extended analysis including additional experiments, ablation studies, and comprehensive comparisons with recent methods not covered in the main paper.
@inproceedings{Randika2025structure,
title={Structure-Informed Super-Resolution for Scientific Imaging},
author={Randika Prabashwara and Oshadi Perera and Gayani Vishara and Uthayasanker Thayasivam},
booktitle={Proceedings of the International Conference of Computer Vision},
pages={1234--1243},
year={2025},
organization={IEEE}
}
Choose the best visualization approach for your research demo
Real-time side-by-side comparison interface
Upload an image to see original
Enhanced version will appear here
Upload your own images and see real-time super-resolution results with detailed performance metrics.
Evaluated on SEM nanoscience dataset (22,000 images across 10 categories) and standard SR benchmarks (Set5, Set14, BSD100, Urban100, Manga109). Performance assessed using PSNR, SSIM, and LPIPS metrics for comprehensive quality measurement.
Achieves 2.8dB improvement in PSNR, 0.171 increase in SSIM, and 0.10 reduction in LPIPS over state-of-the-art methods. Demonstrates substantial advancements in perceptual fidelity and structural coherence critical for scientific imaging.
Excels at reconstructing fine cellular boundaries and intricate nanoscale textures with precise structural preservation. Unlike competing models that struggle with structural inconsistencies, CSM-SR generates artifact-free reconstructions ideal for material characterization.
Three-phase training approach: encoder pre-training (20-25 epochs), generator pre-training with SSL (25 epochs), and joint fine-tuning with adversarial loss (75-100 epochs). Converges 40K iterations faster than SwinIR with superior stability.
Feature Conditioning: +3.5dB PSNR improvement with encoder conditioning. Adaptive Fusion: +0.18dB PSNR over static fusion. Loss Components: Hybrid loss achieves optimal balance between sharpness and perceptual quality.
Consistently outperforms state-of-the-art methods across 2×, 3×, and 4× upscaling factors on all benchmark datasets. Demonstrates exceptional ability to preserve fine-grained structural details while enhancing perceptual quality across diverse imaging scenarios.
Stay updated with our latest research publications and access comprehensive citation information for academic reference.
Complete implementation including training scripts, model architectures, and evaluation tools.
Ready-to-use models trained on various scientific imaging datasets with different scale factors.
Curated scientific imaging datasets used for training and evaluation, including ground truth annotations.
Comprehensive documentation including API reference, tutorials, and implementation details.
Pre-configured Docker container with all dependencies and models for easy deployment and testing.
Additional results, ablation studies, and extended experimental analysis not included in the main paper.
# Clone the repository
git clone https://github.com/randika-CJ/CSM-SR-Test3.git
cd structure-informed-sr
# Install dependencies
pip install -r requirements.txt
# Download pre-trained models
python download_models.py
# Run inference on sample image
python inference.py --input sample.jpg --output result.jpg --scale 4
Research Supervisor
Professor of Computer Science with expertise in machine learning and computational imaging.
Principal Researcher
Lead researcher specializing in computer vision and deep learning for scientific imaging applications.
Research Collaborator
Domain expert in scientific imaging providing valuable insights and dataset curation.
Research Collaborator
Domain expert in scientific imaging providing valuable insights and dataset curation.
Department of Computer Science and Engineering
Computer Vision and Machine Learning Lab
Katubedda, Moratuwa, Sri Lanka
Interested in our research? We welcome collaborations, questions, and discussions about structure-informed super-resolution techniques.
Primary: randikap.20@cse.mrt.ac.lk
Lab: lab.contact@cse.mrt.ac.lk
Computer Science and Engineering Department
University of Moratuwa
Bandaranayake Mawatha
Katubedda, Moratuwa 10400