SlideShare a Scribd company logo
Enhanced Deep Residual Networks
for Single Image Super-Resolution
Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, and Kyoung Mu Lee
Computer Vision Lab.
Dept. of ECE, ASRI, Seoul National University
http://cv.snu.ac.kr
SISR (Single Image Super Resolution)
Goal: Restoring a HR image from a single LR image
Low-resolution
image
High-resolution
image
Super-Resolution
Lessons from Recent Studies
 Skip connections
 Global and local skip connections enable deep architecture & stable training
 Upscaling methods
 Post-upscaling using sub-pixel convolution is more efficient than pre-upscaling
 However, they are limited that only single-scale SR is possible
SRResNet (CVPR2017)VDSR (CVPR2016)
EDSR MDSR
4 Techniques for Better SR
Need Batch-Normalization?
Increasing model size
Better loss function
Geometric self-ensemble
EDSR
Need Batch-Normalization?
Empirical tests show that removing Batch-Normalization improves the
performance!
Need Batch-Normalization?
 Unlike classification problem,
input and output have similar
distributions
 In SR, normalizing intermediate
features may not be desirable
 Also, can save ~40% of memory
→ Can enlarge the model size
Increasing Model Size
 Empirical test show that
increasing #features is
better than increasing
depth
 Instability occurs when
#features increased up to
256
Given a limited memory, which design is better?
Increasing Model Size
 Residual Scaling Layer
 Increasing #features (up to 256) results instability during training
 Constant scaling layers after each residual path prevents such
instability
Proposed in (Szegedy 2016), “Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning”
Loss Function: L1 vs L2
 Is MSE (L2 loss) the best choice?
 Comparison between different loss functions
 EDSR baseline(16 res-blocks), scale=2, tested on DIV2K images (791~800)
→ MSE is not a good choice!
Geometric Self-Ensemble
 Motivation
 Model ensemble is nice, but expensive!
 How can we achieve an ensemble effect
while avoiding training new models?
Proposed in (Timofte 2016), “Seven ways to improve example-based single-image super-
resolution”
 Method
 Transform test image 8 times with flips and rotations (x8)
 Build 8 outputs and inverse-transform correspondingly
 Average 8 results
Geometric Self-Ensemble
EDSR Summary
 Deeper & Wider: 32 ResBlocks and 256 channels
 Global-local skip connections
 Post-upscaling
 No Batch-Normalization
 Residual scaling
 L1 loss function
 Geometric self-ensemble (EDSR+)
EDSR MDSR
Motivation
 VDSR: Multi-scale SR in a single model
 Multi-scale knowledge transfer
Efficient Multi-Scale Model
 Designing MDSR
 Single vs. Multi-scale learning
 Train & Test method
 EDSR vs. MDSR
MDSR
Motivation
SRCNN, VDSR: A single architecture regardless of upscaling factor
⇨ Multi-scale SR in a single model (VDSR)
FSRCNN, ESPCN, SRResNet: Fast & Efficient, (late upsampling)
but cannot deal with the multiple scales in a single model.
Motivation
FSRCNN, ESPCN, SRResNet
⇨ Different models for different scales?
 Heavy training burden
 Waste of parameters for similar tasks
 Redundancy
Motivation
 Pre-trained scale x2 networks
greatly helps training scale x3
and x4 networks.
 Super-resolution at multiple
scales are inter-related tasks!
Multi-scale knowledge transfer
Designing MDSR
How to make EDSR (post-upscaling) to handle multiscale SR as VDSR?
Requirements
1. Reduce the variance between the different
scales
2. Most parameters are shared across scales
3. For efficiency: Post-upscaling
⇨ Scale-specific pre-processing modules
⇨ main branch
⇨ Scale-specific up-samplers
Train and Test Method
1. Train
 Only one of 3 scale-specific
branches is activated at
each iteration
 A mini-batch consists of
single-scale patches
2. Test
 Select one of the paths
(①~③) according to the
desired SR scale
EDSR vs. MDSR
 Performance:
MDSR ≲ EDSR
 # Parameters:
MDSR << EDSR
(Almost ⅕! + MDSR can handle the multiple scales in a single model)
 Stability:
MDSR << EDSR
(We failed to increase #features
even with residual scaling)
MDSR Summary
 Very deep architecture: 80 ResBlocks
 Most parameters are shared in main branch
 Scale-specific pre-processing modules and up-samplers
 Post-upscaling
 No Batch-Normalization
 L1 loss function
 Geometric self-ensemble (MDSR+)
Results
Training Details
Quantitative Results
Qualitative Results
Qualitative Results
Qualitative Results
Qualitative Results
Qualitative Results
Unknown Track (Challenge)
Unknown Track (Challenge)
Extreme SR (up to x64)
1/64 Scale!
How about extreme cases?
Extreme SR (up to x64)
Bicubic EDSRNN
Extreme SR (up to x64)
Bicubic EDSRNN
Conclusion
1. State-of-the-art single image super-resolution system using
better ResNet structure
2. Techniques to build & train extremely large model
3. A single network to deal with multi-scale SR problem
Thank you!
http://cv.snu.ac.kr

More Related Content

PDF
Single Image Super Resolution Overview
PPTX
Deep learning for image super resolution
PPTX
Dbscan algorithom
PDF
Unlocking the Power of ChatGPT and AI in Testing - A Real-World Look, present...
PDF
Chiller ppt
PPTX
Vision and Scope Document For Library Management System
PDF
Tutorial de Google Classroom
PPT
Digital certificates
Single Image Super Resolution Overview
Deep learning for image super resolution
Dbscan algorithom
Unlocking the Power of ChatGPT and AI in Testing - A Real-World Look, present...
Chiller ppt
Vision and Scope Document For Library Management System
Tutorial de Google Classroom
Digital certificates

What's hot (20)

PDF
Super resolution in deep learning era - Jaejun Yoo
PDF
Deep-Learning Based Stereo Super-Resolution
PDF
Deep learning super resolution
PDF
[OSGeo-KR Tech Workshop] Deep Learning for Single Image Super-Resolution
PPTX
A Deep Journey into Super-resolution
PPTX
Super resolution from a single image
PPTX
Super Resolution
PPTX
Image super resolution
PDF
Super resolution
PDF
Master Thesis of Computer Engineering SuperResoluton Giuseppe Caliendo
PPTX
Artificial Intelligence, Machine Learning and Deep Learning
PDF
Image Enhancement
PPTX
AGE AND GENDER DETECTION.pptx
PPTX
Image Segmentation Using Deep Learning : A survey
PPTX
Transfer Learning and Fine-tuning Deep Neural Networks
PDF
Introduction to XGBoost
PDF
Architecture Design for Deep Neural Networks I
PPTX
Face spoofing detection using texture analysis
PDF
4.intensity transformations
PPTX
Machine Learning - Splitting Datasets
Super resolution in deep learning era - Jaejun Yoo
Deep-Learning Based Stereo Super-Resolution
Deep learning super resolution
[OSGeo-KR Tech Workshop] Deep Learning for Single Image Super-Resolution
A Deep Journey into Super-resolution
Super resolution from a single image
Super Resolution
Image super resolution
Super resolution
Master Thesis of Computer Engineering SuperResoluton Giuseppe Caliendo
Artificial Intelligence, Machine Learning and Deep Learning
Image Enhancement
AGE AND GENDER DETECTION.pptx
Image Segmentation Using Deep Learning : A survey
Transfer Learning and Fine-tuning Deep Neural Networks
Introduction to XGBoost
Architecture Design for Deep Neural Networks I
Face spoofing detection using texture analysis
4.intensity transformations
Machine Learning - Splitting Datasets
Ad

Similar to Enhanced Deep Residual Networks for Single Image Super-Resolution (20)

PDF
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
PPTX
A Fully Progressive approach to Single image super-resolution
DOCX
Image super resolution based on
PPTX
Upscaling Image using Fast Super Resolution Convolution Natural Network
PDF
Rethinking Data Augmentation for Image Super-resolution: A Comprehensive Anal...
PDF
ASE2010
PDF
Research Paper of Image Recognition .02.pdf
PPTX
ABayesianApproachToLocalizedMultiKernelLearningUsingTheRelevanceVectorMachine...
PDF
Review : Multi-Domain Image Completion for Random Missing Input Data [cdm]
PPTX
FINAL REVIEW PPT UpdatedFINAL REVIEW PPT Updated
PDF
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
PPT
Face alignment by deep convolutional network with adaptive learning rate
PPT
lec6a.ppt
PDF
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
PPTX
Regression vs Deep Neural net vs SVM
PPTX
ML_in_QM_JC_02-10-18
PDF
CUDA Accelerated Face Recognition
PDF
Real-Time Face Tracking with GPU Acceleration
PDF
"Energy-efficient Hardware for Embedded Vision and Deep Convolutional Neural ...
PPTX
Building distributed deep learning engine
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
A Fully Progressive approach to Single image super-resolution
Image super resolution based on
Upscaling Image using Fast Super Resolution Convolution Natural Network
Rethinking Data Augmentation for Image Super-resolution: A Comprehensive Anal...
ASE2010
Research Paper of Image Recognition .02.pdf
ABayesianApproachToLocalizedMultiKernelLearningUsingTheRelevanceVectorMachine...
Review : Multi-Domain Image Completion for Random Missing Input Data [cdm]
FINAL REVIEW PPT UpdatedFINAL REVIEW PPT Updated
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)
Face alignment by deep convolutional network with adaptive learning rate
lec6a.ppt
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
Regression vs Deep Neural net vs SVM
ML_in_QM_JC_02-10-18
CUDA Accelerated Face Recognition
Real-Time Face Tracking with GPU Acceleration
"Energy-efficient Hardware for Embedded Vision and Deep Convolutional Neural ...
Building distributed deep learning engine
Ad

More from NAVER Engineering (20)

PDF
React vac pattern
PDF
디자인 시스템에 직방 ZUIX
PDF
진화하는 디자인 시스템(걸음마 편)
PDF
서비스 운영을 위한 디자인시스템 프로젝트
PDF
BPL(Banksalad Product Language) 무야호
PDF
이번 생에 디자인 시스템은 처음이라
PDF
날고 있는 여러 비행기 넘나 들며 정비하기
PDF
쏘카프레임 구축 배경과 과정
PDF
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
PDF
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
PDF
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
PDF
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
PDF
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
PDF
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
PDF
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
PDF
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
PDF
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
PDF
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
PDF
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
PDF
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기
React vac pattern
디자인 시스템에 직방 ZUIX
진화하는 디자인 시스템(걸음마 편)
서비스 운영을 위한 디자인시스템 프로젝트
BPL(Banksalad Product Language) 무야호
이번 생에 디자인 시스템은 처음이라
날고 있는 여러 비행기 넘나 들며 정비하기
쏘카프레임 구축 배경과 과정
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기

Recently uploaded (20)

PDF
gpt5_lecture_notes_comprehensive_20250812015547.pdf
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
DOCX
The AUB Centre for AI in Media Proposal.docx
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PPTX
ACSFv1EN-58255 AWS Academy Cloud Security Foundations.pptx
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PPT
Teaching material agriculture food technology
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Empathic Computing: Creating Shared Understanding
PDF
cuic standard and advanced reporting.pdf
PDF
Review of recent advances in non-invasive hemoglobin estimation
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
gpt5_lecture_notes_comprehensive_20250812015547.pdf
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
The AUB Centre for AI in Media Proposal.docx
Per capita expenditure prediction using model stacking based on satellite ima...
Mobile App Security Testing_ A Comprehensive Guide.pdf
ACSFv1EN-58255 AWS Academy Cloud Security Foundations.pptx
Network Security Unit 5.pdf for BCA BBA.
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Digital-Transformation-Roadmap-for-Companies.pptx
“AI and Expert System Decision Support & Business Intelligence Systems”
Dropbox Q2 2025 Financial Results & Investor Presentation
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Teaching material agriculture food technology
Diabetes mellitus diagnosis method based random forest with bat algorithm
Empathic Computing: Creating Shared Understanding
cuic standard and advanced reporting.pdf
Review of recent advances in non-invasive hemoglobin estimation
Building Integrated photovoltaic BIPV_UPV.pdf

Enhanced Deep Residual Networks for Single Image Super-Resolution

  • 1. Enhanced Deep Residual Networks for Single Image Super-Resolution Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, and Kyoung Mu Lee Computer Vision Lab. Dept. of ECE, ASRI, Seoul National University http://cv.snu.ac.kr
  • 2. SISR (Single Image Super Resolution) Goal: Restoring a HR image from a single LR image Low-resolution image High-resolution image Super-Resolution
  • 3. Lessons from Recent Studies  Skip connections  Global and local skip connections enable deep architecture & stable training  Upscaling methods  Post-upscaling using sub-pixel convolution is more efficient than pre-upscaling  However, they are limited that only single-scale SR is possible SRResNet (CVPR2017)VDSR (CVPR2016)
  • 5. 4 Techniques for Better SR Need Batch-Normalization? Increasing model size Better loss function Geometric self-ensemble EDSR
  • 6. Need Batch-Normalization? Empirical tests show that removing Batch-Normalization improves the performance!
  • 7. Need Batch-Normalization?  Unlike classification problem, input and output have similar distributions  In SR, normalizing intermediate features may not be desirable  Also, can save ~40% of memory → Can enlarge the model size
  • 8. Increasing Model Size  Empirical test show that increasing #features is better than increasing depth  Instability occurs when #features increased up to 256 Given a limited memory, which design is better?
  • 9. Increasing Model Size  Residual Scaling Layer  Increasing #features (up to 256) results instability during training  Constant scaling layers after each residual path prevents such instability Proposed in (Szegedy 2016), “Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning”
  • 10. Loss Function: L1 vs L2  Is MSE (L2 loss) the best choice?  Comparison between different loss functions  EDSR baseline(16 res-blocks), scale=2, tested on DIV2K images (791~800) → MSE is not a good choice!
  • 11. Geometric Self-Ensemble  Motivation  Model ensemble is nice, but expensive!  How can we achieve an ensemble effect while avoiding training new models? Proposed in (Timofte 2016), “Seven ways to improve example-based single-image super- resolution”  Method  Transform test image 8 times with flips and rotations (x8)  Build 8 outputs and inverse-transform correspondingly  Average 8 results
  • 13. EDSR Summary  Deeper & Wider: 32 ResBlocks and 256 channels  Global-local skip connections  Post-upscaling  No Batch-Normalization  Residual scaling  L1 loss function  Geometric self-ensemble (EDSR+)
  • 15. Motivation  VDSR: Multi-scale SR in a single model  Multi-scale knowledge transfer Efficient Multi-Scale Model  Designing MDSR  Single vs. Multi-scale learning  Train & Test method  EDSR vs. MDSR MDSR
  • 16. Motivation SRCNN, VDSR: A single architecture regardless of upscaling factor ⇨ Multi-scale SR in a single model (VDSR) FSRCNN, ESPCN, SRResNet: Fast & Efficient, (late upsampling) but cannot deal with the multiple scales in a single model.
  • 17. Motivation FSRCNN, ESPCN, SRResNet ⇨ Different models for different scales?  Heavy training burden  Waste of parameters for similar tasks  Redundancy
  • 18. Motivation  Pre-trained scale x2 networks greatly helps training scale x3 and x4 networks.  Super-resolution at multiple scales are inter-related tasks! Multi-scale knowledge transfer
  • 19. Designing MDSR How to make EDSR (post-upscaling) to handle multiscale SR as VDSR? Requirements 1. Reduce the variance between the different scales 2. Most parameters are shared across scales 3. For efficiency: Post-upscaling ⇨ Scale-specific pre-processing modules ⇨ main branch ⇨ Scale-specific up-samplers
  • 20. Train and Test Method 1. Train  Only one of 3 scale-specific branches is activated at each iteration  A mini-batch consists of single-scale patches 2. Test  Select one of the paths (①~③) according to the desired SR scale
  • 21. EDSR vs. MDSR  Performance: MDSR ≲ EDSR  # Parameters: MDSR << EDSR (Almost ⅕! + MDSR can handle the multiple scales in a single model)  Stability: MDSR << EDSR (We failed to increase #features even with residual scaling)
  • 22. MDSR Summary  Very deep architecture: 80 ResBlocks  Most parameters are shared in main branch  Scale-specific pre-processing modules and up-samplers  Post-upscaling  No Batch-Normalization  L1 loss function  Geometric self-ensemble (MDSR+)
  • 33. Extreme SR (up to x64) 1/64 Scale! How about extreme cases?
  • 34. Extreme SR (up to x64) Bicubic EDSRNN
  • 35. Extreme SR (up to x64) Bicubic EDSRNN
  • 36. Conclusion 1. State-of-the-art single image super-resolution system using better ResNet structure 2. Techniques to build & train extremely large model 3. A single network to deal with multi-scale SR problem