SlideShare a Scribd company logo
“ ”
Perception and Intelligence Lab,
School of ECE, ASRI, Seoul National University,
Seoul, Korea
Presenter :
2019.03.15
Backbone Can Not be Trained at Once:
Rolling Back to Pre-trained Network
for Person Re-identification
AAAI-19
Youngmin Ro , Jongwon Choi , Dae Ung Jo ,
Byeongho Heo , Jongin Lim and Jin Young Choi
Youngmin Ro
• From the non-overlapped cameras,
• Find who is the same person.
• Simply call it ReID
Perception and Intelligence Lab., Copyright © 2016 2
What is Person Re-identification
http://www.liangzheng.com.cn/Project/project_prw.html
Query
ReID dataset
Gallery
• There are two strategies for ReID
• Triplet loss
• Cross entropy loss
Perception and Intelligence Lab., Copyright © 2016 3
How to solve ReID
CNN
Anchor Positive Negative
Anchor
Positive
Negative
Negative pair
Positive pair
CNN
Class: 1 2 3 4 5
010000
labelprediction
§ Consider each identities to classes
Person Re-identification: Past, Present and Future(2016 arXiv Liang Zheng et al.)
−
• Using Pre-trained on the large dataset (e.g., ImageNet à 1000class)
Re-define Classifier to target dataset (e.g., Market-1501 à 751class)
1. Only train Classifier with freezing Feature extractor
2. Train Classifier and Feature extractor together.
Perception and Intelligence Lab., Copyright © 2016 4
Fine tuning in ReID
conv
conv
conv
Block1 Block2 Block3
conv
conv
Block4
conv
conv
GAP
Block5
conv
conv
FC
Feature à 1D vector
Classify target classes
Feature extractor Classifier
• The superficial issue of ReID is to solve the following problems;
pose variation, misalignment, background clutter, occlusion,
missing body parts, illumination change
Perception and Intelligence Lab., Copyright © 2016 5
Issues of ReID
Market-1501 dataset
Perception and Intelligence Lab., Copyright © 2016 6
Another issue of ReID
• In ReID, all dataset suffer from the insufficient training set
• Insufficient training set can yield overfitting problem.
• In Market-1501 dataset,
training set has 751 classes(id) and about 10 images per class(id).
(MNIST 10 class /5000 per class)
(CIFAR 100 class/ 500 per class)
Unlabeled samples generated by gan improve the person re-identification
baseline in vitro (ICCV2017 Zhedong Zheng et al)
• An effective way to solve overfitting problems during network training
Perception and Intelligence Lab., Copyright © 2016 7
What is the Rollback?
Drop learning rate
• Generally, ReID uses the ImageNet pre-trained network.
• We need to sufficiently fine-tune the low-level layers to improve the
discriminant power for the specific class ‘person’
Perception and Intelligence Lab., Copyright © 2016 8
Why Rollback?
“cherry”“container ship”
“mite” “grille” “mushroom”
“person”“leopard”
Fine tune
• Low-level layers contain detail information.
• High-level layers contain semantic information
Perception and Intelligence Lab., Copyright © 2016 9
The roll of the conv filter
conv
conv
conv
Block1 Block2 Block3
conv
conv
Block4
conv
conv
FC
Block5
conv
conv
Feature extractor Classifier
low level layers
[edge, blob/texture]
high level layers
[object parts]
FC layers
[object classes]
• Rollback to the pre-trained weight.
• Rollback gives a chance for low-level layers to trained more.
Perception and Intelligence Lab., Copyright © 2016 10
What is the Rollback?
Block1 Block2 Block3 Block4
FC
Block5
conv
conv
conv
conv
conv
conv
conv
conv
conv
Feature extractor Classifier
conv
conv
conv
conv
conv
conv
conv
FC
conv
conv
Fine-tuned
Network
(once train starts from pre-trained network)
Pre-trained
Network
From ImageNet
High-level blocks are rollbacked
to weights of pre-trained one
• Rollback to the pre-trained weight.
• Rollback gives a chance for low-level layers to trained more.
Perception and Intelligence Lab., Copyright © 2016 11
What is the Rollback?
Block1 Block2 Block3 Block4
FC
Block5
conv
conv
conv
conv
conv
conv
conv
conv
conv
conv
conv
conv
Block1 Block2 Block3
conv
conv
Block4
conv
conv
FC
Block5
conv
conv
Fine-tuned
Network
(once train starts from pre-trained network)
Pre-trained
Network
From ImageNet
Strengthened!!
Perception and Intelligence Lab., Copyright © 2016 12
Analysis of Rollback (Rollback w/o one block)
0.0005
0.005
0.05
0 10 20 30
baseline
Block1
Block2
Block3
Block4
Block5
epochs
trainloss
Remaining low-level blocks (Block1, Block2, Block3)
shows slower convergence, but better generalization
performance
• When train loss does not decrease anymore, rollback high-level layers
Perception and Intelligence Lab., Copyright © 2016 13
When/Where the Rollback?
Remain blocks Rollback blocks
1st B1 B2+B3+B4+B5
2nd B1+B2 B3+B4+B5
3rd B1+B2+B3 B4+B5
• Where (In the incremental manner)
0.0003
0.003
0.03
0.3
0 10 20 30 40
epochs
trainloss
learning rate: 0.01
Does not decrease
learning rate: 0.001
Perception and Intelligence Lab., Copyright © 2016 14
Power of the Rollback
• Rollback and Rollback again, then performance grows up!!
Red line(our)
Blue line(base)
Perception and Intelligence Lab., Copyright © 2016 15
Generalization of Rollback
• Rollback can be adopted to various architecture and various task.
Perception and Intelligence Lab., Copyright © 2016 16
This Simple Rollback achieved S-O-T-A
• No additional information!
• No additional structure!
• No data augmentation! (only horizontal flip)
CVPR2018
Discrimination ability
• Rollback can generate a more distinguishable feature than the baseline.
Perception and Intelligence Lab., Copyright © 2016 18
Conclusion
• In this paper, we proposed a refine tuning method with a rolling-back
scheme which further enhances the backbone network.
• The key idea of the rolling-back scheme is to restore the weights in a part
of the backbone network to the pre-trained weights when the fine-tuning
converges at a premature state.
• To escape from the premature state, we adopt an incremental refine tuning
strategy by applying the fine tuning repeatedly, along with the rolling-back.
• According to the experimental results, the rolling-back scheme makes a
meaningful contribution to enhancement of the backbone network where it
derives the convergence to a local basin of a good generalization
performance.
• As a result, our method without any add-on scheme could outperform the
state-of-the-arts with help of add-on scheme.
19
Thank you!!
youngmin4920@gmail.com

More Related Content

PDF
Learning to learn unlearned feature for segmentation
PDF
Learning loss for active learning
PDF
Human uncertainty makes classification more robust, ICCV 2019 Review
PDF
Cross-domain complementary learning with synthetic data for multi-person part...
PPTX
Deep Learning in Robotics
PPTX
InfoGAIL
PDF
Step zhedong
PDF
Image Translation with GAN
Learning to learn unlearned feature for segmentation
Learning loss for active learning
Human uncertainty makes classification more robust, ICCV 2019 Review
Cross-domain complementary learning with synthetic data for multi-person part...
Deep Learning in Robotics
InfoGAIL
Step zhedong
Image Translation with GAN

What's hot (20)

PDF
Deep Learning for Computer Vision: A comparision between Convolutional Neural...
PDF
161209 Unsupervised Learning of Video Representations using LSTMs
PDF
Unsupervised visual representation learning overview: Toward Self-Supervision
PDF
Scene classification using Convolutional Neural Networks - Jayani Withanawasam
PDF
Deep Learning for Information Retrieval: Models, Progress, & Opportunities
PDF
do adversarially robust image net models transfer better
PDF
Recent Trends in Deep Learning
PDF
A Neural Grammatical Error Correction built on Better Pre-training and Sequen...
PPTX
Deep Learning in Computer Vision
PDF
Improving neural question generation using answer separation
PDF
[PR12] Generative Models as Distributions of Functions
PPTX
Deep Learning Models for Question Answering
PPTX
AlexNet
PDF
Model-Based Reinforcement Learning @NIPS2017
PDF
Variants of GANs - Jaejun Yoo
PDF
Convolutional Neural Networks (CNN)
PDF
[CVPR2020] Simple but effective image enhancement techniques
PPTX
Few shot learning/ one shot learning/ machine learning
PPTX
One shot learning
PDF
Unsupervised Cross-Domain Image Generation
Deep Learning for Computer Vision: A comparision between Convolutional Neural...
161209 Unsupervised Learning of Video Representations using LSTMs
Unsupervised visual representation learning overview: Toward Self-Supervision
Scene classification using Convolutional Neural Networks - Jayani Withanawasam
Deep Learning for Information Retrieval: Models, Progress, & Opportunities
do adversarially robust image net models transfer better
Recent Trends in Deep Learning
A Neural Grammatical Error Correction built on Better Pre-training and Sequen...
Deep Learning in Computer Vision
Improving neural question generation using answer separation
[PR12] Generative Models as Distributions of Functions
Deep Learning Models for Question Answering
AlexNet
Model-Based Reinforcement Learning @NIPS2017
Variants of GANs - Jaejun Yoo
Convolutional Neural Networks (CNN)
[CVPR2020] Simple but effective image enhancement techniques
Few shot learning/ one shot learning/ machine learning
One shot learning
Unsupervised Cross-Domain Image Generation
Ad

Similar to Backbone can not be trained at once rolling back to pre trained network for person re-identification (20)

PDF
ct_meeting_final_jcy (1).pdf
PDF
2019 cvpr paper overview by Ho Seong Lee
PDF
2019 cvpr paper_overview
PDF
Learning where to look: focus and attention in deep vision
PDF
Deep learning for person re-identification
PDF
The Opportunities and Challenges of Putting the Latest Computer Vision and De...
PDF
Tutorial inns2019 full
PPTX
Human age and gender Detection
PPTX
Deep learning and computer vision
PDF
物件偵測與辨識技術
PDF
[CVPR 2018] Utilizing unlabeled or noisy labeled data (classification, detect...
PDF
Information to Wisdom: Commonsense Knowledge Extraction and Compilation - Part 3
PDF
Sparse representation based human action recognition using an action region-a...
PDF
Practical computer vision-- A problem-driven approach towards learning CV/ML/DL
PDF
Introduction to the Artificial Intelligence and Computer Vision revolution
PDF
Introduction talk to Computer Vision
PDF
PR-207: YOLOv3: An Incremental Improvement
PPTX
Fa19_P1.pptx
PDF
Computer vision for transportation
PDF
Lecture 2.B: Computer Vision Applications - Full Stack Deep Learning - Spring...
ct_meeting_final_jcy (1).pdf
2019 cvpr paper overview by Ho Seong Lee
2019 cvpr paper_overview
Learning where to look: focus and attention in deep vision
Deep learning for person re-identification
The Opportunities and Challenges of Putting the Latest Computer Vision and De...
Tutorial inns2019 full
Human age and gender Detection
Deep learning and computer vision
物件偵測與辨識技術
[CVPR 2018] Utilizing unlabeled or noisy labeled data (classification, detect...
Information to Wisdom: Commonsense Knowledge Extraction and Compilation - Part 3
Sparse representation based human action recognition using an action region-a...
Practical computer vision-- A problem-driven approach towards learning CV/ML/DL
Introduction to the Artificial Intelligence and Computer Vision revolution
Introduction talk to Computer Vision
PR-207: YOLOv3: An Incremental Improvement
Fa19_P1.pptx
Computer vision for transportation
Lecture 2.B: Computer Vision Applications - Full Stack Deep Learning - Spring...
Ad

More from NAVER Engineering (20)

PDF
React vac pattern
PDF
디자인 시스템에 직방 ZUIX
PDF
진화하는 디자인 시스템(걸음마 편)
PDF
서비스 운영을 위한 디자인시스템 프로젝트
PDF
BPL(Banksalad Product Language) 무야호
PDF
이번 생에 디자인 시스템은 처음이라
PDF
날고 있는 여러 비행기 넘나 들며 정비하기
PDF
쏘카프레임 구축 배경과 과정
PDF
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
PDF
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
PDF
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
PDF
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
PDF
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
PDF
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
PDF
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
PDF
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
PDF
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
PDF
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
PDF
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
PDF
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기
React vac pattern
디자인 시스템에 직방 ZUIX
진화하는 디자인 시스템(걸음마 편)
서비스 운영을 위한 디자인시스템 프로젝트
BPL(Banksalad Product Language) 무야호
이번 생에 디자인 시스템은 처음이라
날고 있는 여러 비행기 넘나 들며 정비하기
쏘카프레임 구축 배경과 과정
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기

Recently uploaded (20)

PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
cuic standard and advanced reporting.pdf
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PPTX
Cloud computing and distributed systems.
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
KodekX | Application Modernization Development
PDF
Review of recent advances in non-invasive hemoglobin estimation
PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
Empathic Computing: Creating Shared Understanding
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PDF
Encapsulation theory and applications.pdf
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
Unlocking AI with Model Context Protocol (MCP)
cuic standard and advanced reporting.pdf
Network Security Unit 5.pdf for BCA BBA.
Chapter 3 Spatial Domain Image Processing.pdf
Cloud computing and distributed systems.
Reach Out and Touch Someone: Haptics and Empathic Computing
KodekX | Application Modernization Development
Review of recent advances in non-invasive hemoglobin estimation
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
NewMind AI Weekly Chronicles - August'25 Week I
The Rise and Fall of 3GPP – Time for a Sabbatical?
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Encapsulation_ Review paper, used for researhc scholars
Empathic Computing: Creating Shared Understanding
20250228 LYD VKU AI Blended-Learning.pptx
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
Encapsulation theory and applications.pdf
Build a system with the filesystem maintained by OSTree @ COSCUP 2025

Backbone can not be trained at once rolling back to pre trained network for person re-identification

  • 1. “ ” Perception and Intelligence Lab, School of ECE, ASRI, Seoul National University, Seoul, Korea Presenter : 2019.03.15 Backbone Can Not be Trained at Once: Rolling Back to Pre-trained Network for Person Re-identification AAAI-19 Youngmin Ro , Jongwon Choi , Dae Ung Jo , Byeongho Heo , Jongin Lim and Jin Young Choi Youngmin Ro
  • 2. • From the non-overlapped cameras, • Find who is the same person. • Simply call it ReID Perception and Intelligence Lab., Copyright © 2016 2 What is Person Re-identification http://www.liangzheng.com.cn/Project/project_prw.html Query ReID dataset Gallery
  • 3. • There are two strategies for ReID • Triplet loss • Cross entropy loss Perception and Intelligence Lab., Copyright © 2016 3 How to solve ReID CNN Anchor Positive Negative Anchor Positive Negative Negative pair Positive pair CNN Class: 1 2 3 4 5 010000 labelprediction § Consider each identities to classes Person Re-identification: Past, Present and Future(2016 arXiv Liang Zheng et al.) −
  • 4. • Using Pre-trained on the large dataset (e.g., ImageNet à 1000class) Re-define Classifier to target dataset (e.g., Market-1501 à 751class) 1. Only train Classifier with freezing Feature extractor 2. Train Classifier and Feature extractor together. Perception and Intelligence Lab., Copyright © 2016 4 Fine tuning in ReID conv conv conv Block1 Block2 Block3 conv conv Block4 conv conv GAP Block5 conv conv FC Feature à 1D vector Classify target classes Feature extractor Classifier
  • 5. • The superficial issue of ReID is to solve the following problems; pose variation, misalignment, background clutter, occlusion, missing body parts, illumination change Perception and Intelligence Lab., Copyright © 2016 5 Issues of ReID Market-1501 dataset
  • 6. Perception and Intelligence Lab., Copyright © 2016 6 Another issue of ReID • In ReID, all dataset suffer from the insufficient training set • Insufficient training set can yield overfitting problem. • In Market-1501 dataset, training set has 751 classes(id) and about 10 images per class(id). (MNIST 10 class /5000 per class) (CIFAR 100 class/ 500 per class) Unlabeled samples generated by gan improve the person re-identification baseline in vitro (ICCV2017 Zhedong Zheng et al)
  • 7. • An effective way to solve overfitting problems during network training Perception and Intelligence Lab., Copyright © 2016 7 What is the Rollback? Drop learning rate
  • 8. • Generally, ReID uses the ImageNet pre-trained network. • We need to sufficiently fine-tune the low-level layers to improve the discriminant power for the specific class ‘person’ Perception and Intelligence Lab., Copyright © 2016 8 Why Rollback? “cherry”“container ship” “mite” “grille” “mushroom” “person”“leopard” Fine tune
  • 9. • Low-level layers contain detail information. • High-level layers contain semantic information Perception and Intelligence Lab., Copyright © 2016 9 The roll of the conv filter conv conv conv Block1 Block2 Block3 conv conv Block4 conv conv FC Block5 conv conv Feature extractor Classifier low level layers [edge, blob/texture] high level layers [object parts] FC layers [object classes]
  • 10. • Rollback to the pre-trained weight. • Rollback gives a chance for low-level layers to trained more. Perception and Intelligence Lab., Copyright © 2016 10 What is the Rollback? Block1 Block2 Block3 Block4 FC Block5 conv conv conv conv conv conv conv conv conv Feature extractor Classifier conv conv conv conv conv conv conv FC conv conv Fine-tuned Network (once train starts from pre-trained network) Pre-trained Network From ImageNet High-level blocks are rollbacked to weights of pre-trained one
  • 11. • Rollback to the pre-trained weight. • Rollback gives a chance for low-level layers to trained more. Perception and Intelligence Lab., Copyright © 2016 11 What is the Rollback? Block1 Block2 Block3 Block4 FC Block5 conv conv conv conv conv conv conv conv conv conv conv conv Block1 Block2 Block3 conv conv Block4 conv conv FC Block5 conv conv Fine-tuned Network (once train starts from pre-trained network) Pre-trained Network From ImageNet Strengthened!!
  • 12. Perception and Intelligence Lab., Copyright © 2016 12 Analysis of Rollback (Rollback w/o one block) 0.0005 0.005 0.05 0 10 20 30 baseline Block1 Block2 Block3 Block4 Block5 epochs trainloss Remaining low-level blocks (Block1, Block2, Block3) shows slower convergence, but better generalization performance
  • 13. • When train loss does not decrease anymore, rollback high-level layers Perception and Intelligence Lab., Copyright © 2016 13 When/Where the Rollback? Remain blocks Rollback blocks 1st B1 B2+B3+B4+B5 2nd B1+B2 B3+B4+B5 3rd B1+B2+B3 B4+B5 • Where (In the incremental manner) 0.0003 0.003 0.03 0.3 0 10 20 30 40 epochs trainloss learning rate: 0.01 Does not decrease learning rate: 0.001
  • 14. Perception and Intelligence Lab., Copyright © 2016 14 Power of the Rollback • Rollback and Rollback again, then performance grows up!! Red line(our) Blue line(base)
  • 15. Perception and Intelligence Lab., Copyright © 2016 15 Generalization of Rollback • Rollback can be adopted to various architecture and various task.
  • 16. Perception and Intelligence Lab., Copyright © 2016 16 This Simple Rollback achieved S-O-T-A • No additional information! • No additional structure! • No data augmentation! (only horizontal flip) CVPR2018
  • 17. Discrimination ability • Rollback can generate a more distinguishable feature than the baseline.
  • 18. Perception and Intelligence Lab., Copyright © 2016 18 Conclusion • In this paper, we proposed a refine tuning method with a rolling-back scheme which further enhances the backbone network. • The key idea of the rolling-back scheme is to restore the weights in a part of the backbone network to the pre-trained weights when the fine-tuning converges at a premature state. • To escape from the premature state, we adopt an incremental refine tuning strategy by applying the fine tuning repeatedly, along with the rolling-back. • According to the experimental results, the rolling-back scheme makes a meaningful contribution to enhancement of the backbone network where it derives the convergence to a local basin of a good generalization performance. • As a result, our method without any add-on scheme could outperform the state-of-the-arts with help of add-on scheme.