SlideShare a Scribd company logo
Interaction Lab. Kumoh National Institute of Technology
Appearance-based gaze estimation using
deep feature and random forest regression
Knowledge-Based System 2016.
Jeong JaeYeop
■Intro
■Method
■Experimental result
■Conclusion
Agenda
Interaction Lab., Kumoh National Institue of Technology 2
Intro
Method
Experimental result
Data Engineering Lab., Kumoh National Institue of Technology 3
■Gaze
 Significant role in understanding human attention, feeling and mind
• Cognitive processes analysis
• Human computer interaction
 Non-intrusive gaze tracking systems
• One single web camera under natural light with free head movement
• Hard to get accuracy
Intro(1/4)
Interaction Lab., Kumoh National Institue of Technology 4
■Gaze estimation methods
 Feature-based method
• Depend on pupil detection with the light sources’ reflections on cornea
• Map from pupil center of geometry model to the gaze calibration points
• High accuracy, fluctuate strongly due to free head movement
 Appearance-based method
• Point representation in high dimensional space
• Learn the mapping relation from this point in given feature space to
screen coordinates
Intro(2/4)
Interaction Lab., Kumoh National Institue of Technology 5
Intro(3/4)
Interaction Lab., Kumoh National Institue of Technology 6
■Focus
 Deep feature extraction and feature forest regression
• Using one web camera under natural light
• Free head movement
Intro(4/4)
Interaction Lab., Kumoh National Institue of Technology 7
Method
Experimental result
Conclusion
Data Engineering Lab., Kumoh National Institue of Technology 8
Method(1/7)
■Eye image data
 Crop eye region from its localization of facial landmarks
• Firstly, the subject’s face which always appears fully in the FOV
■ Using Viola-Jones method
• Get Eye region
■ SDM(Supervised Descent Method)
■ Facial landmark detector
▪ Eye corner and other fiducial points
▪ Eye region are shown as red dots
• “Face in the wild” datasets
■ Robust under natural light with different illumination conditions
■ Significantly fast and accurate for face align in the wild with large head rotary
Method(2/7)
Interaction Lab., Kumoh National Institue of Technology 10
■Eye image data
 Cropped eye images without background
• About 4% of all cases are discarded
• Images dataset delivers eye images without background noise
Method(3/7)
Interaction Lab., Kumoh National Institue of Technology 11
■Deep feature learning
 Task of CNN
• Learning deep feature from the input eye images
• Last hidden layer
■ Fully connected
■ 160 dimension deep feature is extracted
Method(4/7)
Interaction Lab., Kumoh National Institue of Technology 12
Method(5/7)
Interaction Lab., Kumoh National Institue of Technology 13
■Random forest regression
• 𝑁 is calibration point number
• 𝑥𝑖 is 𝑖𝑡ℎ deep feature vector in feature space
• 𝑦𝑖 is 𝑖𝑡ℎ two-dimensional gaze vector
Method(6/7)
Interaction Lab., Kumoh National Institue of Technology 14
■Random forest regression
Method(7/7)
Interaction Lab., Kumoh National Institue of Technology 15
Experimental result
Conclusion
Data Engineering Lab., Kumoh National Institue of Technology 16
■Dataset and experimental setup
 Desktop computer using USB camera
• The image resolution is 640 by 480
 Data
• Collected different time
■ Under normal natural light, strong sun light, weak screen light at night
• 22 groups of video from 6 subjects
■ Different light illumination conditions
■ 16 groups wearing glasses
■ Calibration step
■ Without head movement
■ Free head movement
• Calibration points
■ 25 points is training
■ 16 points is test
Experimental result(1/10)
■Dataset and experimental setup
 Distribution of head pose angle in the free head movement
• Using POSIT(Pose from Orthography and Scaling with Iterations)
• Pitch : [0°, 15°], Yaw : [-10°, 10°]
 Mean grayscale intensity of images
• Darker light condition has a lower mean grayscale intensity
Experimental result(2/10)
Interaction Lab., Kumoh National Institue of Technology 18
■Deep gaze feature
 The deep CNN was implemented based on theano
Experimental result(3/10)
Interaction Lab., Kumoh National Institue of Technology 19
■Comparison with other methods
 MAE(Mean Absolute Error)
• Gaze angle on different regression approaches and different features
Experimental result(4/10)
Interaction Lab., Kumoh National Institue of Technology 20
■Comparison with other methods
 Time cost(𝑚𝑠)
• Single image testing
■ Feature extraction and feature regression
Experimental result(5/10)
Interaction Lab., Kumoh National Institue of Technology 21
■Comparison with other methods
 MAE(Mean Absolute Error)
• Gaze angle estimation
Experimental result(6/10)
Interaction Lab., Kumoh National Institue of Technology 22
■Comparison on illumination condition
 Gaze estimation under natural light
• The red points and curves
■ Error of training data
• The blue points and curves
■ Error of test data
Experimental result(7/10)
Interaction Lab., Kumoh National Institue of Technology 23
■Comparison on test points
 MAE(Mean Absolute Error)
• Gaze angle on test points using different calibration points
Experimental result(8/10)
Interaction Lab., Kumoh National Institue of Technology 24
■Comparison on person-independent
 MAE(Mean Absolute Error)
• Gaze angle for cross subject gaze estimation
• Low resolution, free head movement and natural light
Experimental result(9/10)
Interaction Lab., Kumoh National Institue of Technology 25
■Comparison on occlusion
Experimental result(10/10)
Interaction Lab., Kumoh National Institue of Technology 26
Conclusion
Data Engineering Lab., Kumoh National Institue of Technology 27
■In this work
 CNN-based regression method for gaze estimation
• Under natural light
• CNN built a sparse active feature space from the training eye images
• Gaze estimation model learned mapping between deep feature and
gaze coordinate by random forest regression
 Deep feature significantly improves the performance
 Hope to apply
• Real time driver gaze tracking
• Gaze estimation in real driving environment with background noise
Conclusion

More Related Content

PPTX
Tablet gaze unconstrained appearance based gaze estimation in mobile tablets
PPTX
Unsupervised representation learning for gaze estimation
PPTX
Gaze estimation using transformer
PPTX
Improving accuracy of binary neural networks using unbalanced activation dist...
PPTX
Deep learning based gaze detection system for automobile drivers using nir ca...
PPTX
Brain signal seminar
PPTX
Ultrafast Optical signal processing
PDF
Vector spaces for information extraction - Random Projection Example
Tablet gaze unconstrained appearance based gaze estimation in mobile tablets
Unsupervised representation learning for gaze estimation
Gaze estimation using transformer
Improving accuracy of binary neural networks using unbalanced activation dist...
Deep learning based gaze detection system for automobile drivers using nir ca...
Brain signal seminar
Ultrafast Optical signal processing
Vector spaces for information extraction - Random Projection Example

Similar to Appearance based gaze estimation using deep features and random forest regression (20)

PPTX
Neural networks for semantic gaze analysis in xr settings
PDF
Gaze detection
PDF
Eye Gaze Tracking With a Web Camera in a Desktop Environment
DOCX
Eye gaze tracking with a web camera
PDF
IRJET-Unconstraint Eye Tracking on Mobile Smartphone
PDF
Movie Recommendation Using CNN
PPTX
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
DOCX
Eye gaze tracking with a web camera in a desktop environment
PDF
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
PDF
An eye gaze detection using low resolution web camera in desktop environment
PDF
ReadingBehaviour_LiteratureReview.pdf
PPTX
EyeGrip: Detecting Targets in a Series of Uni-directional Moving Objects Usin...
PPTX
Age Estimation And Gender Prediction Using Convolutional Neural Network.pptx
PPTX
[NS][Lab_Seminar_240626]GazeGNN: A Gaze-Guided Graph Neural Network for Chest...
PDF
IRJET- Smart Mirror using Eye Gaze Tracking
PDF
“Eye Tracking for the Future,” a Presentation from Parallel Rules
PPTX
Human computer interaction_ 23CSM1R19.pptx
PDF
"Approaches for Vision-based Driver Monitoring," a Presentation from PathPart...
PDF
Faro An Interactive Interface For Remote Administration Of Clinical Tests Bas...
Neural networks for semantic gaze analysis in xr settings
Gaze detection
Eye Gaze Tracking With a Web Camera in a Desktop Environment
Eye gaze tracking with a web camera
IRJET-Unconstraint Eye Tracking on Mobile Smartphone
Movie Recommendation Using CNN
Toward Accurate and Robust Cross-Ratio based Gaze Trackers Through Learning F...
Eye gaze tracking with a web camera in a desktop environment
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
An eye gaze detection using low resolution web camera in desktop environment
ReadingBehaviour_LiteratureReview.pdf
EyeGrip: Detecting Targets in a Series of Uni-directional Moving Objects Usin...
Age Estimation And Gender Prediction Using Convolutional Neural Network.pptx
[NS][Lab_Seminar_240626]GazeGNN: A Gaze-Guided Graph Neural Network for Chest...
IRJET- Smart Mirror using Eye Gaze Tracking
“Eye Tracking for the Future,” a Presentation from Parallel Rules
Human computer interaction_ 23CSM1R19.pptx
"Approaches for Vision-based Driver Monitoring," a Presentation from PathPart...
Faro An Interactive Interface For Remote Administration Of Clinical Tests Bas...
Ad

More from Jaey Jeong (10)

PPTX
Mlp mixer an all-mlp architecture for vision
PPTX
핵심 딥러닝 입문 4장 RNN
PPTX
Gaze supported 3 d object manipulation in virtual reality
PPTX
hands on machine learning Chapter 4 model training
PPTX
hands on machine learning Chapter 6&7 decision tree, ensemble and random forest
PPTX
deep learning from scratch chapter 7.cnn
PPTX
deep learning from scratch chapter 5.learning related skills
PPTX
deep learning from scratch chapter 6.backpropagation
PPTX
deep learning from scratch chapter 4.neural network learing
PPTX
deep learning from scratch chapter 3 neural network
Mlp mixer an all-mlp architecture for vision
핵심 딥러닝 입문 4장 RNN
Gaze supported 3 d object manipulation in virtual reality
hands on machine learning Chapter 4 model training
hands on machine learning Chapter 6&7 decision tree, ensemble and random forest
deep learning from scratch chapter 7.cnn
deep learning from scratch chapter 5.learning related skills
deep learning from scratch chapter 6.backpropagation
deep learning from scratch chapter 4.neural network learing
deep learning from scratch chapter 3 neural network
Ad

Recently uploaded (20)

PDF
iTop VPN 6.5.0 Crack + License Key 2025 (Premium Version)
PPTX
history of c programming in notes for students .pptx
PDF
17 Powerful Integrations Your Next-Gen MLM Software Needs
PDF
Salesforce Agentforce AI Implementation.pdf
PDF
Complete Guide to Website Development in Malaysia for SMEs
PDF
Digital Systems & Binary Numbers (comprehensive )
PPTX
CHAPTER 2 - PM Management and IT Context
PDF
How to Make Money in the Metaverse_ Top Strategies for Beginners.pdf
PPTX
Monitoring Stack: Grafana, Loki & Promtail
PDF
Nekopoi APK 2025 free lastest update
PDF
Cost to Outsource Software Development in 2025
PDF
Adobe Illustrator 28.6 Crack My Vision of Vector Design
PPTX
Computer Software and OS of computer science of grade 11.pptx
PPTX
Why Generative AI is the Future of Content, Code & Creativity?
PDF
Autodesk AutoCAD Crack Free Download 2025
PPTX
Weekly report ppt - harsh dattuprasad patel.pptx
PDF
Designing Intelligence for the Shop Floor.pdf
PDF
Product Update: Alluxio AI 3.7 Now with Sub-Millisecond Latency
PPTX
Oracle Fusion HCM Cloud Demo for Beginners
PDF
Design an Analysis of Algorithms II-SECS-1021-03
iTop VPN 6.5.0 Crack + License Key 2025 (Premium Version)
history of c programming in notes for students .pptx
17 Powerful Integrations Your Next-Gen MLM Software Needs
Salesforce Agentforce AI Implementation.pdf
Complete Guide to Website Development in Malaysia for SMEs
Digital Systems & Binary Numbers (comprehensive )
CHAPTER 2 - PM Management and IT Context
How to Make Money in the Metaverse_ Top Strategies for Beginners.pdf
Monitoring Stack: Grafana, Loki & Promtail
Nekopoi APK 2025 free lastest update
Cost to Outsource Software Development in 2025
Adobe Illustrator 28.6 Crack My Vision of Vector Design
Computer Software and OS of computer science of grade 11.pptx
Why Generative AI is the Future of Content, Code & Creativity?
Autodesk AutoCAD Crack Free Download 2025
Weekly report ppt - harsh dattuprasad patel.pptx
Designing Intelligence for the Shop Floor.pdf
Product Update: Alluxio AI 3.7 Now with Sub-Millisecond Latency
Oracle Fusion HCM Cloud Demo for Beginners
Design an Analysis of Algorithms II-SECS-1021-03

Appearance based gaze estimation using deep features and random forest regression

  • 1. Interaction Lab. Kumoh National Institute of Technology Appearance-based gaze estimation using deep feature and random forest regression Knowledge-Based System 2016. Jeong JaeYeop
  • 3. Intro Method Experimental result Data Engineering Lab., Kumoh National Institue of Technology 3
  • 4. ■Gaze  Significant role in understanding human attention, feeling and mind • Cognitive processes analysis • Human computer interaction  Non-intrusive gaze tracking systems • One single web camera under natural light with free head movement • Hard to get accuracy Intro(1/4) Interaction Lab., Kumoh National Institue of Technology 4
  • 5. ■Gaze estimation methods  Feature-based method • Depend on pupil detection with the light sources’ reflections on cornea • Map from pupil center of geometry model to the gaze calibration points • High accuracy, fluctuate strongly due to free head movement  Appearance-based method • Point representation in high dimensional space • Learn the mapping relation from this point in given feature space to screen coordinates Intro(2/4) Interaction Lab., Kumoh National Institue of Technology 5
  • 6. Intro(3/4) Interaction Lab., Kumoh National Institue of Technology 6
  • 7. ■Focus  Deep feature extraction and feature forest regression • Using one web camera under natural light • Free head movement Intro(4/4) Interaction Lab., Kumoh National Institue of Technology 7
  • 8. Method Experimental result Conclusion Data Engineering Lab., Kumoh National Institue of Technology 8
  • 10. ■Eye image data  Crop eye region from its localization of facial landmarks • Firstly, the subject’s face which always appears fully in the FOV ■ Using Viola-Jones method • Get Eye region ■ SDM(Supervised Descent Method) ■ Facial landmark detector ▪ Eye corner and other fiducial points ▪ Eye region are shown as red dots • “Face in the wild” datasets ■ Robust under natural light with different illumination conditions ■ Significantly fast and accurate for face align in the wild with large head rotary Method(2/7) Interaction Lab., Kumoh National Institue of Technology 10
  • 11. ■Eye image data  Cropped eye images without background • About 4% of all cases are discarded • Images dataset delivers eye images without background noise Method(3/7) Interaction Lab., Kumoh National Institue of Technology 11
  • 12. ■Deep feature learning  Task of CNN • Learning deep feature from the input eye images • Last hidden layer ■ Fully connected ■ 160 dimension deep feature is extracted Method(4/7) Interaction Lab., Kumoh National Institue of Technology 12
  • 13. Method(5/7) Interaction Lab., Kumoh National Institue of Technology 13
  • 14. ■Random forest regression • 𝑁 is calibration point number • 𝑥𝑖 is 𝑖𝑡ℎ deep feature vector in feature space • 𝑦𝑖 is 𝑖𝑡ℎ two-dimensional gaze vector Method(6/7) Interaction Lab., Kumoh National Institue of Technology 14
  • 15. ■Random forest regression Method(7/7) Interaction Lab., Kumoh National Institue of Technology 15
  • 16. Experimental result Conclusion Data Engineering Lab., Kumoh National Institue of Technology 16
  • 17. ■Dataset and experimental setup  Desktop computer using USB camera • The image resolution is 640 by 480  Data • Collected different time ■ Under normal natural light, strong sun light, weak screen light at night • 22 groups of video from 6 subjects ■ Different light illumination conditions ■ 16 groups wearing glasses ■ Calibration step ■ Without head movement ■ Free head movement • Calibration points ■ 25 points is training ■ 16 points is test Experimental result(1/10)
  • 18. ■Dataset and experimental setup  Distribution of head pose angle in the free head movement • Using POSIT(Pose from Orthography and Scaling with Iterations) • Pitch : [0°, 15°], Yaw : [-10°, 10°]  Mean grayscale intensity of images • Darker light condition has a lower mean grayscale intensity Experimental result(2/10) Interaction Lab., Kumoh National Institue of Technology 18
  • 19. ■Deep gaze feature  The deep CNN was implemented based on theano Experimental result(3/10) Interaction Lab., Kumoh National Institue of Technology 19
  • 20. ■Comparison with other methods  MAE(Mean Absolute Error) • Gaze angle on different regression approaches and different features Experimental result(4/10) Interaction Lab., Kumoh National Institue of Technology 20
  • 21. ■Comparison with other methods  Time cost(𝑚𝑠) • Single image testing ■ Feature extraction and feature regression Experimental result(5/10) Interaction Lab., Kumoh National Institue of Technology 21
  • 22. ■Comparison with other methods  MAE(Mean Absolute Error) • Gaze angle estimation Experimental result(6/10) Interaction Lab., Kumoh National Institue of Technology 22
  • 23. ■Comparison on illumination condition  Gaze estimation under natural light • The red points and curves ■ Error of training data • The blue points and curves ■ Error of test data Experimental result(7/10) Interaction Lab., Kumoh National Institue of Technology 23
  • 24. ■Comparison on test points  MAE(Mean Absolute Error) • Gaze angle on test points using different calibration points Experimental result(8/10) Interaction Lab., Kumoh National Institue of Technology 24
  • 25. ■Comparison on person-independent  MAE(Mean Absolute Error) • Gaze angle for cross subject gaze estimation • Low resolution, free head movement and natural light Experimental result(9/10) Interaction Lab., Kumoh National Institue of Technology 25
  • 26. ■Comparison on occlusion Experimental result(10/10) Interaction Lab., Kumoh National Institue of Technology 26
  • 27. Conclusion Data Engineering Lab., Kumoh National Institue of Technology 27
  • 28. ■In this work  CNN-based regression method for gaze estimation • Under natural light • CNN built a sparse active feature space from the training eye images • Gaze estimation model learned mapping between deep feature and gaze coordinate by random forest regression  Deep feature significantly improves the performance  Hope to apply • Real time driver gaze tracking • Gaze estimation in real driving environment with background noise Conclusion