SlideShare a Scribd company logo
Interaction Lab. Kumoh National Institute of Technology
TabletGaze : Unconstrained Appearance-based
Gaze Estimation in Mobile Tablets
:Computer Vision and Pattern Recognition 2016
Jeong JaeYeop
■Intro
■Rice TabletGaze dataset
■TabletGaze algorithms
■Results and analysis
■Discussion and conclusion
Agenda
Interaction Lab., Kumoh National Institue of Technology 2
Intro
Rice TabletGaze dataset
TabletGaze algorithms
Data Engineering Lab., Kumoh National Institue of Technology 3
■Gaze estimation in Mobile Tablets
 Commonplace connected mobile computing device
 User-tablet interaction
• Touch and sound
• Gaze is an emerging proxy of the user’s attention and intention
■ Hands-free human device interaction
■ Behavior studies
■ User authentication
Intro(1/4)
Interaction Lab., Kumoh National Institue of Technology 4
■Gaze estimation in Mobile Tablets
 Gaze estimation of tablets
• Without requiring any additional hardware
• Front-facing cameras
• Appearance-based methods
■ Not calibration stage
■ Mapping from appearance of eye region to gaze direction
Intro(2/4)
Interaction Lab., Kumoh National Institue of Technology 5
■A key challenge in tablet gaze estimation
 No constraint on
• How people use the tablet
• What kind of body posture people have when using tablet
• The user of the tablet
Intro(3/4)
Interaction Lab., Kumoh National Institue of Technology 6
■Tablet gaze estimation problem in three steps
 Collect dataset an unconstrained mobile gaze dataset of tablet
• 51 subjects
• Rice TabletGaze dataset
 TabletGaze Algorithms
• Feature extraction
• Dimensionality reduction
• Regression
 Analysis
Intro(4/4)
Interaction Lab., Kumoh National Institue of Technology 7
Rice TabletGaze dataset
TabletGaze algorithms
Results and analysis
Data Engineering Lab., Kumoh National Institue of Technology 8
■Rice TabletGaze dataset
 Unique, unconstrained characteristics in the mobile environment
• 51 subjects, each with 4 different body postures
• Released online
 The learned model can be used for other devices
• Transfer learning, domain adaptation
Rice TabletGaze dataset(1/10)
■Data collection
 Setup
• Samsung Galaxy Tab S 10.5
■ Screen size of 22.62 x 14.14 com (8.90 x 5.57 inches)
■ 35 gaze locations (points)
■ The raw data(videos) captured by the front-camera
■ Resolution – 1280 x 720
• 51 subjects
■ 12 female and 39 male
■ 26 of them wearing prescription glasses
■ 28 of the subjects are Caucasians, and the remaining 23 are Asians
■ The ages of subjects range from 20 to 40
Rice TabletGaze dataset(2/10)
Interaction Lab., Kumoh National Institue of Technology 10
■Data collection
 Four body postures
• Standing
• Sitting
• Slouching
• Lying
 Four recording session and four body posture
• 16 video sequences
• No restriction on
■ How the subject held the tablet
■ How they performed each body posture
 Natural lit office environment
Rice TabletGaze dataset(3/10)
Interaction Lab., Kumoh National Institue of Technology 11
■Data collection
 One data collection session
• Front-facing camera of the tablet begin recording a video
• Beep sound notified the beginning of the video
• Dot changing its location every three seconds and focus
■ Dot Randomize among 35 possible points
■ Free to blink
Rice TabletGaze dataset(4/10)
Interaction Lab., Kumoh National Institue of Technology 12
■Observations on the Rice TabletGaze dataset
 The entire face may not be visible in most of the image frames
• To quantify the extent of facial visibility, label each video in the dataset
■ The whole face
■ From mouth and above
■ From nose and above
■ From eyes and above
■ Even the eyes are no visible
• Manually reviewed 4 images
Rice TabletGaze dataset(5/10)
Interaction Lab., Kumoh National Institue of Technology 13
■Observations on the Rice TabletGaze dataset
 Body posture and facial visibility extent appear to be correlated
Rice TabletGaze dataset(6/10)
Interaction Lab., Kumoh National Institue of Technology 14
■Observations on the Rice TabletGaze dataset
 Glasses can cause reflection, and in many instances, the reflection
can be significant
Rice TabletGaze dataset(7/10)
Interaction Lab., Kumoh National Institue of Technology 15
■Sub-dataset Labeling
 The total amount of raw data
• 51 x 16 = 816 video sequences
• A portion of the data is not usable
■ Loss of concentration of subjects
■ Eye detector failure
■ Involuntary eye blinks and large motion blur
 Sub-dataset of 41 subjects to be used in experiments
Rice TabletGaze dataset(8/10)
Interaction Lab., Kumoh National Institue of Technology 16
■Sub-dataset Labeling
 Loss of concentration of subjects
• Refocus time
• Extract after 1.5 to 2.5 seconds when dot appears at a new location
• For the 35 videos chunk from extracted from each video
■ Inspect gaze drift more than 5 video chunk, if so, abandon the data
• Hard to determine the true gaze location
Rice TabletGaze dataset(9/10)
Interaction Lab., Kumoh National Institue of Technology 17
■Sub-dataset Labeling
 Eye detector failures
• Eyes are not visible in the image frame
• Strong reflection from glasses
• Occlusion from hair
• Poor illumination
 Use LoG(Laplacian of Gaussian) value
• Images of closed eyes
■ Higher mean intensity value given the disappearance of the dark pupil
• Blurred eye region image
■ Lower mean intensity value because motion blur weakens the edge information
Rice TabletGaze dataset(10/10)
Interaction Lab., Kumoh National Institue of Technology 18
TabletGaze algorithms
Results and analysis
Discussion and conclusion
Data Engineering Lab., Kumoh National Institue of Technology 19
■Overview
TabletGaze algorithms(1/6)
■Preprocessing
 Eye detector
• Two Harr feature CART-tree based cascade detectors
• False positive bounding boxes
■ Establish threshold for the size of the box (nostril)
■ Symmetric locations of the boxes (mouth)
• 100 x 15 resize
TabletGaze algorithms(2/6)
Interaction Lab., Kumoh National Institue of Technology 21
■Feature Calculation
 Feature extraction
• Contrast normalized pixel intensities
• LoG (Laplacian of Gaussian)
• LBP (Local Binary Pattern)
• HoG (Histogram of Oriented Gradients)
• mHoG (multilevel HoG)
■ Concatenate HoG features at different scales
TabletGaze algorithms(3/6)
Interaction Lab., Kumoh National Institue of Technology 22
■Feature Calculation
 Dimensionality reduction
• Feature is High dimensional and compromised by noise
• Mapping the features to a lower dimensional space
• CNN pooling
• LDA (Linear Discriminant Analysis)
• PCA (Principal Component Analysis)
TabletGaze algorithms(4/6)
Interaction Lab., Kumoh National Institue of Technology 23
■Feature Calculation
 LDA (Linear Discriminant Analysis)
• Intra-class scatter is maximized
• Inter-class scatter is minimized
 PCA(Principal Component Analysis)
TabletGaze algorithms(5/6)
Interaction Lab., Kumoh National Institue of Technology 24
■Regression
 The gaze labels of the data include two parts
• Horizontal and vertical coordinates on the tablet screen (x, y)
 Methods
• k-NN (k-Nearest Neighbors)
• RF (Random Forest)
■ Set of weak binary tree regressors
■ 100 trees
• GPR (Gaussian Process Regression)
• SVR (Support Vector Regression)
TabletGaze algorithms(6/6)
Interaction Lab., Kumoh National Institue of Technology 25
Results and analysis
Discussion and conclusion
Data Engineering Lab., Kumoh National Institue of Technology 26
■Error Metrics
 Previous works
• Angular error
■ Arctangent of the ratio between the distance from the subject’s eyes to screen
 This work
• Mean Error (ME)
■ 2D location on the tablet
Results and analysis(1/8)
■Comparisons for different features + regressors
 Use 100,000 images from 41 subjects using cross validation
• GPR, SVR are only used 15 subject data
Results and analysis(2/8)
Interaction Lab., Kumoh National Institue of Technology 28
■ Person-dependent and person-independent performance comparison
 in prior works appearance-based gaze estimation methods
• Person and session dependent
Results and analysis(3/8)
Interaction Lab., Kumoh National Institue of Technology 29
■Comparison with prior results
Results and analysis(4/8)
Interaction Lab., Kumoh National Institue of Technology 30
■Effect of training data size
 Groups of different sizes 𝐾
• 𝐾 : 2 ~ 41
 Use 𝐾 − 1 for training, one is test
 Randomly selecting a subset of data
 Repeat the same process 5 times and average
Results and analysis(5/8)
Interaction Lab., Kumoh National Institue of Technology 31
■Eyeglasses, race and posture
 Three experiments setups
• Experiment 1
■ The dataset was partitioned into 2 groups of wearing glasses and not
■ Training and testing are done separately for each group
• Experiment 2
■ Leave-one-subject-out cross validation for all data
■ ME is separated for each group
• Experiment 3
■ Combine data of half each group
Results and analysis(6/8)
Interaction Lab., Kumoh National Institue of Technology 32
■Eyeglasses, race and posture
Results and analysis(7/8)
Interaction Lab., Kumoh National Institue of Technology 33
■Continuous gaze tracking from videos
 Use temporal information
• Bilateral filter
Results and analysis(8/8)
Interaction Lab., Kumoh National Institue of Technology 34
Discussion and conclusion
Data Engineering Lab., Kumoh National Institue of Technology 35
■Discussion
 All of the evaluations of the algorithms are conducted on a desktop
 Pre-trained off-line and loaded onto the device
 RF + mHoG feature
• Real time gaze estimation
Discussion and conclusion(1/2)
■Conclusion
 Unconstrained mobile gaze estimation problem
• A large dataset is collected in an unconstrained environment
■ First dataset
■ Four different postures
• ME : 3.17±2.10 cm
Discussion and conclusion(2/2)
Interaction Lab., Kumoh National Institue of Technology 37
Q&A
Data Engineering Lab., Kumoh National Institue of Technology 38

More Related Content

PPTX
Improving accuracy of binary neural networks using unbalanced activation dist...
PPTX
Unsupervised representation learning for gaze estimation
PPTX
Gaze estimation using transformer
PPTX
hands on machine learning Chapter 4 model training
PPTX
Appearance based gaze estimation using deep features and random forest regres...
PDF
A Re-evaluation of Pedestrian Detection on Riemannian Manifolds
PDF
DeepXplore: Automated Whitebox Testing of Deep Learning Systems
PPTX
Online learning in estimation of distribution algorithms for dynamic environm...
Improving accuracy of binary neural networks using unbalanced activation dist...
Unsupervised representation learning for gaze estimation
Gaze estimation using transformer
hands on machine learning Chapter 4 model training
Appearance based gaze estimation using deep features and random forest regres...
A Re-evaluation of Pedestrian Detection on Riemannian Manifolds
DeepXplore: Automated Whitebox Testing of Deep Learning Systems
Online learning in estimation of distribution algorithms for dynamic environm...

What's hot (18)

PDF
Centertrack and naver airush 2020 review
PPTX
deep learning from scratch chapter 6.backpropagation
PPTX
Deep learning based gaze detection system for automobile drivers using nir ca...
PDF
Object Tracking By Online Discriminative Feature Selection Algorithm
PPTX
nnU-Net: a self-configuring method for deep learning-based biomedical image s...
PPTX
Face detection system design seminar
PDF
Testing Machine Learning-enabled Systems: A Personal Perspective
ODP
End of Year Presentation
PDF
Can we predict the quality of spectrum-based fault localization?
PDF
Keynote SBST 2014 - Search-Based Testing
PDF
Enabling Automated Software Testing with Artificial Intelligence
PDF
A comparison of SIFT, PCA-SIFT and SURF
PDF
Scalable and Cost-Effective Model-Based Software Verification and Testing
PDF
Practical Constraint Solving for Generating System Test Data
PDF
A practical guide for using Statistical Tests to assess Randomized Algorithms...
PDF
Automatic Test Suite Generation for Key-Points Detection DNNs using Many-Obje...
PDF
AI in SE: A 25-year Journey
PDF
Ag04606202206
Centertrack and naver airush 2020 review
deep learning from scratch chapter 6.backpropagation
Deep learning based gaze detection system for automobile drivers using nir ca...
Object Tracking By Online Discriminative Feature Selection Algorithm
nnU-Net: a self-configuring method for deep learning-based biomedical image s...
Face detection system design seminar
Testing Machine Learning-enabled Systems: A Personal Perspective
End of Year Presentation
Can we predict the quality of spectrum-based fault localization?
Keynote SBST 2014 - Search-Based Testing
Enabling Automated Software Testing with Artificial Intelligence
A comparison of SIFT, PCA-SIFT and SURF
Scalable and Cost-Effective Model-Based Software Verification and Testing
Practical Constraint Solving for Generating System Test Data
A practical guide for using Statistical Tests to assess Randomized Algorithms...
Automatic Test Suite Generation for Key-Points Detection DNNs using Many-Obje...
AI in SE: A 25-year Journey
Ag04606202206
Ad

Similar to Tablet gaze unconstrained appearance based gaze estimation in mobile tablets (20)

PDF
Spakov.2011.comparison of gaze to-objects mapping algorithms
PPTX
EyeGrip: Detecting Targets in a Series of Uni-directional Moving Objects Usin...
PDF
IRJET-Unconstraint Eye Tracking on Mobile Smartphone
PDF
Evaluation of eye gaze interaction
PDF
Real Time Eye Tracking and Application
PDF
IRJET- Smart Mirror using Eye Gaze Tracking
PDF
MGaze: Multi-Gaze Interactions
PPTX
Neural networks for semantic gaze analysis in xr settings
PDF
Stellmach.2011.designing gaze supported multimodal interactions for the explo...
PDF
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
PDF
Eye Gaze Tracking With a Web Camera in a Desktop Environment
PDF
Its About Time: Analyzing Temporal MicroLevel Behavioral Patterns
PDF
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
PDF
IRJET- Mobile Eye Friend Application
PDF
Inferring visual behaviour from user interaction data on a medical dashboard
DOCX
Eye gaze tracking with a web camera in a desktop environment
PPTX
Why isnt gaze a main stream hmi
PDF
Kandemir Inferring Object Relevance From Gaze In Dynamic Scenes
PDF
Gaze detection
DOCX
Eye gaze tracking with a web camera
Spakov.2011.comparison of gaze to-objects mapping algorithms
EyeGrip: Detecting Targets in a Series of Uni-directional Moving Objects Usin...
IRJET-Unconstraint Eye Tracking on Mobile Smartphone
Evaluation of eye gaze interaction
Real Time Eye Tracking and Application
IRJET- Smart Mirror using Eye Gaze Tracking
MGaze: Multi-Gaze Interactions
Neural networks for semantic gaze analysis in xr settings
Stellmach.2011.designing gaze supported multimodal interactions for the explo...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Eye Gaze Tracking With a Web Camera in a Desktop Environment
Its About Time: Analyzing Temporal MicroLevel Behavioral Patterns
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Mobile Eye Friend Application
Inferring visual behaviour from user interaction data on a medical dashboard
Eye gaze tracking with a web camera in a desktop environment
Why isnt gaze a main stream hmi
Kandemir Inferring Object Relevance From Gaze In Dynamic Scenes
Gaze detection
Eye gaze tracking with a web camera
Ad

More from Jaey Jeong (8)

PPTX
Mlp mixer an all-mlp architecture for vision
PPTX
핵심 딥러닝 입문 4장 RNN
PPTX
Gaze supported 3 d object manipulation in virtual reality
PPTX
hands on machine learning Chapter 6&7 decision tree, ensemble and random forest
PPTX
deep learning from scratch chapter 7.cnn
PPTX
deep learning from scratch chapter 5.learning related skills
PPTX
deep learning from scratch chapter 4.neural network learing
PPTX
deep learning from scratch chapter 3 neural network
Mlp mixer an all-mlp architecture for vision
핵심 딥러닝 입문 4장 RNN
Gaze supported 3 d object manipulation in virtual reality
hands on machine learning Chapter 6&7 decision tree, ensemble and random forest
deep learning from scratch chapter 7.cnn
deep learning from scratch chapter 5.learning related skills
deep learning from scratch chapter 4.neural network learing
deep learning from scratch chapter 3 neural network

Recently uploaded (20)

PPTX
Computer Software and OS of computer science of grade 11.pptx
PDF
Internet Downloader Manager (IDM) Crack 6.42 Build 42 Updates Latest 2025
PDF
Design an Analysis of Algorithms II-SECS-1021-03
PPTX
Operating system designcfffgfgggggggvggggggggg
PPTX
Agentic AI Use Case- Contract Lifecycle Management (CLM).pptx
PDF
Nekopoi APK 2025 free lastest update
PDF
Cost to Outsource Software Development in 2025
PDF
CCleaner Pro 6.38.11537 Crack Final Latest Version 2025
PDF
AutoCAD Professional Crack 2025 With License Key
PPTX
Reimagine Home Health with the Power of Agentic AI​
PDF
EN-Survey-Report-SAP-LeanIX-EA-Insights-2025.pdf
PPTX
Oracle Fusion HCM Cloud Demo for Beginners
PPTX
Patient Appointment Booking in Odoo with online payment
PPTX
Why Generative AI is the Future of Content, Code & Creativity?
PDF
How AI/LLM recommend to you ? GDG meetup 16 Aug by Fariman Guliev
PDF
Navsoft: AI-Powered Business Solutions & Custom Software Development
PDF
Odoo Companies in India – Driving Business Transformation.pdf
PPTX
CHAPTER 2 - PM Management and IT Context
PDF
Design an Analysis of Algorithms I-SECS-1021-03
PDF
Website Design Services for Small Businesses.pdf
Computer Software and OS of computer science of grade 11.pptx
Internet Downloader Manager (IDM) Crack 6.42 Build 42 Updates Latest 2025
Design an Analysis of Algorithms II-SECS-1021-03
Operating system designcfffgfgggggggvggggggggg
Agentic AI Use Case- Contract Lifecycle Management (CLM).pptx
Nekopoi APK 2025 free lastest update
Cost to Outsource Software Development in 2025
CCleaner Pro 6.38.11537 Crack Final Latest Version 2025
AutoCAD Professional Crack 2025 With License Key
Reimagine Home Health with the Power of Agentic AI​
EN-Survey-Report-SAP-LeanIX-EA-Insights-2025.pdf
Oracle Fusion HCM Cloud Demo for Beginners
Patient Appointment Booking in Odoo with online payment
Why Generative AI is the Future of Content, Code & Creativity?
How AI/LLM recommend to you ? GDG meetup 16 Aug by Fariman Guliev
Navsoft: AI-Powered Business Solutions & Custom Software Development
Odoo Companies in India – Driving Business Transformation.pdf
CHAPTER 2 - PM Management and IT Context
Design an Analysis of Algorithms I-SECS-1021-03
Website Design Services for Small Businesses.pdf

Tablet gaze unconstrained appearance based gaze estimation in mobile tablets

  • 1. Interaction Lab. Kumoh National Institute of Technology TabletGaze : Unconstrained Appearance-based Gaze Estimation in Mobile Tablets :Computer Vision and Pattern Recognition 2016 Jeong JaeYeop
  • 2. ■Intro ■Rice TabletGaze dataset ■TabletGaze algorithms ■Results and analysis ■Discussion and conclusion Agenda Interaction Lab., Kumoh National Institue of Technology 2
  • 3. Intro Rice TabletGaze dataset TabletGaze algorithms Data Engineering Lab., Kumoh National Institue of Technology 3
  • 4. ■Gaze estimation in Mobile Tablets  Commonplace connected mobile computing device  User-tablet interaction • Touch and sound • Gaze is an emerging proxy of the user’s attention and intention ■ Hands-free human device interaction ■ Behavior studies ■ User authentication Intro(1/4) Interaction Lab., Kumoh National Institue of Technology 4
  • 5. ■Gaze estimation in Mobile Tablets  Gaze estimation of tablets • Without requiring any additional hardware • Front-facing cameras • Appearance-based methods ■ Not calibration stage ■ Mapping from appearance of eye region to gaze direction Intro(2/4) Interaction Lab., Kumoh National Institue of Technology 5
  • 6. ■A key challenge in tablet gaze estimation  No constraint on • How people use the tablet • What kind of body posture people have when using tablet • The user of the tablet Intro(3/4) Interaction Lab., Kumoh National Institue of Technology 6
  • 7. ■Tablet gaze estimation problem in three steps  Collect dataset an unconstrained mobile gaze dataset of tablet • 51 subjects • Rice TabletGaze dataset  TabletGaze Algorithms • Feature extraction • Dimensionality reduction • Regression  Analysis Intro(4/4) Interaction Lab., Kumoh National Institue of Technology 7
  • 8. Rice TabletGaze dataset TabletGaze algorithms Results and analysis Data Engineering Lab., Kumoh National Institue of Technology 8
  • 9. ■Rice TabletGaze dataset  Unique, unconstrained characteristics in the mobile environment • 51 subjects, each with 4 different body postures • Released online  The learned model can be used for other devices • Transfer learning, domain adaptation Rice TabletGaze dataset(1/10)
  • 10. ■Data collection  Setup • Samsung Galaxy Tab S 10.5 ■ Screen size of 22.62 x 14.14 com (8.90 x 5.57 inches) ■ 35 gaze locations (points) ■ The raw data(videos) captured by the front-camera ■ Resolution – 1280 x 720 • 51 subjects ■ 12 female and 39 male ■ 26 of them wearing prescription glasses ■ 28 of the subjects are Caucasians, and the remaining 23 are Asians ■ The ages of subjects range from 20 to 40 Rice TabletGaze dataset(2/10) Interaction Lab., Kumoh National Institue of Technology 10
  • 11. ■Data collection  Four body postures • Standing • Sitting • Slouching • Lying  Four recording session and four body posture • 16 video sequences • No restriction on ■ How the subject held the tablet ■ How they performed each body posture  Natural lit office environment Rice TabletGaze dataset(3/10) Interaction Lab., Kumoh National Institue of Technology 11
  • 12. ■Data collection  One data collection session • Front-facing camera of the tablet begin recording a video • Beep sound notified the beginning of the video • Dot changing its location every three seconds and focus ■ Dot Randomize among 35 possible points ■ Free to blink Rice TabletGaze dataset(4/10) Interaction Lab., Kumoh National Institue of Technology 12
  • 13. ■Observations on the Rice TabletGaze dataset  The entire face may not be visible in most of the image frames • To quantify the extent of facial visibility, label each video in the dataset ■ The whole face ■ From mouth and above ■ From nose and above ■ From eyes and above ■ Even the eyes are no visible • Manually reviewed 4 images Rice TabletGaze dataset(5/10) Interaction Lab., Kumoh National Institue of Technology 13
  • 14. ■Observations on the Rice TabletGaze dataset  Body posture and facial visibility extent appear to be correlated Rice TabletGaze dataset(6/10) Interaction Lab., Kumoh National Institue of Technology 14
  • 15. ■Observations on the Rice TabletGaze dataset  Glasses can cause reflection, and in many instances, the reflection can be significant Rice TabletGaze dataset(7/10) Interaction Lab., Kumoh National Institue of Technology 15
  • 16. ■Sub-dataset Labeling  The total amount of raw data • 51 x 16 = 816 video sequences • A portion of the data is not usable ■ Loss of concentration of subjects ■ Eye detector failure ■ Involuntary eye blinks and large motion blur  Sub-dataset of 41 subjects to be used in experiments Rice TabletGaze dataset(8/10) Interaction Lab., Kumoh National Institue of Technology 16
  • 17. ■Sub-dataset Labeling  Loss of concentration of subjects • Refocus time • Extract after 1.5 to 2.5 seconds when dot appears at a new location • For the 35 videos chunk from extracted from each video ■ Inspect gaze drift more than 5 video chunk, if so, abandon the data • Hard to determine the true gaze location Rice TabletGaze dataset(9/10) Interaction Lab., Kumoh National Institue of Technology 17
  • 18. ■Sub-dataset Labeling  Eye detector failures • Eyes are not visible in the image frame • Strong reflection from glasses • Occlusion from hair • Poor illumination  Use LoG(Laplacian of Gaussian) value • Images of closed eyes ■ Higher mean intensity value given the disappearance of the dark pupil • Blurred eye region image ■ Lower mean intensity value because motion blur weakens the edge information Rice TabletGaze dataset(10/10) Interaction Lab., Kumoh National Institue of Technology 18
  • 19. TabletGaze algorithms Results and analysis Discussion and conclusion Data Engineering Lab., Kumoh National Institue of Technology 19
  • 21. ■Preprocessing  Eye detector • Two Harr feature CART-tree based cascade detectors • False positive bounding boxes ■ Establish threshold for the size of the box (nostril) ■ Symmetric locations of the boxes (mouth) • 100 x 15 resize TabletGaze algorithms(2/6) Interaction Lab., Kumoh National Institue of Technology 21
  • 22. ■Feature Calculation  Feature extraction • Contrast normalized pixel intensities • LoG (Laplacian of Gaussian) • LBP (Local Binary Pattern) • HoG (Histogram of Oriented Gradients) • mHoG (multilevel HoG) ■ Concatenate HoG features at different scales TabletGaze algorithms(3/6) Interaction Lab., Kumoh National Institue of Technology 22
  • 23. ■Feature Calculation  Dimensionality reduction • Feature is High dimensional and compromised by noise • Mapping the features to a lower dimensional space • CNN pooling • LDA (Linear Discriminant Analysis) • PCA (Principal Component Analysis) TabletGaze algorithms(4/6) Interaction Lab., Kumoh National Institue of Technology 23
  • 24. ■Feature Calculation  LDA (Linear Discriminant Analysis) • Intra-class scatter is maximized • Inter-class scatter is minimized  PCA(Principal Component Analysis) TabletGaze algorithms(5/6) Interaction Lab., Kumoh National Institue of Technology 24
  • 25. ■Regression  The gaze labels of the data include two parts • Horizontal and vertical coordinates on the tablet screen (x, y)  Methods • k-NN (k-Nearest Neighbors) • RF (Random Forest) ■ Set of weak binary tree regressors ■ 100 trees • GPR (Gaussian Process Regression) • SVR (Support Vector Regression) TabletGaze algorithms(6/6) Interaction Lab., Kumoh National Institue of Technology 25
  • 26. Results and analysis Discussion and conclusion Data Engineering Lab., Kumoh National Institue of Technology 26
  • 27. ■Error Metrics  Previous works • Angular error ■ Arctangent of the ratio between the distance from the subject’s eyes to screen  This work • Mean Error (ME) ■ 2D location on the tablet Results and analysis(1/8)
  • 28. ■Comparisons for different features + regressors  Use 100,000 images from 41 subjects using cross validation • GPR, SVR are only used 15 subject data Results and analysis(2/8) Interaction Lab., Kumoh National Institue of Technology 28
  • 29. ■ Person-dependent and person-independent performance comparison  in prior works appearance-based gaze estimation methods • Person and session dependent Results and analysis(3/8) Interaction Lab., Kumoh National Institue of Technology 29
  • 30. ■Comparison with prior results Results and analysis(4/8) Interaction Lab., Kumoh National Institue of Technology 30
  • 31. ■Effect of training data size  Groups of different sizes 𝐾 • 𝐾 : 2 ~ 41  Use 𝐾 − 1 for training, one is test  Randomly selecting a subset of data  Repeat the same process 5 times and average Results and analysis(5/8) Interaction Lab., Kumoh National Institue of Technology 31
  • 32. ■Eyeglasses, race and posture  Three experiments setups • Experiment 1 ■ The dataset was partitioned into 2 groups of wearing glasses and not ■ Training and testing are done separately for each group • Experiment 2 ■ Leave-one-subject-out cross validation for all data ■ ME is separated for each group • Experiment 3 ■ Combine data of half each group Results and analysis(6/8) Interaction Lab., Kumoh National Institue of Technology 32
  • 33. ■Eyeglasses, race and posture Results and analysis(7/8) Interaction Lab., Kumoh National Institue of Technology 33
  • 34. ■Continuous gaze tracking from videos  Use temporal information • Bilateral filter Results and analysis(8/8) Interaction Lab., Kumoh National Institue of Technology 34
  • 35. Discussion and conclusion Data Engineering Lab., Kumoh National Institue of Technology 35
  • 36. ■Discussion  All of the evaluations of the algorithms are conducted on a desktop  Pre-trained off-line and loaded onto the device  RF + mHoG feature • Real time gaze estimation Discussion and conclusion(1/2)
  • 37. ■Conclusion  Unconstrained mobile gaze estimation problem • A large dataset is collected in an unconstrained environment ■ First dataset ■ Four different postures • ME : 3.17±2.10 cm Discussion and conclusion(2/2) Interaction Lab., Kumoh National Institue of Technology 37
  • 38. Q&A Data Engineering Lab., Kumoh National Institue of Technology 38