SlideShare a Scribd company logo
Compressed Learning for Time
Series Classification
Shueh-Han Shih
Department of Computer Science and Information
Engineering, National Taiwan University of Science and
Technology
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
Motivation
Image: http://www.aeris.com/
Motivation (cont’d)
• The key to handle time series data effectively is
choosing a suitable representation
• Transmission and storage issues are critical in IoT
scenario
• To provide interpretable result for human is
important
 Time series sparse representation - envelope
Time series data type
1. Symbolic sequence
2. Complex symbolic sequence
3. Simple time series
4. Multivariate time series
A brief survey on sequence classification Z Xing, J Pei, E Keogh - ACM SIGKDD , 2010
Classification of time series
• Assigning instances to one of the predefined classes.
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 1
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 2
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 3
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 4
Time series classification approaches
• Feature based
• Sequence distance based
• Model based
A brief survey on sequence classification Z Xing, J Pei, E Keogh - ACM SIGKDD, 2010
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
Conventional approach
• Number of sample needed for compressed sensing is
much more lower than Nyquist frequency.
Image: http://www.ni.com/
Main idea
• Most real-world signals are sparse in some basis
A𝑥 = 𝑦, A ∈ ℝ 𝑝×𝑛 𝑎𝑛𝑑 𝑝 ≪ 𝑛
• Dramatically reduce the transmission loading
a measure
Requirements of compressed sensing
1. 𝑥 should be a 𝑘-sparse signal
 1 to 1 relation between data and compressed domain
2. A must satisfies the restricted isometry property
 (1 − δ 𝑝)ǁ𝑥ǁ2
2
≤ ǁA𝑥ǁ2
2
≤ (1 + δ 𝑝)ǁ𝑥ǁ2
2
A =
𝑟𝑎𝑛𝑑𝑛 𝑝,𝑛
𝑛
(mean=0, 𝜎 =
1
𝑛
)
for some constant 𝛿 𝑝 ∈ (0, 1
Image: Mostafa Mohsenvand Projects
Basic routine
𝑨 = 𝑸
𝑨′ = 𝑸𝑷
𝒚 = 𝑨𝒙(𝒐𝒓 𝑨′
𝒔)
transmission
• Postpone the computational cost to recovery stage
Learning in the compressed domain
• Perform task without recovery
• SVM can keep the learnability
in compressed domain
• Reduce model complexity
Image: Compressed learning: Universal sparse dimensionality reduction and learning in the
measurement domain. R Calderbank, S Jafarpour, R Schapire - preprint, 2009 - dsp.rice.edu
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
The origin
• The ‘envelope’ in finance
Image: http://www.investopedia.com/
Preliminaries
• A time series
– T = 𝑡1, 𝑡2, … 𝑡𝑗 … 𝑡 𝑛 , 𝑡𝑗 ∈ ℝ
• A time series dataset
– D = T 𝑖 | T 𝑖 = 𝑡1
𝑖
, 𝑡2
𝑖
, … 𝑡 𝑛
𝑖 , 𝑖 = 1 𝑡𝑜 𝑚
• Well-synchronized with the same length
– A set of random sample from random variables 𝐓1, 𝐓2, … 𝐓𝑗, … 𝐓𝑛
Envelope creation
• Given 𝐷, envelope with size 𝑘
– E 𝑘 = 𝑍 𝑍 = 𝑧1, 𝑧2, … 𝑧 𝑛 , 𝑧𝑗 − 𝜇 𝑗 ≤ 𝑘 ∙ 𝑠𝑡𝑑𝑗 , ∀ 𝑧𝑗 ∈ ℝ}
• 𝜇 𝑗 = 𝑚𝑒𝑎𝑛(𝑻𝑗) , 𝑠𝑡𝑑𝑗 = 𝑠𝑡𝑑(𝑻𝑗)
– Profiling the time series dataset
Envelope encoding
• Encoding time series T as a sparse series S
• Sparsity indicates the similarity of a time series and 𝐷
𝑠𝑗 = 1, 𝑖𝑓 𝑡𝑗 > 𝜇 𝑗 + 𝑘 ∙ 𝑠𝑡𝑑𝑗
𝑠𝑗 = −1, 𝑖𝑓 𝑡𝑗 < 𝜇 𝑗 − 𝑘 ∙ 𝑠𝑡𝑑𝑗
𝑠𝑗 = 0, 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒
, 𝑓𝑜𝑟 𝑗 = 1 𝑡𝑜 𝑛
Guarantee of sparsity
• According to Chebyshev’s inequality,
– Pr(|X − 𝜇| ≤ 𝑘𝜎) ≥ 1 −
1
𝑘2
– No matter what kind of distribution for 𝑻𝑗
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
Determination of 𝑘
• 𝑘 affects the effectiveness of envelope representation
Determination of 𝑘 (cont’d)
• Focus on time series multi-class classification
– Envelope representation should be discriminative
𝑘∗ = arg max
𝑘
(−𝑎 𝑘 + 𝜆 ∙ 𝑏 𝑘)
– 𝑘 : tradeoff between sparsity and distinguishability
Encoding result visualization
• Sparsity indicates similarity
Sparse property
⟹ transmission
efficiency
• Encoding results
are interpretable
ECGFiveDays from UCR
Envelope representation workflow
• From raw series to feature
Overall workflow
• Classification scheme
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
1. Proposed method vs. state-of-art method on
classification task
2. Compressibility of envelope representation
with compressed sensing
3. Noise resistance of envelope representation
4. Time efficiency
5. Space efficiency
Classification performance
• Benchmark dataset from UCR (5/42)
Dataset/ Algorithm Number of
classes
Size of training
set
Size of testing
set
Time
series Length
CBF 3 30 900 128
Coffee 2 28 28 286
ECGFiveDays 2 23 861 136
ItalyPowerDemand 2 67 1029 24
Sony II 2 27 953 65
Classification performance (cont’d)
• Result on benchmark dataset (5/42)
– Win:9 / lose:18 / between:15 (close:12)
– Not the case with IoT scenario, never lack of data
Dataset/ Algorithm 1NN-Euclidean 1NN-DTW (best, noWin) Envelope+
linearSVM
CBF 85.2(0.9357) 99.6/99.7 90.66
Coffee 75(0.031608) 82.1/82.1 85.71
ECGFiveDays 79.7(0.8758) 79.7/76.8 88.38
ItalyPowerDemand 95.5(1.0661) 95.5/95 97.08
Sony II 69.5(0.9986) 69.5/72.5 82.79
Influence of compression ratio
• Compression ratio = 𝑝/𝑛
(Number of measurements) / (data dimension)
Influence of compression ratio (cont’d)
• Using nearly
1
3
datasets from UCR
– Some datasets have excellent compressibility
Influence of compression ratio (cont’d)
• Result on benchmark dataset (5/42)
Dataset/ Algorithm 1NN-
Euclidean
1NN-DTW
(best, noWin)
Compression
ratio=10%
Compression
ratio=20%
Compression
ratio=50%
CBF 85.2 99.6/99.7 80.44 88.22 88.44
Coffee 75 82.1/82.1 71.42 82.14 89.28
ECGFiveDays 79.7 79.7/76.8 78.86 81.3 81.64
ItalyPowerDemand 95.5 95.5/95 86.58 91.73 93.97
Sony II 69.5 69.5/72.5 76.91 78.38 80.06
Robustness to noise
• Noise level - SNR
Image: documentation.meraki.com
Robustness to noise (cont’d)
• Using ECG200 dataset as example
The original envelope The envelope with noise level SNR=10.
Robustness to noise (cont’d)
• Envelope representation is noise-resistant
– Can even ignore denoising stage
Envelope built/SVM trained with clean data Envelope built/SVM trained with noisy data
Time efficiency
1. Building envelope takes O(m*n)
2. Encoding each instance takes O(n)
3. Linear SVM, expects to be O(m2)
 Linear time in prediction
0 2 4 6 8 10 12 14 16
0
2
4
6
8
10
12
14
16
Execution time (testing)
envelope (Sec.)
KNN+ED(Sec.)
Space efficiency
1. 32 to (2 ∗ #𝑐𝑙𝑎𝑠𝑠) ratio of reduction
2. 32 to (32 ∗ #𝑐𝑙𝑎𝑠𝑠 ∗ 𝑐𝑜𝑚𝑝𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝑟𝑎𝑡𝑖𝑜) ratio of
reduction through compressed sensing
3. Run length encoding
Outline
• Introduction
• Compressed sensing
• Sparse representation - envelope
• Classification framework
• Experimental results
• Case study
• Conclusion
Smart home project
• Passive user identification
Image: www.bitronvideo.eu
Using sensor
• Data collection
– EcoBT Mini
– 33Hz
50 100 150 200 250
-0.1
-0.05
0
0.05
0.1
0.15
0.2
0.25
50 100 150 200 250
-0.4
-0.3
-0.2
-0.1
0
0.1
0.2
0.3
50 100 150 200 250
-0.4
-0.3
-0.2
-0.1
0
0.1
0.2
0.3
50 100 150 200 250
-80
-60
-40
-20
0
20
40
60
80
50 100 150 200 250
-200
-150
-100
-50
0
50
100
150
200
50 100 150 200 250
-200
-150
-100
-50
0
50
100
150
Door opening recognition
• User identification
Door opening recognition (cont’d)
• Recognition performance
– Left: axis 5 Right: axis 1&5
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 1
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 2
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 3
0 20 40 60 80 100 120 140 160 180 200
-200
-150
-100
-50
0
50
100
150
200
Class 4
Gait recognition
• Via slipper
Image: www.footbionics.com
Gait recognition(cont’d)
• Recognition capability
– Left: axis 2 Right: axis 2&3&4
0 5 10 15 20 25 30 35 40
-1
-0.5
0
0.5
1
1.5
2
Class 1
0 5 10 15 20 25 30 35 40
-1
-0.5
0
0.5
1
1.5
2
Class 2
0 5 10 15 20 25 30 35 40
-1
-0.5
0
0.5
1
1.5
2
Class 3
0 5 10 15 20 25 30 35 40
-1
-0.5
0
0.5
1
1.5
2
Class 4
Demonstration
• Demo workflow
Demonstration(cont’d)
Conclusion
• Propose a sparse representation for time series
• Propose a heuristic to determine envelope size 𝑘
• Effectiveness, efficiency, robustness verification
• Real-world use case

More Related Content

PDF
Neural Networks: Least Mean Square (LSM) Algorithm
PDF
Defense_Talk
PDF
Estimating Future Initial Margin with Machine Learning
PPTX
Gradient descent optimizer
PDF
Decision Forests and discriminant analysis
PDF
Andres hernandez ai_machine_learning_london_nov2017
PPTX
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
PDF
CUDA and Caffe for deep learning
Neural Networks: Least Mean Square (LSM) Algorithm
Defense_Talk
Estimating Future Initial Margin with Machine Learning
Gradient descent optimizer
Decision Forests and discriminant analysis
Andres hernandez ai_machine_learning_london_nov2017
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
CUDA and Caffe for deep learning

What's hot (19)

PDF
Una introducción a la minería de series temporales
PDF
Sergei Vassilvitskii, Research Scientist, Google at MLconf NYC - 4/15/16
PPTX
ACM 2013-02-25
PPTX
Fast Single-pass K-means Clusterting at Oxford
PDF
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
PDF
Neural Networks: Support Vector machines
PDF
Svm map reduce_slides
PDF
Stochastic gradient descent and its tuning
PDF
Anil Thomas - Object recognition
PDF
Large scale logistic regression and linear support vector machines using spark
PPTX
Support Vector Machines Simply
PDF
Stochastic Alternating Direction Method of Multipliers
PDF
MLHEP Lectures - day 2, basic track
PPTX
presentation_btp
PDF
Paper Study: Melding the data decision pipeline
PDF
Neural Networks: Model Building Through Linear Regression
PDF
MLHEP Lectures - day 1, basic track
Una introducción a la minería de series temporales
Sergei Vassilvitskii, Research Scientist, Google at MLconf NYC - 4/15/16
ACM 2013-02-25
Fast Single-pass K-means Clusterting at Oxford
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
Neural Networks: Support Vector machines
Svm map reduce_slides
Stochastic gradient descent and its tuning
Anil Thomas - Object recognition
Large scale logistic regression and linear support vector machines using spark
Support Vector Machines Simply
Stochastic Alternating Direction Method of Multipliers
MLHEP Lectures - day 2, basic track
presentation_btp
Paper Study: Melding the data decision pipeline
Neural Networks: Model Building Through Linear Regression
MLHEP Lectures - day 1, basic track
Ad

Viewers also liked (14)

PDF
期末最終報告 NTUST 大學專題
PPT
人工智慧 & 機器學習 bacon number
PPTX
Presentation for interview
PDF
Pmi servizio pca
PDF
Pmi servizio mb
PPT
Las libertades del aire grupo 5
PDF
Pmi servizio sbr
PDF
What Makes Great Infographics
PDF
Masters of SlideShare
PDF
STOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
PDF
You Suck At PowerPoint!
PDF
10 Ways to Win at SlideShare SEO & Presentation Optimization
PDF
How To Get More From SlideShare - Super-Simple Tips For Content Marketing
PDF
How to Make Awesome SlideShares: Tips & Tricks
期末最終報告 NTUST 大學專題
人工智慧 & 機器學習 bacon number
Presentation for interview
Pmi servizio pca
Pmi servizio mb
Las libertades del aire grupo 5
Pmi servizio sbr
What Makes Great Infographics
Masters of SlideShare
STOP! VIEW THIS! 10-Step Checklist When Uploading to Slideshare
You Suck At PowerPoint!
10 Ways to Win at SlideShare SEO & Presentation Optimization
How To Get More From SlideShare - Super-Simple Tips For Content Marketing
How to Make Awesome SlideShares: Tips & Tricks
Ad

Similar to Compressed learning for time series classification (20)

PDF
Gaussian Dictionary for Compressive Sensing of the ECG Signal
PDF
IRJET- Reconstruction of Sparse Signals(Speech) Using Compressive Sensing
PDF
MASSS_Presentation_20160209
PDF
Image compression based on
PDF
Performance analysis of compressive sensing recovery algorithms for image pr...
PDF
Support Recovery with Sparsely Sampled Free Random Matrices for Wideband Cogn...
PDF
Comparative Study of Compressive Sensing Techniques For Image Enhancement
PDF
Compressed sensing system for efficient ecg signal
PDF
Energy Efficient Compression of Shock Data using Compressed Sensing
PDF
50120140501009
PDF
Compressed sensing techniques for sensor data using unsupervised learning
PDF
(eBook PDF) Compressed Sensing in Radar Signal Processing
PDF
(eBook PDF) Compressed Sensing in Radar Signal Processing
PDF
CS-ChapterI.pdf
PDF
(eBook PDF) Compressed Sensing in Radar Signal Processing
PDF
(eBook PDF) Compressed Sensing in Radar Signal Processing
PDF
Provably secure and efficient audio compression based on compressive sensing
PDF
Matrix Padding Method for Sparse Signal Reconstruction
Gaussian Dictionary for Compressive Sensing of the ECG Signal
IRJET- Reconstruction of Sparse Signals(Speech) Using Compressive Sensing
MASSS_Presentation_20160209
Image compression based on
Performance analysis of compressive sensing recovery algorithms for image pr...
Support Recovery with Sparsely Sampled Free Random Matrices for Wideband Cogn...
Comparative Study of Compressive Sensing Techniques For Image Enhancement
Compressed sensing system for efficient ecg signal
Energy Efficient Compression of Shock Data using Compressed Sensing
50120140501009
Compressed sensing techniques for sensor data using unsupervised learning
(eBook PDF) Compressed Sensing in Radar Signal Processing
(eBook PDF) Compressed Sensing in Radar Signal Processing
CS-ChapterI.pdf
(eBook PDF) Compressed Sensing in Radar Signal Processing
(eBook PDF) Compressed Sensing in Radar Signal Processing
Provably secure and efficient audio compression based on compressive sensing
Matrix Padding Method for Sparse Signal Reconstruction

Recently uploaded (20)

PPTX
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PPTX
IB Computer Science - Internal Assessment.pptx
PDF
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
PDF
Lecture1 pattern recognition............
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PPTX
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
PPTX
Database Infoormation System (DBIS).pptx
PPTX
Business Acumen Training GuidePresentation.pptx
PDF
.pdf is not working space design for the following data for the following dat...
PPTX
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
PPT
ISS -ESG Data flows What is ESG and HowHow
PDF
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
PPTX
Qualitative Qantitative and Mixed Methods.pptx
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PDF
Mega Projects Data Mega Projects Data
PPTX
Supervised vs unsupervised machine learning algorithms
PDF
Fluorescence-microscope_Botany_detailed content
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
IB Computer Science - Internal Assessment.pptx
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
Lecture1 pattern recognition............
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
Data_Analytics_and_PowerBI_Presentation.pptx
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
Database Infoormation System (DBIS).pptx
Business Acumen Training GuidePresentation.pptx
.pdf is not working space design for the following data for the following dat...
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
ISS -ESG Data flows What is ESG and HowHow
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
Qualitative Qantitative and Mixed Methods.pptx
Galatica Smart Energy Infrastructure Startup Pitch Deck
Mega Projects Data Mega Projects Data
Supervised vs unsupervised machine learning algorithms
Fluorescence-microscope_Botany_detailed content

Compressed learning for time series classification

  • 1. Compressed Learning for Time Series Classification Shueh-Han Shih Department of Computer Science and Information Engineering, National Taiwan University of Science and Technology
  • 2. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion
  • 3. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion
  • 5. Motivation (cont’d) • The key to handle time series data effectively is choosing a suitable representation • Transmission and storage issues are critical in IoT scenario • To provide interpretable result for human is important  Time series sparse representation - envelope
  • 6. Time series data type 1. Symbolic sequence 2. Complex symbolic sequence 3. Simple time series 4. Multivariate time series A brief survey on sequence classification Z Xing, J Pei, E Keogh - ACM SIGKDD , 2010
  • 7. Classification of time series • Assigning instances to one of the predefined classes. 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 1 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 2 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 3 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 4
  • 8. Time series classification approaches • Feature based • Sequence distance based • Model based A brief survey on sequence classification Z Xing, J Pei, E Keogh - ACM SIGKDD, 2010
  • 9. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion
  • 10. Conventional approach • Number of sample needed for compressed sensing is much more lower than Nyquist frequency. Image: http://www.ni.com/
  • 11. Main idea • Most real-world signals are sparse in some basis A𝑥 = 𝑦, A ∈ ℝ 𝑝×𝑛 𝑎𝑛𝑑 𝑝 ≪ 𝑛 • Dramatically reduce the transmission loading a measure
  • 12. Requirements of compressed sensing 1. 𝑥 should be a 𝑘-sparse signal  1 to 1 relation between data and compressed domain 2. A must satisfies the restricted isometry property  (1 − δ 𝑝)ǁ𝑥ǁ2 2 ≤ ǁA𝑥ǁ2 2 ≤ (1 + δ 𝑝)ǁ𝑥ǁ2 2 A = 𝑟𝑎𝑛𝑑𝑛 𝑝,𝑛 𝑛 (mean=0, 𝜎 = 1 𝑛 ) for some constant 𝛿 𝑝 ∈ (0, 1 Image: Mostafa Mohsenvand Projects
  • 13. Basic routine 𝑨 = 𝑸 𝑨′ = 𝑸𝑷 𝒚 = 𝑨𝒙(𝒐𝒓 𝑨′ 𝒔) transmission • Postpone the computational cost to recovery stage
  • 14. Learning in the compressed domain • Perform task without recovery • SVM can keep the learnability in compressed domain • Reduce model complexity Image: Compressed learning: Universal sparse dimensionality reduction and learning in the measurement domain. R Calderbank, S Jafarpour, R Schapire - preprint, 2009 - dsp.rice.edu
  • 15. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion
  • 16. The origin • The ‘envelope’ in finance Image: http://www.investopedia.com/
  • 17. Preliminaries • A time series – T = 𝑡1, 𝑡2, … 𝑡𝑗 … 𝑡 𝑛 , 𝑡𝑗 ∈ ℝ • A time series dataset – D = T 𝑖 | T 𝑖 = 𝑡1 𝑖 , 𝑡2 𝑖 , … 𝑡 𝑛 𝑖 , 𝑖 = 1 𝑡𝑜 𝑚 • Well-synchronized with the same length – A set of random sample from random variables 𝐓1, 𝐓2, … 𝐓𝑗, … 𝐓𝑛
  • 18. Envelope creation • Given 𝐷, envelope with size 𝑘 – E 𝑘 = 𝑍 𝑍 = 𝑧1, 𝑧2, … 𝑧 𝑛 , 𝑧𝑗 − 𝜇 𝑗 ≤ 𝑘 ∙ 𝑠𝑡𝑑𝑗 , ∀ 𝑧𝑗 ∈ ℝ} • 𝜇 𝑗 = 𝑚𝑒𝑎𝑛(𝑻𝑗) , 𝑠𝑡𝑑𝑗 = 𝑠𝑡𝑑(𝑻𝑗) – Profiling the time series dataset
  • 19. Envelope encoding • Encoding time series T as a sparse series S • Sparsity indicates the similarity of a time series and 𝐷 𝑠𝑗 = 1, 𝑖𝑓 𝑡𝑗 > 𝜇 𝑗 + 𝑘 ∙ 𝑠𝑡𝑑𝑗 𝑠𝑗 = −1, 𝑖𝑓 𝑡𝑗 < 𝜇 𝑗 − 𝑘 ∙ 𝑠𝑡𝑑𝑗 𝑠𝑗 = 0, 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒 , 𝑓𝑜𝑟 𝑗 = 1 𝑡𝑜 𝑛
  • 20. Guarantee of sparsity • According to Chebyshev’s inequality, – Pr(|X − 𝜇| ≤ 𝑘𝜎) ≥ 1 − 1 𝑘2 – No matter what kind of distribution for 𝑻𝑗
  • 21. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion
  • 22. Determination of 𝑘 • 𝑘 affects the effectiveness of envelope representation
  • 23. Determination of 𝑘 (cont’d) • Focus on time series multi-class classification – Envelope representation should be discriminative 𝑘∗ = arg max 𝑘 (−𝑎 𝑘 + 𝜆 ∙ 𝑏 𝑘) – 𝑘 : tradeoff between sparsity and distinguishability
  • 24. Encoding result visualization • Sparsity indicates similarity Sparse property ⟹ transmission efficiency • Encoding results are interpretable ECGFiveDays from UCR
  • 25. Envelope representation workflow • From raw series to feature
  • 27. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion 1. Proposed method vs. state-of-art method on classification task 2. Compressibility of envelope representation with compressed sensing 3. Noise resistance of envelope representation 4. Time efficiency 5. Space efficiency
  • 28. Classification performance • Benchmark dataset from UCR (5/42) Dataset/ Algorithm Number of classes Size of training set Size of testing set Time series Length CBF 3 30 900 128 Coffee 2 28 28 286 ECGFiveDays 2 23 861 136 ItalyPowerDemand 2 67 1029 24 Sony II 2 27 953 65
  • 29. Classification performance (cont’d) • Result on benchmark dataset (5/42) – Win:9 / lose:18 / between:15 (close:12) – Not the case with IoT scenario, never lack of data Dataset/ Algorithm 1NN-Euclidean 1NN-DTW (best, noWin) Envelope+ linearSVM CBF 85.2(0.9357) 99.6/99.7 90.66 Coffee 75(0.031608) 82.1/82.1 85.71 ECGFiveDays 79.7(0.8758) 79.7/76.8 88.38 ItalyPowerDemand 95.5(1.0661) 95.5/95 97.08 Sony II 69.5(0.9986) 69.5/72.5 82.79
  • 30. Influence of compression ratio • Compression ratio = 𝑝/𝑛 (Number of measurements) / (data dimension)
  • 31. Influence of compression ratio (cont’d) • Using nearly 1 3 datasets from UCR – Some datasets have excellent compressibility
  • 32. Influence of compression ratio (cont’d) • Result on benchmark dataset (5/42) Dataset/ Algorithm 1NN- Euclidean 1NN-DTW (best, noWin) Compression ratio=10% Compression ratio=20% Compression ratio=50% CBF 85.2 99.6/99.7 80.44 88.22 88.44 Coffee 75 82.1/82.1 71.42 82.14 89.28 ECGFiveDays 79.7 79.7/76.8 78.86 81.3 81.64 ItalyPowerDemand 95.5 95.5/95 86.58 91.73 93.97 Sony II 69.5 69.5/72.5 76.91 78.38 80.06
  • 33. Robustness to noise • Noise level - SNR Image: documentation.meraki.com
  • 34. Robustness to noise (cont’d) • Using ECG200 dataset as example The original envelope The envelope with noise level SNR=10.
  • 35. Robustness to noise (cont’d) • Envelope representation is noise-resistant – Can even ignore denoising stage Envelope built/SVM trained with clean data Envelope built/SVM trained with noisy data
  • 36. Time efficiency 1. Building envelope takes O(m*n) 2. Encoding each instance takes O(n) 3. Linear SVM, expects to be O(m2)  Linear time in prediction 0 2 4 6 8 10 12 14 16 0 2 4 6 8 10 12 14 16 Execution time (testing) envelope (Sec.) KNN+ED(Sec.)
  • 37. Space efficiency 1. 32 to (2 ∗ #𝑐𝑙𝑎𝑠𝑠) ratio of reduction 2. 32 to (32 ∗ #𝑐𝑙𝑎𝑠𝑠 ∗ 𝑐𝑜𝑚𝑝𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝑟𝑎𝑡𝑖𝑜) ratio of reduction through compressed sensing 3. Run length encoding
  • 38. Outline • Introduction • Compressed sensing • Sparse representation - envelope • Classification framework • Experimental results • Case study • Conclusion
  • 39. Smart home project • Passive user identification Image: www.bitronvideo.eu
  • 40. Using sensor • Data collection – EcoBT Mini – 33Hz 50 100 150 200 250 -0.1 -0.05 0 0.05 0.1 0.15 0.2 0.25 50 100 150 200 250 -0.4 -0.3 -0.2 -0.1 0 0.1 0.2 0.3 50 100 150 200 250 -0.4 -0.3 -0.2 -0.1 0 0.1 0.2 0.3 50 100 150 200 250 -80 -60 -40 -20 0 20 40 60 80 50 100 150 200 250 -200 -150 -100 -50 0 50 100 150 200 50 100 150 200 250 -200 -150 -100 -50 0 50 100 150
  • 41. Door opening recognition • User identification
  • 42. Door opening recognition (cont’d) • Recognition performance – Left: axis 5 Right: axis 1&5 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 1 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 2 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 3 0 20 40 60 80 100 120 140 160 180 200 -200 -150 -100 -50 0 50 100 150 200 Class 4
  • 43. Gait recognition • Via slipper Image: www.footbionics.com
  • 44. Gait recognition(cont’d) • Recognition capability – Left: axis 2 Right: axis 2&3&4 0 5 10 15 20 25 30 35 40 -1 -0.5 0 0.5 1 1.5 2 Class 1 0 5 10 15 20 25 30 35 40 -1 -0.5 0 0.5 1 1.5 2 Class 2 0 5 10 15 20 25 30 35 40 -1 -0.5 0 0.5 1 1.5 2 Class 3 0 5 10 15 20 25 30 35 40 -1 -0.5 0 0.5 1 1.5 2 Class 4
  • 47. Conclusion • Propose a sparse representation for time series • Propose a heuristic to determine envelope size 𝑘 • Effectiveness, efficiency, robustness verification • Real-world use case

Editor's Notes

  • #3: Time series classification (dis)advantage of CS Time series representation method Then experiments Last,
  • #5: Data collected continuously, time correlated series; communication & storage issue Reduce dimension & recover with little loss Extract info. From TS for ML model www.aeris.com www.comp-engine.org www.ceremade.dauphine.fr
  • #6: In order to ~~~ we propose envelope…
  • #7: alphabet of symbols e.g. DNA complex symbolic sequence e.g. transaction simple time series e.g. electric meter multivariate time series e.g. EEG www.bios.net www.apps.rus.mto.gov.on.ca www.rowetel.com www.dianliwenmi.com
  • #8: Assign new instance to certain class based on given data Using the example from case study
  • #9: model each part as one state; the mean of the state is the mean estimated from that part For a gesture recognizer we build multiple of these models, one for each gesture. a training set to estimate the parameters of models. During recognition we simply pick the model describing the data best.
  • #10: CS
  • #11: Lower transmission loading & recover well 取樣率 (f s) > 2 *受測訊號的最高頻率部份, 否則高頻的內容會失真(Aliasing)
  • #12: The goal of compressed sensing is to provide measurement matrix A, with the number of measurements m as small as possible M is the #sample, which is << Nyquist
  • #13: Normal Random matrix A generated with specific parameters is usually good enough for most real world applications.
  • #14: Most efforts are cost in recovery stage (discard this)
  • #15: The erroe of SVM in the measurement domain is with high probability close to the error of the best linear classifier in the data domain
  • #17: The idea of ‘envelope’ has been applied in finance for a long time used by investors and traders to help identify extreme overbought and oversold conditions. (兩端分別是由開盤價和收盤價)
  • #18: a vector in temporal order D is well-synchronized with the same length regarded as random samples from a set of random variables
  • #19: A set of values covered mu+- k*std. Envelope is the profile of the dataset
  • #21: possibility of applying CS for further boost
  • #23: 𝑘 is a critical issue, which directly affect the distinguishability /sparsity
  • #24: Propose a heuristic to make envelope distinguishable Large/small k will affect distinguishability
  • #26: Options for concise format
  • #27: libSVM with 1 to 1 (shorter training time)
  • #29: few datasets from UCR comparing the performance of proposed method with state-of-art
  • #30: Envelope may have inferior performance due to the lack of training instances
  • #31: possible to reduce the size to about 20%~10% and still keep the classification performance, which is very promising.
  • #33: Still keep good performance after compression
  • #34: 為訊號功率(Power of Signal)。 為雜訊功率(Power of Noise)。 為訊號振幅(Amplitude of Signal)。 為雜訊振幅(Amplitude of Noise)。
  • #35: Robustness of proposed method to noise
  • #36: proposed method is noise-resisting.
  • #37: Faster than Knn+ED
  • #38: Space-saving
  • #40: Using multiple devices with weak models integrate them to get better performance
  • #42: Using BLE for transmission
  • #43: identify users from door opening trajectories.
  • #44: treat each step as a time series instance
  • #45: the proposed method is also suitable for distinct cases. Integrate results of multi-steps to get better performance
  • #46: Door and slippers make distinct predictions
  • #47: Demo
  • #48: Intro. of TS classification, Pros & cons of CS Supervised feature extraction technique Heuristic Benchmark, noise, compression Real-world cases