SlideShare a Scribd company logo
Sparse Distributed Representations:
Our Brain’s Data Structure
Numenta Workshop
October 17, 2014
Subutai Ahmad, VP Research
sahmad@numenta.com
Sparse Distributed Representations: Our Brain's Data Structure
Sparse Distributed Representations: Our Brain's Data Structure
Sparse Distributed Representations:
Our Brain’s Data Structure
Numenta Workshop
October 17, 2014
Subutai Ahmad, VP Research
sahmad@numenta.com
The Role of Sparse Distributed Representations in Cortex
1) Sensory perception
3) Motor control
4) Prediction
2) Planning
5) Attention
Sparse Distribution Representations (SDRs) are the foundation for all these
functions, across all sensory modalities
Analysis of this common cortical data structure can provide a rigorous
foundation for cortical computing
Talk Outline
1) Introduction to Sparse Distributed Representations (SDRs)
2) Fundamental properties of SDRs
– Error bounds
– Scaling laws
From: Prof. Hasan, Max-Planck-
Institut for Research
Basics Attributes of SDRs
1) Only a small number of neurons are firing
at any point in time
3) Every cell represents something and has
meaning
4) Information is distributed and no single
neuron is critical
2) There are a very large number of neurons
5) Every neuron only connects to a subset of
other neurons
6) SDRs enable extremely fast computation
7) SDRs are binary
x = 0100000000000000000100000000000110000000
Multiple input SDR’s Single bit in an output SDR
How Does a Single Neuron Operate on SDRs?
Proximal segments
represent dozens of
separate patterns in a single
segment
How Does a Single Neuron Operate on SDRs?
Hundreds of distal segments each detect a
unique SDR using a threshold
Feedback SDR
Context SDR
Bottom-up input SDR
In both cases each synapse corresponds to one bit in
the incoming high dimensional SDR
• Extremely high capacity
• Recognize patterns in the presence of noise
• Robust to random deletions
• Represent dynamic set of patterns in a single fixed structure
• Extremely efficient
Fundamental Properties of SDRs
Notation
• We represent a SDR vector as a vector with n binary values
where each bit represents the activity of a single neuron:
• s = percent of ON bits, w = number of ON bits
x =[b0,… ,bn-1]
wx = s ´ n = x 1
Example
• n = 40, s = 0.1, w = 4
• Typical range of numbers in HTM implementations:
n = 2048 to 65,536 s = 0.05% to 2% w = 40
y =1000000000000000000100000000000110000000
x = 0100000000000000000100000000000110000000
SDRs Have Extremely High Capacity
• The number of unique patterns that can be represented is:
• This is far smaller than 2n, but far larger than any reasonable need
• Example: with n = 2048 and w = 40,
the number of unique patterns is > 1084 >> # atoms in universe
• Chance that two random vectors are identical is essentially zero:
n
w
æ
èç
ö
ø÷ =
n!
w! n - w( )!
1/
n
w
æ
èç
ö
ø÷
• Extremely high capacity
• Recognize patterns in the presence of noise
• Robust to random deletions
• Represent multiple patterns in a single fixed structure
• Extremely efficient
Fundamental Properties of SDRs
Similarity Metric for Recognition of SDR Patterns
• We don’t use typical vector similarities
– Neurons cannot compute Euclidean or Hamming distance between SDRs
– Any p-norm requires full connectivity
• Compute similarity using an overlap metric
– The overlap is simply the number of bits in common
– Requires only minimal connectivity
– Mathematically, take the AND of two vectors and compute its length
• Detecting a “Match”
– Two SDR vectors “match” if their overlap meets a minimum threshold
overlap(x,y) º x Ù y
match(x,y) º overlap(x,y) ³q
q
Overlap example
• N=40, s=0.1, w=4
• The two vectors have an overlap of 3, so they “match” if the
threshold is 3.
y =1000000000000000000100000000000110000000
x = 0100000000000000000100000000000110000000
How Accurate is Matching With Noise?
• As you decrease the match threshold , you decrease sensitivity and increase
robustness to noise
• You also increase the chance of false positives
Decrease
q
q
How Many Vectors Match When You Decrease the Threshold?
• Define the “overlap set of x” to be the set of
vectors with exactly b bits of overlap with x
• The number of such vectors is:
Wx (n,w,b) =
wx
b
æ
èç
ö
ø÷ ´
n - wx
w - b
æ
èç
ö
ø÷
Wx (n,w,b)
Number subsets of x with
exactly b bits ON
Number patterns occupying the rest
of the vector with exactly w-b bits ON
Error Bound for Classification with Noise
• Give a single stored pattern, probability of false positive is:
• Given M patterns, probability of a false positive is:
fpw
n
(q) =
Wx (n,w,b)
b=q
w
å
n
w
æ
èç
ö
ø÷
fpX (q) £ fpwxi
n
(q)
i=0
M-1
å
What Does This Mean in Practice?
• With SDRs you can classify a huge number of patterns with substantial noise
(if n and w are large enough)
Examples
• n = 2048, w = 40
With up to 14 bits of noise (33%), you can classify a quadrillion
patterns with an error rate of less than 10-24
With up to 20 bits of noise (50%), you can classify a quadrillion
patterns with an error rate of less than 10-11
• n = 64, w=12
With up to 4 bits of noise (33%), you can classify 10 patterns
with an error rate of 0.04%
Neurons Are Highly Robust Pattern Recognizers
Hundreds of distal segments each detect a
unique SDR using a threshold
You can have tens of thousands of neurons examining a single input SDR, and very
robustly matching complex patterns
• Extremely high capacity
• Recognize patterns in the presence of noise
• Robust to random deletions
• Represent multiple patterns in a single fixed structure
• Extremely efficient
Fundamental Properties of SDRs
SDRs are Robust to Random Deletions
• In cortex bits in an SDR can randomly disappear
– Synapses can be quite unreliable
– Individual neurons can die
– A patch of cortex can be damaged
• The analysis for random deletions is very similar to noise
• SDRs can naturally handle fairly significant random failures
– Failures are tolerated in any SDR and in any part of the system
• This is a great property for those building HTM based hardware
– The probability of failures can be exactly characterized
• Extremely high capacity
• Recognize patterns in the presence of noise
• Robust to random deletions
• Represent multiple patterns in a single fixed structure
• Extremely efficient
Fundamental Properties of SDRs
Representing Multiple Patterns in a Single SDR
• There are situations where we want to store multiple patterns within a single SDR
and match them
• In temporal inference the system might make multiple predictions about the future
Example
Unions of SDRs
• We can store a set of patterns in a single fixed representation by taking the OR of
all the individual patterns
• The vector representing the union is also going to match a large number of other
patterns that were not one of the original 10
• How many such patterns can we store reliably, without a high chance of false
positives?
Is this SDR
a member?
1)
2)
3)
….
10)
2%
< 20%Union
Error Bounds for Unions
• Expected number of ON bits:
• Give a union of M patterns, the expected probability of a false positive (with
noise) is:
What Does This Mean in Practice?
• You can form reliable unions of a reasonable number of patterns (assuming
large enough n and w)
Examples
• n = 2048, w = 40
The union of 50 patterns leads to an error rate of 10-9
• n = 512, w=10
The union of 50 patterns leads to an error rate of 0.9%
• Extremely high capacity
• Recognize patterns in the presence of noise
• Robust to random deletions
• Represent multiple patterns in a single fixed structure
• Extremely efficient
Fundamental Properties of SDRs
SDRs Enable Highly Efficient Operations
• In cortex complex operations are carried out rapidly
– Visual system can perform object recognition in 100-150 msecs
• SDR vectors are large, but all operations are O(w) and independent of
vector size
– No loops or optimization process required
• Matching a pattern against a dynamic list (unions) is O(w) and
independent of the number of items in the list
• Enables a tiny dendritic segment to perform robust pattern recognition
• We can simulate 200,000 neurons in software at about 25-50Hz
Summary
• SDR’s are the common data structure in the cortex
• SDR’s enable flexible recognition systems that have very high capacity, and are
robust to a large amount of noise
• The union property allows a fixed representation to encode a dynamically
changing set of patterns
• The analysis of SDR’s provides a principled foundation for characterizing the
behavior of the HTM learning algorithms and all cognitive functions
• Sparse memory (Kanerva), Sparse coding (Olshausen), Bloom filters (Broder)
Related work
Questions? Math jokes?
Follow us on Twitter @numenta
Sign up for our newsletter at www.numenta.com
Subutai Ahmad
sahmad@numenta.com
nupic-theory mailing list
numenta.org/lists

More Related Content

PDF
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
PPTX
Why Neurons have thousands of synapses? A model of sequence memory in the brain
PDF
Numenta Brain Theory Discoveries of 2016/2017 by Jeff Hawkins
PDF
Does the neocortex use grid cell-like mechanisms to learn the structure of ob...
PDF
The Predictive Neuron: How Active Dendrites Enable Spatiotemporal Computation...
PDF
BAAI Conference 2021: The Thousand Brains Theory - A Roadmap for Creating Mac...
PPT
HTM Theory
PPTX
Have We Missed Half of What the Neocortex Does? by Jeff Hawkins (12/15/2017)
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
Why Neurons have thousands of synapses? A model of sequence memory in the brain
Numenta Brain Theory Discoveries of 2016/2017 by Jeff Hawkins
Does the neocortex use grid cell-like mechanisms to learn the structure of ob...
The Predictive Neuron: How Active Dendrites Enable Spatiotemporal Computation...
BAAI Conference 2021: The Thousand Brains Theory - A Roadmap for Creating Mac...
HTM Theory
Have We Missed Half of What the Neocortex Does? by Jeff Hawkins (12/15/2017)

What's hot (20)

PDF
Jeff Hawkins NAISys 2020: How the Brain Uses Reference Frames, Why AI Needs t...
PPTX
The Predictive Neuron: How Active Dendrites Enable Spatiotemporal Computation...
PDF
Could A Model Of Predictive Voting Explain Many Long-Range Connections? by Su...
PPTX
Jeff Hawkins Human Brain Project Summit Keynote: "Location, Location, Locatio...
PPTX
Location, Location, Location - A Framework for Intelligence and Cortical Comp...
PDF
Sparsity In The Neocortex, And Its Implications For Machine Learning
PPTX
Locations in the Neocortex: A Theory of Sensorimotor Prediction Using Cortica...
PPTX
Have We Missed Half of What the Neocortex Does? A New Predictive Framework ...
PPTX
Brains, Data, and Machine Intelligence (2014 04 14 London Meetup)
PDF
CVPR 2020 Workshop: Sparsity in the neocortex, and its implications for conti...
PDF
RNN Explore
PDF
Neural networks
PPTX
Artificial neural networks and its applications
PPTX
Artificial Neural Networks: Pointers
PDF
Intro to Neural Networks
PPTX
Neural network & its applications
PDF
Artificial neural networks and its application
PPT
Neural Networks
PPT
NEURAL NETWORKS
PPT
Neural networks
Jeff Hawkins NAISys 2020: How the Brain Uses Reference Frames, Why AI Needs t...
The Predictive Neuron: How Active Dendrites Enable Spatiotemporal Computation...
Could A Model Of Predictive Voting Explain Many Long-Range Connections? by Su...
Jeff Hawkins Human Brain Project Summit Keynote: "Location, Location, Locatio...
Location, Location, Location - A Framework for Intelligence and Cortical Comp...
Sparsity In The Neocortex, And Its Implications For Machine Learning
Locations in the Neocortex: A Theory of Sensorimotor Prediction Using Cortica...
Have We Missed Half of What the Neocortex Does? A New Predictive Framework ...
Brains, Data, and Machine Intelligence (2014 04 14 London Meetup)
CVPR 2020 Workshop: Sparsity in the neocortex, and its implications for conti...
RNN Explore
Neural networks
Artificial neural networks and its applications
Artificial Neural Networks: Pointers
Intro to Neural Networks
Neural network & its applications
Artificial neural networks and its application
Neural Networks
NEURAL NETWORKS
Neural networks
Ad

Viewers also liked (10)

PDF
Predictive Analytics with Numenta Machine Intelligence
PPTX
a tour of several popular tensorflow models
PPTX
What the Brain says about Machine Intelligence
PDF
Recognizing Locations on Objects by Marcus Lewis
PDF
The Biological Path Towards Strong AI Strange Loop 2017, St. Louis
PPTX
Getting Started with Numenta Technology
PPTX
Applications of Hierarchical Temporal Memory (HTM)
PPTX
Beginner's Guide to NuPIC
PPTX
HTM Spatial Pooler
PDF
TouchNet preview at Numenta
Predictive Analytics with Numenta Machine Intelligence
a tour of several popular tensorflow models
What the Brain says about Machine Intelligence
Recognizing Locations on Objects by Marcus Lewis
The Biological Path Towards Strong AI Strange Loop 2017, St. Louis
Getting Started with Numenta Technology
Applications of Hierarchical Temporal Memory (HTM)
Beginner's Guide to NuPIC
HTM Spatial Pooler
TouchNet preview at Numenta
Ad

Similar to Sparse Distributed Representations: Our Brain's Data Structure (20)

PPTX
Hierarchical Temporal Memory for Real-time Anomaly Detection
PDF
is2015_poster
PPTX
CNN for modeling sentence
PDF
ARTIFICIAL-NEURAL-NETWORKMACHINELEARNING
PDF
Lect#2_Pattern_Recognition_PGIT204D_By_Dr_TSSinha.pdf
PPTX
Introduction to artificial neural network.pptx
PDF
Artificial Neural Networks Lect1: Introduction & neural computation
PPTX
Introduction to spred spectrum and CDMA
PPTX
Wits presentation 6_28072015
PDF
SF Big Analytics20170706: What the brain tells us about the future of streami...
PPTX
• An attacker’s aim for carrying out a CSRF attack is to force the user to su...
PDF
Ted Willke - The Brain’s Guide to Dealing with Context in Language Understanding
PDF
Artificial Neural Network (draft)
PPTX
Lecture on Deep Learning
PPTX
Neural Networks for Machine Learning and Deep Learning
PPTX
10.0 SequenceModeling-merged-compressed_edited.pptx
PDF
Lecture 11 Neural network and fuzzy system
PPTX
Information Retrieval Dynamic Time Warping - Interspeech 2013 presentation
PDF
From neural networks to deep learning
PDF
Why do we need machine learning? Neural Networks for Machine Learning Lectur...
Hierarchical Temporal Memory for Real-time Anomaly Detection
is2015_poster
CNN for modeling sentence
ARTIFICIAL-NEURAL-NETWORKMACHINELEARNING
Lect#2_Pattern_Recognition_PGIT204D_By_Dr_TSSinha.pdf
Introduction to artificial neural network.pptx
Artificial Neural Networks Lect1: Introduction & neural computation
Introduction to spred spectrum and CDMA
Wits presentation 6_28072015
SF Big Analytics20170706: What the brain tells us about the future of streami...
• An attacker’s aim for carrying out a CSRF attack is to force the user to su...
Ted Willke - The Brain’s Guide to Dealing with Context in Language Understanding
Artificial Neural Network (draft)
Lecture on Deep Learning
Neural Networks for Machine Learning and Deep Learning
10.0 SequenceModeling-merged-compressed_edited.pptx
Lecture 11 Neural network and fuzzy system
Information Retrieval Dynamic Time Warping - Interspeech 2013 presentation
From neural networks to deep learning
Why do we need machine learning? Neural Networks for Machine Learning Lectur...

More from Numenta (13)

PDF
Deep learning at the edge: 100x Inference improvement on edge devices
PDF
Brains@Bay Meetup: A Primer on Neuromodulatory Systems - Srikanth Ramaswamy
PDF
Brains@Bay Meetup: How to Evolve Your Own Lab Rat - Thomas Miconi
PDF
Brains@Bay Meetup: The Increasing Role of Sensorimotor Experience in Artifici...
PDF
Brains@Bay Meetup: Open-ended Skill Acquisition in Humans and Machines: An Ev...
PDF
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
PDF
SBMT 2021: Can Neuroscience Insights Transform AI? - Lawrence Spracklen
PDF
FPGA Conference 2021: Breaking the TOPS ceiling with sparse neural networks -...
PDF
OpenAI’s GPT 3 Language Model - guest Steve Omohundro
PDF
CVPR 2020 Workshop: Sparsity in the neocortex, and its implications for conti...
PDF
The Thousand Brains Theory: A Framework for Understanding the Neocortex and B...
PDF
The Biological Path Toward Strong AI by Matt Taylor (05/17/18)
PDF
Biological path toward strong AI
Deep learning at the edge: 100x Inference improvement on edge devices
Brains@Bay Meetup: A Primer on Neuromodulatory Systems - Srikanth Ramaswamy
Brains@Bay Meetup: How to Evolve Your Own Lab Rat - Thomas Miconi
Brains@Bay Meetup: The Increasing Role of Sensorimotor Experience in Artifici...
Brains@Bay Meetup: Open-ended Skill Acquisition in Humans and Machines: An Ev...
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
SBMT 2021: Can Neuroscience Insights Transform AI? - Lawrence Spracklen
FPGA Conference 2021: Breaking the TOPS ceiling with sparse neural networks -...
OpenAI’s GPT 3 Language Model - guest Steve Omohundro
CVPR 2020 Workshop: Sparsity in the neocortex, and its implications for conti...
The Thousand Brains Theory: A Framework for Understanding the Neocortex and B...
The Biological Path Toward Strong AI by Matt Taylor (05/17/18)
Biological path toward strong AI

Recently uploaded (20)

PPTX
Introduction-to-Cloud-ComputingFinal.pptx
PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PPT
Miokarditis (Inflamasi pada Otot Jantung)
PDF
Introduction to Data Science and Data Analysis
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PPTX
Introduction to Knowledge Engineering Part 1
PDF
Mega Projects Data Mega Projects Data
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PPTX
STUDY DESIGN details- Lt Col Maksud (21).pptx
PDF
annual-report-2024-2025 original latest.
PPTX
climate analysis of Dhaka ,Banglades.pptx
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PPT
ISS -ESG Data flows What is ESG and HowHow
PPTX
Introduction to machine learning and Linear Models
PPT
Reliability_Chapter_ presentation 1221.5784
PPT
Quality review (1)_presentation of this 21
PPTX
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
PPTX
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
Introduction-to-Cloud-ComputingFinal.pptx
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
Miokarditis (Inflamasi pada Otot Jantung)
Introduction to Data Science and Data Analysis
Galatica Smart Energy Infrastructure Startup Pitch Deck
Introduction to Knowledge Engineering Part 1
Mega Projects Data Mega Projects Data
Acceptance and paychological effects of mandatory extra coach I classes.pptx
STUDY DESIGN details- Lt Col Maksud (21).pptx
annual-report-2024-2025 original latest.
climate analysis of Dhaka ,Banglades.pptx
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
oil_refinery_comprehensive_20250804084928 (1).pptx
ISS -ESG Data flows What is ESG and HowHow
Introduction to machine learning and Linear Models
Reliability_Chapter_ presentation 1221.5784
Quality review (1)_presentation of this 21
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
Business Ppt On Nestle.pptx huunnnhhgfvu

Sparse Distributed Representations: Our Brain's Data Structure

  • 1. Sparse Distributed Representations: Our Brain’s Data Structure Numenta Workshop October 17, 2014 Subutai Ahmad, VP Research sahmad@numenta.com
  • 4. Sparse Distributed Representations: Our Brain’s Data Structure Numenta Workshop October 17, 2014 Subutai Ahmad, VP Research sahmad@numenta.com
  • 5. The Role of Sparse Distributed Representations in Cortex 1) Sensory perception 3) Motor control 4) Prediction 2) Planning 5) Attention Sparse Distribution Representations (SDRs) are the foundation for all these functions, across all sensory modalities Analysis of this common cortical data structure can provide a rigorous foundation for cortical computing
  • 6. Talk Outline 1) Introduction to Sparse Distributed Representations (SDRs) 2) Fundamental properties of SDRs – Error bounds – Scaling laws
  • 7. From: Prof. Hasan, Max-Planck- Institut for Research
  • 8. Basics Attributes of SDRs 1) Only a small number of neurons are firing at any point in time 3) Every cell represents something and has meaning 4) Information is distributed and no single neuron is critical 2) There are a very large number of neurons 5) Every neuron only connects to a subset of other neurons 6) SDRs enable extremely fast computation 7) SDRs are binary x = 0100000000000000000100000000000110000000
  • 9. Multiple input SDR’s Single bit in an output SDR How Does a Single Neuron Operate on SDRs?
  • 10. Proximal segments represent dozens of separate patterns in a single segment How Does a Single Neuron Operate on SDRs? Hundreds of distal segments each detect a unique SDR using a threshold Feedback SDR Context SDR Bottom-up input SDR In both cases each synapse corresponds to one bit in the incoming high dimensional SDR
  • 11. • Extremely high capacity • Recognize patterns in the presence of noise • Robust to random deletions • Represent dynamic set of patterns in a single fixed structure • Extremely efficient Fundamental Properties of SDRs
  • 12. Notation • We represent a SDR vector as a vector with n binary values where each bit represents the activity of a single neuron: • s = percent of ON bits, w = number of ON bits x =[b0,… ,bn-1] wx = s ´ n = x 1 Example • n = 40, s = 0.1, w = 4 • Typical range of numbers in HTM implementations: n = 2048 to 65,536 s = 0.05% to 2% w = 40 y =1000000000000000000100000000000110000000 x = 0100000000000000000100000000000110000000
  • 13. SDRs Have Extremely High Capacity • The number of unique patterns that can be represented is: • This is far smaller than 2n, but far larger than any reasonable need • Example: with n = 2048 and w = 40, the number of unique patterns is > 1084 >> # atoms in universe • Chance that two random vectors are identical is essentially zero: n w æ èç ö ø÷ = n! w! n - w( )! 1/ n w æ èç ö ø÷
  • 14. • Extremely high capacity • Recognize patterns in the presence of noise • Robust to random deletions • Represent multiple patterns in a single fixed structure • Extremely efficient Fundamental Properties of SDRs
  • 15. Similarity Metric for Recognition of SDR Patterns • We don’t use typical vector similarities – Neurons cannot compute Euclidean or Hamming distance between SDRs – Any p-norm requires full connectivity • Compute similarity using an overlap metric – The overlap is simply the number of bits in common – Requires only minimal connectivity – Mathematically, take the AND of two vectors and compute its length • Detecting a “Match” – Two SDR vectors “match” if their overlap meets a minimum threshold overlap(x,y) º x Ù y match(x,y) º overlap(x,y) ³q q
  • 16. Overlap example • N=40, s=0.1, w=4 • The two vectors have an overlap of 3, so they “match” if the threshold is 3. y =1000000000000000000100000000000110000000 x = 0100000000000000000100000000000110000000
  • 17. How Accurate is Matching With Noise? • As you decrease the match threshold , you decrease sensitivity and increase robustness to noise • You also increase the chance of false positives Decrease q q
  • 18. How Many Vectors Match When You Decrease the Threshold? • Define the “overlap set of x” to be the set of vectors with exactly b bits of overlap with x • The number of such vectors is: Wx (n,w,b) = wx b æ èç ö ø÷ ´ n - wx w - b æ èç ö ø÷ Wx (n,w,b) Number subsets of x with exactly b bits ON Number patterns occupying the rest of the vector with exactly w-b bits ON
  • 19. Error Bound for Classification with Noise • Give a single stored pattern, probability of false positive is: • Given M patterns, probability of a false positive is: fpw n (q) = Wx (n,w,b) b=q w å n w æ èç ö ø÷ fpX (q) £ fpwxi n (q) i=0 M-1 å
  • 20. What Does This Mean in Practice? • With SDRs you can classify a huge number of patterns with substantial noise (if n and w are large enough) Examples • n = 2048, w = 40 With up to 14 bits of noise (33%), you can classify a quadrillion patterns with an error rate of less than 10-24 With up to 20 bits of noise (50%), you can classify a quadrillion patterns with an error rate of less than 10-11 • n = 64, w=12 With up to 4 bits of noise (33%), you can classify 10 patterns with an error rate of 0.04%
  • 21. Neurons Are Highly Robust Pattern Recognizers Hundreds of distal segments each detect a unique SDR using a threshold You can have tens of thousands of neurons examining a single input SDR, and very robustly matching complex patterns
  • 22. • Extremely high capacity • Recognize patterns in the presence of noise • Robust to random deletions • Represent multiple patterns in a single fixed structure • Extremely efficient Fundamental Properties of SDRs
  • 23. SDRs are Robust to Random Deletions • In cortex bits in an SDR can randomly disappear – Synapses can be quite unreliable – Individual neurons can die – A patch of cortex can be damaged • The analysis for random deletions is very similar to noise • SDRs can naturally handle fairly significant random failures – Failures are tolerated in any SDR and in any part of the system • This is a great property for those building HTM based hardware – The probability of failures can be exactly characterized
  • 24. • Extremely high capacity • Recognize patterns in the presence of noise • Robust to random deletions • Represent multiple patterns in a single fixed structure • Extremely efficient Fundamental Properties of SDRs
  • 25. Representing Multiple Patterns in a Single SDR • There are situations where we want to store multiple patterns within a single SDR and match them • In temporal inference the system might make multiple predictions about the future Example
  • 26. Unions of SDRs • We can store a set of patterns in a single fixed representation by taking the OR of all the individual patterns • The vector representing the union is also going to match a large number of other patterns that were not one of the original 10 • How many such patterns can we store reliably, without a high chance of false positives? Is this SDR a member? 1) 2) 3) …. 10) 2% < 20%Union
  • 27. Error Bounds for Unions • Expected number of ON bits: • Give a union of M patterns, the expected probability of a false positive (with noise) is:
  • 28. What Does This Mean in Practice? • You can form reliable unions of a reasonable number of patterns (assuming large enough n and w) Examples • n = 2048, w = 40 The union of 50 patterns leads to an error rate of 10-9 • n = 512, w=10 The union of 50 patterns leads to an error rate of 0.9%
  • 29. • Extremely high capacity • Recognize patterns in the presence of noise • Robust to random deletions • Represent multiple patterns in a single fixed structure • Extremely efficient Fundamental Properties of SDRs
  • 30. SDRs Enable Highly Efficient Operations • In cortex complex operations are carried out rapidly – Visual system can perform object recognition in 100-150 msecs • SDR vectors are large, but all operations are O(w) and independent of vector size – No loops or optimization process required • Matching a pattern against a dynamic list (unions) is O(w) and independent of the number of items in the list • Enables a tiny dendritic segment to perform robust pattern recognition • We can simulate 200,000 neurons in software at about 25-50Hz
  • 31. Summary • SDR’s are the common data structure in the cortex • SDR’s enable flexible recognition systems that have very high capacity, and are robust to a large amount of noise • The union property allows a fixed representation to encode a dynamically changing set of patterns • The analysis of SDR’s provides a principled foundation for characterizing the behavior of the HTM learning algorithms and all cognitive functions • Sparse memory (Kanerva), Sparse coding (Olshausen), Bloom filters (Broder) Related work
  • 32. Questions? Math jokes? Follow us on Twitter @numenta Sign up for our newsletter at www.numenta.com Subutai Ahmad sahmad@numenta.com nupic-theory mailing list numenta.org/lists