SlideShare a Scribd company logo
Embeddings
Presented by
Featured Speaker
Jocelyn Matthews
Head of Community, Pinecone
Presented by
jocelyn@pinecone.io
3
What are embeddings?
Embeddings are numerical representations that capture the essential
features and relationships of discrete objects, like words or documents,
in a continuous vector space.
Embeddings:
● Are dynamic and context-sensitive.
● Capture the essence of the data they represent
● Are influenced by the context in which they are used
● Adaptability makes them powerful
Humans think in sensations, words, ideas.
Computers think in numbers
You don’t need to memorize this now
Vector: a list of numbers that tell us about something
Vector space: an environment in which vectors exist
Semantics: the study of meaning communicated through language
Vectors
A vector is a mathematical structure
with a size and a direction. For
example, we can think of the vector
as a point in space, with the
“direction” being an arrow from
(0,0,0) to that point in the vector
space.
Vectors
As developers, it might be easier to
think of a vector as an array
containing numerical values. For
example:
vector = [0,-2,...4]
Vectors
When we look at a bunch of vectors
in one space, we can say that some
are closer to one another, while
others are far apart. Some vectors
can seem to cluster together, while
others could be sparsely distributed
in the space.
An example you can bank on
🏦 Where is the Bank of England?
🌱 Where is the grassy bank?​
🛩️ How does a plane bank?
🐝 “the bees decided to have a mutiny against their queen”
🐝 “flying stinging insects rebelled in opposition to the
matriarch”
Polysemy and homonyms
Embeddings visualized
apidays Paris 2024 - Embeddings: Core Concepts for Developers, Jocelyn Matthews, Pinecone
Owning the concepts
Word arithmetic
king – man + woman = queen
Image, Peter Sutor, “Metaconcepts: Isolating Context in Word Embeddings”
Word arithmetic
king – man + woman = queen
“Distributed Representations of Words and Phrases and their Compositionality”
Word arithmetic
king – man + woman = queen
“adding the vectors associated with the words king
and woman while subtracting man is equal to the
vector associated with queen. This describes a gender
relationship.”
– MIT Technology Review, 2015
Word arithmetic
Paris - France + Poland = Warsaw
Word arithmetic
Paris - France + Poland = Warsaw
“In this case, the vector difference between Paris and
France captures the concept of capital city.”
– MIT Technology Review, 2015
Proximity
Together and apart
Coffee
Hospital
Music
Restaurant
School
Together and apart
Coffee
Hospital
Music
Restaurant
School
Cup
Caffeine
Morning
Galaxy
Dinosaur
Doctor
Patient
Surgery
Volcano
Unicorn
Song
Melody
Instrument
Asteroid
Bacteria
Food
Menu
Waiter
Nebula
Dragon
Teacher
Classroom
Student
Volcano
Spaceship
Exam
Dimensionality!
Coffee
Hospital
Music
Restaurant
School
Cup
Caffeine
Morning
Galaxy
Dinosaur
Doctor
Patient
Surgery
Volcano
Unicorn
Song
Melody
Instrument
Asteroid
Bacteria
Food
Menu
Waiter
Nebula
Dragon
Teacher
Classroom
Student
Volcano
Spaceship
Exam
apidays Paris 2024 - Embeddings: Core Concepts for Developers, Jocelyn Matthews, Pinecone
Green is to blue
green blue
As orange is to…
green
orange
blue
As orange is to…yep!
green
orange
blue
red
apidays Paris 2024 - Embeddings: Core Concepts for Developers, Jocelyn Matthews, Pinecone
What’s The Fallacy?
Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool
• Simplicity of relationships
• Linear vs nuanced
• Lack of Context
• How are the words used?
• Dimensionality
• 3D vs 100s of D
• Oversimplification
What’s The Fallacy?
Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool
• Simplicity of relationships
• Linear vs nuanced
• Lack of Context
• How are the words used?
• Dimensionality
• 3D vs 100s of D
• Oversimplification
What’s The Fallacy?
Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool
• Simplicity of relationships
• Linear vs nuanced
• Lack of Context
• How are the words used?
• Dimensionality
• 3D vs 100s of D
• Oversimplification
What’s The Fallacy?
Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool
• Simplicity of relationships
• Linear vs nuanced
• Lack of Context
• How are the words used?
• Dimensionality
• 3D vs 100s of D
• Oversimplification
Life is Like a Box of…
(Or, ”Check the Vectors”)
Check the vectors
The distance between red and orange
is incredibly similar to blue and green…
But when we tested things trying to verify,
we got interesting results which show the
"understanding" of the relationship
This actually yields this
# Find a term that has the same distance and direction blue has from green, but starting from
blue
target_distance = distance_green_blue
target_direction = direction_green_blue
# Define a list of terms to compare
terms = ["red", "orange", "yellow", "green", "blue", "purple", "pink", "black", "white", "gray"]
# Get the embedding for each term
term_embeddings = {term: get_embedding(term) for term in terms}
# Find the term with the closest distance and same direction to the target distance and direction
closest_term = None
closest_distance = float('inf')
start_term = "red"
start_embedding = get_embedding(start_term)
for term, embedding in term_embeddings.items():
if term == start_term:
continue
distance, direction = cosine_distance_and_direction(start_embedding, embedding)
if direction == target_direction and abs(distance - target_distance) < closest_distance:
closest_distance = abs(distance - target_distance)
closest_term = term
closest_term, closest_distance
Check the vectors
The distance between red and orange
is incredibly similar to blue and green…
But when we played around to verify, we
got interesting results revealing the
semantic "understanding" of the
relationship
This actually yields this
# Find a term that has the same distance and direction blue
has from green, but starting from blue
target_distance = distance_green_blue
target_direction = direction_green_blue
# Define a list of terms to compare
terms = ["red", "orange", "yellow", "green", "blue",
"purple", "pink", "black", "white", "gray"]
# Get the embedding for each term
term_embeddings = {term: get_embedding(term) for term in
terms}
# Find the term with the closest distance and same direction
to the target distance and direction
closest_term = None
closest_distance = float('inf')
start_term = "red"
start_embedding = get_embedding(start_term)
for term, embedding in term_embeddings.items():
if term == start_term:
continue
distance, direction =
cosine_distance_and_direction(start_embedding, embedding)
if direction == target_direction and abs(distance -
target_distance) < closest_distance:
closest_distance = abs(distance - target_distance)
closest_term = term
closest_term, closest_distance
('purple', np.float64(0.006596347059928065))
Purple
Why not 'orange'?
The code's result of ('purple',
np.float64(0.006596347059928065))
suggests that, in the embedding space used by
the model, "red" and "purple" have a closer
semantic relationship than "red" and "orange".
The embedding model used in the code has
determined that "red" and "purple" are closer
semantically. This is likely due to the specific
contexts and relationships captured by the model
during training.
It yields 'purple' instead of orange because the
cosine distance and direction calculations
between the embeddings of "red" and other terms
result in "purple" being the closest match to the
target distance and direction from "green" to
"blue".
# Find a term that has the same distance and direction blue
has from green, but starting from blue
target_distance = distance_green_blue
target_direction = direction_green_blue
# Define a list of terms to compare
terms = ["red", "orange", "yellow", "green", "blue",
"purple", "pink", "black", "white", "gray"]
# Get the embedding for each term
term_embeddings = {term: get_embedding(term) for term in
terms}
# Find the term with the closest distance and same direction
to the target distance and direction
closest_term = None
closest_distance = float('inf')
start_term = "red"
start_embedding = get_embedding(start_term)
for term, embedding in term_embeddings.items():
if term == start_term:
continue
distance, direction =
cosine_distance_and_direction(start_embedding, embedding)
if direction == target_direction and abs(distance -
target_distance) < closest_distance:
closest_distance = abs(distance - target_distance)
closest_term = term
closest_term, closest_distance
Embeddings
TL;DR
What are embeddings?
Embeddings are numerical representations that capture the essential
features and relationships of discrete objects, like words or documents,
in a continuous vector space.
The most important thing to understand
Embeddings are numerical representations of data that:
capture semantic meaning
and
allow for efficient comparison of similarity.
Key points about embeddings
1. They can represent various data types, not just text.
2. Dimensionality
3. Context sensitivity affects interpretation and application.
Applications of embeddings include:
- Semantic search
- Question-answering applications
- Image search
- Audio search
- Recommender systems
- Anomaly detection
“Generate your own embeddings”
(Inference API)
Sample app
Legal Semantic Search
Sample app
Shop the Look
© 2024 Pinecone – All rights reserved 45
1. Questions?
#hallwaytrack
2. Recording?
YouTube!
3. Slides?
Ask me
Thank you!
jocelyn@pinecone.io
apidays Paris 2024 - Embeddings: Core Concepts for Developers, Jocelyn Matthews, Pinecone
© 2024 Pinecone – All rights reserved 48

More Related Content

PDF
Embeddings the geometry of relational algebra
PPTX
A Simple Introduction to Word Embeddings
PDF
Word Embeddings - Introduction
PDF
Power of Visualizing Embeddings
PDF
Interactive Analysis of Word Vector Embeddings
PPTX
Interpreting Embeddings with Comparison
PDF
word2vec, node2vec, graph2vec, X2vec: Towards a Theory of Vector Embeddings o...
PDF
Words in Space - Rebecca Bilbro
Embeddings the geometry of relational algebra
A Simple Introduction to Word Embeddings
Word Embeddings - Introduction
Power of Visualizing Embeddings
Interactive Analysis of Word Vector Embeddings
Interpreting Embeddings with Comparison
word2vec, node2vec, graph2vec, X2vec: Towards a Theory of Vector Embeddings o...
Words in Space - Rebecca Bilbro

Similar to apidays Paris 2024 - Embeddings: Core Concepts for Developers, Jocelyn Matthews, Pinecone (20)

PDF
A Visual Exploration of Distance, Documents, and Distributions
PDF
Words in space
PPT
[PPT]
PPTX
Word vectors
PPT
lecture_mooney.ppt
PPTX
The Neural Search Frontier - Doug Turnbull, OpenSource Connections
PPT
Using binary classifiers
PPTX
Natural Language Processing in R (rNLP)
PDF
Similarity and Contrast on Conceptual Spaces for Pertinent Description Genera...
DOC
learningIntro.doc
DOC
learningIntro.doc
PPTX
Neural Models for Information Retrieval
PPTX
Vector Space Word Representations - Rani Nelken PhD
PPTX
Haystack 2019 - Search with Vectors - Simon Hughes
PPTX
Searching with vectors
PPTX
Interpreting Embeddings with Comparison
PPT
Support Vector Machines Support Vector Machines
PDF
Bijaya Zenchenko - An Embedding is Worth 1000 Words - Start Using Word Embedd...
PPTX
L6.pptxsdv dfbdfjftj hgjythgfvfhjyggunghb fghtffn
PPTX
Vectors in Search - Towards More Semantic Matching
A Visual Exploration of Distance, Documents, and Distributions
Words in space
[PPT]
Word vectors
lecture_mooney.ppt
The Neural Search Frontier - Doug Turnbull, OpenSource Connections
Using binary classifiers
Natural Language Processing in R (rNLP)
Similarity and Contrast on Conceptual Spaces for Pertinent Description Genera...
learningIntro.doc
learningIntro.doc
Neural Models for Information Retrieval
Vector Space Word Representations - Rani Nelken PhD
Haystack 2019 - Search with Vectors - Simon Hughes
Searching with vectors
Interpreting Embeddings with Comparison
Support Vector Machines Support Vector Machines
Bijaya Zenchenko - An Embedding is Worth 1000 Words - Start Using Word Embedd...
L6.pptxsdv dfbdfjftj hgjythgfvfhjyggunghb fghtffn
Vectors in Search - Towards More Semantic Matching
Ad

More from apidays (20)

PDF
apidays Munich 2025 - The Physics of Requirement Sciences Through Application...
PDF
apidays Munich 2025 - Developer Portals, API Catalogs, and Marketplaces, Miri...
PDF
apidays Munich 2025 - Making Sense of AI-Ready APIs in a Buzzword World, Andr...
PDF
apidays Munich 2025 - Integrate Your APIs into the New AI Marketplace, Senthi...
PDF
apidays Munich 2025 - The Double Life of the API Product Manager, Emmanuel Pa...
PDF
apidays Munich 2025 - Let’s build, debug and test a magic MCP server in Postm...
PDF
apidays Munich 2025 - The life-changing magic of great API docs, Jens Fischer...
PDF
apidays Munich 2025 - Automating Operations Without Reinventing the Wheel, Ma...
PDF
apidays Munich 2025 - Geospatial Artificial Intelligence (GeoAI) with OGC API...
PPTX
apidays Munich 2025 - GraphQL 101: I won't REST, until you GraphQL, Surbhi Si...
PPTX
apidays Munich 2025 - Effectively incorporating API Security into the overall...
PPTX
apidays Munich 2025 - Federated API Management and Governance, Vince Baker (D...
PPTX
apidays Munich 2025 - Agentic AI: A Friend or Foe?, Merja Kajava (Aavista Oy)
PPTX
apidays Munich 2025 - Streamline & Secure LLM Traffic with APISIX AI Gateway ...
PPTX
apidays Munich 2025 - Building Telco-Aware Apps with Open Gateway APIs, Subhr...
PPTX
apidays Munich 2025 - Building an AWS Serverless Application with Terraform, ...
PDF
apidays Helsinki & North 2025 - REST in Peace? Hunting the Dominant Design fo...
PDF
apidays Helsinki & North 2025 - Monetizing AI APIs: The New API Economy, Alla...
PDF
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
PDF
apidays Helsinki & North 2025 - APIs in the healthcare sector: hospitals inte...
apidays Munich 2025 - The Physics of Requirement Sciences Through Application...
apidays Munich 2025 - Developer Portals, API Catalogs, and Marketplaces, Miri...
apidays Munich 2025 - Making Sense of AI-Ready APIs in a Buzzword World, Andr...
apidays Munich 2025 - Integrate Your APIs into the New AI Marketplace, Senthi...
apidays Munich 2025 - The Double Life of the API Product Manager, Emmanuel Pa...
apidays Munich 2025 - Let’s build, debug and test a magic MCP server in Postm...
apidays Munich 2025 - The life-changing magic of great API docs, Jens Fischer...
apidays Munich 2025 - Automating Operations Without Reinventing the Wheel, Ma...
apidays Munich 2025 - Geospatial Artificial Intelligence (GeoAI) with OGC API...
apidays Munich 2025 - GraphQL 101: I won't REST, until you GraphQL, Surbhi Si...
apidays Munich 2025 - Effectively incorporating API Security into the overall...
apidays Munich 2025 - Federated API Management and Governance, Vince Baker (D...
apidays Munich 2025 - Agentic AI: A Friend or Foe?, Merja Kajava (Aavista Oy)
apidays Munich 2025 - Streamline & Secure LLM Traffic with APISIX AI Gateway ...
apidays Munich 2025 - Building Telco-Aware Apps with Open Gateway APIs, Subhr...
apidays Munich 2025 - Building an AWS Serverless Application with Terraform, ...
apidays Helsinki & North 2025 - REST in Peace? Hunting the Dominant Design fo...
apidays Helsinki & North 2025 - Monetizing AI APIs: The New API Economy, Alla...
apidays Helsinki & North 2025 - How (not) to run a Graphql Stewardship Group,...
apidays Helsinki & North 2025 - APIs in the healthcare sector: hospitals inte...
Ad

Recently uploaded (20)

PDF
[EN] Industrial Machine Downtime Prediction
PDF
Introduction to Data Science and Data Analysis
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PDF
.pdf is not working space design for the following data for the following dat...
PPT
Miokarditis (Inflamasi pada Otot Jantung)
PPTX
01_intro xxxxxxxxxxfffffffffffaaaaaaaaaaafg
PPTX
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
PPT
Quality review (1)_presentation of this 21
PDF
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
PDF
annual-report-2024-2025 original latest.
PPTX
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
PPTX
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
PDF
Introduction to the R Programming Language
PPTX
Qualitative Qantitative and Mixed Methods.pptx
PDF
Fluorescence-microscope_Botany_detailed content
PPTX
SAP 2 completion done . PRESENTATION.pptx
PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PPTX
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
[EN] Industrial Machine Downtime Prediction
Introduction to Data Science and Data Analysis
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
.pdf is not working space design for the following data for the following dat...
Miokarditis (Inflamasi pada Otot Jantung)
01_intro xxxxxxxxxxfffffffffffaaaaaaaaaaafg
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
Quality review (1)_presentation of this 21
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
annual-report-2024-2025 original latest.
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
Introduction to the R Programming Language
Qualitative Qantitative and Mixed Methods.pptx
Fluorescence-microscope_Botany_detailed content
SAP 2 completion done . PRESENTATION.pptx
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
Business Ppt On Nestle.pptx huunnnhhgfvu

apidays Paris 2024 - Embeddings: Core Concepts for Developers, Jocelyn Matthews, Pinecone

  • 2. Featured Speaker Jocelyn Matthews Head of Community, Pinecone Presented by jocelyn@pinecone.io 3
  • 3. What are embeddings? Embeddings are numerical representations that capture the essential features and relationships of discrete objects, like words or documents, in a continuous vector space.
  • 4. Embeddings: ● Are dynamic and context-sensitive. ● Capture the essence of the data they represent ● Are influenced by the context in which they are used ● Adaptability makes them powerful Humans think in sensations, words, ideas. Computers think in numbers
  • 5. You don’t need to memorize this now Vector: a list of numbers that tell us about something Vector space: an environment in which vectors exist Semantics: the study of meaning communicated through language
  • 6. Vectors A vector is a mathematical structure with a size and a direction. For example, we can think of the vector as a point in space, with the “direction” being an arrow from (0,0,0) to that point in the vector space.
  • 7. Vectors As developers, it might be easier to think of a vector as an array containing numerical values. For example: vector = [0,-2,...4]
  • 8. Vectors When we look at a bunch of vectors in one space, we can say that some are closer to one another, while others are far apart. Some vectors can seem to cluster together, while others could be sparsely distributed in the space.
  • 9. An example you can bank on 🏦 Where is the Bank of England? 🌱 Where is the grassy bank?​ 🛩️ How does a plane bank? 🐝 “the bees decided to have a mutiny against their queen” 🐝 “flying stinging insects rebelled in opposition to the matriarch”
  • 14. Word arithmetic king – man + woman = queen Image, Peter Sutor, “Metaconcepts: Isolating Context in Word Embeddings”
  • 15. Word arithmetic king – man + woman = queen “Distributed Representations of Words and Phrases and their Compositionality”
  • 16. Word arithmetic king – man + woman = queen “adding the vectors associated with the words king and woman while subtracting man is equal to the vector associated with queen. This describes a gender relationship.” – MIT Technology Review, 2015
  • 17. Word arithmetic Paris - France + Poland = Warsaw
  • 18. Word arithmetic Paris - France + Poland = Warsaw “In this case, the vector difference between Paris and France captures the concept of capital city.” – MIT Technology Review, 2015
  • 24. Green is to blue green blue
  • 25. As orange is to… green orange blue
  • 26. As orange is to…yep! green orange blue red
  • 28. What’s The Fallacy? Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool • Simplicity of relationships • Linear vs nuanced • Lack of Context • How are the words used? • Dimensionality • 3D vs 100s of D • Oversimplification
  • 29. What’s The Fallacy? Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool • Simplicity of relationships • Linear vs nuanced • Lack of Context • How are the words used? • Dimensionality • 3D vs 100s of D • Oversimplification
  • 30. What’s The Fallacy? Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool • Simplicity of relationships • Linear vs nuanced • Lack of Context • How are the words used? • Dimensionality • 3D vs 100s of D • Oversimplification
  • 31. What’s The Fallacy? Why "Green : Blue :: Orange : Red" is Imperfect as a Teaching Tool • Simplicity of relationships • Linear vs nuanced • Lack of Context • How are the words used? • Dimensionality • 3D vs 100s of D • Oversimplification
  • 32. Life is Like a Box of… (Or, ”Check the Vectors”)
  • 33. Check the vectors The distance between red and orange is incredibly similar to blue and green… But when we tested things trying to verify, we got interesting results which show the "understanding" of the relationship This actually yields this # Find a term that has the same distance and direction blue has from green, but starting from blue target_distance = distance_green_blue target_direction = direction_green_blue # Define a list of terms to compare terms = ["red", "orange", "yellow", "green", "blue", "purple", "pink", "black", "white", "gray"] # Get the embedding for each term term_embeddings = {term: get_embedding(term) for term in terms} # Find the term with the closest distance and same direction to the target distance and direction closest_term = None closest_distance = float('inf') start_term = "red" start_embedding = get_embedding(start_term) for term, embedding in term_embeddings.items(): if term == start_term: continue distance, direction = cosine_distance_and_direction(start_embedding, embedding) if direction == target_direction and abs(distance - target_distance) < closest_distance: closest_distance = abs(distance - target_distance) closest_term = term closest_term, closest_distance
  • 34. Check the vectors The distance between red and orange is incredibly similar to blue and green… But when we played around to verify, we got interesting results revealing the semantic "understanding" of the relationship This actually yields this # Find a term that has the same distance and direction blue has from green, but starting from blue target_distance = distance_green_blue target_direction = direction_green_blue # Define a list of terms to compare terms = ["red", "orange", "yellow", "green", "blue", "purple", "pink", "black", "white", "gray"] # Get the embedding for each term term_embeddings = {term: get_embedding(term) for term in terms} # Find the term with the closest distance and same direction to the target distance and direction closest_term = None closest_distance = float('inf') start_term = "red" start_embedding = get_embedding(start_term) for term, embedding in term_embeddings.items(): if term == start_term: continue distance, direction = cosine_distance_and_direction(start_embedding, embedding) if direction == target_direction and abs(distance - target_distance) < closest_distance: closest_distance = abs(distance - target_distance) closest_term = term closest_term, closest_distance ('purple', np.float64(0.006596347059928065)) Purple
  • 35. Why not 'orange'? The code's result of ('purple', np.float64(0.006596347059928065)) suggests that, in the embedding space used by the model, "red" and "purple" have a closer semantic relationship than "red" and "orange". The embedding model used in the code has determined that "red" and "purple" are closer semantically. This is likely due to the specific contexts and relationships captured by the model during training. It yields 'purple' instead of orange because the cosine distance and direction calculations between the embeddings of "red" and other terms result in "purple" being the closest match to the target distance and direction from "green" to "blue". # Find a term that has the same distance and direction blue has from green, but starting from blue target_distance = distance_green_blue target_direction = direction_green_blue # Define a list of terms to compare terms = ["red", "orange", "yellow", "green", "blue", "purple", "pink", "black", "white", "gray"] # Get the embedding for each term term_embeddings = {term: get_embedding(term) for term in terms} # Find the term with the closest distance and same direction to the target distance and direction closest_term = None closest_distance = float('inf') start_term = "red" start_embedding = get_embedding(start_term) for term, embedding in term_embeddings.items(): if term == start_term: continue distance, direction = cosine_distance_and_direction(start_embedding, embedding) if direction == target_direction and abs(distance - target_distance) < closest_distance: closest_distance = abs(distance - target_distance) closest_term = term closest_term, closest_distance
  • 37. What are embeddings? Embeddings are numerical representations that capture the essential features and relationships of discrete objects, like words or documents, in a continuous vector space.
  • 38. The most important thing to understand Embeddings are numerical representations of data that: capture semantic meaning and allow for efficient comparison of similarity.
  • 39. Key points about embeddings 1. They can represent various data types, not just text. 2. Dimensionality 3. Context sensitivity affects interpretation and application.
  • 40. Applications of embeddings include: - Semantic search - Question-answering applications - Image search - Audio search - Recommender systems - Anomaly detection “Generate your own embeddings” (Inference API)
  • 43. © 2024 Pinecone – All rights reserved 45 1. Questions? #hallwaytrack 2. Recording? YouTube! 3. Slides? Ask me
  • 46. © 2024 Pinecone – All rights reserved 48