SlideShare a Scribd company logo
Chapter 5: Information
Retrieval and Web Search
An introduction
CS583, Bing Liu, UIC 2
Introduction
 Text mining refers to data mining using text
documents as data.
 Most text mining tasks use Information
Retrieval (IR) methods to pre-process text
documents.
 These methods are quite different from
traditional data pre-processing methods
used for relational tables.
 Web search also has its root in IR.
CS583, Bing Liu, UIC 3
Information Retrieval (IR)
 Conceptually, IR is the study of finding needed
information. I.e., IR helps users find information
that matches their information needs.
 Expressed as queries
 Historically, IR is about document retrieval,
emphasizing document as the basic unit.
 Finding documents relevant to user queries
 Technically, IR studies the acquisition,
organization, storage, retrieval, and distribution of
information.
CS583, Bing Liu, UIC 4
IR architecture
CS583, Bing Liu, UIC 5
IR queries
 Keyword queries
 Boolean queries (using AND, OR, NOT)
 Phrase queries
 Proximity queries
 Full document queries
 Natural language questions
CS583, Bing Liu, UIC 6
Information retrieval models
 An IR model governs how a document and a
query are represented and how the relevance
of a document to a user query is defined.
 Main models:
 Boolean model
 Vector space model
 Statistical language model
 etc
CS583, Bing Liu, UIC 7
Boolean model
 Each document or query is treated as a
“bag” of words or terms. Word sequence is
not considered.
 Given a collection of documents D, let V = {t1,
t2, ..., t|V|} be the set of distinctive words/terms
in the collection. V is called the vocabulary.
 A weight wij > 0 is associated with each term ti
of a document dj ∈ D. For a term that does
not appear in document dj, wij = 0.
dj = (w1j, w2j, ..., w|V|j),
CS583, Bing Liu, UIC 8
Boolean model (contd)
 Query terms are combined logically using the
Boolean operators AND, OR, and NOT.
 E.g., ((data AND mining) AND (NOT text))
 Retrieval
 Given a Boolean query, the system retrieves
every document that makes the query logically
true.
 Called exact match.
 The retrieval results are usually quite poor
because term frequency is not considered.
CS583, Bing Liu, UIC 9
Vector space model
 Documents are also treated as a “bag” of words or
terms.
 Each document is represented as a vector.
 However, the term weights are no longer 0 or 1.
Each term weight is computed based on some
variations of TF or TF-IDF scheme.
 Term Frequency (TF) Scheme: The weight of a term
ti in document dj is the number of times that ti
appears in dj, denoted by fij. Normalization may also
be applied.
CS583, Bing Liu, UIC 10
TF-IDF term weighting scheme
 The most well known
weighting scheme
 TF: still term frequency
 IDF: inverse document
frequency.
N: total number of docs
dfi: the number of docs that ti
appears.
 The final TF-IDF term
weight is:
CS583, Bing Liu, UIC 11
Retrieval in vector space model
 Query q is represented in the same way or slightly
differently.
 Relevance of di to q: Compare the similarity of
query q and document di.
 Cosine similarity (the cosine of the angle between
the two vectors)
 Cosine is also commonly used in text clustering
CS583, Bing Liu, UIC 12
An Example
 A document space is defined by three terms:
 hardware, software, users
 the vocabulary
 A set of documents are defined as:
 A1=(1, 0, 0), A2=(0, 1, 0), A3=(0, 0, 1)
 A4=(1, 1, 0), A5=(1, 0, 1), A6=(0, 1, 1)
 A7=(1, 1, 1) A8=(1, 0, 1). A9=(0, 1, 1)
 If the Query is “hardware and software”
 what documents should be retrieved?
CS583, Bing Liu, UIC 13
An Example (cont.)
 In Boolean query matching:
 document A4, A7 will be retrieved (“AND”)
 retrieved: A1, A2, A4, A5, A6, A7, A8, A9 (“OR”)
 In similarity matching (cosine):
 q=(1, 1, 0)
 S(q, A1)=0.71, S(q, A2)=0.71, S(q, A3)=0
 S(q, A4)=1, S(q, A5)=0.5, S(q, A6)=0.5
 S(q, A7)=0.82, S(q, A8)=0.5, S(q, A9)=0.5
 Document retrieved set (with ranking)=
 {A4, A7, A1, A2, A5, A6, A8, A9}
CS583, Bing Liu, UIC 14
Okapi relevance method
 Another way to assess the degree of relevance is to
directly compute a relevance score for each
document to the query.
 The Okapi method and its variations are popular
techniques in this setting.
CS583, Bing Liu, UIC 15
Relevance feedback
 Relevance feedback is one of the techniques for
improving retrieval effectiveness. The steps:
 the user first identifies some relevant (Dr) and irrelevant
documents (Dir) in the initial list of retrieved documents
 the system expands the query q by extracting some
additional terms from the sample relevant and irrelevant
documents to produce qe
 Perform a second round of retrieval.
 Rocchio method (α, β and γ are parameters)
CS583, Bing Liu, UIC 16
Rocchio text classifier
 In fact, a variation of the Rocchio method above,
called the Rocchio classification method, can be
used to improve retrieval effectiveness too
 so are other machine learning methods. Why?
 Rocchio classifier is constructed by producing a
prototype vector ci for each class i (relevant or
irrelevant in this case):
 In classification, cosine is used.
CS583, Bing Liu, UIC 17
Text pre-processing
 Word (term) extraction: easy
 Stopwords removal
 Stemming
 Frequency counts and computing TF-IDF
term weights.
CS583, Bing Liu, UIC 18
Stopwords removal
 Many of the most frequently used words in English are useless
in IR and text mining – these words are called stop words.
 the, of, and, to, ….
 Typically about 400 to 500 such words
 For an application, an additional domain specific stopwords list
may be constructed
 Why do we need to remove stopwords?
 Reduce indexing (or data) file size
 stopwords accounts 20-30% of total word counts.
 Improve efficiency and effectiveness
 stopwords are not useful for searching or text mining
 they may also confuse the retrieval system.
CS583, Bing Liu, UIC 19
Stemming
 Techniques used to find out the root/stem of a
word. E.g.,
 user engineering
 users engineered
 used engineer
 using
 stem: use engineer
Usefulness:
 improving effectiveness of IR and text mining
 matching similar words
 Mainly improve recall
 reducing indexing size
 combing words with same roots may reduce indexing
size as much as 40-50%.
CS583, Bing Liu, UIC 20
Basic stemming methods
Using a set of rules. E.g.,
 remove ending
 if a word ends with a consonant other than s,
followed by an s, then delete s.
 if a word ends in es, drop the s.
 if a word ends in ing, delete the ing unless the remaining word
consists only of one letter or of th.
 If a word ends with ed, preceded by a consonant, delete the ed
unless this leaves only a single letter.
 …...
 transform words
 if a word ends with “ies” but not “eies” or “aies” then “ies --> y.”
CS583, Bing Liu, UIC 21
Frequency counts + TF-IDF
 Counts the number of times a word occurred
in a document.
 Using occurrence frequencies to indicate relative
importance of a word in a document.
 if a word appears often in a document, the document
likely “deals with” subjects related to the word.
 Counts the number of documents in the
collection that contains each word
 TF-IDF can be computed.
CS583, Bing Liu, UIC 22
Evaluation: Precision and Recall
 Given a query:
 Are all retrieved documents relevant?
 Have all the relevant documents been retrieved?
 Measures for system performance:
 The first question is about the precision of the
search
 The second is about the completeness (recall) of
the search.
CS583, Bing Liu, UIC 23
Precision-recall curve
CS583, Bing Liu, UIC 24
Compare different retrieval algorithms
CS583, Bing Liu, UIC 25
Compare with multiple queries
 Compute the average precision at each recall
level.
 Draw precision recall curves
 Do not forget the F-score evaluation measure.
CS583, Bing Liu, UIC 26
Rank precision
 Compute the precision values at some
selected rank positions.
 Mainly used in Web search evaluation.
 For a Web search engine, we can compute
precisions for the top 5, 10, 15, 20, 25 and 30
returned pages
 as the user seldom looks at more than 30 pages.
 Recall is not very meaningful in Web search.
 Why?
CS583, Bing Liu, UIC 27
Web Search as a huge IR system
 A Web crawler (robot) crawls the Web to
collect all the pages.
 Servers establish a huge inverted indexing
database and other indexing databases
 At query (search) time, search engines
conduct different types of vector query
matching.
CS583, Bing Liu, UIC 28
Inverted index
 The inverted index of a document collection
is basically a data structure that
 attaches each distinctive term with a list of all
documents that contains the term.
 Thus, in retrieval, it takes constant time to
 find the documents that contains a query term.
 multiple query terms are also easy handle as we
will see soon.
CS583, Bing Liu, UIC 29
An example
CS583, Bing Liu, UIC 30
Index construction
 Easy! See the example,
CS583, Bing Liu, UIC 31
Search using inverted index
Given a query q, search has the following steps:
 Step 1 (vocabulary search): find each
term/word in q in the inverted index.
 Step 2 (results merging): Merge results to
find documents that contain all or some of the
words/terms in q.
 Step 3 (Rank score computation): To rank
the resulting documents/pages, using,
 content-based ranking
 link-based ranking
CS583, Bing Liu, UIC 32
Different search engines
 The real differences among different search
engines are
 their index weighting schemes
 Including location of terms, e.g., title, body,
emphasized words, etc.
 their query processing methods (e.g., query
classification, expansion, etc)
 their ranking algorithms
 Few of these are published by any of the search
engine companies. They aretightly guarded
secrets.
CS583, Bing Liu, UIC 33
Summary
 We only give a VERY brief introduction to IR. There
are a large number of other topics, e.g.,
 Statistical language model
 Latent semantic indexing (LSI and SVD).
 (read an IR book or take an IR course)
 Many other interesting topics are not covered, e.g.,
 Web search
 Index compression
 Ranking: combining contents and hyperlinks
 Web page pre-processing
 Combining multiple rankings and meta search
 Web spamming
 Want to know more? Read the textbook

More Related Content

PPTX
Aws ppt
PPTX
Data mining
PPT
Data Visualization
PPTX
AzureOpenAI.pptx
PDF
[Machine Learning 15minutes! #61] Azure OpenAI Service
PPT
8. column oriented databases
PDF
Everything You Need To Know About ChatGPT
PPTX
Cloud computing presentation
Aws ppt
Data mining
Data Visualization
AzureOpenAI.pptx
[Machine Learning 15minutes! #61] Azure OpenAI Service
8. column oriented databases
Everything You Need To Know About ChatGPT
Cloud computing presentation

What's hot (20)

PPT
Cloud computing
PPTX
OLAP & DATA WAREHOUSE
PDF
Best Practice on using Azure OpenAI Service
PPTX
Introduction to cloud computing
PDF
Conversational AI is Now the Heart of Customer Experience.pdf
PPT
What is OLAP -Data Warehouse Concepts - IT Online Training @ Newyorksys
PPTX
PDF
Trend and Future of Cloud Computing
PPTX
OpenStack Introduction
PPTX
web mining
PPT
Graph Analytics for big data
PPTX
Data science ppt
PPT
Cloud computing ppt
PDF
OpenAI API crash course
PDF
Prompt Engineering
PDF
Introduction to Azure Synapse Webinar
PPTX
What is ChatGPT
PPTX
introduction Azure OpenAI by Usama wahab khan
DOCX
Seminar Report On Amazon Web Service
Cloud computing
OLAP & DATA WAREHOUSE
Best Practice on using Azure OpenAI Service
Introduction to cloud computing
Conversational AI is Now the Heart of Customer Experience.pdf
What is OLAP -Data Warehouse Concepts - IT Online Training @ Newyorksys
Trend and Future of Cloud Computing
OpenStack Introduction
web mining
Graph Analytics for big data
Data science ppt
Cloud computing ppt
OpenAI API crash course
Prompt Engineering
Introduction to Azure Synapse Webinar
What is ChatGPT
introduction Azure OpenAI by Usama wahab khan
Seminar Report On Amazon Web Service
Ad

Similar to Cs583 info-retrieval (20)

PPT
Information Retrieval and Storage Systems
PPTX
IRT Unit_ 2.pptx
PPTX
unit -4MODELING AND RETRIEVAL EVALUATION
PDF
Information Retrieval and Map-Reduce Implementations
PPT
Web search engines
PPTX
Text mining
PDF
Information Retrieval
PDF
Information_Retrieval_Models_Nfaoui_El_Habib
PPT
Information Retrieval
PDF
An Introduction to Information Retrieval.pdf
PPT
3392413.ppt information retreival systems
PPT
chapter 5 Information Retrieval Models.ppt
PDF
Tutorial 1 (information retrieval basics)
PPTX
Information retrival system and PageRank algorithm
PPT
Information retrieval is the process of accessing data resources. Usually doc...
PPT
Ir models
PPTX
Information Retrieval and Extraction - Module 7
PDF
Information retrieval systems irt ppt do
PPTX
Information retrieval introduction
Information Retrieval and Storage Systems
IRT Unit_ 2.pptx
unit -4MODELING AND RETRIEVAL EVALUATION
Information Retrieval and Map-Reduce Implementations
Web search engines
Text mining
Information Retrieval
Information_Retrieval_Models_Nfaoui_El_Habib
Information Retrieval
An Introduction to Information Retrieval.pdf
3392413.ppt information retreival systems
chapter 5 Information Retrieval Models.ppt
Tutorial 1 (information retrieval basics)
Information retrival system and PageRank algorithm
Information retrieval is the process of accessing data resources. Usually doc...
Ir models
Information Retrieval and Extraction - Module 7
Information retrieval systems irt ppt do
Information retrieval introduction
Ad

More from Borseshweta (7)

PPT
Cs621 lect7-si-13aug07
PPT
Cs583 link-analysis
PPT
Cs583 intro
PPT
Cs583 information-integration
PPT
Cs583 association-sequential-patterns
PPT
Cs511 data-extraction
PPT
Web crawlingchapter
Cs621 lect7-si-13aug07
Cs583 link-analysis
Cs583 intro
Cs583 information-integration
Cs583 association-sequential-patterns
Cs511 data-extraction
Web crawlingchapter

Recently uploaded (20)

PPTX
Construction Project Organization Group 2.pptx
PPT
Project quality management in manufacturing
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PPTX
Sustainable Sites - Green Building Construction
PDF
Well-logging-methods_new................
PPTX
Internet of Things (IOT) - A guide to understanding
PPTX
UNIT 4 Total Quality Management .pptx
PPTX
Lecture Notes Electrical Wiring System Components
PPTX
web development for engineering and engineering
PPTX
additive manufacturing of ss316l using mig welding
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
PPTX
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
PPTX
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
PPTX
CYBER-CRIMES AND SECURITY A guide to understanding
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PPTX
Geodesy 1.pptx...............................................
Construction Project Organization Group 2.pptx
Project quality management in manufacturing
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
bas. eng. economics group 4 presentation 1.pptx
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Sustainable Sites - Green Building Construction
Well-logging-methods_new................
Internet of Things (IOT) - A guide to understanding
UNIT 4 Total Quality Management .pptx
Lecture Notes Electrical Wiring System Components
web development for engineering and engineering
additive manufacturing of ss316l using mig welding
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
Model Code of Practice - Construction Work - 21102022 .pdf
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
CYBER-CRIMES AND SECURITY A guide to understanding
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
Geodesy 1.pptx...............................................

Cs583 info-retrieval

  • 1. Chapter 5: Information Retrieval and Web Search An introduction
  • 2. CS583, Bing Liu, UIC 2 Introduction  Text mining refers to data mining using text documents as data.  Most text mining tasks use Information Retrieval (IR) methods to pre-process text documents.  These methods are quite different from traditional data pre-processing methods used for relational tables.  Web search also has its root in IR.
  • 3. CS583, Bing Liu, UIC 3 Information Retrieval (IR)  Conceptually, IR is the study of finding needed information. I.e., IR helps users find information that matches their information needs.  Expressed as queries  Historically, IR is about document retrieval, emphasizing document as the basic unit.  Finding documents relevant to user queries  Technically, IR studies the acquisition, organization, storage, retrieval, and distribution of information.
  • 4. CS583, Bing Liu, UIC 4 IR architecture
  • 5. CS583, Bing Liu, UIC 5 IR queries  Keyword queries  Boolean queries (using AND, OR, NOT)  Phrase queries  Proximity queries  Full document queries  Natural language questions
  • 6. CS583, Bing Liu, UIC 6 Information retrieval models  An IR model governs how a document and a query are represented and how the relevance of a document to a user query is defined.  Main models:  Boolean model  Vector space model  Statistical language model  etc
  • 7. CS583, Bing Liu, UIC 7 Boolean model  Each document or query is treated as a “bag” of words or terms. Word sequence is not considered.  Given a collection of documents D, let V = {t1, t2, ..., t|V|} be the set of distinctive words/terms in the collection. V is called the vocabulary.  A weight wij > 0 is associated with each term ti of a document dj ∈ D. For a term that does not appear in document dj, wij = 0. dj = (w1j, w2j, ..., w|V|j),
  • 8. CS583, Bing Liu, UIC 8 Boolean model (contd)  Query terms are combined logically using the Boolean operators AND, OR, and NOT.  E.g., ((data AND mining) AND (NOT text))  Retrieval  Given a Boolean query, the system retrieves every document that makes the query logically true.  Called exact match.  The retrieval results are usually quite poor because term frequency is not considered.
  • 9. CS583, Bing Liu, UIC 9 Vector space model  Documents are also treated as a “bag” of words or terms.  Each document is represented as a vector.  However, the term weights are no longer 0 or 1. Each term weight is computed based on some variations of TF or TF-IDF scheme.  Term Frequency (TF) Scheme: The weight of a term ti in document dj is the number of times that ti appears in dj, denoted by fij. Normalization may also be applied.
  • 10. CS583, Bing Liu, UIC 10 TF-IDF term weighting scheme  The most well known weighting scheme  TF: still term frequency  IDF: inverse document frequency. N: total number of docs dfi: the number of docs that ti appears.  The final TF-IDF term weight is:
  • 11. CS583, Bing Liu, UIC 11 Retrieval in vector space model  Query q is represented in the same way or slightly differently.  Relevance of di to q: Compare the similarity of query q and document di.  Cosine similarity (the cosine of the angle between the two vectors)  Cosine is also commonly used in text clustering
  • 12. CS583, Bing Liu, UIC 12 An Example  A document space is defined by three terms:  hardware, software, users  the vocabulary  A set of documents are defined as:  A1=(1, 0, 0), A2=(0, 1, 0), A3=(0, 0, 1)  A4=(1, 1, 0), A5=(1, 0, 1), A6=(0, 1, 1)  A7=(1, 1, 1) A8=(1, 0, 1). A9=(0, 1, 1)  If the Query is “hardware and software”  what documents should be retrieved?
  • 13. CS583, Bing Liu, UIC 13 An Example (cont.)  In Boolean query matching:  document A4, A7 will be retrieved (“AND”)  retrieved: A1, A2, A4, A5, A6, A7, A8, A9 (“OR”)  In similarity matching (cosine):  q=(1, 1, 0)  S(q, A1)=0.71, S(q, A2)=0.71, S(q, A3)=0  S(q, A4)=1, S(q, A5)=0.5, S(q, A6)=0.5  S(q, A7)=0.82, S(q, A8)=0.5, S(q, A9)=0.5  Document retrieved set (with ranking)=  {A4, A7, A1, A2, A5, A6, A8, A9}
  • 14. CS583, Bing Liu, UIC 14 Okapi relevance method  Another way to assess the degree of relevance is to directly compute a relevance score for each document to the query.  The Okapi method and its variations are popular techniques in this setting.
  • 15. CS583, Bing Liu, UIC 15 Relevance feedback  Relevance feedback is one of the techniques for improving retrieval effectiveness. The steps:  the user first identifies some relevant (Dr) and irrelevant documents (Dir) in the initial list of retrieved documents  the system expands the query q by extracting some additional terms from the sample relevant and irrelevant documents to produce qe  Perform a second round of retrieval.  Rocchio method (α, β and γ are parameters)
  • 16. CS583, Bing Liu, UIC 16 Rocchio text classifier  In fact, a variation of the Rocchio method above, called the Rocchio classification method, can be used to improve retrieval effectiveness too  so are other machine learning methods. Why?  Rocchio classifier is constructed by producing a prototype vector ci for each class i (relevant or irrelevant in this case):  In classification, cosine is used.
  • 17. CS583, Bing Liu, UIC 17 Text pre-processing  Word (term) extraction: easy  Stopwords removal  Stemming  Frequency counts and computing TF-IDF term weights.
  • 18. CS583, Bing Liu, UIC 18 Stopwords removal  Many of the most frequently used words in English are useless in IR and text mining – these words are called stop words.  the, of, and, to, ….  Typically about 400 to 500 such words  For an application, an additional domain specific stopwords list may be constructed  Why do we need to remove stopwords?  Reduce indexing (or data) file size  stopwords accounts 20-30% of total word counts.  Improve efficiency and effectiveness  stopwords are not useful for searching or text mining  they may also confuse the retrieval system.
  • 19. CS583, Bing Liu, UIC 19 Stemming  Techniques used to find out the root/stem of a word. E.g.,  user engineering  users engineered  used engineer  using  stem: use engineer Usefulness:  improving effectiveness of IR and text mining  matching similar words  Mainly improve recall  reducing indexing size  combing words with same roots may reduce indexing size as much as 40-50%.
  • 20. CS583, Bing Liu, UIC 20 Basic stemming methods Using a set of rules. E.g.,  remove ending  if a word ends with a consonant other than s, followed by an s, then delete s.  if a word ends in es, drop the s.  if a word ends in ing, delete the ing unless the remaining word consists only of one letter or of th.  If a word ends with ed, preceded by a consonant, delete the ed unless this leaves only a single letter.  …...  transform words  if a word ends with “ies” but not “eies” or “aies” then “ies --> y.”
  • 21. CS583, Bing Liu, UIC 21 Frequency counts + TF-IDF  Counts the number of times a word occurred in a document.  Using occurrence frequencies to indicate relative importance of a word in a document.  if a word appears often in a document, the document likely “deals with” subjects related to the word.  Counts the number of documents in the collection that contains each word  TF-IDF can be computed.
  • 22. CS583, Bing Liu, UIC 22 Evaluation: Precision and Recall  Given a query:  Are all retrieved documents relevant?  Have all the relevant documents been retrieved?  Measures for system performance:  The first question is about the precision of the search  The second is about the completeness (recall) of the search.
  • 23. CS583, Bing Liu, UIC 23 Precision-recall curve
  • 24. CS583, Bing Liu, UIC 24 Compare different retrieval algorithms
  • 25. CS583, Bing Liu, UIC 25 Compare with multiple queries  Compute the average precision at each recall level.  Draw precision recall curves  Do not forget the F-score evaluation measure.
  • 26. CS583, Bing Liu, UIC 26 Rank precision  Compute the precision values at some selected rank positions.  Mainly used in Web search evaluation.  For a Web search engine, we can compute precisions for the top 5, 10, 15, 20, 25 and 30 returned pages  as the user seldom looks at more than 30 pages.  Recall is not very meaningful in Web search.  Why?
  • 27. CS583, Bing Liu, UIC 27 Web Search as a huge IR system  A Web crawler (robot) crawls the Web to collect all the pages.  Servers establish a huge inverted indexing database and other indexing databases  At query (search) time, search engines conduct different types of vector query matching.
  • 28. CS583, Bing Liu, UIC 28 Inverted index  The inverted index of a document collection is basically a data structure that  attaches each distinctive term with a list of all documents that contains the term.  Thus, in retrieval, it takes constant time to  find the documents that contains a query term.  multiple query terms are also easy handle as we will see soon.
  • 29. CS583, Bing Liu, UIC 29 An example
  • 30. CS583, Bing Liu, UIC 30 Index construction  Easy! See the example,
  • 31. CS583, Bing Liu, UIC 31 Search using inverted index Given a query q, search has the following steps:  Step 1 (vocabulary search): find each term/word in q in the inverted index.  Step 2 (results merging): Merge results to find documents that contain all or some of the words/terms in q.  Step 3 (Rank score computation): To rank the resulting documents/pages, using,  content-based ranking  link-based ranking
  • 32. CS583, Bing Liu, UIC 32 Different search engines  The real differences among different search engines are  their index weighting schemes  Including location of terms, e.g., title, body, emphasized words, etc.  their query processing methods (e.g., query classification, expansion, etc)  their ranking algorithms  Few of these are published by any of the search engine companies. They aretightly guarded secrets.
  • 33. CS583, Bing Liu, UIC 33 Summary  We only give a VERY brief introduction to IR. There are a large number of other topics, e.g.,  Statistical language model  Latent semantic indexing (LSI and SVD).  (read an IR book or take an IR course)  Many other interesting topics are not covered, e.g.,  Web search  Index compression  Ranking: combining contents and hyperlinks  Web page pre-processing  Combining multiple rankings and meta search  Web spamming  Want to know more? Read the textbook

Editor's Notes

  • #18: <number>