SlideShare a Scribd company logo
Advances in 
Image Search and Retrieval	

Oge Marques	

Florida Atlantic University	

Boca Raton, FL - USA
Take-home message	

•  Visual Information Retrieval (VIR) is a fascinating
research field with many open challenges and
opportunities which have the potential to impact
the way we organize, annotate, and retrieve visual
data (images and videos).	

•  In this tutorial we present some of the latest and
most representative advances in image search and
retrieval.
Disclaimer #1	

•  Visual Information Retrieval (VIR) is a highly
interdisciplinary field, but …	

Visual
Information
Retrieval
Image and
Video
Processing
(Multimedia)
Database
Systems
Information
Retrieval
Machine
Learning
Computer
Vision
Data Mining
Visual data
modeling and
representation
Human Visual
Perception
Disclaimer #2	

•  There are many things that I believe…	

•  … but cannot prove
Background and Motivation	

What is it that we’re trying to do 
and 
why is it so difficult? 	

– Taking pictures and storing, sharing, and publishing
them has never been so easy and inexpensive. 	

– If only we could say the same about finding the images
we want and retrieving them…
Background and Motivation	

The “big mismatch”
Background and Motivation	

•  Q: What do you do when you need to find an image
(on the Web)?
•  A1: Google (image search), of course!
Background and Motivation	

Google image search results for “sydney opera house”
Background and Motivation	

Google image search results for “opera”
Background and Motivation	

•  Q: What do you do when you need to find an
image (on the Web)?
•  A2: Other (so-called specialized) image search
engines
•  http://images.search.yahoo.com/
•  http://pictures.ask.com
•  http://www.bing.com/images
•  http://pixsy.com/
Yahoo!
Ask
Bing
Pixsy – several years ago
Pixsy – several hours ago
Background and Motivation	

•  Q:What do you do when you need to find an
image (on the Web)?	

•  A3: Search directly on large photo repositories:	

– Flickr	

– Webshots	

– Shutterstock
Background and Motivation	

Flickr image search results for “opera”
Background and Motivation	

Webshots image search results for “opera”
Background and Motivation	

Shutterstock image search results for “opera”
Background and Motivation	

•  Are you happy with the results so far?
Background and Motivation	

•  Back to our original (two-part) question:	

– What is it that we’re trying to do?	

– We are trying to create 
automated solutions to the problem of 
finding and retrieving visual information, 
from (large, unstructured) repositories, 
in a way that satisfies search criteria specified by users,
relying (primarily) on the visual contents of the media.
Background and Motivation	

•  Why is it so difficult?	

•  There are many challenges, among them:	

– The elusive notion of similarity 	

– The semantic gap	

– Large datasets and broad domains	

– Combination of visual and textual information	

– The users (and how to make them happy)
Outline	

•  Part I – Core concepts, techniques, and tools	

– Design, implementation, and evaluation aspects	

•  Part II – Medical image retrieval	

– Challenges, resources, and opportunities	

•  Part III – Applications and related areas	

– Mobile visual search, social networks, and more	

•  Part IV – Where is image search headed? 	

– Advice for young researchers
Part I	

Core concepts, techniques, and
tools
Core concepts, techniques, and tools	

•  Design	

– Challenges	

– Principles	

– Concepts	

•  Implementation	

– Languages and tools	

•  Evaluation 	

– Datasets	

– Benchmarks
Design challenges	

•  Capturing and measuring similarity	

•  Semantic gap (and other gaps)	

•  Large datasets and broad domains	

•  Users’ needs and intentions	

•  Growing up (as a field)
The elusive notion of similarity	

•  Are these two images similar?
The elusive notion of similarity	

•  Are these two images similar?
The elusive notion of similarity	

•  Is the second or the third image more similar to
the first?
The elusive notion of similarity	

•  Which image fits better to the first two: the third
or the fourth?
The semantic gap	

•  The semantic gap is the lack of coincidence
between the information that one can extract
from the visual data and the interpretation that
the same data have for a user in a given situation.	

•  The pivotal point in content-based retrieval is that the user
seeks semantic similarity, but the database can only provide
similarity by data processing.This is what we called the
semantic gap. [Smeulders et al., 2000]
Alipr
Alipr
Alipr
Alipr
Google similarity search
Google similarity search
Google sort by subject	

http://www.google.com/landing/imagesorting/
Google image swirl	

http://image-swirl.googlelabs.com/
How I see it…	

•  The semantic gap problem has not been solved (and
maybe will never be…)	

•  What are the alternatives?	

–  Treat visual similarity and semantic relatedness differently	

•  Examples:Alipr, Google (or Bing) similarity search, etc.	

–  Improve both (text-based and visual) search methods
independently 	

–  Combine visual and textual information in a meaningful
way	

–  Trust the user 	

•  Collaborative filtering, crowdsourcing, games.
•  But, wait… There
are other gaps!	

– Just when you
thought the
semantic gap was
your only
problem…	

Source: [Deserno, Antani, and Long, 2009]
Large datasets and broad domains	

•  Large datasets bring additional challenges in all
aspects of the system:	

– Storage requirements: images, metadata, and “visual
signatures”	

– Computational cost of indexing, searching, retrieving,
and displaying images	

– Network and latency issues
Large datasets and broad domains
Challenge: users’ needs and intentions	

•  Users and developers have quite different views	

•  Cultural and contextual information should be
taken into account	

•  User intentions are hard to infer	

– Privacy issues	

– Users themselves don’t always know what they want	

– Who misses the MS Office paper clip?
Challenge: users’ needs and intentions	

•  The user’s
perspective	

– What do they
want? 	

– Where do
they want to
search?	

– In what form
do they
express their
query?
Challenge: users’ needs and intentions	

•  The image
retrieval system
should be able to
be mindful of: 	

–  How users wish
the results to be
presented 	

–  Where users
desire to search	

–  The nature of
user input/
interaction.
Challenge: users’ needs and intentions	

•  Each application has
different users (with
different intent, needs,
background, cultural bias,
etc.) and different visual
assets.
Challenge: growing up (as a field)	

•  It’s been 10 years since the “end of the early years”	

–  Are the challenges from 2000 still relevant?	

–  Are the directions and guidelines from 2000 still
appropriate?	

–  Have we grown up (at all)?	

–  Let’s revisit the ‘Concluding Remarks’ from that paper…
Revisiting [Smeulders et al. 2000]	

What they said	

•  Driving forces	

–  “[…] content-based image
retrieval (CBIR) will continue
to grow in every direction:
new audiences, new purposes,
new styles of use, new modes
of interaction, larger data sets,
and new methods to solve
the problems.”	

How I see it	

•  Yes, we have seen many new
audiences, new purposes, new
styles of use, and new modes
of interaction emerge.	

•  Each of these usually requires
new methods to solve the
problems that they bring.	

•  However, not too many
researchers see them as a
driving force (as they should).
Revisiting [Smeulders et al. 2000]	

What they said	

•  Heritage of computer vision	

–  “An important obstacle to
overcome […] is to realize
that image retrieval does not
entail solving the general
image understanding
problem.”	

How I see it	

•  I’m afraid I have bad news…	

–  Computer vision hasn’t made
so much progress during the
past 10 years.	

–  Some classical problems 
(including image 
understanding)
remain unresolved.	

–  Similarly, CBIR from a 
pure computer vision
perspective didn’t work 
too well either.
Revisiting [Smeulders et al. 2000]	

What they said	

•  Influence on computer
vision	

–  “[…] CBIR offers a different
look at traditional computer
vision problems: large data
sets, no reliance on strong
segmentation, and revitalized
interest in color image
processing and invariance.”	

How I see it	

•  The adoption of large data sets
became standard practice in
computer vision.	

•  No reliance on strong
segmentation (still unresolved) led
to new areas of research, e.g.,
automatic ROI extraction and RBIR.	

•  Color image processing and color
descriptors became incredibly
popular, useful, and (to some
degree) effective.	

•  Invariance still a huge problem	

–  But it’s cheaper than ever to have
multiple views.
Revisiting [Smeulders et al. 2000]	

What they said	

•  Similarity and learning	

–  “We make a pledge for the
importance of human- based
similarity rather than general
similarity.Also, the connection
between image semantics,
image data, and query context
will have to be made clearer
in the future.”	

–  “[…] in order to bring
semantics to the user, learning
is inevitable.” 	

How I see it	

•  The authors were pointing in the
right direction (human in the
loop, role of context, benefits
from learning,…)	

•  However:	

–  Similarity is a tough problem to
crack and model. 	

•  Even the understanding of how
humans judge image similarity is
very limited.	

–  Machine learning is almost
inevitable…	

•  … but sometimes it can be
abused.
Revisiting [Smeulders et al. 2000]	

What they said	

•  Interaction	

–  Better visualization options,
more control to the user,
ability to provide feedback
[…]	

How I see it	

•  Significant progress on
visualization interfaces and
devices.	

•  Relevance Feedback: still a
very tricky tradeoff (effort
vs. perceived benefit), but
more popular than ever
(rating, thumbs up/down,
etc.)
Revisiting [Smeulders et al. 2000]	

What they said	

•  Need for databases	

–  “The connection between
CBIR and database research is
likely to increase in the
future. […] problems like the
definition of suitable query
languages, efficient search in
high dimensional feature
space, search in the presence
of changing similarity
measures are largely unsolved
[…]”	

How I see it	

•  Very little progress	

–  Image search and retrieval has
benefited much more from
document information
retrieval than from database
research.
Revisiting [Smeulders et al. 2000]	

What they said	

•  The problem of evaluation	

–  CBIR could use a reference
standard against which new
algorithms could be evaluated
(similar to TREC in the field of
text recognition). 	

–  “A comprehensive and publicly
available collection of images,
sorted by class and retrieval
purposes, together with a
protocol to standardize
experimental practices, will be
instrumental in the next phase
of CBIR.”	

How I see it	

•  Significant progress on
benchmarks, standardized
datasets, etc. 	

–  ImageCLEF	

–  PascalVOC Challenge	

–  MSRA dataset	

–  Simplicity dataset	

–  UCID dataset and ground truth
(GT)	

–  Accio / SIVAL dataset and GT	

–  Caltech 101, Caltech 256	

–  LabelMe
Revisiting [Smeulders et al. 2000]	

What they said	

•  Semantic gap and other
sources	

–  “A critical point in the
advancement of CBIR is the
semantic gap, where the
meaning of an image is rarely
self-evident. […] One way to
resolve the semantic gap
comes from sources outside
the image by integrating other
sources of information about the
image in the query.”	

How I see it	

•  The semantic gap problem
has not been solved (and
maybe will never be…)	

•  But the idea about using
other sources was right on
the spot!	

–  Geographical context	

–  Social networks	

–  Tags
Visual Information Retrieval (VIR)	

Query / Search
Engine
User
User interface (Querying, Browsing,
Viewing)
Digital Image and
Video Archive
Visual summaries Indexes
Digitization +
Compression
Cataloguing / Feature
extraction
Image or Video
Designing aVIR system: a mind map
Tools and resources	

•  Visual descriptors and machine learning algorithms
have become commodities.	

•  Examples of publicly available implementation and
tools:	

– Visual descriptors:	

•  img(Rummager) by Savvas Chatzichristofis	

•  Caliph  Emir and Lire by Mathias Lux	

– Machine Learning:	

•  Weka
Part II	

Medical Image Retrieval
Medical image retrieval	

•  Challenges	

– We’re entering a new country… 	

•  How much can we bring?	

•  Do we speak the language?	

•  Do we know their culture?	

•  Do they understand us and where we come from?	

•  Opportunities	

– They use images (extensively)	

– They have expert knowledge	

– Domains are narrow (almost by definition)	

– Fewer clients, but potentially more $$
Medical image retrieval	

•  Selected challenges:	

– Different terminology	

– Standards	

– Modality dependencies	

•  Other challenges:	

– Equipment dependencies	

– Privacy issues	

– Proprietary data
Different terminology	

•  Be prepared for:	

– New acronyms	

•  CBMIR (Content-Based Medical Image Retrieval)	

•  PACS (Picture Archiving and Communication System)	

•  DICOM (Digital Imaging and COmmunication in Medicine)	

•  Hospital Information Systems (HIS)	

•  Radiological Information Systems (RIS)	

– New phrases	

•  Imaging informatics	

– Lots of technical medical terms
Standards	

•  DICOM (http://medical.nema.org/)	

–  Global IT standard, created in 1993, used in virtually all
hospitals worldwide. 	

–  Designed to ensure the interoperability of different
systems and manage related workflow.	

–  Will be required by all EHR systems that include imaging
information as an integral part of the patient record. 	

–  750+ technical and medical experts participate in 20+
active DICOM working groups.	

–  Standard is updated 4-5 times per year.	

–  Many available tools! (see http://www.idoimaging.com/)
Medical image modalities	

•  The IRMA code [Lehmann et al., 2003]	

–  4 axes with 3 to 4 positions, each in {0,...9,a,...,z}, where 0
denotes unspecified to determine the end of a path along an
axis. 	

•  Technical code (T) describes the imaging modality	

•  Directional code (D) models body orientations	

•  Anatomical code (A) refers to the body region examined	

•  Biological code (B) describes the biological system
examined.
Medical image modalities	

•  The IRMA code [Lehmann et al., 2003]	

–  The entire code results in a character string of 14
characters (IRMA:TTTT – DDD – AAA – BBB).	

Example: “x-ray, projection radiography,
analog, high energy – sagittal, left lateral
decubitus, inspiration – chest, lung –
respiratory system, lung”
Source: [Lehmann et al., 2003]
Medical image modalities	

•  The IRMA code
[Lehmann et al.,
2003]	

–  The companion
tool…	

Source: [Lehmann et al., 2004]
CBMIR vs. text-based MIR	

•  Most current retrieval systems in clinical use rely on
text keywords such as DICOM header information to
perform retrieval.	

•  CBIR has been widely researched in a variety of
domains and provides an intuitive and expressive
method for querying visual data using features, e.g.
color, shape, and texture.	

•  However, current CBIR systems:	

–  are not easily integrated into the healthcare environment; 	

–  have not been widely evaluated using a large dataset; and 	

–  lack the ability to perform relevance feedback to refine
retrieval results.	

Source: [Hsu et al., 2009]
Who are the main players?	

•  USA	

– NIH (National Institutes of Health)	

•  NIBIB - National Institute of Biomedical Imaging and
Bioengineering	

•  NCI - National Cancer Institute	

•  NLM – National Libraries of Medicine	

– Several universities and hospitals	

•  Europe	

– Aachen University (Germany)	

– Geneva University (Switzerland)	

•  Big companies (Siemens, GE, etc.)
Medical image retrieval systems: examples	

•  IRMA (Image Retrieval in Medical Applications)	

–  Aachen University (Germany)	

•  http://ganymed.imib.rwth-aachen.de/irma/	

–  3 online demos:	

•  IRMA Query demo: allows the evaluation of CBIR on several
databases.	

•  IRMA Extended Query Refinement demo: CBIR from the IRMA
database (a subset of 10,000 images). 	

•  Spine Pathology and Image Retrieval Systems (SPIRS) designed by the
NLM/NIH (USA): holds information of ~17,000 spine x-rays.
Medical image retrieval systems: examples	

•  MedGIFT (GNU Image Finding Tool)	

– Geneva University (Switzerland)	

•  http://www.sim.hcuge.ch/medgift/	

– Large effort, including projects such as:	

•  Talisman (lung image retrieval) 	

•  Case-based fracture image retrieval system	

•  Onco-Media: medical image retrieval + grid computing	

•  ImageCLEF: evaluation and validation	

•  medSearch
Medical image retrieval systems: examples	

•  WebMIRS	

– NIH / NLM (USA)	

•  http://archive.nlm.nih.gov/proj/webmirs/index.php 	

– Query by text + navigation by categories	

– Uses datasets and related x-ray images from the
National Health and Nutrition Examination Survey
(NHANES)
Medical image retrieval systems: examples	

•  SPIRS (Spine Pathology  Image Retrieval System):
Web-based image retrieval system for large
biomedical databases	

– NIH / UCLA (USA)	

– Representative case study on highly specialized CBMIR	

Source: [Hsu et al., 2009]
Medical image retrieval systems: examples	

•  National Biomedical Imaging Archive (NBIA)	

– NCI / NIH (USA)	

•  https://imaging.nci.nih.gov/ 	

– Search based on metadata (DICOM fields)	

– 3 search options:	

•  Simple	

•  Advanced	

•  Dynamic
Medical image retrieval systems: examples	

•  ARSS Goldminer 	

– American Roentgen Ray Society (USA)	

•  http://goldminer.arrs.org/ 	

– Query by text	

– Results can be filtered by:	

•  Modality	

•  Age	

•  Sex
Medical image retrieval systems: examples	

•  Yottalook Images	

–  iVirtuoso (USA)	

•  http://www.yottalook.com/ 	

–  Developed and maintained by four radiologists	

–  Query by text	

–  Claims to use 4 “core technologies”:	

•  natural query analysis”	

•  semantic ontology”	

•  “relevance algorithm” 	

•  a specialized content delivery system that provides high yield
content based on the search term.
Evaluation: ImageCLEF Medical Image Retrieval 	

•  ImageCLEF Medical Image 
Retrieval 	

•  http://www.imageclef.org/2011/medical 	

– Dataset: 77,000+ images from articles published in
medical journals including text of the captions and link
to the html of the full text articles. 	

– 3 types of tasks:	

•  Modality Classification: given an image, return its modality	

•  Ad-hoc retrieval: classic medical retrieval task, with 3
“flavors”: textual, mixed and semantic queries 	

•  Case-based retrieval: retrieve cases including images that
might best suit the provided case description.
Evaluation: ImageCLEF Medical Image Retrieval	

•  ImageCLEF Medical Image Retrieval 2011 	

– Modality Classification
Evaluation: ImageCLEF Medical Image Retrieval 	

•  ImageCLEF Medical Image Retrieval 2011 	

– Modality Classification – FAU Team 	

•  Personnel: 4 grad students + 2 undergrads + advisor 	

•  Strategy:	

–  Textual classification using Lucene and associated tools and libraries 	

–  Visual classification using 8 contemporary descriptors and 3
different families of classifiers, implemented using Weka and
associated tools and libraries	

•  Supporting tools:	

–  Manual annotation tool (http://imageclef.mlab.ceecs.fau.edu/)	

–  Training set visualization tool 
(http://imageclef.mlab.ceecs.fau.edu/classification/)
Evaluation: ImageCLEF Medical Image Retrieval 	

•  ImageCLEF
Medical Image
Retrieval 2011 	

– Modality
Classification
Results – FAU
Team (textual)
Evaluation: ImageCLEF Medical Image Retrieval	

•  ImageCLEF Medical Image Retrieval 2011 	

– Modality Classification Results – FAU Team (visual)
Evaluation: ImageCLEF Medical Image Retrieval	

•  ImageCLEF
Medical
Image
Retrieval
2011 	

•  Modality
Classification
Results –
FAU Team
(visual)
Medical Image Retrieval: promising directions	

•  Better user interfaces (responsive, highly interactive,
and capable of supporting relevance feedback)	

•  New applications of CBMIR, including:	

–  Teaching	

–  Research 	

–  Diagnosis 	

–  PACS and Electronic Patient Records	

•  CBMIR evaluation using medical experts	

•  Integration of local and global features	

•  New visual descriptors
Medical Image Retrieval: promising directions	

•  New devices
Part III	

Applications and related areas
Applications and related areas	

•  New devices and services	

•  Mobile visual search	

•  Image search and retrieval in the age of social
networks	

•  Games!	

•  Other related areas	

•  Our recent work (highlights)
New devices and services	

•  Flickr (b. 2004)	

•  YouTube (b. 2005)	

•  Flip video cameras (b. 2006)	

•  iPhone (b. 2007)	

•  iPad (b. 2010)
Mobile visual search	

•  Driving factors	

– Capable devices	

Source: http://www.apple.com/iphone/specs.html 	

1 GHz ARM
Cortex-A8
processor,
PowerVR
SGX535GPU,
Apple A4 chipset
Mobile visual search	

•  Driving factors	

– Motivated users: image taking and image sharing are
huge!	

–  Source: http://www.onlinemarketing-trends.com/2011/03/facebook-photo-statistics-and-insights.html
Mobile visual search	

•  Facebook for iPhone	

–  Source: http://statistics.allfacebook.com/applications/single/facebook-for-iphone/6628568379/
Mobile visual search	

•  Instagram: 2 million registered (although not necessarily
active) users, who upload ~300,000 photos per day	

•  Several apps based on it!	

–  http://iphone.appstorm.net/roundups/photography/5-cool-apps-for-getting-the-most-out-of-instagram/
Mobile visual search	

•  Food photo
sharing!
Mobile visual search	

•  Driving factors	

– Legitimate (or not quite…) needs and use cases	

–  Source: http://www.slideshare.net/dtunkelang/search-by-sight-google-goggles
Mobile visual search	

•  Driving factors	

– Smart phone market
Mobile visual search	

•  Smart phone market	

Source: http://www.cellular-news.com/story/48647.php?s=h
Mobile visual search	

•  Examples of applications	

– Google Goggles	

– oMoby (and the IQ Engines API)	

– Others (kooaba, Fetch!, Gazopa, etc.)
Mobile visual search	

•  Google Goggles 	

– Android and iPhone	

– Narrow-domain search and retrieval
Mobile visual search	

•  oMoby (and the IQ Engines API)	

– iPhone
Mobile visual search	

•  oMoby (and the IQ Engines API)
Image search and retrieval  social networks	

•  The [so-called] Web 2.0 has brought about:	

–  New data sources	

–  New usage patterns	

–  New understanding about the users, their needs,
habits, preferences	

–  New opportunities	

–  Lots of metadata!	

–  A chance to experience a true paradigm shift	

•  Before: image annotation is tedious, labor-intensive,
expensive	

•  After: image annotation is fun!
Games!	

◦  Google Image Labeler	

◦  Games with a purpose (GWAP)	

  The ESP Game	

  Squigl	

  Matchin
Other related areas	

•  Semi-automatic image annotation	

•  Tag recommendation systems	

•  Story annotation engines	

•  Content-based image filtering	

•  Copyright detection	

•  Watermark detection 	

– and many more
Our recent work (highlights)	

•  PRISM	

– Image Genius	

•  Unsupervised ROI extraction from an image	

– Crazy Collage	

•  MEDIX and associated tools	

•  Callisto: a content-based tag recommendation
tool
Research Team
PRISM	

	

With Liam Mayron, Harris Corp., USA
Image Genius	

	

With Asif Rahman, FAU, USA
Unsupervised ROI extraction	

	

With
Gustavo B.
Borba and
Humberto
R. Gamba,
UTFPR,
Brazil
Crazy Collage	

	

Gustavo B.
Borba et
al., UTFPR,
Brazil
MEDIX	

•  Medical image retrieval system with DICOM
capabilities	

	

With Asif Rahman, FAU, USA
Callisto	

	

With Mathias Lux and
Arthur Pitman, Klagenfurt
University,Austria
Part IV	

Where is image search headed?
Where is image search headed? 	

•  Advice for [young] researchers	

– In this last part, I’ve compiled pieces and bits of advice
that I believe might help researchers who are entering
the field.	

– They focus on research avenues that I personally
consider to be the most promising.
Advice for [young] researchers	

• LOOK	

• THINK	

• UNDERSTAND	

• CREATE
Advice for [young] researchers	

• LOOK…	

– at yourself (how do you search for images and videos?)	

– around (related areas and how they have grown)	

– at Google (and other major players)
Advice for [young] researchers	

• THINK…	

– mobile devices	

– new devices and services	

– social networks	

– games
Advice for [young] researchers	

• UNDERSTAND…	

– human intentions and emotions	

– the context of the search	

– user’s preferences and needs
Advice for [young] researchers	

• CREATE…	

– better interfaces	

– better user experience	

– new business opportunities (added value)
Concluding thoughts	

–  I believe (but cannot prove…) that successfulVIR
solutions will:	

•  combine content-based image retrieval (CBIR) with
metadata (high-level semantic-based image retrieval)	

•  only be truly successful in narrow domains	

•  include the user in the loop	

– Relevance Feedback (RF)	

– Collaborative efforts (tagging, rating, annotating)	

•  provide friendly, intuitive interfaces	

•  incorporate results and insights from cognitive science,
particularly human visual attention, perception, and
memory
Concluding thoughts	

–  I believe (but cannot prove…) that successfulVIR
solutions will:	

•  combine content-based image retrieval (CBIR) with
metadata (high-level semantic-based image retrieval)	

•  only be truly successful in narrow domains	

•  include the user in the loop	

– Relevance Feedback (RF)	

– Collaborative efforts (tagging, rating, annotating)	

•  provide friendly, intuitive interfaces	

•  incorporate results and insights from cognitive science,
particularly human visual attention, perception, and
memory
Concluding thoughts	

–  I believe (but cannot prove…) that successfulVIR
solutions will:	

•  combine content-based image retrieval (CBIR) with
metadata (high-level semantic-based image retrieval)	

•  only be truly successful in narrow domains	

•  include the user in the loop	

– Relevance Feedback (RF)	

– Collaborative efforts (tagging, rating, annotating)	

•  provide friendly, intuitive interfaces	

•  incorporate results and insights from cognitive science,
particularly human visual attention, perception, and
memory
Concluding thoughts	

–  I believe (but cannot prove…) that successfulVIR
solutions will:	

•  combine content-based image retrieval (CBIR) with
metadata (high-level semantic-based image retrieval)	

•  only be truly successful in narrow domains	

•  include the user in the loop	

– Relevance Feedback (RF)	

– Collaborative efforts (tagging, rating, annotating)	

•  provide friendly, intuitive interfaces	

•  incorporate results and insights from cognitive science,
particularly human visual attention, perception, and
memory
Concluding thoughts	

–  I believe (but cannot prove…) that successfulVIR
solutions will:	

•  combine content-based image retrieval (CBIR) with
metadata (high-level semantic-based image retrieval)	

•  only be truly successful in narrow domains	

•  include the user in the loop	

– Relevance Feedback (RF)	

– Collaborative efforts (tagging, rating, annotating)	

•  provide friendly, intuitive interfaces	

•  incorporate results and insights from cognitive science,
particularly human visual attention, perception, and
memory
Concluding thoughts	

•  “Image search and retrieval” is not a problem, but
rather a collection of related problems that look like
one.	

•  There is a great need for good solutions to specific
problems. 	

•  10 years after “the end of the early years”, research in
visual information retrieval still has many open
problems, challenges, and opportunities.
Thanks!	

•  Questions?	

•  For additional information: omarques@fau.edu

More Related Content

PDF
Visual Information Retrieval: Advances, Challenges and Opportunities
PDF
Advances and Challenges in Visual Information Search and Retrieval (WVC 2012 ...
PDF
Oge Marques (FAU) - invited talk at WISMA 2010 (Barcelona, May 2010)
PDF
Image Processing and Computer Vision in iOS
PDF
Recent advances in visual information retrieval marques klu june 2010
PDF
Jia-Bin Huang's Curriculum Vitae
PPT
Multimedia Information Retrieval: What is it, and why isn't ...
PDF
16 ijcse-01237
Visual Information Retrieval: Advances, Challenges and Opportunities
Advances and Challenges in Visual Information Search and Retrieval (WVC 2012 ...
Oge Marques (FAU) - invited talk at WISMA 2010 (Barcelona, May 2010)
Image Processing and Computer Vision in iOS
Recent advances in visual information retrieval marques klu june 2010
Jia-Bin Huang's Curriculum Vitae
Multimedia Information Retrieval: What is it, and why isn't ...
16 ijcse-01237

What's hot (7)

PDF
Recent Advances in Computer Vision
PPTX
Computer Vision Crash Course
PDF
Lecture 1 computer vision introduction
DOC
Elegant Resume
PDF
Information Architecture - Part 1 - Spring 2013 - Class 1
PDF
Lecture 5: Personalization on the Social Web (2013)
PPTX
Technology & robotics dem
Recent Advances in Computer Vision
Computer Vision Crash Course
Lecture 1 computer vision introduction
Elegant Resume
Information Architecture - Part 1 - Spring 2013 - Class 1
Lecture 5: Personalization on the Social Web (2013)
Technology & robotics dem
Ad

Viewers also liked (20)

PDF
CBIR by deep learning
PDF
CBIR in the Era of Deep Learning
PDF
PPTX
Une approche d’indexation et de recherche d’images pulmonaires TDM par le con...
PDF
Une Approche d’Indexation et de Recherche d’Images Pulmonaires TDM par le Con...
PDF
Image retrieval: challenges and opportunities
PPT
Content based image retrieval(cbir)
PDF
Indexation image
PPT
Multimodal Analysis for Bridging Semantic Gap with Biologically Inspired Algo...
PDF
A Comparative Study of Content Based Image Retrieval Trends and Approaches
PDF
Image Search Engine Frequently Asked Questions
PDF
UNSUPERVISED VISUAL HASHING WITH SEMANTIC ASSISTANT FOR CONTENT-BASED IMAGE R...
PPTX
Image search engine
PPS
The Real Problem of Bridging the Multimedia “Semantic Gap”
PPTX
Cbir final ppt
PPTX
Content based image retrieval
PDF
Building Knowledge Graphs in DIG
PPTX
DARPA Project Memex Erodes Privacy
PDF
Literature Review on Content Based Image Retrieval
PPT
Visual Search
CBIR by deep learning
CBIR in the Era of Deep Learning
Une approche d’indexation et de recherche d’images pulmonaires TDM par le con...
Une Approche d’Indexation et de Recherche d’Images Pulmonaires TDM par le Con...
Image retrieval: challenges and opportunities
Content based image retrieval(cbir)
Indexation image
Multimodal Analysis for Bridging Semantic Gap with Biologically Inspired Algo...
A Comparative Study of Content Based Image Retrieval Trends and Approaches
Image Search Engine Frequently Asked Questions
UNSUPERVISED VISUAL HASHING WITH SEMANTIC ASSISTANT FOR CONTENT-BASED IMAGE R...
Image search engine
The Real Problem of Bridging the Multimedia “Semantic Gap”
Cbir final ppt
Content based image retrieval
Building Knowledge Graphs in DIG
DARPA Project Memex Erodes Privacy
Literature Review on Content Based Image Retrieval
Visual Search
Ad

Similar to Advances in Image Search and Retrieval (20)

PDF
PDF
IRJET- Image Seeker:Finding Similar Images
PPTX
Content Based Image and Video Retrieval Algorithm
PPSX
Image Search: Then and Now
PPT
Yang (1) Image Information Retrieval.ppt
PPT
viretrieval2.ppt chain codes Multimedia Information Retrieval
PPTX
Content Based Image Retrieval
PPTX
Invited Talk OAGM Workshop Salzburg, May 2015
PDF
Content-Based Image Retrieval Features: A Survey
PDF
Interactive Video Search: Where is the User in the Age of Deep Learning?
PDF
Robust and Radial Image Comparison Using Reverse Image Search
PPT
Yang.ppt
PDF
Paper id 25201471
PPT
Visual Search
PDF
An Impact on Content Based Image Retrival A Perspective View
PDF
Defense applications white paper
PDF
Introduction about TRECVID (The Video Retreival benchmark)
PPT
Final Year Major Project Report ( Year 2010-2014 Batch )
PPTX
CBMI 2013 Presentation: User Intentions in Multimedia
IRJET- Image Seeker:Finding Similar Images
Content Based Image and Video Retrieval Algorithm
Image Search: Then and Now
Yang (1) Image Information Retrieval.ppt
viretrieval2.ppt chain codes Multimedia Information Retrieval
Content Based Image Retrieval
Invited Talk OAGM Workshop Salzburg, May 2015
Content-Based Image Retrieval Features: A Survey
Interactive Video Search: Where is the User in the Age of Deep Learning?
Robust and Radial Image Comparison Using Reverse Image Search
Yang.ppt
Paper id 25201471
Visual Search
An Impact on Content Based Image Retrival A Perspective View
Defense applications white paper
Introduction about TRECVID (The Video Retreival benchmark)
Final Year Major Project Report ( Year 2010-2014 Batch )
CBMI 2013 Presentation: User Intentions in Multimedia

Recently uploaded (20)

PDF
LDMMIA Reiki Yoga Finals Review Spring Summer
PPTX
Final Presentation General Medicine 03-08-2024.pptx
PPTX
History, Philosophy and sociology of education (1).pptx
PPTX
Unit 4 Skeletal System.ppt.pptxopresentatiom
PDF
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
PPTX
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
PDF
Computing-Curriculum for Schools in Ghana
PDF
Empowerment Technology for Senior High School Guide
PDF
RTP_AR_KS1_Tutor's Guide_English [FOR REPRODUCTION].pdf
PPTX
Digestion and Absorption of Carbohydrates, Proteina and Fats
PDF
Paper A Mock Exam 9_ Attempt review.pdf.
PDF
Classroom Observation Tools for Teachers
PPTX
CHAPTER IV. MAN AND BIOSPHERE AND ITS TOTALITY.pptx
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PDF
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
PDF
LNK 2025 (2).pdf MWEHEHEHEHEHEHEHEHEHEHE
PDF
1_English_Language_Set_2.pdf probationary
PDF
IGGE1 Understanding the Self1234567891011
PDF
Hazard Identification & Risk Assessment .pdf
PDF
Trump Administration's workforce development strategy
LDMMIA Reiki Yoga Finals Review Spring Summer
Final Presentation General Medicine 03-08-2024.pptx
History, Philosophy and sociology of education (1).pptx
Unit 4 Skeletal System.ppt.pptxopresentatiom
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
Computing-Curriculum for Schools in Ghana
Empowerment Technology for Senior High School Guide
RTP_AR_KS1_Tutor's Guide_English [FOR REPRODUCTION].pdf
Digestion and Absorption of Carbohydrates, Proteina and Fats
Paper A Mock Exam 9_ Attempt review.pdf.
Classroom Observation Tools for Teachers
CHAPTER IV. MAN AND BIOSPHERE AND ITS TOTALITY.pptx
Supply Chain Operations Speaking Notes -ICLT Program
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
LNK 2025 (2).pdf MWEHEHEHEHEHEHEHEHEHEHE
1_English_Language_Set_2.pdf probationary
IGGE1 Understanding the Self1234567891011
Hazard Identification & Risk Assessment .pdf
Trump Administration's workforce development strategy

Advances in Image Search and Retrieval

  • 1. Advances in Image Search and Retrieval Oge Marques Florida Atlantic University Boca Raton, FL - USA
  • 2. Take-home message •  Visual Information Retrieval (VIR) is a fascinating research field with many open challenges and opportunities which have the potential to impact the way we organize, annotate, and retrieve visual data (images and videos). •  In this tutorial we present some of the latest and most representative advances in image search and retrieval.
  • 3. Disclaimer #1 •  Visual Information Retrieval (VIR) is a highly interdisciplinary field, but … Visual Information Retrieval Image and Video Processing (Multimedia) Database Systems Information Retrieval Machine Learning Computer Vision Data Mining Visual data modeling and representation Human Visual Perception
  • 4. Disclaimer #2 •  There are many things that I believe… •  … but cannot prove
  • 5. Background and Motivation What is it that we’re trying to do and why is it so difficult? – Taking pictures and storing, sharing, and publishing them has never been so easy and inexpensive. – If only we could say the same about finding the images we want and retrieving them…
  • 6. Background and Motivation The “big mismatch”
  • 7. Background and Motivation •  Q: What do you do when you need to find an image (on the Web)? •  A1: Google (image search), of course!
  • 8. Background and Motivation Google image search results for “sydney opera house”
  • 9. Background and Motivation Google image search results for “opera”
  • 10. Background and Motivation •  Q: What do you do when you need to find an image (on the Web)? •  A2: Other (so-called specialized) image search engines •  http://images.search.yahoo.com/ •  http://pictures.ask.com •  http://www.bing.com/images •  http://pixsy.com/
  • 12. Ask
  • 13. Bing
  • 14. Pixsy – several years ago
  • 15. Pixsy – several hours ago
  • 16. Background and Motivation •  Q:What do you do when you need to find an image (on the Web)? •  A3: Search directly on large photo repositories: – Flickr – Webshots – Shutterstock
  • 17. Background and Motivation Flickr image search results for “opera”
  • 18. Background and Motivation Webshots image search results for “opera”
  • 19. Background and Motivation Shutterstock image search results for “opera”
  • 20. Background and Motivation •  Are you happy with the results so far?
  • 21. Background and Motivation •  Back to our original (two-part) question: – What is it that we’re trying to do? – We are trying to create automated solutions to the problem of finding and retrieving visual information, from (large, unstructured) repositories, in a way that satisfies search criteria specified by users, relying (primarily) on the visual contents of the media.
  • 22. Background and Motivation •  Why is it so difficult? •  There are many challenges, among them: – The elusive notion of similarity – The semantic gap – Large datasets and broad domains – Combination of visual and textual information – The users (and how to make them happy)
  • 23. Outline •  Part I – Core concepts, techniques, and tools – Design, implementation, and evaluation aspects •  Part II – Medical image retrieval – Challenges, resources, and opportunities •  Part III – Applications and related areas – Mobile visual search, social networks, and more •  Part IV – Where is image search headed? – Advice for young researchers
  • 24. Part I Core concepts, techniques, and tools
  • 25. Core concepts, techniques, and tools •  Design – Challenges – Principles – Concepts •  Implementation – Languages and tools •  Evaluation – Datasets – Benchmarks
  • 26. Design challenges •  Capturing and measuring similarity •  Semantic gap (and other gaps) •  Large datasets and broad domains •  Users’ needs and intentions •  Growing up (as a field)
  • 27. The elusive notion of similarity •  Are these two images similar?
  • 28. The elusive notion of similarity •  Are these two images similar?
  • 29. The elusive notion of similarity •  Is the second or the third image more similar to the first?
  • 30. The elusive notion of similarity •  Which image fits better to the first two: the third or the fourth?
  • 31. The semantic gap •  The semantic gap is the lack of coincidence between the information that one can extract from the visual data and the interpretation that the same data have for a user in a given situation. •  The pivotal point in content-based retrieval is that the user seeks semantic similarity, but the database can only provide similarity by data processing.This is what we called the semantic gap. [Smeulders et al., 2000]
  • 32. Alipr
  • 33. Alipr
  • 34. Alipr
  • 35. Alipr
  • 38. Google sort by subject http://www.google.com/landing/imagesorting/
  • 40. How I see it… •  The semantic gap problem has not been solved (and maybe will never be…) •  What are the alternatives? –  Treat visual similarity and semantic relatedness differently •  Examples:Alipr, Google (or Bing) similarity search, etc. –  Improve both (text-based and visual) search methods independently –  Combine visual and textual information in a meaningful way –  Trust the user •  Collaborative filtering, crowdsourcing, games.
  • 41. •  But, wait… There are other gaps! – Just when you thought the semantic gap was your only problem… Source: [Deserno, Antani, and Long, 2009]
  • 42. Large datasets and broad domains •  Large datasets bring additional challenges in all aspects of the system: – Storage requirements: images, metadata, and “visual signatures” – Computational cost of indexing, searching, retrieving, and displaying images – Network and latency issues
  • 43. Large datasets and broad domains
  • 44. Challenge: users’ needs and intentions •  Users and developers have quite different views •  Cultural and contextual information should be taken into account •  User intentions are hard to infer – Privacy issues – Users themselves don’t always know what they want – Who misses the MS Office paper clip?
  • 45. Challenge: users’ needs and intentions •  The user’s perspective – What do they want? – Where do they want to search? – In what form do they express their query?
  • 46. Challenge: users’ needs and intentions •  The image retrieval system should be able to be mindful of: –  How users wish the results to be presented –  Where users desire to search –  The nature of user input/ interaction.
  • 47. Challenge: users’ needs and intentions •  Each application has different users (with different intent, needs, background, cultural bias, etc.) and different visual assets.
  • 48. Challenge: growing up (as a field) •  It’s been 10 years since the “end of the early years” –  Are the challenges from 2000 still relevant? –  Are the directions and guidelines from 2000 still appropriate? –  Have we grown up (at all)? –  Let’s revisit the ‘Concluding Remarks’ from that paper…
  • 49. Revisiting [Smeulders et al. 2000] What they said •  Driving forces –  “[…] content-based image retrieval (CBIR) will continue to grow in every direction: new audiences, new purposes, new styles of use, new modes of interaction, larger data sets, and new methods to solve the problems.” How I see it •  Yes, we have seen many new audiences, new purposes, new styles of use, and new modes of interaction emerge. •  Each of these usually requires new methods to solve the problems that they bring. •  However, not too many researchers see them as a driving force (as they should).
  • 50. Revisiting [Smeulders et al. 2000] What they said •  Heritage of computer vision –  “An important obstacle to overcome […] is to realize that image retrieval does not entail solving the general image understanding problem.” How I see it •  I’m afraid I have bad news… –  Computer vision hasn’t made so much progress during the past 10 years. –  Some classical problems (including image understanding) remain unresolved. –  Similarly, CBIR from a pure computer vision perspective didn’t work too well either.
  • 51. Revisiting [Smeulders et al. 2000] What they said •  Influence on computer vision –  “[…] CBIR offers a different look at traditional computer vision problems: large data sets, no reliance on strong segmentation, and revitalized interest in color image processing and invariance.” How I see it •  The adoption of large data sets became standard practice in computer vision. •  No reliance on strong segmentation (still unresolved) led to new areas of research, e.g., automatic ROI extraction and RBIR. •  Color image processing and color descriptors became incredibly popular, useful, and (to some degree) effective. •  Invariance still a huge problem –  But it’s cheaper than ever to have multiple views.
  • 52. Revisiting [Smeulders et al. 2000] What they said •  Similarity and learning –  “We make a pledge for the importance of human- based similarity rather than general similarity.Also, the connection between image semantics, image data, and query context will have to be made clearer in the future.” –  “[…] in order to bring semantics to the user, learning is inevitable.” How I see it •  The authors were pointing in the right direction (human in the loop, role of context, benefits from learning,…) •  However: –  Similarity is a tough problem to crack and model. •  Even the understanding of how humans judge image similarity is very limited. –  Machine learning is almost inevitable… •  … but sometimes it can be abused.
  • 53. Revisiting [Smeulders et al. 2000] What they said •  Interaction –  Better visualization options, more control to the user, ability to provide feedback […] How I see it •  Significant progress on visualization interfaces and devices. •  Relevance Feedback: still a very tricky tradeoff (effort vs. perceived benefit), but more popular than ever (rating, thumbs up/down, etc.)
  • 54. Revisiting [Smeulders et al. 2000] What they said •  Need for databases –  “The connection between CBIR and database research is likely to increase in the future. […] problems like the definition of suitable query languages, efficient search in high dimensional feature space, search in the presence of changing similarity measures are largely unsolved […]” How I see it •  Very little progress –  Image search and retrieval has benefited much more from document information retrieval than from database research.
  • 55. Revisiting [Smeulders et al. 2000] What they said •  The problem of evaluation –  CBIR could use a reference standard against which new algorithms could be evaluated (similar to TREC in the field of text recognition). –  “A comprehensive and publicly available collection of images, sorted by class and retrieval purposes, together with a protocol to standardize experimental practices, will be instrumental in the next phase of CBIR.” How I see it •  Significant progress on benchmarks, standardized datasets, etc. –  ImageCLEF –  PascalVOC Challenge –  MSRA dataset –  Simplicity dataset –  UCID dataset and ground truth (GT) –  Accio / SIVAL dataset and GT –  Caltech 101, Caltech 256 –  LabelMe
  • 56. Revisiting [Smeulders et al. 2000] What they said •  Semantic gap and other sources –  “A critical point in the advancement of CBIR is the semantic gap, where the meaning of an image is rarely self-evident. […] One way to resolve the semantic gap comes from sources outside the image by integrating other sources of information about the image in the query.” How I see it •  The semantic gap problem has not been solved (and maybe will never be…) •  But the idea about using other sources was right on the spot! –  Geographical context –  Social networks –  Tags
  • 57. Visual Information Retrieval (VIR) Query / Search Engine User User interface (Querying, Browsing, Viewing) Digital Image and Video Archive Visual summaries Indexes Digitization + Compression Cataloguing / Feature extraction Image or Video
  • 59. Tools and resources •  Visual descriptors and machine learning algorithms have become commodities. •  Examples of publicly available implementation and tools: – Visual descriptors: •  img(Rummager) by Savvas Chatzichristofis •  Caliph Emir and Lire by Mathias Lux – Machine Learning: •  Weka
  • 61. Medical image retrieval •  Challenges – We’re entering a new country… •  How much can we bring? •  Do we speak the language? •  Do we know their culture? •  Do they understand us and where we come from? •  Opportunities – They use images (extensively) – They have expert knowledge – Domains are narrow (almost by definition) – Fewer clients, but potentially more $$
  • 62. Medical image retrieval •  Selected challenges: – Different terminology – Standards – Modality dependencies •  Other challenges: – Equipment dependencies – Privacy issues – Proprietary data
  • 63. Different terminology •  Be prepared for: – New acronyms •  CBMIR (Content-Based Medical Image Retrieval) •  PACS (Picture Archiving and Communication System) •  DICOM (Digital Imaging and COmmunication in Medicine) •  Hospital Information Systems (HIS) •  Radiological Information Systems (RIS) – New phrases •  Imaging informatics – Lots of technical medical terms
  • 64. Standards •  DICOM (http://medical.nema.org/) –  Global IT standard, created in 1993, used in virtually all hospitals worldwide. –  Designed to ensure the interoperability of different systems and manage related workflow. –  Will be required by all EHR systems that include imaging information as an integral part of the patient record. –  750+ technical and medical experts participate in 20+ active DICOM working groups. –  Standard is updated 4-5 times per year. –  Many available tools! (see http://www.idoimaging.com/)
  • 65. Medical image modalities •  The IRMA code [Lehmann et al., 2003] –  4 axes with 3 to 4 positions, each in {0,...9,a,...,z}, where 0 denotes unspecified to determine the end of a path along an axis. •  Technical code (T) describes the imaging modality •  Directional code (D) models body orientations •  Anatomical code (A) refers to the body region examined •  Biological code (B) describes the biological system examined.
  • 66. Medical image modalities •  The IRMA code [Lehmann et al., 2003] –  The entire code results in a character string of 14 characters (IRMA:TTTT – DDD – AAA – BBB). Example: “x-ray, projection radiography, analog, high energy – sagittal, left lateral decubitus, inspiration – chest, lung – respiratory system, lung” Source: [Lehmann et al., 2003]
  • 67. Medical image modalities •  The IRMA code [Lehmann et al., 2003] –  The companion tool… Source: [Lehmann et al., 2004]
  • 68. CBMIR vs. text-based MIR •  Most current retrieval systems in clinical use rely on text keywords such as DICOM header information to perform retrieval. •  CBIR has been widely researched in a variety of domains and provides an intuitive and expressive method for querying visual data using features, e.g. color, shape, and texture. •  However, current CBIR systems: –  are not easily integrated into the healthcare environment; –  have not been widely evaluated using a large dataset; and –  lack the ability to perform relevance feedback to refine retrieval results. Source: [Hsu et al., 2009]
  • 69. Who are the main players? •  USA – NIH (National Institutes of Health) •  NIBIB - National Institute of Biomedical Imaging and Bioengineering •  NCI - National Cancer Institute •  NLM – National Libraries of Medicine – Several universities and hospitals •  Europe – Aachen University (Germany) – Geneva University (Switzerland) •  Big companies (Siemens, GE, etc.)
  • 70. Medical image retrieval systems: examples •  IRMA (Image Retrieval in Medical Applications) –  Aachen University (Germany) •  http://ganymed.imib.rwth-aachen.de/irma/ –  3 online demos: •  IRMA Query demo: allows the evaluation of CBIR on several databases. •  IRMA Extended Query Refinement demo: CBIR from the IRMA database (a subset of 10,000 images). •  Spine Pathology and Image Retrieval Systems (SPIRS) designed by the NLM/NIH (USA): holds information of ~17,000 spine x-rays.
  • 71. Medical image retrieval systems: examples •  MedGIFT (GNU Image Finding Tool) – Geneva University (Switzerland) •  http://www.sim.hcuge.ch/medgift/ – Large effort, including projects such as: •  Talisman (lung image retrieval) •  Case-based fracture image retrieval system •  Onco-Media: medical image retrieval + grid computing •  ImageCLEF: evaluation and validation •  medSearch
  • 72. Medical image retrieval systems: examples •  WebMIRS – NIH / NLM (USA) •  http://archive.nlm.nih.gov/proj/webmirs/index.php – Query by text + navigation by categories – Uses datasets and related x-ray images from the National Health and Nutrition Examination Survey (NHANES)
  • 73. Medical image retrieval systems: examples •  SPIRS (Spine Pathology Image Retrieval System): Web-based image retrieval system for large biomedical databases – NIH / UCLA (USA) – Representative case study on highly specialized CBMIR Source: [Hsu et al., 2009]
  • 74. Medical image retrieval systems: examples •  National Biomedical Imaging Archive (NBIA) – NCI / NIH (USA) •  https://imaging.nci.nih.gov/ – Search based on metadata (DICOM fields) – 3 search options: •  Simple •  Advanced •  Dynamic
  • 75. Medical image retrieval systems: examples •  ARSS Goldminer – American Roentgen Ray Society (USA) •  http://goldminer.arrs.org/ – Query by text – Results can be filtered by: •  Modality •  Age •  Sex
  • 76. Medical image retrieval systems: examples •  Yottalook Images –  iVirtuoso (USA) •  http://www.yottalook.com/ –  Developed and maintained by four radiologists –  Query by text –  Claims to use 4 “core technologies”: •  natural query analysis” •  semantic ontology” •  “relevance algorithm” •  a specialized content delivery system that provides high yield content based on the search term.
  • 77. Evaluation: ImageCLEF Medical Image Retrieval •  ImageCLEF Medical Image Retrieval •  http://www.imageclef.org/2011/medical – Dataset: 77,000+ images from articles published in medical journals including text of the captions and link to the html of the full text articles. – 3 types of tasks: •  Modality Classification: given an image, return its modality •  Ad-hoc retrieval: classic medical retrieval task, with 3 “flavors”: textual, mixed and semantic queries •  Case-based retrieval: retrieve cases including images that might best suit the provided case description.
  • 78. Evaluation: ImageCLEF Medical Image Retrieval •  ImageCLEF Medical Image Retrieval 2011 – Modality Classification
  • 79. Evaluation: ImageCLEF Medical Image Retrieval •  ImageCLEF Medical Image Retrieval 2011 – Modality Classification – FAU Team •  Personnel: 4 grad students + 2 undergrads + advisor •  Strategy: –  Textual classification using Lucene and associated tools and libraries –  Visual classification using 8 contemporary descriptors and 3 different families of classifiers, implemented using Weka and associated tools and libraries •  Supporting tools: –  Manual annotation tool (http://imageclef.mlab.ceecs.fau.edu/) –  Training set visualization tool (http://imageclef.mlab.ceecs.fau.edu/classification/)
  • 80. Evaluation: ImageCLEF Medical Image Retrieval •  ImageCLEF Medical Image Retrieval 2011 – Modality Classification Results – FAU Team (textual)
  • 81. Evaluation: ImageCLEF Medical Image Retrieval •  ImageCLEF Medical Image Retrieval 2011 – Modality Classification Results – FAU Team (visual)
  • 82. Evaluation: ImageCLEF Medical Image Retrieval •  ImageCLEF Medical Image Retrieval 2011 •  Modality Classification Results – FAU Team (visual)
  • 83. Medical Image Retrieval: promising directions •  Better user interfaces (responsive, highly interactive, and capable of supporting relevance feedback) •  New applications of CBMIR, including: –  Teaching –  Research –  Diagnosis –  PACS and Electronic Patient Records •  CBMIR evaluation using medical experts •  Integration of local and global features •  New visual descriptors
  • 84. Medical Image Retrieval: promising directions •  New devices
  • 85. Part III Applications and related areas
  • 86. Applications and related areas •  New devices and services •  Mobile visual search •  Image search and retrieval in the age of social networks •  Games! •  Other related areas •  Our recent work (highlights)
  • 87. New devices and services •  Flickr (b. 2004) •  YouTube (b. 2005) •  Flip video cameras (b. 2006) •  iPhone (b. 2007) •  iPad (b. 2010)
  • 88. Mobile visual search •  Driving factors – Capable devices Source: http://www.apple.com/iphone/specs.html 1 GHz ARM Cortex-A8 processor, PowerVR SGX535GPU, Apple A4 chipset
  • 89. Mobile visual search •  Driving factors – Motivated users: image taking and image sharing are huge! –  Source: http://www.onlinemarketing-trends.com/2011/03/facebook-photo-statistics-and-insights.html
  • 90. Mobile visual search •  Facebook for iPhone –  Source: http://statistics.allfacebook.com/applications/single/facebook-for-iphone/6628568379/
  • 91. Mobile visual search •  Instagram: 2 million registered (although not necessarily active) users, who upload ~300,000 photos per day •  Several apps based on it! –  http://iphone.appstorm.net/roundups/photography/5-cool-apps-for-getting-the-most-out-of-instagram/
  • 92. Mobile visual search •  Food photo sharing!
  • 93. Mobile visual search •  Driving factors – Legitimate (or not quite…) needs and use cases –  Source: http://www.slideshare.net/dtunkelang/search-by-sight-google-goggles
  • 94. Mobile visual search •  Driving factors – Smart phone market
  • 95. Mobile visual search •  Smart phone market Source: http://www.cellular-news.com/story/48647.php?s=h
  • 96. Mobile visual search •  Examples of applications – Google Goggles – oMoby (and the IQ Engines API) – Others (kooaba, Fetch!, Gazopa, etc.)
  • 97. Mobile visual search •  Google Goggles – Android and iPhone – Narrow-domain search and retrieval
  • 98. Mobile visual search •  oMoby (and the IQ Engines API) – iPhone
  • 99. Mobile visual search •  oMoby (and the IQ Engines API)
  • 100. Image search and retrieval social networks •  The [so-called] Web 2.0 has brought about: –  New data sources –  New usage patterns –  New understanding about the users, their needs, habits, preferences –  New opportunities –  Lots of metadata! –  A chance to experience a true paradigm shift •  Before: image annotation is tedious, labor-intensive, expensive •  After: image annotation is fun!
  • 101. Games! ◦  Google Image Labeler ◦  Games with a purpose (GWAP)   The ESP Game   Squigl   Matchin
  • 102. Other related areas •  Semi-automatic image annotation •  Tag recommendation systems •  Story annotation engines •  Content-based image filtering •  Copyright detection •  Watermark detection – and many more
  • 103. Our recent work (highlights) •  PRISM – Image Genius •  Unsupervised ROI extraction from an image – Crazy Collage •  MEDIX and associated tools •  Callisto: a content-based tag recommendation tool
  • 105. PRISM With Liam Mayron, Harris Corp., USA
  • 106. Image Genius With Asif Rahman, FAU, USA
  • 107. Unsupervised ROI extraction With Gustavo B. Borba and Humberto R. Gamba, UTFPR, Brazil
  • 108. Crazy Collage Gustavo B. Borba et al., UTFPR, Brazil
  • 109. MEDIX •  Medical image retrieval system with DICOM capabilities With Asif Rahman, FAU, USA
  • 110. Callisto With Mathias Lux and Arthur Pitman, Klagenfurt University,Austria
  • 111. Part IV Where is image search headed?
  • 112. Where is image search headed? •  Advice for [young] researchers – In this last part, I’ve compiled pieces and bits of advice that I believe might help researchers who are entering the field. – They focus on research avenues that I personally consider to be the most promising.
  • 113. Advice for [young] researchers • LOOK • THINK • UNDERSTAND • CREATE
  • 114. Advice for [young] researchers • LOOK… – at yourself (how do you search for images and videos?) – around (related areas and how they have grown) – at Google (and other major players)
  • 115. Advice for [young] researchers • THINK… – mobile devices – new devices and services – social networks – games
  • 116. Advice for [young] researchers • UNDERSTAND… – human intentions and emotions – the context of the search – user’s preferences and needs
  • 117. Advice for [young] researchers • CREATE… – better interfaces – better user experience – new business opportunities (added value)
  • 118. Concluding thoughts –  I believe (but cannot prove…) that successfulVIR solutions will: •  combine content-based image retrieval (CBIR) with metadata (high-level semantic-based image retrieval) •  only be truly successful in narrow domains •  include the user in the loop – Relevance Feedback (RF) – Collaborative efforts (tagging, rating, annotating) •  provide friendly, intuitive interfaces •  incorporate results and insights from cognitive science, particularly human visual attention, perception, and memory
  • 119. Concluding thoughts –  I believe (but cannot prove…) that successfulVIR solutions will: •  combine content-based image retrieval (CBIR) with metadata (high-level semantic-based image retrieval) •  only be truly successful in narrow domains •  include the user in the loop – Relevance Feedback (RF) – Collaborative efforts (tagging, rating, annotating) •  provide friendly, intuitive interfaces •  incorporate results and insights from cognitive science, particularly human visual attention, perception, and memory
  • 120. Concluding thoughts –  I believe (but cannot prove…) that successfulVIR solutions will: •  combine content-based image retrieval (CBIR) with metadata (high-level semantic-based image retrieval) •  only be truly successful in narrow domains •  include the user in the loop – Relevance Feedback (RF) – Collaborative efforts (tagging, rating, annotating) •  provide friendly, intuitive interfaces •  incorporate results and insights from cognitive science, particularly human visual attention, perception, and memory
  • 121. Concluding thoughts –  I believe (but cannot prove…) that successfulVIR solutions will: •  combine content-based image retrieval (CBIR) with metadata (high-level semantic-based image retrieval) •  only be truly successful in narrow domains •  include the user in the loop – Relevance Feedback (RF) – Collaborative efforts (tagging, rating, annotating) •  provide friendly, intuitive interfaces •  incorporate results and insights from cognitive science, particularly human visual attention, perception, and memory
  • 122. Concluding thoughts –  I believe (but cannot prove…) that successfulVIR solutions will: •  combine content-based image retrieval (CBIR) with metadata (high-level semantic-based image retrieval) •  only be truly successful in narrow domains •  include the user in the loop – Relevance Feedback (RF) – Collaborative efforts (tagging, rating, annotating) •  provide friendly, intuitive interfaces •  incorporate results and insights from cognitive science, particularly human visual attention, perception, and memory
  • 123. Concluding thoughts •  “Image search and retrieval” is not a problem, but rather a collection of related problems that look like one. •  There is a great need for good solutions to specific problems. •  10 years after “the end of the early years”, research in visual information retrieval still has many open problems, challenges, and opportunities.
  • 124. Thanks! •  Questions? •  For additional information: omarques@fau.edu