SlideShare a Scribd company logo
1 of 47
Content-Based Retrieval
in Digital Libraries
Lecture Slides
By
Israa
Chapter 20
Content-Based Retrieval in Digital Libraries
2
1. How Should We Retrieve Images?
2. Synopsis of Early CBIR Systems
3. C-BIRD — A Case Study
4. Quantifying Search Results
5. Key Technologies in Current CBIR Systems
6. Querying on Videos
7. Querying on Videos Based on Human Activity
8. Quality-Aware Mobile Visual Search
20.1 H ow Should W e Retrieve Images?
• Text-based search will do the best job, provided the multi- media
database is fully indexed with proper keywords.
• Most multimedia retrieval schemes, however, have moved to- ward
an approach favoring multimedia content itself) content-based(
• Many existing systems retrieve images with the following
image features and/or their variants:
– Color histogram: 3-dimensional array that counts pixels with specific
Red, Green, and Blue values in an image.
– Color layout: a simple sketch of where in a checkerboard grid cov-
ering the image to look for blue skies or orange sunsets, say.
– Texture: various texture descriptors, typically based on edges in the
image.
3
Fig. 20.1 How can we best characterize the information content of an
image?
4
20.2 Synopsis of E a r l y C B I R Systems
The following provides examples of some early CBIR systems. It
is by no means a complete synopsis. Most of these engines are
experimental, but all those included here are interesting in
some way.
– QBIC (Query By Image Content(
– Chabot
– Blobworld
– WebSEEk
– Photobook and FourEyes
– Informedia
– UC Santa Barbara Search Engines
– MARS
– Virage
5
20.3 C - B I R D — A Case Study
• C-BIRD (Content-Base Image Retrieval from Digital libraries):
an image database search engine devised by one of the authors of
this text.
C-BIRD GUI: The C-BIRD image database contains
approximately 5,000 images, many of them key frames from
videos. The database can be searched using a selection of tools:
– Text annotations
– Color histogram
– Color layout
– Texture layout
– Illumination Invariance
– Object Model
6
Color Histogram
• A color histogram counts pixels with a given pixel value in
Red, Green, and Blue ( R G B ( .
• An example of histogram that has 2563 bins, for images with 8-
bit values in each of R, G, B :
int hist[256][256][256]; // reset to 0
//image is an appropriate struct with byte fields red, green, blue
7
for i=0..(MAX_Y-1)
for j=0..(MAX_X-1)
{
R = image[i][j].red;
G = image[i][j].green;
B = image[i][j].blue;
hist[R][G][B]++;
}
20.3.1
Color Histogram (Cont’d(
• Image search is done by matching feature-vector (here color
histogram) for the sample image with feature-vector for im-
ages in the database.
• In C-BIRD, a color histogram is calculated for each target
image as a preprocessing step, and then referenced in the
database for each user query image.
• For example, Fig. 20.2 shows that the user has selected a
particular image — one of a red flower on a green foliage
background.
The result obtained, from a database of some 5,000 images, is a
set of 60 matching images.
8
Fig. 20.2: Search by color histogram results.
9
Histogram Intersection
• Histogram intersection: The standard measure of similarity used for
color histograms:
– A color histogram Hi is generated for each image i in the database –
feature vector.
– The histogram is normalized so that its sum (now a double) equals
unity – effectively removes the size of the image.
– The histogram is then stored in the database.
– Now suppose we select a model image – the new image to match
against all possible targets in the database.
– Its histogram Hm is intersected with all database image histograms
Hi according to the equation
) 20.3)
j – histogram bin, n – total number of bins for each histogram
– The closer the intersection value is to 1, the better the images match.
10
Color D ensity and Color L ayout
• To specify the desired colors by their density, the user
selects the percentage of the image having any particular
color or set of colors is selected by the user, using a color
picker and sliders.
• User can choose from either conjunction (ANDing) or dis-
junction (ORing) a simple color percentage specification.
• This is a very coarse search method.
11
20.3.2
Color D ensity
Color L ayout
• The user can set up a scheme of how colors should appear in
the image, in terms of coarse blocks of color.
The user has a choice of four grid sizes: 1 × 1, 2 × 2, 4 × 4
and 8 × 8.
• Search is specified on one of the grid sizes, and the grid can
be filled with any R G B color value or no color value at all to
indicate the cell should not be considered.
• Every database image is partitioned into windows four times,
once for every window size.
– A clustered color histogram is used inside each window and
the five most frequent colors are stored in the database
– Position and size for each query cell correspond to the
position and size of a window in the image
• Fig. 20.3 shows how this layout scheme is used.
12
Fig. 20.3: Color layout grid.
13
T exture A nalysis Details
-1 0 1 1 2 1
dx : -2 0 2 dy : 0 0 0
-1 0 1 -1 -2 -1
1. Edge-based texture histogram
• A2-dimensional texture histogram is used based on edge directionality
φ, and separation ξ (closely related to repetitiveness (
• To extract an edge-map for the image, the image is first converted to
luminance Y via Y = 0.299R + 0.587tt + 0.114B.
• A Sobel edge operator is applied to the Y -image by sliding the fol- lowing 3
× 3 weighting matrices (convolution masks) over the image.
(20.4)
• The edge magnitude D and the edge gradient φ are given by
x yD =
.
d 2 + d2 , φ = arctan
dy
dx
(20.5)
14
20.3.4
Texture Analysis Details (Cont’d (
2. Preparation for creation of texture histogram
• The edges are thinned by suppressing all but maximum
values.
If a pixel i with edge gradient φi and edge magnitude
Di has a neighbor pixel j along the direction of φi with gradient φj
≈ φi and edge magnitude D j > Di then pixel i is suppressed to
0.
• To make a binary edge image, set all pixels with D greater than a
threshold value to 1 and all others to 0.
• For edge separation ξ, for each edge pixel i we measure the
distance along its gradient φi to the nearest pixel j having φj ≈
φi within 15◦.
• If such a pixel j doesn’t exist, then the separation is con- sidered
infinite.
15
Texture Analysis Details (Cont’d)
3. Having created edge directionality and edge separation maps, a 2D
texture histogram of ξ versus φ is constructed.
• The initial histogram size is 193 × 180, where separation value ξ
= 193 is reserved for a separation of infinity (as well as any ξ >
192).
• The histogram is “smoothed” by replacing each pixel with a
weighted sum of its neighbors, and then reduced to size 7 × 8,
separation value 7 reserved for infinity.
• Finally, the texture histogram is normalized by dividing by the
number of pixels in the image segment. It will then be used for
matching.
16
Search by Illumination Invariance
• To deal with illumination change from the query image to dif- ferent
database images, each color channel band of each im- age is first
normalized, and then compressed to a 36-vector.
• A 2-dimensional color histogram is then created by using the
chromaticity, which is the set of band ratios
{R, G }/ (R + G + B)
• To further reduce the number of vector components, the
D C T coefficients for the smaller histogram are calculated and
placed in zigzag order, and then all but 36 components dropped.
• Matching is performed in the compressed domain by taking the
Euclidean distance between two DCT-compressed 36- component
feature vectors.
• Fig. 20.5 shows the results of such a search.
17
20.3.5
Fig. 20.5: Search with illumination invariance.
18
Search by Object M odel
• This search type proceeds by the user selecting a thumbnail and
clicking the Model tab to enter object selection mode.
– An image region can be selected by using primitive shapes such
as a rectangle or an ellipse, a magic wand tool that is basically a
seed-based flooding algorithm, an active con- tour (a “snake”(,
or a brush tool where the painted region is selected.
– An object is then interactively selected as a portion of the image.
– Multiple regions can be dragged to the selection pane, but only
the active object in the selection pane will be searched on.
• A sample object selection is shown in Fig. 20.6.
19
20.3.6
Fig. 20.6 C-BIRD interface showing object selection using an ellipse
primitive.
20
Details of Search by Object M odel
1. The user-selected model image is processed and its features localized
(i.e., generate color locales [see below]).
chro-2. Color histogram intersection, based on the reduced
maticity histogram, is then applied as a first screen.
3. Estimate the pose (scale, translation, rotation) of the object inside a
target image from the database.
4. Verification by intersection of texture histograms, and then a final
check using an efficient version of a Generalized Hough Transform
for shape verification.
21
Fig. 20.7: Block diagram of object matching steps.
22
M o del Image and T arget Images
• A possible model image and one of the target images in the
database might be as in Fig. 20.8.
Fig. 20.8 : Model and target images. (a): Sample model image.
(b): Sample database image containing the model book.
23
20.4 Quantifying Search Results
• Precision is the percentage of relevant documents retrieved
compared to the number of all the documents retrieved.
P recision =
Relevant images returned
All retrieved images
• Recall is the percentage of relevant documents retrieved out of all
relevant documents.
Recall =
Relevant images returned
All relevant images
)20.6(
• These measures are affected by the database size and the amount
of similar information in the database, and as well they do not
consider fuzzy matching or search result ordering.
24
)20.6)
20.4 Quantifying Search Results (Cont’d)
25
Alternatively, they may also be written as:
where TP (True Positives) is the number of relevant images returned, FP
(False Positives) is the number of irrelevant images returned, and FN (False
Negatives) is the number of relevant images not returned.
20.5 Key Technologies in Current CBIR Systems
 The field of CBIR has witnessed rapid growth and progress in the new
millennium.
 Unlike in the early years, users are no longer satisfied with query results
returning scenes containing “blue skies,” or red blobs that may or may not be
flowers.
 They are more and more looking for objects, people, and often search for
human activities as well.
 Lew et al. and Datta et al. provide comprehensive surveys of the CBIR field.
In this section, we will briefly describe some technologies and issues that are key
to the success of current and future CBIR systems
26
Key Technologies in Current CBIR Systems
20.5.1. Robust Image Features and Their Representation
20.5.2. Relevance Feedback
20.5.3. Other Post-processing Techniques
20.5.4. Visual Concept Search
20.5.5. The Role of Users in Interactive CBIR Systems
20.5.1.Robust Image Features and Their Representation
Many feature descriptors have been developed beyond the ones specified in
MPEG-7. Here, we will only discuss SIFT and Visual Words.
SIFT (Scale Invariant Feature Transform)
In addition to global features such as color and texture histograms, local features
characterized by SIFT (Scale Invariant Feature Transform) have been developed
because they are more suitable for searches emphasizing visual objects. SIFT has
been shown to be robust against image noise; it also has a fair degree of invariance
to translation, rotation, and scale change.
The process of extracting SIFT features from an image starts with building a
multiresolution scale space, as shown in Fig. 20.11. At each scale, stacks of images
are generated by applying Gaussian smoothing (filtering) operations.
Visual words, as used in image retrieval systems, refer to small parts of an image
which carry some kind of information related to the features (such as the color,
shape or texture), or changes occurring in the pixels such as the filtering, low-level
feature descriptors (SIFT , SURF , ...etc.).
27
20.5.2.Relevance Feedback
• Relevance Feedback: involve the user in a loop, whereby
images retrieved are used in further rounds of convergence
onto correct returns.
• Relevance Feedback Approaches
– The usual situation: the user identifies images as good, bad, or
don’t care, and weighting systems are updated according to this
user guidance.
– Another approach is to move the query towards positively marked
content.
– An even more interesting idea is to move every data point in a
disciplined way, by warping the space of feature points.
28
20.5.3.Other Post-processing Techniques
Beside Relevance Feedback, other post-processing methods have been developed
after the initial query results are in.
Spatial Verification
The quality of the query results can be verified and improved by Spatial Verification.
The spatial verification consists in verify a spatial correlation between certain points
of a pair of images.
The main problem is that outliers (that does not fit or does not match the selected
model) affect adjustment called least squares
Modern cameras often provide information about the location where pictures were
taken, so it is trivial to check for consistency or relevance in certain query results if
the user is only interested in images from certain locations, e.g., Paris or Rome.
Advantages
•Effective when one is able to find safe features without clutter.
•Good results for correspondence in specific instances.
Disadvantages
•The scaling models.
•The spatial verification can not be used as post-processing.
29
Query Expansion
 Query Expansion is a well-known method in Information Retrieval in
which some combination of high-ranked relevant documents can be
presented as a new query to boost the performance of the retrieval. The
combination could simply be averages (means) of the feature
descriptors from the returned documents.
30
 The problem is that if one of the high-ranked document is a false-positive,
then it will immediately harm the performance of the new query. Some
robust statistical methods or even simply taking the median (instead of
the mean) can help.
Question-answering (QA) tries to replace the large number of images or
multimedia content in general, that is returned by a search query, by
making use of media content, knowledge about the search domain, and
also linguistic analysis to return hits based on users’ natural-language
questions.
QA Paradigm
20.5.4.Visual Concept Search
31
• Search by concept is an automated information retrieval method and it is
another major step in closing the semantic gap. Typically, after local features
are turned into words, the words are converted to semantic concepts with the
aid of machine learning algorithms.
• In an interesting work presented by Wang et al. They show that such a
system performs better than others that directly measure image similarity with
simple low-level visual features such as color, texture, etc.
• To assist in the development and testing of these systems, researchers and
practitioners have developed many multimedia databases and benchmarks.
The best known is perhaps theTRECVideo Retrieval Evaluation (TRECVID)
benchmark. TRECVID started in 2003.
• Initially, TRECVID provided video data from professional sources such as
broadcast news, TV programs, and surveillance systems, thus having limited
styles and content:
For example, TRECVID 2013 evaluated the following tasks:
• Semantic indexing
• Interactive surveillance event detection
• Instance search
• Multimedia event detection
• Multimedia event recounting
20.5.5 The Role of Users in Interactive CBIR Systems
 Beside mechanisms such as Relevance Feedback, it has been argued that
users should play a more active role in CBIR systems.
 The authors again raised the question: Where is the User in Multimedia
Retrieval?
 They pointed out the dominance of MAP, common in evaluations such as
TRECVID, may have hindered the development of better and more useful
multimedia retrieval systems.
 One way to comprehend how people view images as similar is to study using
user groups just what forms our basis of perception of image similarity.
 The function used in this approach can be a type of “perceptual similarity
measure” and is learned by finding the best set of features (color, texture, etc.)
to capture “similarity”
as defined via the groups of similar images identified.
 Another way to understand users is to talk to them and carefully analyze their
search patterns.
32
20.6 Querying on V ideos
• Video indexing can make use of motion as the salient fea- ture
of temporally changing images for various types of query.
• Inverse Hollywood: can we recover thevideo director’s “flowchart”?
– Dividing the video into shots, where each shot consists roughly
of the video frames between the on and off clicks of the record
button.
– Detection of shot boundaries is usually not simple as fade-in,
fade- out, dissolve, wipe, etc. may often be involved.
33
• In dealing with digital video, it is desirable to avoid uncom-
pressing M P E G files.
– A simple approach to this idea is to uncompress only enough to re-
cover just the DC term, thus generating a thumbnail that is 64 times
as small as the original.
– Once DC frames are obtained from the whole video, many different
approaches have been used for finding shot boundaries – based on
features such as color, texture, and motion vectors.
34
An Example of Querying on V ideo
• Fig) 20.12: (a shows a selection of frames from a video of beach
activity.
Here the keyframes in( Fig. 20.12 (b. are selected based mainly on
color information (but being careful with respect to the changes
incurred by changing illumination conditions when videos are shot (
.
• A more difficult problem arises when changes between shots are
gradual, and when colors are rather similar overall, as in Fig)
20.13(.a
The keyframes in Fig) 20.13(b. are sufficient to show the
development of the whole video sequence.
35
(a)
(a)
Fig. 20.12: Digital video and associated keyframes, beach video. (a):
Frames from a digital video. (b): Keyframes selected.
36
(a)
)b)
(a): F ramesfrom a digital video.Fig. 20.13: Garden video.
(b): Keyframes selected.
37
20.7 Querying on Videos Based on Human Activity
 Thousands of hours of video are being captured every day by CCTV
cameras, web cameras, broadcast cameras, etc. However, most of the
activities of interest (e.g., a soccer player scores a goal) only occur in a
relatively small region along the spatial and temporal extent of the video. In
this scenario, effective retrieval of a small spatial/temporal segment of the
video containing a particular activity from a large collection of videos is very
important.
38
 Lan et al.’s work on activity retrieval is directly inspired by the application
of searching for activities of interest from broadcast sports videos. For example,
consider the scene shown in Fig. 20.14.
we can explain the scene at multiple levels of detail such as low-level actions
(e.g., standing and jogging) and mid-level social roles (e.g., attacker and first
defenders).
The social roles are denoted by different colors. In this example, we use
magenta, blue and white to represent attacker, man-marking and players in the
same team as the attacker respectively.
39
Fig.20.14 An example of human activity in realistic scenes . Beyond
the general scene-level activity description (e.g. Free Hit).
20.7.1 Modeling Human Activity Structures
Here we introduce the model . To illustrate, we describe an instantiation
applicable to modeling field hockey videos.
We first describe the labeling. Assume an image has been pre-processed, so
the location of the goal and persons in the image have been found. We separate
the players into two teams according to their color histograms. Each person is
associated with two labels: action and social role.
40
The model is hierarchical, and includes various levels of detail: low-level
actions, mid-level social roles, and high-level events. The relationships and
interactions between these are included in the model. We define the score of
interpreting a video frame with the hierarchical event representation as::
41
Fig. 20.15 Graphical illustration of the model.
Different types of potentials are denoted by lines with different colors.
Details of the potential functions are
contained in Eq. 20.10
20.7.2 Experimental Results
42
 The challenging Broadcast Field Hockey Dataset that consists
of sporting activities captured from broadcast cameras is used
to demonstrate the efficacy of the model. The model can carry
out different inferences based on a user’s queries. The method
is directly applicable to multilevel human activity recognition.
 The goal is to predict events, social roles or actions of each
person. In this case, the query doesn’t specify a particular event
or social roles, but consists of more general questions, such as
what is the overall game situation, or what are the social roles of
each player.
20.8 Quality-Aware Mobile Visual Search
43
 With the increasing popularity of mobile phones and tablets,
Mobile Visual Search has attracted growing interest in the field of
content-based image retrieval (CBIR).
 On the mobile-client side, a query image is compressed to
a certain quality level to accommodate the network
conditions and then uploaded onto a server with its quality
level transferred as side information.
 On the server side, a set of features are extracted from the
query image and then compared against the features of the
images in the database.
 Mobile Visual Search enables people to look for visually
similar products or find information about movies or CDs online
by initiating a search request from a camera phone.
Quality-Aware Mobile Visual Search … (Cont’d)
44
1. A query image is transmitted to the server, and then feature
extraction and retrieval are done on the server.
2. The mobile client extracts some features from the query image
and uploads only the features to the server. The retrieval is
performed on the server.
3. The mobile client maintains a cache of the image database.
The retrieval is done locally on the client. Only if no match is
found does the client sends a query to the server.
 Depending on which part of the visual search is performed
on the mobile devices, there are several possible client-server
architectures :
Quality-Aware Method
45
We outline a query quality-dependent fusion approach for mobile visual search.
Specifically, it is based on the fusion of five common image features :
• Tiny Images: This is the most trivial descriptor that compares images directly in
color space after reducing the image dimensions drastically.
• Color Histograms: The histograms have 16 bins for each component of RGB
color space, yielding a total of 48 dimensions.
• GIST: This descriptor computes the output of a bank of 24 Gabor-like filters tuned
to eight orientations at four scales. The squared output of each filter is then
averaged on a 4 × 4 grid.
• Texton Histograms: For each image build a 512-dimensional histogram using a
512-entry universal texton dictionary.
• SSIM : The self-similarity descriptors are quantized into 300 visual words by
k-means.
Quality-Aware Method
46
Fig. 20.18 Map results of different quality levels
Thank yo
u
for liste
ning

More Related Content

What's hot

Multimedia data and file format
Multimedia data and file formatMultimedia data and file format
Multimedia data and file formatNiketa Jain
 
Color Image Processing
Color Image ProcessingColor Image Processing
Color Image Processingkiruthiammu
 
Presentation of Lossy compression
Presentation of Lossy compressionPresentation of Lossy compression
Presentation of Lossy compressionOmar Ghazi
 
Chapter 3 - Multimedia System Design
Chapter 3 - Multimedia System DesignChapter 3 - Multimedia System Design
Chapter 3 - Multimedia System DesignPratik Pradhan
 
Noise recognition in digital image
Noise recognition in digital imageNoise recognition in digital image
Noise recognition in digital imageReyad Hossain
 
Introduction to Image Compression
Introduction to Image CompressionIntroduction to Image Compression
Introduction to Image CompressionKalyan Acharjya
 
World wide web with multimedia
World wide web with multimediaWorld wide web with multimedia
World wide web with multimediaAfaq Siddiqui
 
Fundamentals and image compression models
Fundamentals and image compression modelsFundamentals and image compression models
Fundamentals and image compression modelslavanya marichamy
 
Basic Introduction To Graphic File Formats
Basic Introduction To Graphic File FormatsBasic Introduction To Graphic File Formats
Basic Introduction To Graphic File FormatsAnkit Mishra
 
Interpixel redundancy
Interpixel redundancyInterpixel redundancy
Interpixel redundancyNaveen Kumar
 
Chapter 8 image compression
Chapter 8 image compressionChapter 8 image compression
Chapter 8 image compressionasodariyabhavesh
 
Lecture5 graphics
Lecture5   graphicsLecture5   graphics
Lecture5 graphicsMr SMAK
 
Graphic hardware and software
Graphic hardware and softwareGraphic hardware and software
Graphic hardware and softwarerafhat
 
Multimedia synchronization
Multimedia synchronizationMultimedia synchronization
Multimedia synchronizationI World Tech
 
1. Introduction of Computer Graphics
1. Introduction of Computer Graphics1. Introduction of Computer Graphics
1. Introduction of Computer GraphicsAparna Joshi
 

What's hot (20)

Multimedia data and file format
Multimedia data and file formatMultimedia data and file format
Multimedia data and file format
 
Color Image Processing
Color Image ProcessingColor Image Processing
Color Image Processing
 
Presentation of Lossy compression
Presentation of Lossy compressionPresentation of Lossy compression
Presentation of Lossy compression
 
Chapter 3 - Multimedia System Design
Chapter 3 - Multimedia System DesignChapter 3 - Multimedia System Design
Chapter 3 - Multimedia System Design
 
Noise recognition in digital image
Noise recognition in digital imageNoise recognition in digital image
Noise recognition in digital image
 
Introduction to Image Compression
Introduction to Image CompressionIntroduction to Image Compression
Introduction to Image Compression
 
Rgb and cmy color model
Rgb and cmy color modelRgb and cmy color model
Rgb and cmy color model
 
Image restoration and reconstruction
Image restoration and reconstructionImage restoration and reconstruction
Image restoration and reconstruction
 
World wide web with multimedia
World wide web with multimediaWorld wide web with multimedia
World wide web with multimedia
 
Fundamentals and image compression models
Fundamentals and image compression modelsFundamentals and image compression models
Fundamentals and image compression models
 
multimedia image.ppt
multimedia image.pptmultimedia image.ppt
multimedia image.ppt
 
Basic Introduction To Graphic File Formats
Basic Introduction To Graphic File FormatsBasic Introduction To Graphic File Formats
Basic Introduction To Graphic File Formats
 
Multimedia compression
Multimedia compressionMultimedia compression
Multimedia compression
 
Interpixel redundancy
Interpixel redundancyInterpixel redundancy
Interpixel redundancy
 
Chapter 8 image compression
Chapter 8 image compressionChapter 8 image compression
Chapter 8 image compression
 
Multimedia chapter 5
Multimedia chapter 5Multimedia chapter 5
Multimedia chapter 5
 
Lecture5 graphics
Lecture5   graphicsLecture5   graphics
Lecture5 graphics
 
Graphic hardware and software
Graphic hardware and softwareGraphic hardware and software
Graphic hardware and software
 
Multimedia synchronization
Multimedia synchronizationMultimedia synchronization
Multimedia synchronization
 
1. Introduction of Computer Graphics
1. Introduction of Computer Graphics1. Introduction of Computer Graphics
1. Introduction of Computer Graphics
 

Viewers also liked

Content based image retrieval
Content based image retrievalContent based image retrieval
Content based image retrievalrubaiyat11
 
Need for Software Engineering
Need for Software EngineeringNeed for Software Engineering
Need for Software EngineeringUpekha Vandebona
 
Pattern recognition voice biometrics
Pattern recognition voice biometricsPattern recognition voice biometrics
Pattern recognition voice biometricsMazin Alwaaly
 
CBIR For Medical Imaging...
CBIR For Medical  Imaging...CBIR For Medical  Imaging...
CBIR For Medical Imaging...Isha Sharma
 
Literature Review on Content Based Image Retrieval
Literature Review on Content Based Image RetrievalLiterature Review on Content Based Image Retrieval
Literature Review on Content Based Image RetrievalUpekha Vandebona
 
Content Based Image and Video Retrieval Algorithm
Content Based Image and Video Retrieval AlgorithmContent Based Image and Video Retrieval Algorithm
Content Based Image and Video Retrieval AlgorithmAkshit Bum
 
Content Based Image Retrieval
Content Based Image RetrievalContent Based Image Retrieval
Content Based Image RetrievalAman Patel
 
Content Based Image Retrieval
Content Based Image RetrievalContent Based Image Retrieval
Content Based Image RetrievalPrem kumar
 
Content based image retrieval (cbir) using
Content based image retrieval (cbir) usingContent based image retrieval (cbir) using
Content based image retrieval (cbir) usingijcsity
 
Cbir final ppt
Cbir final pptCbir final ppt
Cbir final pptrinki nag
 
Content Based Image Retrieval
Content Based Image RetrievalContent Based Image Retrieval
Content Based Image RetrievalSOURAV KAR
 
Content based image retrieval using clustering Algorithm(CBIR)
Content based image retrieval using clustering Algorithm(CBIR)Content based image retrieval using clustering Algorithm(CBIR)
Content based image retrieval using clustering Algorithm(CBIR)Raja Sekar
 
Content Based Image Retrieval
Content Based Image Retrieval Content Based Image Retrieval
Content Based Image Retrieval Swati Chauhan
 
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)Universitat Politècnica de Catalunya
 

Viewers also liked (20)

Content based image retrieval
Content based image retrievalContent based image retrieval
Content based image retrieval
 
Need for Software Engineering
Need for Software EngineeringNeed for Software Engineering
Need for Software Engineering
 
Pattern recognition voice biometrics
Pattern recognition voice biometricsPattern recognition voice biometrics
Pattern recognition voice biometrics
 
CBIR For Medical Imaging...
CBIR For Medical  Imaging...CBIR For Medical  Imaging...
CBIR For Medical Imaging...
 
Literature Review on Content Based Image Retrieval
Literature Review on Content Based Image RetrievalLiterature Review on Content Based Image Retrieval
Literature Review on Content Based Image Retrieval
 
Content Based Image and Video Retrieval Algorithm
Content Based Image and Video Retrieval AlgorithmContent Based Image and Video Retrieval Algorithm
Content Based Image and Video Retrieval Algorithm
 
Ga
GaGa
Ga
 
Content Based Image Retrieval
Content Based Image RetrievalContent Based Image Retrieval
Content Based Image Retrieval
 
Content Based Image Retrieval
Content Based Image RetrievalContent Based Image Retrieval
Content Based Image Retrieval
 
Content-based Image Retrieval
Content-based Image RetrievalContent-based Image Retrieval
Content-based Image Retrieval
 
Image retrieval
Image retrievalImage retrieval
Image retrieval
 
Cbir ‐ features
Cbir ‐ featuresCbir ‐ features
Cbir ‐ features
 
Content based image retrieval (cbir) using
Content based image retrieval (cbir) usingContent based image retrieval (cbir) using
Content based image retrieval (cbir) using
 
Cbir final ppt
Cbir final pptCbir final ppt
Cbir final ppt
 
CBIR
CBIRCBIR
CBIR
 
Content Based Image Retrieval
Content Based Image RetrievalContent Based Image Retrieval
Content Based Image Retrieval
 
Content based image retrieval using clustering Algorithm(CBIR)
Content based image retrieval using clustering Algorithm(CBIR)Content based image retrieval using clustering Algorithm(CBIR)
Content based image retrieval using clustering Algorithm(CBIR)
 
CBIR
CBIRCBIR
CBIR
 
Content Based Image Retrieval
Content Based Image Retrieval Content Based Image Retrieval
Content Based Image Retrieval
 
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
Content-Based Image Retrieval (D2L6 Insight@DCU Machine Learning Workshop 2017)
 

Similar to Multimedia content based retrieval in digital libraries

Similar to Multimedia content based retrieval in digital libraries (20)

Cj36511514
Cj36511514Cj36511514
Cj36511514
 
regions
regionsregions
regions
 
IJET-V2I6P17
IJET-V2I6P17IJET-V2I6P17
IJET-V2I6P17
 
Multimedia graphics and image data representation
Multimedia graphics and image data representationMultimedia graphics and image data representation
Multimedia graphics and image data representation
 
ImSeg04.ppt
ImSeg04.pptImSeg04.ppt
ImSeg04.ppt
 
Im seg04
Im seg04Im seg04
Im seg04
 
ImSeg04 (2).ppt
ImSeg04 (2).pptImSeg04 (2).ppt
ImSeg04 (2).ppt
 
ModuleII.ppt
ModuleII.pptModuleII.ppt
ModuleII.ppt
 
ModuleII.ppt
ModuleII.pptModuleII.ppt
ModuleII.ppt
 
ModuleII.ppt
ModuleII.pptModuleII.ppt
ModuleII.ppt
 
B01460713
B01460713B01460713
B01460713
 
Research Inventy : International Journal of Engineering and Science
Research Inventy : International Journal of Engineering and ScienceResearch Inventy : International Journal of Engineering and Science
Research Inventy : International Journal of Engineering and Science
 
Digital Image Processing
Digital Image ProcessingDigital Image Processing
Digital Image Processing
 
Week06 bme429-cbir
Week06 bme429-cbirWeek06 bme429-cbir
Week06 bme429-cbir
 
I07015261
I07015261I07015261
I07015261
 
Presentation1.pptx
Presentation1.pptxPresentation1.pptx
Presentation1.pptx
 
leaf diseses.pptx
leaf diseses.pptxleaf diseses.pptx
leaf diseses.pptx
 
B0310408
B0310408B0310408
B0310408
 
Noise tolerant color image segmentation using support vector machine
Noise tolerant color image segmentation using support vector machineNoise tolerant color image segmentation using support vector machine
Noise tolerant color image segmentation using support vector machine
 
Quality and size assessment of quantized images using K-Means++ clustering
Quality and size assessment of quantized images using K-Means++ clusteringQuality and size assessment of quantized images using K-Means++ clustering
Quality and size assessment of quantized images using K-Means++ clustering
 

More from Mazin Alwaaly

Pattern recognition palm print authentication system
Pattern recognition palm print authentication systemPattern recognition palm print authentication system
Pattern recognition palm print authentication systemMazin Alwaaly
 
Pattern recognition on line signature
Pattern recognition on line signaturePattern recognition on line signature
Pattern recognition on line signatureMazin Alwaaly
 
Pattern recognition multi biometrics using face and ear
Pattern recognition multi biometrics using face and earPattern recognition multi biometrics using face and ear
Pattern recognition multi biometrics using face and earMazin Alwaaly
 
Pattern recognition IRIS recognition
Pattern recognition IRIS recognitionPattern recognition IRIS recognition
Pattern recognition IRIS recognitionMazin Alwaaly
 
Pattern recognition hand vascular pattern recognition
Pattern recognition hand vascular pattern recognitionPattern recognition hand vascular pattern recognition
Pattern recognition hand vascular pattern recognitionMazin Alwaaly
 
Pattern recognition Hand Geometry
Pattern recognition Hand GeometryPattern recognition Hand Geometry
Pattern recognition Hand GeometryMazin Alwaaly
 
Pattern recognition forensic dental identification
Pattern recognition forensic dental identificationPattern recognition forensic dental identification
Pattern recognition forensic dental identificationMazin Alwaaly
 
Pattern recognition fingerprints
Pattern recognition fingerprintsPattern recognition fingerprints
Pattern recognition fingerprintsMazin Alwaaly
 
Pattern recognition facial recognition
Pattern recognition facial recognitionPattern recognition facial recognition
Pattern recognition facial recognitionMazin Alwaaly
 
Pattern recognition ear as a biometric
Pattern recognition ear as a biometricPattern recognition ear as a biometric
Pattern recognition ear as a biometricMazin Alwaaly
 
Pattern recognition 3d face recognition
Pattern recognition 3d face recognitionPattern recognition 3d face recognition
Pattern recognition 3d face recognitionMazin Alwaaly
 
Multimedia multimedia over wireless and mobile networks
Multimedia multimedia over wireless and mobile networksMultimedia multimedia over wireless and mobile networks
Multimedia multimedia over wireless and mobile networksMazin Alwaaly
 
Multimedia network services and protocols for multimedia communications
Multimedia network services and protocols for multimedia communicationsMultimedia network services and protocols for multimedia communications
Multimedia network services and protocols for multimedia communicationsMazin Alwaaly
 
Multimedia lossy compression algorithms
Multimedia lossy compression algorithmsMultimedia lossy compression algorithms
Multimedia lossy compression algorithmsMazin Alwaaly
 
Multimedia lossless compression algorithms
Multimedia lossless compression algorithmsMultimedia lossless compression algorithms
Multimedia lossless compression algorithmsMazin Alwaaly
 
Multimedia basic video compression techniques
Multimedia basic video compression techniquesMultimedia basic video compression techniques
Multimedia basic video compression techniquesMazin Alwaaly
 
Multimedia image compression standards
Multimedia image compression standardsMultimedia image compression standards
Multimedia image compression standardsMazin Alwaaly
 
Multimedia fundamental concepts in video
Multimedia fundamental concepts in videoMultimedia fundamental concepts in video
Multimedia fundamental concepts in videoMazin Alwaaly
 
Information and data security steganography and watermarking
Information and data security steganography and watermarkingInformation and data security steganography and watermarking
Information and data security steganography and watermarkingMazin Alwaaly
 
Information and data security cryptanalysis method
Information and data security cryptanalysis methodInformation and data security cryptanalysis method
Information and data security cryptanalysis methodMazin Alwaaly
 

More from Mazin Alwaaly (20)

Pattern recognition palm print authentication system
Pattern recognition palm print authentication systemPattern recognition palm print authentication system
Pattern recognition palm print authentication system
 
Pattern recognition on line signature
Pattern recognition on line signaturePattern recognition on line signature
Pattern recognition on line signature
 
Pattern recognition multi biometrics using face and ear
Pattern recognition multi biometrics using face and earPattern recognition multi biometrics using face and ear
Pattern recognition multi biometrics using face and ear
 
Pattern recognition IRIS recognition
Pattern recognition IRIS recognitionPattern recognition IRIS recognition
Pattern recognition IRIS recognition
 
Pattern recognition hand vascular pattern recognition
Pattern recognition hand vascular pattern recognitionPattern recognition hand vascular pattern recognition
Pattern recognition hand vascular pattern recognition
 
Pattern recognition Hand Geometry
Pattern recognition Hand GeometryPattern recognition Hand Geometry
Pattern recognition Hand Geometry
 
Pattern recognition forensic dental identification
Pattern recognition forensic dental identificationPattern recognition forensic dental identification
Pattern recognition forensic dental identification
 
Pattern recognition fingerprints
Pattern recognition fingerprintsPattern recognition fingerprints
Pattern recognition fingerprints
 
Pattern recognition facial recognition
Pattern recognition facial recognitionPattern recognition facial recognition
Pattern recognition facial recognition
 
Pattern recognition ear as a biometric
Pattern recognition ear as a biometricPattern recognition ear as a biometric
Pattern recognition ear as a biometric
 
Pattern recognition 3d face recognition
Pattern recognition 3d face recognitionPattern recognition 3d face recognition
Pattern recognition 3d face recognition
 
Multimedia multimedia over wireless and mobile networks
Multimedia multimedia over wireless and mobile networksMultimedia multimedia over wireless and mobile networks
Multimedia multimedia over wireless and mobile networks
 
Multimedia network services and protocols for multimedia communications
Multimedia network services and protocols for multimedia communicationsMultimedia network services and protocols for multimedia communications
Multimedia network services and protocols for multimedia communications
 
Multimedia lossy compression algorithms
Multimedia lossy compression algorithmsMultimedia lossy compression algorithms
Multimedia lossy compression algorithms
 
Multimedia lossless compression algorithms
Multimedia lossless compression algorithmsMultimedia lossless compression algorithms
Multimedia lossless compression algorithms
 
Multimedia basic video compression techniques
Multimedia basic video compression techniquesMultimedia basic video compression techniques
Multimedia basic video compression techniques
 
Multimedia image compression standards
Multimedia image compression standardsMultimedia image compression standards
Multimedia image compression standards
 
Multimedia fundamental concepts in video
Multimedia fundamental concepts in videoMultimedia fundamental concepts in video
Multimedia fundamental concepts in video
 
Information and data security steganography and watermarking
Information and data security steganography and watermarkingInformation and data security steganography and watermarking
Information and data security steganography and watermarking
 
Information and data security cryptanalysis method
Information and data security cryptanalysis methodInformation and data security cryptanalysis method
Information and data security cryptanalysis method
 

Recently uploaded

Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfSumit Kumar yadav
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxRizalinePalanog2
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLkantirani197
 
Forensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfForensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfrohankumarsinghrore1
 
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...Lokesh Kothari
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICESAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICEayushi9330
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticssakshisoni2385
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Sérgio Sacani
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
American Type Culture Collection (ATCC).pptx
American Type Culture Collection (ATCC).pptxAmerican Type Culture Collection (ATCC).pptx
American Type Culture Collection (ATCC).pptxabhishekdhamu51
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)Areesha Ahmad
 
Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Silpa
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptxAlMamun560346
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...chandars293
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Servicenishacall1
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformationAreesha Ahmad
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .Poonam Aher Patil
 

Recently uploaded (20)

Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdf
 
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptxSCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
SCIENCE-4-QUARTER4-WEEK-4-PPT-1 (1).pptx
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
 
Forensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfForensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdf
 
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICESAMASTIPUR CALL GIRL 7857803690  LOW PRICE  ESCORT SERVICE
SAMASTIPUR CALL GIRL 7857803690 LOW PRICE ESCORT SERVICE
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
American Type Culture Collection (ATCC).pptx
American Type Culture Collection (ATCC).pptxAmerican Type Culture Collection (ATCC).pptx
American Type Culture Collection (ATCC).pptx
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)
 
Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptx
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
 
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
9999266834 Call Girls In Noida Sector 22 (Delhi) Call Girl Service
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformation
 
Clean In Place(CIP).pptx .
Clean In Place(CIP).pptx                 .Clean In Place(CIP).pptx                 .
Clean In Place(CIP).pptx .
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .
 

Multimedia content based retrieval in digital libraries

  • 1. Content-Based Retrieval in Digital Libraries Lecture Slides By Israa
  • 2. Chapter 20 Content-Based Retrieval in Digital Libraries 2 1. How Should We Retrieve Images? 2. Synopsis of Early CBIR Systems 3. C-BIRD — A Case Study 4. Quantifying Search Results 5. Key Technologies in Current CBIR Systems 6. Querying on Videos 7. Querying on Videos Based on Human Activity 8. Quality-Aware Mobile Visual Search
  • 3. 20.1 H ow Should W e Retrieve Images? • Text-based search will do the best job, provided the multi- media database is fully indexed with proper keywords. • Most multimedia retrieval schemes, however, have moved to- ward an approach favoring multimedia content itself) content-based( • Many existing systems retrieve images with the following image features and/or their variants: – Color histogram: 3-dimensional array that counts pixels with specific Red, Green, and Blue values in an image. – Color layout: a simple sketch of where in a checkerboard grid cov- ering the image to look for blue skies or orange sunsets, say. – Texture: various texture descriptors, typically based on edges in the image. 3
  • 4. Fig. 20.1 How can we best characterize the information content of an image? 4
  • 5. 20.2 Synopsis of E a r l y C B I R Systems The following provides examples of some early CBIR systems. It is by no means a complete synopsis. Most of these engines are experimental, but all those included here are interesting in some way. – QBIC (Query By Image Content( – Chabot – Blobworld – WebSEEk – Photobook and FourEyes – Informedia – UC Santa Barbara Search Engines – MARS – Virage 5
  • 6. 20.3 C - B I R D — A Case Study • C-BIRD (Content-Base Image Retrieval from Digital libraries): an image database search engine devised by one of the authors of this text. C-BIRD GUI: The C-BIRD image database contains approximately 5,000 images, many of them key frames from videos. The database can be searched using a selection of tools: – Text annotations – Color histogram – Color layout – Texture layout – Illumination Invariance – Object Model 6
  • 7. Color Histogram • A color histogram counts pixels with a given pixel value in Red, Green, and Blue ( R G B ( . • An example of histogram that has 2563 bins, for images with 8- bit values in each of R, G, B : int hist[256][256][256]; // reset to 0 //image is an appropriate struct with byte fields red, green, blue 7 for i=0..(MAX_Y-1) for j=0..(MAX_X-1) { R = image[i][j].red; G = image[i][j].green; B = image[i][j].blue; hist[R][G][B]++; } 20.3.1
  • 8. Color Histogram (Cont’d( • Image search is done by matching feature-vector (here color histogram) for the sample image with feature-vector for im- ages in the database. • In C-BIRD, a color histogram is calculated for each target image as a preprocessing step, and then referenced in the database for each user query image. • For example, Fig. 20.2 shows that the user has selected a particular image — one of a red flower on a green foliage background. The result obtained, from a database of some 5,000 images, is a set of 60 matching images. 8
  • 9. Fig. 20.2: Search by color histogram results. 9
  • 10. Histogram Intersection • Histogram intersection: The standard measure of similarity used for color histograms: – A color histogram Hi is generated for each image i in the database – feature vector. – The histogram is normalized so that its sum (now a double) equals unity – effectively removes the size of the image. – The histogram is then stored in the database. – Now suppose we select a model image – the new image to match against all possible targets in the database. – Its histogram Hm is intersected with all database image histograms Hi according to the equation ) 20.3) j – histogram bin, n – total number of bins for each histogram – The closer the intersection value is to 1, the better the images match. 10
  • 11. Color D ensity and Color L ayout • To specify the desired colors by their density, the user selects the percentage of the image having any particular color or set of colors is selected by the user, using a color picker and sliders. • User can choose from either conjunction (ANDing) or dis- junction (ORing) a simple color percentage specification. • This is a very coarse search method. 11 20.3.2 Color D ensity
  • 12. Color L ayout • The user can set up a scheme of how colors should appear in the image, in terms of coarse blocks of color. The user has a choice of four grid sizes: 1 × 1, 2 × 2, 4 × 4 and 8 × 8. • Search is specified on one of the grid sizes, and the grid can be filled with any R G B color value or no color value at all to indicate the cell should not be considered. • Every database image is partitioned into windows four times, once for every window size. – A clustered color histogram is used inside each window and the five most frequent colors are stored in the database – Position and size for each query cell correspond to the position and size of a window in the image • Fig. 20.3 shows how this layout scheme is used. 12
  • 13. Fig. 20.3: Color layout grid. 13
  • 14. T exture A nalysis Details -1 0 1 1 2 1 dx : -2 0 2 dy : 0 0 0 -1 0 1 -1 -2 -1 1. Edge-based texture histogram • A2-dimensional texture histogram is used based on edge directionality φ, and separation ξ (closely related to repetitiveness ( • To extract an edge-map for the image, the image is first converted to luminance Y via Y = 0.299R + 0.587tt + 0.114B. • A Sobel edge operator is applied to the Y -image by sliding the fol- lowing 3 × 3 weighting matrices (convolution masks) over the image. (20.4) • The edge magnitude D and the edge gradient φ are given by x yD = . d 2 + d2 , φ = arctan dy dx (20.5) 14 20.3.4
  • 15. Texture Analysis Details (Cont’d ( 2. Preparation for creation of texture histogram • The edges are thinned by suppressing all but maximum values. If a pixel i with edge gradient φi and edge magnitude Di has a neighbor pixel j along the direction of φi with gradient φj ≈ φi and edge magnitude D j > Di then pixel i is suppressed to 0. • To make a binary edge image, set all pixels with D greater than a threshold value to 1 and all others to 0. • For edge separation ξ, for each edge pixel i we measure the distance along its gradient φi to the nearest pixel j having φj ≈ φi within 15◦. • If such a pixel j doesn’t exist, then the separation is con- sidered infinite. 15
  • 16. Texture Analysis Details (Cont’d) 3. Having created edge directionality and edge separation maps, a 2D texture histogram of ξ versus φ is constructed. • The initial histogram size is 193 × 180, where separation value ξ = 193 is reserved for a separation of infinity (as well as any ξ > 192). • The histogram is “smoothed” by replacing each pixel with a weighted sum of its neighbors, and then reduced to size 7 × 8, separation value 7 reserved for infinity. • Finally, the texture histogram is normalized by dividing by the number of pixels in the image segment. It will then be used for matching. 16
  • 17. Search by Illumination Invariance • To deal with illumination change from the query image to dif- ferent database images, each color channel band of each im- age is first normalized, and then compressed to a 36-vector. • A 2-dimensional color histogram is then created by using the chromaticity, which is the set of band ratios {R, G }/ (R + G + B) • To further reduce the number of vector components, the D C T coefficients for the smaller histogram are calculated and placed in zigzag order, and then all but 36 components dropped. • Matching is performed in the compressed domain by taking the Euclidean distance between two DCT-compressed 36- component feature vectors. • Fig. 20.5 shows the results of such a search. 17 20.3.5
  • 18. Fig. 20.5: Search with illumination invariance. 18
  • 19. Search by Object M odel • This search type proceeds by the user selecting a thumbnail and clicking the Model tab to enter object selection mode. – An image region can be selected by using primitive shapes such as a rectangle or an ellipse, a magic wand tool that is basically a seed-based flooding algorithm, an active con- tour (a “snake”(, or a brush tool where the painted region is selected. – An object is then interactively selected as a portion of the image. – Multiple regions can be dragged to the selection pane, but only the active object in the selection pane will be searched on. • A sample object selection is shown in Fig. 20.6. 19 20.3.6
  • 20. Fig. 20.6 C-BIRD interface showing object selection using an ellipse primitive. 20
  • 21. Details of Search by Object M odel 1. The user-selected model image is processed and its features localized (i.e., generate color locales [see below]). chro-2. Color histogram intersection, based on the reduced maticity histogram, is then applied as a first screen. 3. Estimate the pose (scale, translation, rotation) of the object inside a target image from the database. 4. Verification by intersection of texture histograms, and then a final check using an efficient version of a Generalized Hough Transform for shape verification. 21
  • 22. Fig. 20.7: Block diagram of object matching steps. 22
  • 23. M o del Image and T arget Images • A possible model image and one of the target images in the database might be as in Fig. 20.8. Fig. 20.8 : Model and target images. (a): Sample model image. (b): Sample database image containing the model book. 23
  • 24. 20.4 Quantifying Search Results • Precision is the percentage of relevant documents retrieved compared to the number of all the documents retrieved. P recision = Relevant images returned All retrieved images • Recall is the percentage of relevant documents retrieved out of all relevant documents. Recall = Relevant images returned All relevant images )20.6( • These measures are affected by the database size and the amount of similar information in the database, and as well they do not consider fuzzy matching or search result ordering. 24 )20.6)
  • 25. 20.4 Quantifying Search Results (Cont’d) 25 Alternatively, they may also be written as: where TP (True Positives) is the number of relevant images returned, FP (False Positives) is the number of irrelevant images returned, and FN (False Negatives) is the number of relevant images not returned.
  • 26. 20.5 Key Technologies in Current CBIR Systems  The field of CBIR has witnessed rapid growth and progress in the new millennium.  Unlike in the early years, users are no longer satisfied with query results returning scenes containing “blue skies,” or red blobs that may or may not be flowers.  They are more and more looking for objects, people, and often search for human activities as well.  Lew et al. and Datta et al. provide comprehensive surveys of the CBIR field. In this section, we will briefly describe some technologies and issues that are key to the success of current and future CBIR systems 26 Key Technologies in Current CBIR Systems 20.5.1. Robust Image Features and Their Representation 20.5.2. Relevance Feedback 20.5.3. Other Post-processing Techniques 20.5.4. Visual Concept Search 20.5.5. The Role of Users in Interactive CBIR Systems
  • 27. 20.5.1.Robust Image Features and Their Representation Many feature descriptors have been developed beyond the ones specified in MPEG-7. Here, we will only discuss SIFT and Visual Words. SIFT (Scale Invariant Feature Transform) In addition to global features such as color and texture histograms, local features characterized by SIFT (Scale Invariant Feature Transform) have been developed because they are more suitable for searches emphasizing visual objects. SIFT has been shown to be robust against image noise; it also has a fair degree of invariance to translation, rotation, and scale change. The process of extracting SIFT features from an image starts with building a multiresolution scale space, as shown in Fig. 20.11. At each scale, stacks of images are generated by applying Gaussian smoothing (filtering) operations. Visual words, as used in image retrieval systems, refer to small parts of an image which carry some kind of information related to the features (such as the color, shape or texture), or changes occurring in the pixels such as the filtering, low-level feature descriptors (SIFT , SURF , ...etc.). 27
  • 28. 20.5.2.Relevance Feedback • Relevance Feedback: involve the user in a loop, whereby images retrieved are used in further rounds of convergence onto correct returns. • Relevance Feedback Approaches – The usual situation: the user identifies images as good, bad, or don’t care, and weighting systems are updated according to this user guidance. – Another approach is to move the query towards positively marked content. – An even more interesting idea is to move every data point in a disciplined way, by warping the space of feature points. 28
  • 29. 20.5.3.Other Post-processing Techniques Beside Relevance Feedback, other post-processing methods have been developed after the initial query results are in. Spatial Verification The quality of the query results can be verified and improved by Spatial Verification. The spatial verification consists in verify a spatial correlation between certain points of a pair of images. The main problem is that outliers (that does not fit or does not match the selected model) affect adjustment called least squares Modern cameras often provide information about the location where pictures were taken, so it is trivial to check for consistency or relevance in certain query results if the user is only interested in images from certain locations, e.g., Paris or Rome. Advantages •Effective when one is able to find safe features without clutter. •Good results for correspondence in specific instances. Disadvantages •The scaling models. •The spatial verification can not be used as post-processing. 29
  • 30. Query Expansion  Query Expansion is a well-known method in Information Retrieval in which some combination of high-ranked relevant documents can be presented as a new query to boost the performance of the retrieval. The combination could simply be averages (means) of the feature descriptors from the returned documents. 30  The problem is that if one of the high-ranked document is a false-positive, then it will immediately harm the performance of the new query. Some robust statistical methods or even simply taking the median (instead of the mean) can help. Question-answering (QA) tries to replace the large number of images or multimedia content in general, that is returned by a search query, by making use of media content, knowledge about the search domain, and also linguistic analysis to return hits based on users’ natural-language questions. QA Paradigm
  • 31. 20.5.4.Visual Concept Search 31 • Search by concept is an automated information retrieval method and it is another major step in closing the semantic gap. Typically, after local features are turned into words, the words are converted to semantic concepts with the aid of machine learning algorithms. • In an interesting work presented by Wang et al. They show that such a system performs better than others that directly measure image similarity with simple low-level visual features such as color, texture, etc. • To assist in the development and testing of these systems, researchers and practitioners have developed many multimedia databases and benchmarks. The best known is perhaps theTRECVideo Retrieval Evaluation (TRECVID) benchmark. TRECVID started in 2003. • Initially, TRECVID provided video data from professional sources such as broadcast news, TV programs, and surveillance systems, thus having limited styles and content: For example, TRECVID 2013 evaluated the following tasks: • Semantic indexing • Interactive surveillance event detection • Instance search • Multimedia event detection • Multimedia event recounting
  • 32. 20.5.5 The Role of Users in Interactive CBIR Systems  Beside mechanisms such as Relevance Feedback, it has been argued that users should play a more active role in CBIR systems.  The authors again raised the question: Where is the User in Multimedia Retrieval?  They pointed out the dominance of MAP, common in evaluations such as TRECVID, may have hindered the development of better and more useful multimedia retrieval systems.  One way to comprehend how people view images as similar is to study using user groups just what forms our basis of perception of image similarity.  The function used in this approach can be a type of “perceptual similarity measure” and is learned by finding the best set of features (color, texture, etc.) to capture “similarity” as defined via the groups of similar images identified.  Another way to understand users is to talk to them and carefully analyze their search patterns. 32
  • 33. 20.6 Querying on V ideos • Video indexing can make use of motion as the salient fea- ture of temporally changing images for various types of query. • Inverse Hollywood: can we recover thevideo director’s “flowchart”? – Dividing the video into shots, where each shot consists roughly of the video frames between the on and off clicks of the record button. – Detection of shot boundaries is usually not simple as fade-in, fade- out, dissolve, wipe, etc. may often be involved. 33
  • 34. • In dealing with digital video, it is desirable to avoid uncom- pressing M P E G files. – A simple approach to this idea is to uncompress only enough to re- cover just the DC term, thus generating a thumbnail that is 64 times as small as the original. – Once DC frames are obtained from the whole video, many different approaches have been used for finding shot boundaries – based on features such as color, texture, and motion vectors. 34
  • 35. An Example of Querying on V ideo • Fig) 20.12: (a shows a selection of frames from a video of beach activity. Here the keyframes in( Fig. 20.12 (b. are selected based mainly on color information (but being careful with respect to the changes incurred by changing illumination conditions when videos are shot ( . • A more difficult problem arises when changes between shots are gradual, and when colors are rather similar overall, as in Fig) 20.13(.a The keyframes in Fig) 20.13(b. are sufficient to show the development of the whole video sequence. 35
  • 36. (a) (a) Fig. 20.12: Digital video and associated keyframes, beach video. (a): Frames from a digital video. (b): Keyframes selected. 36
  • 37. (a) )b) (a): F ramesfrom a digital video.Fig. 20.13: Garden video. (b): Keyframes selected. 37
  • 38. 20.7 Querying on Videos Based on Human Activity  Thousands of hours of video are being captured every day by CCTV cameras, web cameras, broadcast cameras, etc. However, most of the activities of interest (e.g., a soccer player scores a goal) only occur in a relatively small region along the spatial and temporal extent of the video. In this scenario, effective retrieval of a small spatial/temporal segment of the video containing a particular activity from a large collection of videos is very important. 38  Lan et al.’s work on activity retrieval is directly inspired by the application of searching for activities of interest from broadcast sports videos. For example, consider the scene shown in Fig. 20.14. we can explain the scene at multiple levels of detail such as low-level actions (e.g., standing and jogging) and mid-level social roles (e.g., attacker and first defenders). The social roles are denoted by different colors. In this example, we use magenta, blue and white to represent attacker, man-marking and players in the same team as the attacker respectively.
  • 39. 39 Fig.20.14 An example of human activity in realistic scenes . Beyond the general scene-level activity description (e.g. Free Hit).
  • 40. 20.7.1 Modeling Human Activity Structures Here we introduce the model . To illustrate, we describe an instantiation applicable to modeling field hockey videos. We first describe the labeling. Assume an image has been pre-processed, so the location of the goal and persons in the image have been found. We separate the players into two teams according to their color histograms. Each person is associated with two labels: action and social role. 40 The model is hierarchical, and includes various levels of detail: low-level actions, mid-level social roles, and high-level events. The relationships and interactions between these are included in the model. We define the score of interpreting a video frame with the hierarchical event representation as::
  • 41. 41 Fig. 20.15 Graphical illustration of the model. Different types of potentials are denoted by lines with different colors. Details of the potential functions are contained in Eq. 20.10
  • 42. 20.7.2 Experimental Results 42  The challenging Broadcast Field Hockey Dataset that consists of sporting activities captured from broadcast cameras is used to demonstrate the efficacy of the model. The model can carry out different inferences based on a user’s queries. The method is directly applicable to multilevel human activity recognition.  The goal is to predict events, social roles or actions of each person. In this case, the query doesn’t specify a particular event or social roles, but consists of more general questions, such as what is the overall game situation, or what are the social roles of each player.
  • 43. 20.8 Quality-Aware Mobile Visual Search 43  With the increasing popularity of mobile phones and tablets, Mobile Visual Search has attracted growing interest in the field of content-based image retrieval (CBIR).  On the mobile-client side, a query image is compressed to a certain quality level to accommodate the network conditions and then uploaded onto a server with its quality level transferred as side information.  On the server side, a set of features are extracted from the query image and then compared against the features of the images in the database.  Mobile Visual Search enables people to look for visually similar products or find information about movies or CDs online by initiating a search request from a camera phone.
  • 44. Quality-Aware Mobile Visual Search … (Cont’d) 44 1. A query image is transmitted to the server, and then feature extraction and retrieval are done on the server. 2. The mobile client extracts some features from the query image and uploads only the features to the server. The retrieval is performed on the server. 3. The mobile client maintains a cache of the image database. The retrieval is done locally on the client. Only if no match is found does the client sends a query to the server.  Depending on which part of the visual search is performed on the mobile devices, there are several possible client-server architectures :
  • 45. Quality-Aware Method 45 We outline a query quality-dependent fusion approach for mobile visual search. Specifically, it is based on the fusion of five common image features : • Tiny Images: This is the most trivial descriptor that compares images directly in color space after reducing the image dimensions drastically. • Color Histograms: The histograms have 16 bins for each component of RGB color space, yielding a total of 48 dimensions. • GIST: This descriptor computes the output of a bank of 24 Gabor-like filters tuned to eight orientations at four scales. The squared output of each filter is then averaged on a 4 × 4 grid. • Texton Histograms: For each image build a 512-dimensional histogram using a 512-entry universal texton dictionary. • SSIM : The self-similarity descriptors are quantized into 300 visual words by k-means.
  • 46. Quality-Aware Method 46 Fig. 20.18 Map results of different quality levels