Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Abstract:- In image search re-ranking, besides the well-known semantic gap, intent gap, which is the gap between the representation of users
query/demand and the real intent of the users, is becoming a major problem restricting the development of image retrieval. To reduce human
effects, in this paper, we use image click-through data, which can be viewed as the implicit feedback from users, to help overcome the intention
gap, and further improve the image search performance. Generally, the hypothesis visually similar images should be close in a ranking list and
the strategy images with higher relevance should be ranked higher than others are widely accepted. To obtain satisfying search results, thus,
image similarity and the level of relevance typicality are determinate factors correspondingly.
Then, based on the learnt click-based image similarity measure, we conduct spectral clustering to group visually and semantically
similar images into same clusters, and get the final re-ranklist by calculating click-based clusters typicality and within clusters click-based image
typicality in descending order. Our experiments conducted on two real-world query-image data sets with diverse representative queries show that
our proposed re-ranking approach can significantly improve initial search results, and outperform several existing re-ranking approaches.
Index TermsImage search, search re-ranking, click-through data, multi-feature similarity, image typicality.
__________________________________________________*****_________________________________________________
1053
IJRITCC | May 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 5 1053 1058
_______________________________________________________________________________________________
approaches respectively, i.e., visually similar images should problem visual semantic web image search engine. The
be close in a ranking list, and images with higher relevance visual semantic web image re-ranking structure, which
should be ranked higher than others. Therefore, image automatically and directly offline studied different visual
similarity and image typicality (the level of image semantic spaces meaning for different search query
relevance) become determinate factors correspondingly to keywords. Then these visual features of images are extended
obtain satisfying re-ranking results. For image similarity to their visual semantic spaces to formed visual semantic
measure, Euclidean distance and cosine distance are signatures.
commonly used due to the success in the bag-of-words Nikit chaudhary, Sunil jadhav in the paper ,Web
models for text. Since image content is extracted and Image Re-Ranking Using Query-specific semantic
expressed in various kinds of features, in order to mine signatures,[3] Proposed in image re-ranking, as an effective
useful information from image content as much as possible, way to improve the results of web based image search, has
it would be better to leverage multiple visual modalities. been adopted by current commercial search engines. Given a
However, when dealing with multiple visual modalities, query keyword, a pool of images are first retrieved by the
there is often no obvious choice of similarity measure. search engine based on textual information. By asking the
Different kinds of features may lead to different forms of user to select a query image from the pool, the remaining
similarity. images are re-ranked based on their visual similarities with
2. Literature review the query image. A major challenge is that the similarities of
Tao Mei, Yongdong Zhang in the paper ,Web visual features do not well correlate with images semantic
Image Search Re-Ranking With Click-Based Similarity and meanings which interpret users search intention.
Typicality,[1]Proposed in image search re-ranking, besides M Sai KumarDr. C. Nalini in the paper, Learning
the well-known semantic gap, intent gap, which is the gap Image Re-Rank: Query-Dependent Image Re-Ranking
between the representation of users query/demand and the Using Semantic Signature,[4] Proposed is an effective way
real intent of the users, is becoming a major problem to improve the results of web-based image search and has
restricting the development of image retrieval. To reduce been adopted by current commercial search engines such as
human effects, image click-through data, which can be Bing and Google. When a query keyword is given, a list of
viewed as the implicit feedback from users, to help images are first retrieved based on textual information given
overcome the intention gap, and further improve the image by the user. By asking the user to select a query image from
search performance. However, when measuring image the pool of images, the remaining images are re-ranked
similarity and typicality, conventional re-ranking based on their index with the query image. A major
approaches only consider visual information and initial challenge is that sometimes semantic meanings may
ranks of images, while overlooking the influence of click- interpret users search intention. Many people recently
through data. Then, based on the learnt click-based image proposed to match images in a semantic space which used
similarity measure, we conduct spectral clustering to group attributes or reference classes closely related to the semantic
visually and semantically similar images into same clusters, meanings of images as basis.
and get the final re-rank list by calculating click-based
clusters typicality. 3. Motivation
Mr. Sandesh Keshav Pawaskar in the paper,Visual Hundreds of thousands of images are uploaded to
semantic web based image re-ranking for effective search the internet with the explosive growth of online social media
engine,[2] Proposed in visual semantic web based Image and the popularity of capture devices thus, building a
search engine is a way using that multiple images are search satisfying image retrieval system is the key to improve user
and matched in semantic space. This matched images we search experience. Due to the success of information
use for image reranking methodology. Image re-ranking is a retrieval, most commercial search engines employ text-
method using that we improve results of web based images based search techniques for image search by using
search. When user search any query keyword on web based associated textual information, such as file name,
search engine, then a set of images are extracted based on surrounding text, URL, etc. For example, The query is
the textual information. User then select a required query shahrukh and guaris son we want to focus only those
image from the set of images and then the others images are images where the son of the shahrukh and guaris son
recomputed or re-ranked based on visual occurrence of the appears in images.
query image. These similarities of visual features do not
well match with visual semantic meanings of images which
normally coordinate users search intention and it is a main
1054
IJRITCC | May 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 5 1053 1058
_______________________________________________________________________________________________
1054
IJRITCC | May 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 5 1053 1058
_______________________________________________________________________________________________
1053
IJRITCC | May 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 5 1053 1058
_______________________________________________________________________________________________
Similarity Graph: G(V,E,W)
V Vertices (Data points)
E Edge if similarity > 0
W -Edge weights (similarities)
=
Step 7: Then find the standard deviation of color
histogram using
=
Fig. 2. HSV representation
1054
IJRITCC | May 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 5 1053 1058
_______________________________________________________________________________________________
8. SIFT:
Scale-invariant feature transform (or SIFT) is an
algorithm in computer vision to detect and describe local
features in images. Applications include object recognition,
robotic mapping and navigation, image stitching, 3D
modeling, gesture recognition, video tracking, individual
identification of wildlife and match moving.
Given SIFT's ability to find distinctive key points
that are invariant to location, scale and rotation, and robust
6. Texture:
to affine transformations (changes in scale, rotation, shear,
An image texture is a set of matrix calculated in
and position) and changes in illumination, they are usable
image processing designed to quantify the perceived texture
for object recognition.
of an image. Image texture gives us information about the
The steps are given below.
spatial arrangement of color or intensities in an image or
First, SIFT features are obtained from the input
selected region of an image.
image using the algorithm described above.
7. Edge Feature extraction:
These features are matched to the SIFT feature
Edge detection techniques includes a variety of
database obtained from the training images. This
mathematical method that aim of identifying point in a
feature matching is done through a Euclidean-
digital image at which the image brightness changes sharply
distance based nearest neighbor approach.
or more formally, has discontinuities change sharply are
Although the distance ratio test described above
typically organized into a set of curved lines segment termed
edge. discards many of the false matches arising from
Support vector machine: background clutter, we still have matches that
In machine learing, a support vector machine belong to different objects.
constructs a hyperplane or set of hyperplane in a For any object in an image, interesting points on
high dimensional space, which can be used for the object can be extracted to provide a "feature description"
classification and regression or other task. of the object. This description, extracted from a training
Step: image, can then be used to identify the object when
1) Find the goal of support vector machine. attempting to locate the object in a test image containing
2) Compute the margin. many other objects. To perform reliable recognition, it is
3) Unconstrained minimization. important that the features extracted from the training image
4) Convex functions. be detectable even under changes in image scale, noise and
5) Duality and lagrange multiplers. illumination. Such points usually lie on high-contrast
An SVM classifies data by finding the best regions of the image, such as object edges.
hyperplane that separates all data points of one Scale Invariant Feature Transform (SIFT) features
class from those of the other class. are features extracted from images to help in reliable
The best hyperplane for an SVM means the one matching between different views of the same object. SIFT
with the largest margin between the two classes. is an image local feature description algorithm based on
scale-space. Due to its strong matching ability, SIFT has
many applications in different fields, such as image
retrieval, image stitching, and machine vision. The extracted
features are invariant to scale and orientation, and are highly
distinctive of the image. They are extracted in four steps.
The first step computes the locations of potential interest
points in the image by detecting the maxima and minima of
a set of Difference of Gaussian filters applied at different
scales all over the image.
Color Descriptor Based on SIFT
1058
IJRITCC | May 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________