Jummave Sillapi Abs

You might also like

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 1

SEMANTIC SIGNATURES USING ON WEB IMAGE

RE-RANKING
ABSTRACT

Image re-ranking, as an effective way to improve the results of web-based


image search, has been adopted by current commercial search engines such as Bing
and Google. Given a query keyword, a pool of images is first retrieved based on
textual information. By asking the user to select a query image from the pool, the
remaining images are re-ranked based on their visual similarities with the query
image. A major challenge is that the similarities of visual features do not well
correlate with images’ semantic meanings which interpret users’ search intention.

Recently people proposed to match images in a semantic space which used


attributes or reference classes closely related to the semantic meanings of images
as basis. However, learning a universal visual semantic space to characterize
highly diverse images from the web is difficult and inefficient. In this thesis, we
propose a novel image re-ranking framework, which automatically offline learns
different semantic spaces for different query keywords. The visual features of
images are projected into their related semantic spaces to get semantic signatures.
At the online stage, images are re-ranked by comparing their semantic signatures
obtained from the semantic space specified by the query keyword.

The proposed query-specific semantic signatures significantly improve both


the accuracy and efficiency of image re-ranking. The original visual features of
thousands of dimensions can be projected to the semantic signatures as short as 25
dimensions. Experimental results show that 25-40 percent relative improvement
has been achieved on re-ranking precisions compared with the state-of-the-art
methods.

You might also like