語義映像標註和檢索 Semantic Image Annotation and Retrieval

來源:互聯網
上載者:User

Semantic Image Annotation and Retrieval

Content-based image retrieval, the problem of searching large image repositories according to their content, has been the subject of a significant amount of computer vision research in the recent past. While early retrieval architectures were based on the query-by-example paradigm, which formulates image retrieval as the search for the best database match to a user-provided query image, it was quickly realized that the design of fully functional retrieval systems would require support for semantic queries. These are systems where the database of images are annotated with semantic keywords, enabling the user to specify the query through a natural language description of the visual concepts of interest. This realization, combined with the cost of manual image labeling, generated significant interest in the problem of automatically extracting semantic descriptors from images.

The earliest efforts in the area were directed to the reliable extraction of specific semantics, e.g. differentiating indoor from outdoor scenes, cities from landscapes, and detecting trees, horses, or buildings, among others. These efforts posed the problem of semantics extraction as one ofsupervised learning: a set of training images with and without the concept of interest was collected and a binary classifier trained to detect the concept of interest. The classifier was then applied to all database of images which were, in this way, annotated with respect to the presence or absence of the concept.

More recently, there has been an effort to solve the problem in its full generality, by resorting tounsupervised learning. The basic idea is to introduce a set of latent variables that encode hidden states of the world, where each state defines a joint distribution on the space of semantic keywords and image appearance descriptors (in the form of local features computed over image neighborhoods). After the annotation model is learned, an image is annotated by finding the most likely keywords given the features of the image.

Both formulations of the semantic labeling problem have strong advantages and disadvantages. In generic terms, unsupervised labeling leads to significantly more scalable (in database size and number of concepts of interest) training procedures, places much weaker demands on the quality of the manual annotations required to bootstrap learning, and produces a natural ranking of keywords for each new image to annotate. On the other hand, it does not explicitly treat semantics as image classes and, therefore, provides little guarantees that the semantic annotations are optimal in a recognition or retrieval sense. That is, instead of annotations that achieve the smallest probability of retrieval error, it simply produces the ones that have largest joint likelihood under the assumed mixture model.

In this work we show that it is possible to combine the advantages of the two formulations through a slight reformulation of the supervised one. This consists of defining an M-ary classification problem where each of the semantic concepts of interest defines an image class. At annotation time, these classes all directly compete for the image to annotate, which no longer faces a sequence of independent binary tests. This supervised multiclass labeling (SML) obviously retains the classification and retrieval optimality of the supervised formulation, but 1) produces a natural ordering of keywords at annotation time, and 2) eliminates the need to compute a "non-class" model for each of the semantic concepts of interest. In result, it has learning complexity equivalent to that of the unsupervised formulation and, like the latter, places much weaker requirements on the quality of manual labels than supervised OVA.

Results: Here are both qualitative and quantitative results of the SML annotation algorithm. Examples of semantic image annotation on Corel5k. Examples of semantic image retrieval on Corel5k. Quantitative comparison of several annotation algorithms.
Databases: We have used the following data-sets for image annotation experiments. Please contact the respective people for information about obtaining the data: Corel5k - 5,000 annotated images; 371 word vocabulary
(contact Kobus Barnard) Corel30k - 31,695 annotated images; 1035 word vocabulary
(contact Nuno Vasconcelos) PSU - 59,895 images, annotated by group; 442 word vocabulary
(contact James Z. Wang) The images from these data-sets are from the Corel image CDs. The annotations for Corel5k and Corel30k are a subset of those from the Berkeley Digital Library project. The full set of annotations for 40,000 Corel images is available here.
Publications: Supervised Learning of Semantic Classes for Image Annotation and Retrieval
G. Carneiro, A. B. Chan, P. J. Moreno, and N. Vasconcelos
IEEE Transactions on Pattern Analysis and Machine Intelligence,
Vol. 29(3), pp. 394-410, March 2006.� IEEE,[pdf] 

Formulating Semantic Image Annotation as a Supervised Learning Problem
G. Carneiro and N. Vasconcelos
Proceedings of IEEE Conference on Computer Vision and Pattern Recognition,
San Diego, 2005.� IEEE,[ps][pdf]

A Database Centric View of Semantic Image Annotation and Retrieval
G. Carneiro and N. Vasconcelos,
Proceedings of ACM Conference on Research and Development in Information Retrieval (ACM SIGIR)
Salvador, Brazil. 2005. [ps][pdf]

Using Statistics to Search and Annotate Pictures: an Evaluation of Semantic Image Annotation and Retrieval on Large Databases
A. B. Chan, P. J. Moreno, and N. Vasconcelos
Proceedings of Joint Statistical Meetings (JSM),
Seattle, 2006.[ps][pdf]

Formulating Semantic Image Annotation as a Supervised Learning Problem
G. Carneiro and N. Vasconcelos,
Technical Report SVCL-TR-2004-03, December 2004. [ps][pdf]

Articles: Better, More Accurate Image Search
Web Article, Technology Review, April 9, 2007. 

New Algorithms from UCSD Improve Automated Image Labeling
Press Release, Jacobs School of Engineering, March 29, 2007. 

Contact: Nuno Vasconcelos

from: http://www.svcl.ucsd.edu/projects/imgnote/

聯繫我們

該頁面正文內容均來源於網絡整理,並不代表阿里雲官方的觀點,該頁面所提到的產品和服務也與阿里云無關,如果該頁面內容對您造成了困擾,歡迎寫郵件給我們,收到郵件我們將在5個工作日內處理。

如果您發現本社區中有涉嫌抄襲的內容,歡迎發送郵件至: info-contact@alibabacloud.com 進行舉報並提供相關證據,工作人員會在 5 個工作天內聯絡您,一經查實,本站將立刻刪除涉嫌侵權內容。

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.