Geo-based Automatic Image Annotation

Abstract : A huge number of user-tagged images are daily uploaded to the web. Recently, a growing number of those images are also geotagged. These provide new opportunities for solutions to automatically tag images so that efficient image management and retrieval can be achieved. In this paper an automatic image annotation approach is proposed. It is based on a statistical model that combines two different kinds of information: high level information represented by user tags of images captured in the same location as a new unlabeled image (input image); and low level information represented by the visual similarity between the input image and the collection of geographically similar images. To maximize the number of images that are visually similar to the input image, an iterative visual matching approach is proposed and evaluated. The results show that a significant recall improvement can be achieved with an increasing number of iterations. The quality of the recommended tags has also been evaluated and an overall good performance has been observed.
Document type :
Conference papers
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-01353097
Contributor : Équipe Gestionnaire Des Publications Si Liris <>
Submitted on : Wednesday, August 10, 2016 - 4:22:24 PM
Last modification on : Friday, January 11, 2019 - 5:09:20 PM

Identifiers

Citation

Hatem Mousselly-Sergieh, Mario Döller, Elod Egyed-Zsigmond, Gabriele Gianini, Harald Kosch, et al.. Geo-based Automatic Image Annotation. 2nd ACM International Conference on Multimedia Retrieval (ICMR '12), Jun 2012, Hong Kong, China. pp.1-8, ⟨10.1145/2324796.2324850⟩. ⟨hal-01353097⟩

Share

Metrics

Record views

185