[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/1816041.1816051acmconferencesArticle/Chapter ViewAbstractPublication PagescivrConference Proceedingsconference-collections
research-article

On the sampling of web images for learning visual concept classifiers

Published: 05 July 2010 Publication History

Abstract

Visual concept learning often requires a large set of training images. In practice, nevertheless, acquiring noise-free training labels with sufficient positive examples is always expensive. A plausible solution for training data collection is by sampling the largely available user-tagged images from social media websites. With the general belief that the probability of correct tagging is higher than that of incorrect tagging, such a solution often sounds feasible, though is not without challenges. First, user-tags can be subjective and, to certain extent, are ambiguous. For instance, an image tagged with "whales" may be simply a picture about ocean museum. Learning concept "whales" with such training samples will not be effective. Second, user-tags can be overly abbreviated. For instance, an image about concept "wedding" may be tagged with "love" or simply the couple's names. As a result, crawling sufficient positive training examples is difficult. This paper empirically studies the impact of exploiting the tagged images towards concept learning, investigating the issue of how the quality of pseudo training images affects concept detection performance. In addition, we propose a simple approach, named semantic field, for predicting the relevance between a target concept and the tag list associated with the images. Specifically, the relevance is determined through concept-tag co-occurrence by exploring external sources such as WordNet and Wikipedia. The proposed approach is shown to be effective in selecting pseudo training examples, exhibiting better performance in concept learning than other approaches such as those based on keyword sampling and tag voting.

References

[1]
M. Ames and M. Naaman. Why we tag: Motivations for annotation in mobile and online media. In ACM SIGCHI, 2007.
[2]
A. Ulges et al. Learning automatic concept detectors from online video. Comput. Vis. Image Understand, 2009.
[3]
D. Liu et al. Tag ranking. In ACM WWW, 2009.
[4]
G.-J. Qi et al. Transductive inference with hierarchical clustering for video annotation. In ICME, 2007.
[5]
J. Tang et al. Inferring semantic concepts from community-contributed images and noisy tags. In ACM MM, 2009.
[6]
K. Bischoff et al. Can all tags be used for search. In ACM CIKM, 2008.
[7]
Q. Tian et al. A new analysis of the value of unlabeled data in semi-supervised learning for image retrieval. In ICME, 2004.
[8]
S.-F. Chang et al. Columbia University/VIREO-CityU/IRIT TRECVID 2008 high-level feature extraction and interactive video search. In TRECVID, 2008.
[9]
T.-S. Chua et al. NUS-WIDE: A real-world web image database from national university of singapore. In CIVR, 2009.
[10]
Y.-G. Jiang et al. Representations of keypoint-based semantic concept detection: A comprehensive study. IEEE Trans. on Multimedia, 12(1):42--53, 2010.
[11]
Y.-G. Jiang, C.-W Ngo, and J. Yang. Towards optimal bag-of-features for object categorization and semantic video retrieval. In CIVR, 2007.
[12]
D. Jurafsky and J. H. Martin. Speech and language processing. Prentice-Hall, 2000.
[13]
L. S. Kennedy, S.-F. Chang, and I. V. Kozintsev. To search or to label. In ACM MIR, 2006.
[14]
L.-J. Li and L. Fei-Fei. OPTIMOL: automatic object picture collection via incremental model learning. Int. J. of Computer Vision, 2009.
[15]
X.-R. Li and C. G. M. Snoek. Visual categorization with negative examples for free. In ACM MM, 2009.
[16]
X.-R. Li, C. G. M. Snoek, and M. Worring. Learning social tag relevance by neighbor voting. IEEE Trans. on MM, 11(7):1310--1322, 2009.
[17]
M. R. Naphade and J. R. Smith. On the detection of semantic concepts at TRECVID. In ACM MM, 2004.
[18]
G. Quénot and S. Ayache. TRECVID 2009 collaborative annotation. http://mrim.imag.fr/tvca/.
[19]
F. Schroff, A. Criminisi, and A. Zisserman. Harvesting image databases from the web. In ICCV, 2007.
[20]
A. T. Setz and C. G. M. Snoek. Can social tagged images aid concept-based video search. In ICME, 2009.
[21]
S. Tong and E. Chang. Support vector machine active learning for image retrieval. In ACM MM, 2001.
[22]
G. Wang, T.-S. Chua, and M. Zhao. Exploring knowledge of sub-domain in a multi-resoluation bootstrapping framework for concept detection in news. In ACM MM, 2008.
[23]
R. Yan, A. G. Hauptmann, and R. Jin. Negative pseudo-relevance feedback in content-based video retrieval. In ACM MM, 2003.
[24]
R. Yan and M. R. Naphade. Semi-supervised cross feature learning for semantic concept detection in video. In CVPR, 2005.

Cited By

View all

Index Terms

  1. On the sampling of web images for learning visual concept classifiers

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    CIVR '10: Proceedings of the ACM International Conference on Image and Video Retrieval
    July 2010
    492 pages
    ISBN:9781450301176
    DOI:10.1145/1816041
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    In-Cooperation

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 05 July 2010

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. concept detection
    2. sampling
    3. web images

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    CIVR' 10
    Sponsor:

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)3
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 22 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2018)Georeferenced Social Multimedia as Volunteered Geographic InformationCyberGIS for Geospatial Discovery and Innovation10.1007/978-94-024-1531-5_12(225-246)Online publication date: 27-Jun-2018
    • (2017)Automation of image categorization with most relevant negativesPattern Recognition and Image Analysis10.1134/S105466181703005127:3(371-379)Online publication date: 1-Jul-2017
    • (2016)Complex-query web image search with concept-based relevance estimationWorld Wide Web10.1007/s11280-015-0357-x19:2(247-264)Online publication date: 1-Mar-2016
    • (2016)How important is location information in saliency detection of natural imagesMultimedia Tools and Applications10.1007/s11042-015-2875-z75:5(2543-2564)Online publication date: 1-Mar-2016
    • (2016)Visual concept detection of web images based on group sparse ensemble learningMultimedia Tools and Applications10.1007/s11042-014-2179-875:3(1409-1425)Online publication date: 1-Feb-2016
    • (2016)Effect of Junk Images on Inter-concept Distance Measurement: Positive or Negative?MultiMedia Modeling10.1007/978-3-319-51814-5_15(173-184)Online publication date: 31-Dec-2016
    • (2015)Soft-assigned bag of features for object trackingMultimedia Systems10.1007/s00530-014-0384-y21:2(189-205)Online publication date: 1-Mar-2015
    • (2015)Inter-Concept Distance Measurement with Adaptively Weighted Multiple Visual FeaturesComputer Vision - ACCV 2014 Workshops10.1007/978-3-319-16634-6_5(56-70)Online publication date: 12-Apr-2015
    • (2015)Cross-Domain Concept Detection with Dictionary Coherence by Leveraging Web ImagesMultiMedia Modeling10.1007/978-3-319-14442-9_47(415-426)Online publication date: 2015
    • (2014)A Cross-Modal Approach for Extracting Semantic Relationships Between Concepts Using Tagged ImagesIEEE Transactions on Multimedia10.1109/TMM.2014.230665516:4(1059-1074)Online publication date: 1-Jun-2014
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media