skip to main content
10.1145/2070481.2070556acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
research-article

Virtual worlds and active learning for human detection

Authors Info & Claims
Published:14 November 2011Publication History

ABSTRACT

Image based human detection is of paramount interest due to its potential applications in fields such as advanced driving assistance, surveillance and media analysis. However, even detecting non-occluded standing humans remains a challenge of intensive research. The most promising human detectors rely on classifiers developed in the discriminative paradigm, i.e. trained with labelled samples. However, labelling is a manual intensive step, especially in cases like human detection where it is necessary to provide at least bounding boxes framing the humans for training. To overcome such problem, some authors have proposed the use of a virtual world where the labels of the different objects are obtained automatically. This means that the human models (classifiers) are learnt using the appearance of rendered images, i.e. using realistic computer graphics. Later, these models are used for human detection in images of the real world. The results of this technique are surprisingly good. However, these are not always as good as the classical approach of training and testing with data coming from the same camera, or similar ones. Accordingly, in this paper we address the challenge of using a virtual world for gathering (while playing a videogame) a large amount of automatically labelled samples (virtual humans and background) and then training a classifier that performs equal, in real-world images, than the one obtained by equally training from manually labelled real-world samples. For doing that, we cast the problem as one of domain adaptation. In doing so, we assume that a small amount of manually labelled samples from real-world images is required. To collect these labelled samples we propose a non-standard active learning technique. Therefore, ultimately our human model is learnt by the combination of virtual and real world labelled samples, which has not been done before.

References

  1. Y. Abramson and Y. Freund. SEmi-automatic VIsuaL LEarning (SEVILLE): a tutorial on active learning for visual object recognition. In IEEE Conf. on Computer Vision and Pattern Recognition, San Diego, CA, USA, 2005.Google ScholarGoogle Scholar
  2. S. Ben-David, J. Blitzer, K. Crammer, A. Kulesza, F. Pereira, and J. Vaughan. A theory of learning from different domains. Machine Learning , 79(1):151--175, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. T. Berg, A. Sorokin, G. Wang, D. Forsyth, D. Hoeiem, I. Endres, and A. Farhadi. It's all about the data. Proceedings of the IEEE , 98(8):1434--1452, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  4. C. Bishop. Pattern Recognition and Machine Learning. Springer, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Broggi, A. Fascioli, P. Grisleri, T. Graf, and M. Meinecke. Model-based validation approaches and matching techniques for automotive vision based pedestrian detection. In IEEE Conf. on Computer Vision and Pattern Recognition, San Diego, CA, USA, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. N. Dalal. Finding people in images and videos. PhD Thesis, Institut National Polytechnique de Grenoble / INRIA Rhone-Alpes, 2006.Google ScholarGoogle Scholar
  7. N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In IEEE Conf. on Computer Vision and Pattern Recognition, San Diego, CA, USA, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. P. Dollár, C. Wojek, B. Schiele, and P. Perona. Pedestrian detection: a benchmark. In IEEE Conf. on Computer Vision and Pattern Recognition, Miami Beach, FL, USA, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  9. M. Enzweiler and D. Gavrila. A mixed generative-discriminative framework for pedestrian classification. In IEEE Conf. on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 2008.Google ScholarGoogle ScholarCross RefCross Ref
  10. M. Enzweiler and D. Gavrila. Monocular pedestrian detection: survey and experiments. IEEE Trans. on Pattern Analysis and Machine Intelligence, 31(12):2179--2195, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. M. Everingham, L. V. Gool, C. Williams, J. Winn, and A. Zisserman. The PASCAL visual object classes (VOC) challenge. Int. Journal on Computer Vision, 88(2):303--338, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. P. Felzenszwalb, D. McAllester, and D. Ramanan. A discriminatively trained, multiscale, deformable part model. In IEEE Conf. on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 2008.Google ScholarGoogle ScholarCross RefCross Ref
  13. D. Gerónimo, A.M. López, A.D. Sappa, and T. Graf. Survey of pedestrian detection for advanced driver assistance systems. IEEE Trans. on Pattern Analysis and Machine Intelligence, 32(7):1239--1258, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. J. Marín, D. Vázquez, D. Gerónimo, and A.M. López. Learning appearance in virtual scenarios for pedestrian detection. In IEEE Conf. on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  15. T. Pouli, D. Cunningham, and E. Reinhard. Image statistics and their applications in computer graphics. In European Computer Graphics Conference and Exhibition, Norrköping, Sweden, 2010.Google ScholarGoogle Scholar
  16. G. Taylor, A. Chosak, and P. Brewer. OVVV: Using virtual worlds to design and evaluate surveillance systems. In IEEE Conf. on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 2007.Google ScholarGoogle ScholarCross RefCross Ref
  17. W. van der Mark and D. M. Gavrila. Real-time dense stereo for intelligent vehicles. IEEE Trans. on Intelligent Transportation Systems , 7(1):38--50, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. L. von Ahn, R. Liu, and M. Blum. Peekaboom: a game for locating objects in images. In ACM SIGCHI Conf. on Human Factors in Computing Systems, Montréal, Québec, Canada, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. S. Walk, N. Majer, K. Schindler, and B. Schiele. New features and insights for pedestrian detection. In IEEE Conf. on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  20. X. Wang, T.X. Han, and S. Yan. An HOG-LBP human detector with partial occlusion handling. In Int. Conf. on Computer Vision, Kyoto, Japan, 2009.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Virtual worlds and active learning for human detection

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ICMI '11: Proceedings of the 13th international conference on multimodal interfaces
      November 2011
      432 pages
      ISBN:9781450306416
      DOI:10.1145/2070481

      Copyright © 2011 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 14 November 2011

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate453of1,080submissions,42%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader