skip to main content
10.1145/1579114.1579130acmotherconferencesArticle/Chapter ViewAbstractPublication PagespetraConference Proceedingsconference-collections
research-article

Towards automated large vocabulary gesture search

Published:09 June 2009Publication History

ABSTRACT

This paper describes work towards designing a computer vision system for helping users look up the meaning of a sign. Sign lookup is treated as a video database retrieval problem. A video database is utilized that contains one or more video examples for each sign, for a large number of signs (close to 1000 in our current experiments). The emphasis of this paper is on evaluating the trade-offs between a non-automated approach, where the user manually specifies hand locations in the input video, and a fully automated approach, where hand locations are determined using a computer vision module, thus introducing inaccuracies into the sign retrieval process. We experimentally evaluate both approaches and present their respective advantages and disadvantages.

References

  1. J. Alon, V. Athitsos, Q. Yuan, and S. Sclaroff. Simultaneous localization and recognition of dynamic hand gestures. In IEEE Motion Workshop, pages 254--260, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. V. Athitsos, C. Neidle, S. Sclaroff, J. Nash, A. Stefan, Q. Yuan, and A. Thangali. The American Sign Language lexicon video dataset. In IEEE Workshop on Computer Vision and Pattern Recognition for Human Communicative Behavior Analysis (CVPR4HB), 2008.Google ScholarGoogle ScholarCross RefCross Ref
  3. B. Bauer and K. Kraiss. Towards an automatic sign language recognition system using subunits. In Gesture Workshop, pages 64--75, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. M. Black and A. Jepson. Recognizing temporal trajectories using the condensation algorithm. In Face and Gesture Recognition, pages 16--21, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. F. Chen, C. Fu, and C. Huang. Hand gesture recognition using a real-time tracking method and Hidden Markov Models. Image and Video Computing, 21(8):745--758, August 2003.Google ScholarGoogle ScholarCross RefCross Ref
  6. A. Corradini. Dynamic time warping for off-line recognition of a small gesture vocabulary. In Recognition, Analysis and Tracking of Faces and Gestures in Real-time Systems (RATFG-RTS), pages 82--89, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Y. Cui and J. Weng. Appearance-based hand sign recognition from intensity image sequences. Computer Vision and Image Understanding, 78(2):157--176, 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. R. Cutler and M. Turk. View-based interpretation of real-time optical flow for gesture recognition. In Face and Gesture Recognition, pages 416--421, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. T. Darrell, I. Essa, and A. Pentland. Task-specific gesture analysis in real-time using interpolated views. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 18(12):1236--1242, 1996. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. J. Deng and H.-T. Tsui. A PCA/MDA scheme for hand posture recognition. In Automatic Face and Gesture Recognition, pages 294--299, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. K. Fujimura and X. Liu. Sign recognition using depth image streams. In Automatic Face and Gesture Recognition, pages 381--386, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. W. Gao, G. Fang, D. Zhao, and Y. Chen. Transition movement models for large vocabulary continuous sign language recognition. In Automatic Face and Gesture Recognition, pages 553--558, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. T. Kadir, R. Bowden, E. Ong, and A. Zisserman. Minimal training, large lexicon, unconstrained sign language recognition. In British Machine Vision Conference (BMVC), volume 2, pages 939--948, 2004.Google ScholarGoogle ScholarCross RefCross Ref
  14. E. Keogh. Exact indexing of dynamic time warping. In International Conference on Very Large Data Bases, pages 406--417, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. J. B. Kruskall and M. Liberman. The symmetric time warping algorithm: From continuous to discrete. In Time Warps. Addison-Wesley, 1983.Google ScholarGoogle Scholar
  16. J. Ma, W. Gao, J. Wu, and C. Wang. A continuous Chinese Sign Language recognition system. In Automatic Face and Gesture Recognition, pages 428--433, 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. J. Martin, V. Devin, and J. Crowley. Active hand tracking. In Face and Gesture Recognition, pages 573--578, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. S. C. W. Ong and S. Ranganath. Automatic sign language analysis: A survey and the future beyond lexical meaning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(6):873--891, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. V. Pavlovic, R. Sharma, and T. Huang. Visual interpretation of hand gestures for human-computer interaction: A review. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 19(7), 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. L. Rabiner. A tutorial on hidden markov models and selected applications in speech recognition. In Proceedings of the IEEE, volume 77:2, 1989.Google ScholarGoogle ScholarCross RefCross Ref
  21. L. Rabiner and B. Juang. Fundamentals of speech recognition. Prentice Hall, 1993. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. H. Rowley, S. Baluja, and T. Kanade. Rotation invariant neural network-based face detection. In CVPR, pages 38--44, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. H. Sagawa and M. Takeuchi. A method for recognizing a sequence of sign language words represented in a Japanese Sign Language sentence. In Automatic Face and Gesture Recognition, pages 434--439, 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. H. Sakoe and S. Chiba. Dynamic programming algorithm optimization for spoken word recognition. In IEEE Transactions on Acoustics, Speech, and Signal Processing, volume 34(1), pages 43--49, 1978.Google ScholarGoogle ScholarCross RefCross Ref
  25. T. Starner and A. Pentland. Real-time American Sign Language recognition using desk and wearable computer based video. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(12):1371--1375, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. C. Valli, editor. The Gallaudet Dictionary of American Sign Language. Gallaudet U. Press, Washington, DC, 2006.Google ScholarGoogle Scholar
  27. P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In IEEE Conference on Computer Vision and Pattern Recognition, volume 1, pages 511--518, 2001.Google ScholarGoogle ScholarCross RefCross Ref
  28. C. Vogler and D. N. Metaxas. Parallel hidden markov models for american sign language recognition. In IEEE International Conference on Computer Vision (ICCV), pages 116--122, 1999.Google ScholarGoogle ScholarCross RefCross Ref
  29. C. Vogler and D. N. Metaxas. Handshapes and movements: Multiple-channel american sign language recognition. In Gesture Workshop, pages 247--258, 2003.Google ScholarGoogle Scholar
  30. C. Wang, S. Shan, and W. Gao. An approach based on phonemes to large vocabulary Chinese Sign Language recognition. In Automatic Face and Gesture Recognition, pages 411--416, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. M. Yang and N. Ahuja. Recognizing hand gesture using motion trajectories. In IEEE Conference on Computer Vision and Pattern Recognition, volume 1, pages 466--472, 1999.Google ScholarGoogle Scholar
  32. G. Yao, H. Yao, X. Liu, and F. Jiang. Real time large vocabulary continuous sign language recognition based on OP/Viterbi algorithm. In International Conference on Pattern Recognition, volume 3, pages 312--315, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Towards automated large vocabulary gesture search

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          PETRA '09: Proceedings of the 2nd International Conference on PErvasive Technologies Related to Assistive Environments
          June 2009
          481 pages
          ISBN:9781605584096
          DOI:10.1145/1579114

          Copyright © 2009 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 9 June 2009

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader