skip to main content
10.1145/3411763.3451523acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
extended-abstract

CopyCat: Using Sign Language Recognition to Help Deaf Children Acquire Language Skills

Authors Info & Claims
Published:08 May 2021Publication History

ABSTRACT

Deaf children born to hearing parents lack continuous access to language, leading to weaker working memory compared to hearing children and deaf children born to Deaf parents. CopyCat is a game where children communicate with the computer via American Sign Language (ASL), and it has been shown to improve language skills and working memory. Previously, CopyCat depended on unscalable hardware such as custom gloves for sign verification, but modern 4K cameras and pose estimators present new opportunities. Before re-creating the CopyCat game for deaf children using off-the-shelf hardware, we evaluate whether current ASL recognition is sufficient. Using Hidden Markov Models (HMMs), user independent word accuracies were 90.6%, 90.5%, and 90.4% for AlphaPose, Kinect, and MediaPipe, respectively. Transformers, a state-of-the-art model in natural language processing, performed 17.0% worse on average. Given these results, we believe our current HMM-based recognizer can be successfully adapted to verify children’s signing while playing CopyCat.

References

  1. Helene Brashear, Valerie Henderson, Kwang-Hyun Park, Harley Hamilton, Seungyon Lee, and Thad Starner. 2006. American sign language recognition in game development for deaf children. In Proceedings of the 8th international ACM SIGACCESS conference on Computers and accessibility. 79–86. https://doi.org/10.1145/1168987.1169002Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Helene Brashear, Thad Starner, Paul Lukowicz, and Holger Junker. 2006. Using multiple sensors for mobile sign language recognition. In Proceedings of the 8th international ACM SIGACCESS conference on Computers and accessibility. 79–86. https://doi.org/10.1109/iswc.2003.1241392Google ScholarGoogle ScholarCross RefCross Ref
  3. P Margaret Brown and Andrew Cornes. 2015. Mental health of deaf and hard-of-hearing adolescents: what the students say. Journal of deaf studies and deaf education 20, 1 (2015), 75–81. https://doi.org/10.1093/deafed/enu031Google ScholarGoogle ScholarCross RefCross Ref
  4. Necati Cihan Camgoz, Oscar Koller, Simon Hadfield, and Richard Bowden. 2020. Sign Language Transformers: Joint End-to-End Sign Language Recognition and Translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr42600.2020.01004Google ScholarGoogle ScholarCross RefCross Ref
  5. Helen Cooper and Richard Bowden. 2010. Sign language recognition using linguistically derived sub-units. In Proceedings of 4th workshop on the representation and processing of sign languages: corpora and sign language technologies. 57–61.Google ScholarGoogle Scholar
  6. Helen Cooper, Eng-Jon Ong, Nicolas Pugeault, and Richard Bowden. 2012. Sign language recognition using sub-units. The Journal of Machine Learning Research 13, 1 (2012), 2205–2231. https://doi.org/10.1007/978-3-319-57021-1_3Google ScholarGoogle ScholarCross RefCross Ref
  7. David Coulter, Yijie Wang, and Phil Meadows. [n.d.]. Azure Kinect body tracking joints. https://docs.microsoft.com/en-us/azure/kinect-dk/body-joints. Accessed: 2020-01-07.Google ScholarGoogle Scholar
  8. M. Ebrahim Al-Ahdal and M. T. Nooritawati. 2012. Review in Sign Language Recognition Systems. In 2012 IEEE Symposium on Computers Informatics (ISCI). 52–57. https://doi.org/10.1109/ISCI.2012.6222666Google ScholarGoogle ScholarCross RefCross Ref
  9. Hao-Shu Fang, Shuqin Xie, Yu-Wing Tai, and Cewu Lu. 2017. RMPE: Regional Multi-person Pose Estimation. In ICCV. https://doi.org/10.1109/iccv.2017.256Google ScholarGoogle ScholarCross RefCross Ref
  10. Barak Freedman, Alexander Shpunt, Meir Machline, and Yoel Arieli. 2013. Depth mapping using projected patterns. US Patent 8,493,496.Google ScholarGoogle Scholar
  11. Enrique Garcia-Ceja, Michael Riegler, Anders K. Kvernberg, and Jim Torresen. 2020. User-adaptive models for activity and emotion recognition using deep transfer learning and data augmentation. User Modeling and User-Adapted Interaction 30, 3 (July 2020), 365–393. https://doi.org/10.1007/s11257-019-09248-1Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. K. Grobel and M. Assan. 1997. Isolated sign language recognition using hidden Markov models. In 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation, Vol. 1. 162–167 vol.1. https://doi.org/10.1109/ICSMC.1997.625742Google ScholarGoogle ScholarCross RefCross Ref
  13. Jie Huang, Wengang Zhou, Qilin Zhang, Houqiang Li, and Weiping Li. 2018. Video-based sign language recognition without temporal segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 32.Google ScholarGoogle ScholarCross RefCross Ref
  14. Hamid Reza Vaezi Joze and Oscar Koller. 2018. Ms-asl: A large-scale data set and benchmark for understanding american sign language. arXiv preprint arXiv:1812.01053(2018).Google ScholarGoogle Scholar
  15. Kourosh Khoshelham and Sander Oude Elberink. 2012. Accuracy and resolution of kinect depth data for indoor mapping applications. Sensors 12, 2 (2012), 1437–1454. https://doi.org/10.3390/s120201437Google ScholarGoogle ScholarCross RefCross Ref
  16. Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. http://arxiv.org/abs/1412.6980 cite arxiv:1412.6980Comment: Published as a conference paper at the 3rd International Conference for Learning Representations, San Diego, 2015.Google ScholarGoogle Scholar
  17. Sang-Ki Ko, Chang Jo Kim, Hyedong Jung, and Choongsang Cho. 2019. Neural sign language translation based on human keypoint estimation. Applied Sciences 9, 13 (2019), 2683. https://doi.org/10.3390/app9132683Google ScholarGoogle ScholarCross RefCross Ref
  18. Oscar Koller, Hermann Ney, and Richard Bowden. 2016. Deep Hand: How to Train a CNN on 1 Million Hand Images When Your Data Is Continuous and Weakly Labelled. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr.2016.412Google ScholarGoogle ScholarCross RefCross Ref
  19. Yong-Lu Li, Liang Xu, Xinpeng Liu, Xijie Huang, Yue Xu, Shiyi Wang, Hao-Shu Fang, Ze Ma, Mingyang Chen, and Cewu Lu. 2020. PaStaNet: Toward Human Activity Knowledge Engine. In CVPR.Google ScholarGoogle Scholar
  20. Camillo Lugaresi, Jiuqiang Tang, Hadon Nash, Chris McClanahan, Esha Uboweja, Michael Hays, Fan Zhang, Chuo-Ling Chang, Ming Guang Yong, Juhyun Lee, Wan-Teh Chang, Wei Hua, Manfred Georg, and Matthias Grundmann. 2019. MediaPipe: A Framework for Building Perception Pipelines. CoRR abs/1906.08172(2019). arxiv:1906.08172http://arxiv.org/abs/1906.08172Google ScholarGoogle Scholar
  21. Rachel I Mayberry and Ellen B Eichen. 1991. The long-lasting advantage of learning sign language in childhood: Another look at the critical period for language acquisition. Journal of memory and language 30, 4 (1991), 486–512. https://doi.org/10.1016/0749-596x(91)90018-fGoogle ScholarGoogle ScholarCross RefCross Ref
  22. S. Mazilu, M. Hardegger, Z. Zhu, D. Roggen, G. Tröster, M. Plotnik, and J. M. Hausdorff. 2012. Online detection of freezing of gait with smartphones and machine learning techniques. In 2012 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops. 123–130. https://doi.org/10.4108/icst.pervasivehealth.2012.248680Google ScholarGoogle ScholarCross RefCross Ref
  23. Ross E Mitchell and MICHAELA KARCHMER. 2004. Chasing the mythical ten percent: Parental hearing status of deaf and hard of hearing students in the United States. Sign language studies 4, 2 (2004), 138–163. https://doi.org/10.1353/sls.2004.0005Google ScholarGoogle ScholarCross RefCross Ref
  24. Elissa L Newport. 1990. Maturational constraints on language learning. Cognitive science 14, 1 (1990), 11–28. https://doi.org/10.1207/s15516709cog1401_2Google ScholarGoogle ScholarCross RefCross Ref
  25. Robin E Perkins-Dock Ph D, Terrilyn R Battle MS, Jaleassia M Edgerton MS, and Jaqueline N McNeill MS. 2015. A survey of barriers to employment for individuals who are deaf. JADARA 49, 2 (2015), 3.Google ScholarGoogle Scholar
  26. Patricia M Sullivan and John F Knutson. 2000. Maltreatment and disabilities: A population-based epidemiological study. Child abuse & neglect 24, 10 (2000), 1257–1273. https://doi.org/10.1016/s0145-2134(00)00190-3Google ScholarGoogle ScholarCross RefCross Ref
  27. Oliver Turner, Kirsten Windfuhr, and Navneet Kapur. 2007. Suicide in deaf populations: a literature review. Annals of General Psychiatry 6, 1 (2007), 26. https://doi.org/10.1186/1744-859X-6-26Google ScholarGoogle ScholarCross RefCross Ref
  28. Kimberly A Weaver, Harley Hamilton, Zahoor Zafrulla, Helene Brashear, Thad Starner, Peter Presti, and Amy Bruckman. 2010. Improving the language ability of deaf signing children through an interactive American Sign Language-based video game. (2010).Google ScholarGoogle Scholar
  29. Yuliang Xiu, Jiefeng Li, Haoyu Wang, Yinghong Fang, and Cewu Lu. 2018. Pose Flow: Efficient Online Pose Tracking. In BMVC.Google ScholarGoogle Scholar
  30. Kayo Yin. 2020. Sign Language Translation with Transformers. ArXiv abs/2004.00588(2020).Google ScholarGoogle Scholar
  31. Zahoor Zafrulla, Helene Brashear, Pei Yin, Peter Presti, Thad Starner, and Harley Hamilton. 2010. American sign language phrase verification in an educational game for deaf children. In 2010 20th International Conference on Pattern Recognition. IEEE, 3846–3849. https://doi.org/10.1109/icpr.2010.937Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. CopyCat: Using Sign Language Recognition to Help Deaf Children Acquire Language Skills
            Index terms have been assigned to the content through auto-classification.

            Recommendations

            Comments

            Login options

            Check if you have access through your login credentials or your institution to get full access on this article.

            Sign in
            • Published in

              cover image ACM Conferences
              CHI EA '21: Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems
              May 2021
              2965 pages
              ISBN:9781450380959
              DOI:10.1145/3411763

              Copyright © 2021 Owner/Author

              Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

              Publisher

              Association for Computing Machinery

              New York, NY, United States

              Publication History

              • Published: 8 May 2021

              Check for updates

              Qualifiers

              • extended-abstract
              • Research
              • Refereed limited

              Acceptance Rates

              Overall Acceptance Rate6,164of23,696submissions,26%

            PDF Format

            View or Download as a PDF file.

            PDF

            eReader

            View online with eReader.

            eReader

            HTML Format

            View this article in HTML Format .

            View HTML Format