Development of Sign Language Recognition Application Using Deep Learning

  • Conference paper
  • First Online:
Next Generation of Internet of Things

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 445))

  • 672 Accesses

Abstract

Deaf and dumb people use a sign language that can only be communicated through hand gestures to express their ideas and views. This coded language is mainly used by people who have speech and/or hearing impairment. The sign language is constructed by various movement of hands, arms, legs, or facial expressions to express their opinions. Meanings are communicated for every movement or position of gesture. Hand gesture plays a significant role to make mother tongue of impairment people for daily communication. The captured image feature can be extracted to translate the hand gesture communication to text\voice format to minimize the gap between the deaf and normal persons. This work considers the images of sign numerals to classify the numbers 0–9 and the alphabets for A–Z (including space).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. ** CM, Omar Z, Jaward MH (2016) A mobile application of american sign language translation via image processing algorithms. In: IEEE Region 10 symposium (TENSYMP), Bali, Indonesia

    Google Scholar 

  2. Gao X, ** Y, Dou Q, Heng PA (2020) Automatic gesture recognition in robot-assisted surgery with reinforcement learning and tree search. ar**v preprint, ar**v:2002.08718

  3. Jiang D, Zheng Z, Li G, Sun Y, Kong J, Jiang G, Liu H (2019) Gesture recognition based on binocular vision. Clust Comput 22(6):13261–13271

    Article  Google Scholar 

  4. Vijayakumar KP, Nair A, Tomar N (2020) Hand gesture to speech and text conversion device. Int J Innov Technol Explor Eng (IJITEE) 9(5)

    Google Scholar 

  5. Kopuklu O, Gunduz A, Kose N, Rigoll G (2019) Real-time hand gesture detection and classification using convolutional neural networks. ar**v:1901.10323v3 [cs.CV]. Accessed 18 Oct 2019

    Google Scholar 

  6. Cheok MJ, Omar Z, Jaward MH (2017) A review of hand gesture and sign language recognition techniques. Springer, GmbH Germany

    Google Scholar 

  7. Liang RH, Ouhyoung M (1998) A real-time continuous gesture recognition system for sign language. In: IEEE international conference on automatic face and gesture recognition, pp 558–567, Japan

    Google Scholar 

  8. Gurav RM, Kadbe PK (2015) Real time finger tracking and contour detection for gesture recognition using OpenCV. In: 2015 international conference on industrial instrumentation and control (ICIC), College of Engineering Pune, India, May 28–30

    Google Scholar 

  9. Ravikiran J, Mahesh K, Mahishi S, Dheeraj R, Sudheender S, Pujari NV (2009) Finger detection for sign language recognition. In: Proceedings of the international multi conference of engineers and computer scientists 2009 Vol I IMECS 2009, March 18–20, 2009, Hong Kong

    Google Scholar 

  10. Rumyantsev O, Merati M, Ramachandran V (2012) Hand sign recognition through palm gesture and movement image processing. EE 368, Spring 2012

    Google Scholar 

  11. Rekha J, Bhattacharaya J, Majumder S (2011) Hand gesture recognition for sign language: a new hybrid approach. In: International conference on vision and pattern recognition, IPCV’11

    Google Scholar 

  12. Singha J, Roy A, Laskar RH (2018) Dynamic hand gesture recognition using vision-based approach for human–computer interaction. Neural Comput Appl 29(4):1129–1141

    Article  Google Scholar 

  13. **e R, Cao J (2016) Accelerometer-based hand gesture recognition by neural network and similarity matching. IEEE Sens J 16(11):4537–4545

    Article  Google Scholar 

  14. Venkatnarayan RH, Shahzad M (2018) Gesturerecognition using ambient light. Proc ACM Interact Mob Wearable Ubiquitous Technol 2(1):40, 1–40

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to N. R. Rajalakshmi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Rajalakshmi, N.R. (2023). Development of Sign Language Recognition Application Using Deep Learning. In: Kumar, R., Pattnaik, P.K., R. S. Tavares, J.M. (eds) Next Generation of Internet of Things. Lecture Notes in Networks and Systems, vol 445. Springer, Singapore. https://doi.org/10.1007/978-981-19-1412-6_24

Download citation

  • DOI: https://doi.org/10.1007/978-981-19-1412-6_24

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-19-1411-9

  • Online ISBN: 978-981-19-1412-6

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics

Navigation