Abstract: Communication is an integral part of human life. But for people who are mute & hearing impaired, communication is a challenge. To understand them, one has to either learn their language i.e. sign language or finger language. The system proposed in this project aims at tackling this problem to some extent. In this paper, the motivation was to create an object tracking application to interact with the computer, and develop a virtual human computer interaction device. The motivation behind this system is two-fold. It has two modes of operation: Teach and Learn. The project uses a webcam to recognize the hand positions and sign made using contour recognition  and outputs the Sign Language in PC onto the gesture made. This will convert the gesture captured via webcam into audio output which will make normal people understand what exactly is being conveyed. Thus our project Sign Language to Speech Converter aims to convert the Sign Language into text and audio.
Keywords: Gesture recognition, image processing, Edge detection, Grey scale image