Need help with sign-language-alphabet-recognizer?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

181 Stars 116 Forks MIT License 19 Commits 25 Opened issues


Simple sign language alphabet recognizer using Python, openCV and tensorflow for training Inception model (CNN classifier).

Services available


Need anything else?

Contributors list

SLR Alphabet Recognizer

This project is a sign language alphabet recognizer using Python, openCV and tensorflow for training InceptionV3 model, a convolutional neural network model for classification.

The framework used for the CNN implementation can be found here:

Simple transfer learning with an Inception V3 architecture model by xuetsing

The project contains the dataset (1Go). If you are only interested in code, you better copy/paste the few files than cloning the entire project.

You can find the demo here



This project uses python 3.5 and the PIP following packages: * opencv * tensorflow * matplotlib * numpy

See requirements.txt and Dockerfile for versions and required APT packages

Using Docker

docker build -t hands-classifier .
docker run -it hands-classifier bash

Install using PIP

pip3 install -r requirements.txt


To train the model, use the following command (see framework github link for more command options):

python3 \
  --bottleneck_dir=logs/bottlenecks \
  --how_many_training_steps=2000 \
  --model_dir=inception \
  --summaries_dir=logs/training_summaries/basic \
  --output_graph=logs/trained_graph.pb \
  --output_labels=logs/trained_labels.txt \
If you're using the provided dataset, it may take up to three hours.


To test classification, use the following command:

python3 path/to/image.jpg

Using webcam (demo)

To use webcam, use the following command:

Your hand must be inside the rectangle. Keep position to write word, see demo for deletions.

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.