首页 > > 详细

Assignment 3 Sign Language Image Classification using Deep Learning

 Assignment 3

Sign Language Image Classification using Deep Learning
In this assignment you will implement different deep learning networks to classify images of hands in poses that correspond to letters in American Sign Language. The dataset is contained in the assignment zip file, along with some images and a text file describing the dataset. It is similar in many ways to other MNIST datasets.
 
The American Sign Language letter database of hand gestures represent a multi-class problem with 24 classes of letters (excluding J and Z which require motion). The dataset format is patterned to match closely with the classic MNIST. Each training and test case represents a label (0-25) as a one-to-one map for each alphabetic letter A-Z (and no cases for 9=J or 25=Z because of gesture motions). The training data (27,455 cases) and test data (7172 cases) are approximately half the size of the standard MNIST but otherwise similar with a header row of label,  𝑝𝑖𝑥𝑒𝑙1
 , 𝑝𝑖𝑥𝑒𝑙2
 …. 𝑝𝑖𝑥𝑒𝑙784
  which represent a single 28x28 pixel image with grayscale values between 0-255.
 
Scenario
A client is interested in having you (or rather the company that you work for) investigate whether it is possible to develop an app that would enable American sign language to be translated for people that do not sign, or those that sign in different languages/styles. They have provided you with a labelled data of images related to signs (hand positions) that represent individual letters in order to do a preliminary test of feasibility.
 
Your manager has asked you to do this feasibility assessment, but subject to a constraint on the computational facilities available. More specifically, you are asked to do no more than 100 training runs in total (including all models and hyperparameter settings that you consider).
 
The task requires you to create a Jupyter Notebook to perform 22 steps. These steps involve loading the dataset, fixing data problems, converting labels to one-hot encoding, plotting sample images, creating, training, and evaluating two sequential models with 20 Dense layers with 100 neurons each, checking for better accuracy using MC Dropout, retraining the first model with performance scheduling, evaluating both models, using transfer learning to create a new model using pre-trained weights, freezing the weights of the pre-trained layers, adding new Dense layers, training and evaluating the new model, predicting and converting sign language to text using the best model.
 
IMPORTANT
Train all the models locally on your own machine. No model training should occur on Gradescope (GS).
After completing the training, upload the trained models' h5 files and their training histories along with your notebook to GS.
best_dnn_bn_model.h5
best_dnn_bn_perf_model.h5
best_dnn_selu_model.h5
best_mobilenet_model.h5
history1
history2
history1_perf
historymb
To avoid any confusion and poor training on GS, please remember to comment out the training code in your notebook before uploading it to GS.
联系我们
  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-21:00
  • 微信:codinghelp
热点标签

联系我们 - QQ: 99515681 微信:codinghelp
程序辅导网!