Skip to content

Latest commit

 

History

History
68 lines (40 loc) · 2.83 KB

File metadata and controls

68 lines (40 loc) · 2.83 KB

Multimodal-Gesture-Recognition-with-LSTMs-and-CTC

This repository contains code for my diploma thesis MULTIMODAL GESTURE RECOGNITION WITH THE USE OF DEEP LEARNING.

Overview

An end-to-end system that performs temporal recognition of gesture sequences using speech and skeletal input. The model combines two LSTM networks with a CTC output layer that spot and classify gestures from two continuous streams.

The basic modules of the model are two bidirectional LSTMs. The first extracts features from speech and the second from skeletal data. Then another bidirectional LSTM combines the uni-modal features and performs the gesture recognition.

Here we provide code for:

a) A BLSTM network for speech recognition.

b) A BLSTM network for skeletal recognition.

c) A BLSTM network that fuses the two uni-modal networks.

d) An implementation of the CTC loss output.

e) Decoders for the different networks.

f) Sample code for skeletal and speech feature extraction.

We used keras and tensorflow to build our model.

This project was built for the ChaLearn 2013 dataset. We trained and tested the model using the dataset of the challenge. The data can be downloaded here. http://sunai.uoc.edu/chalearn/#tabs-2

This model achieves 94% accuracy on the test set of the ChaLearn 2013 challenge.

Usage

In order to train the models provided here you need to preprocess the data:

  1. MFCC features need to be extracted from the audio .wav files. We used 13 MFCC features as well as the first and second order derivatives (total 39 features). We used the HTK toolkit to extract the features. Here we just provide the configuration file for HCopy (the feature extraction tool for HTK). If you want to use HTK for this purpose you can find it here http://htk.eng.cam.ac.uk/

  2. Once the MFCC features are extracted just put the training data all in one big csv file along with the labels (same for the validation and test data) and you are ready to train the speech lstm network.

  3. For the skeletal features you should provide the joint positions for each file in one csv file each and run the following scripts:

    a) extract_activity_feats.py

    b) gather_skeletal.py

    c) skeletal_feature_extraction.py

  4. Run util/mix_data.py to mix some of the dev data to the training set.

  5. Now you are ready to train the skeletal lstm network.

  6. Once both networks are trained you can train the multimodal fusion network.

  7. Use the sequence_decoding.py script to evaluate the trained model with test data.

The training of the complete system takes approximately 100 hours in an nvidia 1060 gtx.

Requirements

Run pip install -r requirements.txt to install the requirements.