This repository provides a PyTorch implementation of the paper Temporal Action Segmentation from Timestamp Supervision.
Tested with:
- PyTorch 1.1.0
- Python 3.6.10
- Download the data folder, which contains the features and the ground truth labels. (~30GB) (try to download it from here))
- Extract it so that you have the
data
folder in the same directory asmain.py
. - The three
.npy
files in 'data/' in this repository are the timestamp annotations. Put each one in corresponding ground truth folder. For example,./data/breakfast/groundTruth/
for Breakfast dataset. - To train the model run
python main.py --action=train --dataset=DS --split=SP
whereDS
isbreakfast
,50salads
orgtea
, andSP
is the split number (1-5) for 50salads and (1-4) for the other datasets. - The output of evaluation is saved in
result/
folder as an excel file. - The
models/
folder saves the trained model and theresults/
folder saves the predicted action labels of each video in test dataset.
Normally we get the prediction and evaluation after training and do not have to run this independently.
In case you want to test the saved model again by prediction and evaluation, please change the time_data
in main.py
and run
python main.py --action=predict --dataset=DS --split=SP
.
The model used in this paper is a refined MS-TCN model. Please refer to the paper MS-TCN: Multi-Stage Temporal Convolutional Network for Action Segmentation.
If you use the code, please cite
Zhe Li, Yazan Abu Farha and Juergen Gall.
Temporal Action Segmentation from Timestamp Supervision.
In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021