Skip to content

Code repository for BMVC paper LARNet: Latent Action Representation for Human Action Synthesis

License

Notifications You must be signed in to change notification settings

aayushjr/larnet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LARNet: Latent Action Representation for Human Action Synthesis

Code repository for BMVC paper LARNet: Latent Action Representation for Human Action Synthesis Paper available at [arxiv] (https://arxiv.org/pdf/2110.10899.pdf)

Project page with more information available in [CRCV webpage] (https://www.crcv.ucf.edu/research/projects/larnet-latent-action-representation-for-human-action-synthesis)

Demo available HERE

Description

This is an implementation of LARNet on NTU-RGB+D 60 dataset. It is built using the PyTorch library. The datasets have to be downloaded separately.

Setup

Please install following prerequisites: Python>=3.6, Pytorch>=1.6

Install remainig libraries from requirements.txt using

pip install -r requirements.txt

Download datasets and set the data_root_dir variable in train.py for each dataset.

Download I3D pretrained weights for Charades / Kinetics and setup the path in i3d_weights_path variable in train.py.

If you are resuming training, set the weights for generator and discriminator and uncomment weight load command for each in train.py.

Training

After doing setup and putting data/model path correctly in train.py, run using:

python train.py

If you enable tensorboard, it will save the generated video for each snapshot interval in the tensorboard log file.

About

Code repository for BMVC paper LARNet: Latent Action Representation for Human Action Synthesis

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages