Prediction of human actions in assembly process by a spatial-temporal end-to-end learning model

Zhujun Zhang, Weitian Wang, Yi Chen, Yunyi Jia, Gaoliang Peng

Research output: Contribution to journalConference articlepeer-review

3 Scopus citations


It's important to predict human actions in the industry assembly process. Foreseeing future actions before they happened is an essential part for flexible human-robot collaboration and crucial to safety issues. Vision-based human action prediction from videos provides intuitive and adequate knowledge for many complex applications. This problem can be interpreted as deducing the next action of people from a short video clip. The history information needs to be considered to learn these relations among time steps for predicting the future steps. However, it is difficult to extract the history information and use it to infer the future situation with traditional methods. In this scenario, a model is needed to handle the spatial and temporal details stored in the past human motions and construct the future action based on limited accessible human demonstrations. In this paper, we apply an autoencoder-based deep learning framework for human action construction, merging into the RNN pipeline for human action prediction. This contrasts with traditional approaches which use hand-crafted features and different domain outputs. We implement the proposed framework on a model vehicle seat assembly task. Our experiment results indicate that the proposed model is effective in capturing the historical details that are necessary for future human action prediction. In addition, the proposed model synthesizes the prior information from human demonstrations and generates the corresponding future action by those spatial-temporal features successfully.

Original languageEnglish
JournalSAE Technical Papers
Issue numberApril
StatePublished - 2 Apr 2019
EventSAE World Congress Experience, WCX 2019 - Detroit, United States
Duration: 9 Apr 201911 Apr 2019


Dive into the research topics of 'Prediction of human actions in assembly process by a spatial-temporal end-to-end learning model'. Together they form a unique fingerprint.

Cite this