Retrieval of spatial–temporal motion topics from 3D skeleton data

Qianhui Men, Howard Leung*

*Corresponding author for this work

Research output: Contribution to journalArticle (Academic Journal)peer-review

5 Citations (Scopus)

Abstract

Retrieval of a specific human motion from 3D skeleton data is intractable because of its articulated complexity. We propose a context-based motion document formation method to reflect geometric variations by calculating covariance descriptors among skeletal joint locations and joint relative distances, and temporal variations by performing a coarse-to-fine segmentation on the motion sequence. The descriptors of query motion traverse all the motion categories to lock its motion words, which can be regarded as the basic units of a motion document. The discrete motion words of different spatiotemporal descriptors are also mapped to divergent index ranges to add prior knowledge of motion with temporal order to latent Dirichlet allocation (LDA). The similarity matching is based on motion-topic distributions from LDA with semantic meanings. The experiments on public datasets show the effectiveness and robustness of the proposed method over existing models.

Original languageEnglish
Pages (from-to)973-984
Number of pages12
JournalVisual Computer
Volume35
Issue number6-8
DOIs
Publication statusPublished - 1 Jun 2019

Bibliographical note

Publisher Copyright:
© 2019, Springer-Verlag GmbH Germany, part of Springer Nature.

Keywords

  • Latent Dirichlet allocation
  • Motion documents
  • Skeleton-based motion retrieval
  • Spatial–temporal descriptors

Fingerprint

Dive into the research topics of 'Retrieval of spatial–temporal motion topics from 3D skeleton data'. Together they form a unique fingerprint.

Cite this