The task of collaborative human pose forecasting stands for predicting the future poses of multiple interacting people, given those in previous frames. Predicting two people in interaction, instead of each separately, promises better performance, due to their body-body motion correlations. But the task has remained so far primarily unexplored. In this paper, we review the progress in human pose forecasting and provide an in-depth assessment of the single-person practices that perform best for 2-body collaborative motion forecasting. Our study confirms the positive impact of frequency input representations, space-time separable and fully-learnable interaction adjacencies for the encoding GCN and FC decoding. Other single-person practices do not transfer to 2-body, so the proposed best ones do not include hierarchical body modeling or attention-based interaction encoding. We further contribute a novel initialization procedure for the 2-body spatial interaction parameters of the encoder, which benefits performance and stability. Altogether, our proposed 2-body pose forecasting best practices yield a performance improvement of 21.9% over the state-of-the-art on the most recent ExPI dataset, whereby the novel initialization accounts for 3.5%.

Best Practices for 2-Body Pose Forecasting / Rahman, MUHAMMAD RAMEEZ UR; Scofano, Luca; DE MATTEIS, Edoardo; Flaborea, Alessandro; Sampieri, Alessio; Galasso, Fabio. - (2023). (Intervento presentato al convegno The 36th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023) tenutosi a Vancouver, Canada) [10.1109/cvprw59228.2023.00369].

Best Practices for 2-Body Pose Forecasting

Muhammad Rameez Ur Rahman;Luca Scofano;Edoardo De Matteis;Alessandro Flaborea;Alessio Sampieri;Fabio Galasso
2023

Abstract

The task of collaborative human pose forecasting stands for predicting the future poses of multiple interacting people, given those in previous frames. Predicting two people in interaction, instead of each separately, promises better performance, due to their body-body motion correlations. But the task has remained so far primarily unexplored. In this paper, we review the progress in human pose forecasting and provide an in-depth assessment of the single-person practices that perform best for 2-body collaborative motion forecasting. Our study confirms the positive impact of frequency input representations, space-time separable and fully-learnable interaction adjacencies for the encoding GCN and FC decoding. Other single-person practices do not transfer to 2-body, so the proposed best ones do not include hierarchical body modeling or attention-based interaction encoding. We further contribute a novel initialization procedure for the 2-body spatial interaction parameters of the encoder, which benefits performance and stability. Altogether, our proposed 2-body pose forecasting best practices yield a performance improvement of 21.9% over the state-of-the-art on the most recent ExPI dataset, whereby the novel initialization accounts for 3.5%.
2023
The 36th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023)
Human Pose Forecasting, Multi-Person Human Pose Forecasting, Graph Convolutional Neural Networks
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Best Practices for 2-Body Pose Forecasting / Rahman, MUHAMMAD RAMEEZ UR; Scofano, Luca; DE MATTEIS, Edoardo; Flaborea, Alessandro; Sampieri, Alessio; Galasso, Fabio. - (2023). (Intervento presentato al convegno The 36th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023) tenutosi a Vancouver, Canada) [10.1109/cvprw59228.2023.00369].
File allegati a questo prodotto
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1686540
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact