The interest in 3D dynamical tracking is growing in fields such as robotics, biology and fluid dynamics. Recently, a major source of progress in 3D tracking has been the study of collective behaviour in biological systems, where the trajectories of individual animals moving within large and dense groups need to be reconstructed to understand the behavioural interaction rules. Experimental data in this field are generally noisy and at low spatial resolution, so that individuals appear as small featureless objects and trajectories must be retrieved by making use of epipolar information only. Moreover, optical occlusions often occur: in a multicamera system one or more objects become indistinguishable in one view, potentially jeopardizing the conservation of identity over long-Time trajectories. The most advanced 3D tracking algorithms overcome optical occlusions making use of set-cover techniques, which however have to solve NP-hard optimization problems. Moreover, current methods are not able to cope with occlusions arising from actual physical proximity of objects in 3D space. Here, we present a new method designed to work directly in 3D space and time, creating (3D + 1) clouds of points representing the full spatio-Temporal evolution of the moving targets. We can then use a simple connected components labeling routine, which is linear in time, to solve optical occlusions, hence lowering from NP to P the complexity of the problem. Finally, we use normalized cut spectral clustering to tackle 3D physical proximity. © 2016 SciTePress. All rights reserved.
Towards a tracking algorithm based on the clustering of spatio-Temporal clouds of points / Cavagna, Andrea; Creato, Chiara; Del Castello, Lorenzo; Melillo, Stefania; Parisi, Leonardo; Viale, Massimiliano. - 3:(2016), pp. 681-687. (Intervento presentato al convegno 11th Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications tenutosi a Roma) [10.5220/0005770106790685].
Towards a tracking algorithm based on the clustering of spatio-Temporal clouds of points
Andrea Cavagna;Stefania Melillo;Leonardo Parisi
;Massimiliano Viale
2016
Abstract
The interest in 3D dynamical tracking is growing in fields such as robotics, biology and fluid dynamics. Recently, a major source of progress in 3D tracking has been the study of collective behaviour in biological systems, where the trajectories of individual animals moving within large and dense groups need to be reconstructed to understand the behavioural interaction rules. Experimental data in this field are generally noisy and at low spatial resolution, so that individuals appear as small featureless objects and trajectories must be retrieved by making use of epipolar information only. Moreover, optical occlusions often occur: in a multicamera system one or more objects become indistinguishable in one view, potentially jeopardizing the conservation of identity over long-Time trajectories. The most advanced 3D tracking algorithms overcome optical occlusions making use of set-cover techniques, which however have to solve NP-hard optimization problems. Moreover, current methods are not able to cope with occlusions arising from actual physical proximity of objects in 3D space. Here, we present a new method designed to work directly in 3D space and time, creating (3D + 1) clouds of points representing the full spatio-Temporal evolution of the moving targets. We can then use a simple connected components labeling routine, which is linear in time, to solve optical occlusions, hence lowering from NP to P the complexity of the problem. Finally, we use normalized cut spectral clustering to tackle 3D physical proximity. © 2016 SciTePress. All rights reserved.File | Dimensione | Formato | |
---|---|---|---|
Cavagna_towards-a-tracking_2016.pdf
solo gestori archivio
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
837 kB
Formato
Adobe PDF
|
837 kB | Adobe PDF | Contatta l'autore |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.