Deep trajectory representation-based clustering for motion pattern extraction in videosBoyle, J. ORCID: https://orcid.org/0000-0002-5785-8046, Nawaz, T. and Ferryman, J. (2017) Deep trajectory representation-based clustering for motion pattern extraction in videos. In: 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), 29 Aug.-1 Sept. 2017, Lecce, Italy.
It is advisable to refer to the publisher's version if you intend to cite from this work. See Guidance on citing. Official URL: http://ieeexplore.ieee.org/document/8078509 Abstract/SummaryWe present a deep trajectory feature representation approach to aid trajectory clustering and motion pattern extraction in videos. The proposed feature representation includes the use of a neural network-based approach that uses the output of the smallest hidden layer of a trained autoencoder to encapsulate trajectory information. The trajectory features are then fed into a mean-shift clustering framework with an adaptive bandwidth parameter computation to yield dominant trajectory clusters. The corresponding motion patterns are extracted based on a distance minimization from the clusters’ centroids. We show the effectiveness of the proposed approach on challenging public datasets involving traffic as well non-traffic scenarios.
Deposit Details University Staff: Request a correction | Centaur Editors: Update this record |