With the advent of ever-fast computing, real-time processing of visual data has been gaining importance in the field of surveillance. Also, automated decision-making by visual surveillance systems has been contributing to a huge leap in the capability of such systems, and of course their relevance in social security.
This special issue aimed to discuss cloud-based architectures of surveillance frameworks as a service. Such systems, especially when deployed to work in real-time, are required to be fast, efficient, and sustainable with a varying load of visual data.
Four papers were selected for inclusion in this special issue.
Wyzykowski et al. present an approach to synthesize realistic, multiresolution and multisensor fingerprints. Based in Anguli, a handcrafted fingerprint generator, they were able to obtain dynamic ridge maps with sweat pores and scratches. Then, a CycleGAN network was trained to transform these maps into realistic fingerprints. Unlike other CNN-based works, this framework is able to generate images with different resolutions and styles for the same identity. Finally, authors conducted a human perception analysis where 60 volunteers could hardly differentiate between real and high-resolution synthesized fingerprints.
Pawar and Attar address the problem of detection and localization of anomalies in surveillance videos, using pipelined deep autoencoders and one-class learning. Specifically, they used a convolutional autoencoder and sequence-to-sequence long short-term memory autoencoder in a pipelined fashion for spatial and temporal learning of the videos, respectively. In this setting, the principle of one-class classification for training the model on normal data and testing it on anomalous testing data was followed.
Tawfik Mohammed et al. describe a framework, implemented in a RAD (Rapid Application Development) paradigm, for performing iris recognition tests, based in the well-known Daugman's processing chain. They start by segmenting the iris ring using the Integro-differential operator, along with an edge-based Hough transform to isolate eyelids and eyelashes. After the normalization of the data (pseudo-polar domain), the features are encoded using 1D log Gabor kernel. Finally, the matching step is carried out using the Hamming distance.
Barra et al. describe an approach for automated head pose estimation that stems from a previous Partitioned Iterated Function Systems (PIFS)-based approach providing state-of-the-art accuracy with high computing cost and improve it by means of two regression models, namely Gradient Boosting Regressor and Extreme Gradient Boosting Regressor, achieving much faster response and an even lower mean absolute error on the yaw and roll axis, as shown by experiments conducted on the BIWI and AFLW2000 datasets.
扫码关注我们
求助内容:
应助结果提醒方式:
