18 results on '"Cristiano Massaroni"'
Search Results
2. A New Descriptor for Keypoint-Based Background Modeling.
- Author
-
Danilo Avola, Marco Bernardi, Marco Cascio, Luigi Cinque, Gian Luca Foresti, and Cristiano Massaroni
- Published
- 2019
- Full Text
- View/download PDF
3. Master and Rookie Networks for Person Re-identification.
- Author
-
Danilo Avola, Marco Cascio, Luigi Cinque, Alessio Fagioli 0001, Gian Luca Foresti, and Cristiano Massaroni
- Published
- 2019
- Full Text
- View/download PDF
4. Combining Keypoint Clustering and Neural Background Subtraction for Real-time Moving Object Detection by PTZ Cameras.
- Author
-
Danilo Avola, Marco Bernardi, Luigi Cinque, Gian Luca Foresti, and Cristiano Massaroni
- Published
- 2018
- Full Text
- View/download PDF
5. A Machine Learning Approach for the Online Separation of Handwriting from Freehand Drawing.
- Author
-
Danilo Avola, Marco Bernardi, Luigi Cinque, Gian Luca Foresti, Marco Raoul Marini, and Cristiano Massaroni
- Published
- 2017
- Full Text
- View/download PDF
6. A multipurpose autonomous robot for target recognition in unknown environments.
- Author
-
Danilo Avola, Gian Luca Foresti, Luigi Cinque, Cristiano Massaroni, Gabriele Vitale, and Luca Lombardi
- Published
- 2016
- Full Text
- View/download PDF
7. Deep Temporal Analysis for Non-Acted Body Affect Recognition
- Author
-
Gian Luca Foresti, Cristiano Massaroni, Danilo Avola, Alessio Fagioli, and Luigi Cinque
- Subjects
FOS: Computer and information sciences ,3D skeleton ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Computer Science - Computer Vision and Pattern Recognition ,Non-acted affective computing ,02 engineering and technology ,Machine learning ,computer.software_genre ,Affect (psychology) ,Field (computer science) ,03 medical and health sciences ,0302 clinical medicine ,0202 electrical engineering, electronic engineering, information engineering ,Emotion recognition ,Set (psychology) ,Body movement ,business.industry ,Deep learning ,Novelty ,020207 software engineering ,Human-Computer Interaction ,Benchmark (computing) ,Long short-term memory (LSTM) ,Artificial intelligence ,Automatic emotion recognition ,Recurrent neural network (RNN) ,business ,computer ,030217 neurology & neurosurgery ,Software - Abstract
Affective computing is a field of great interest in many computer vision applications, including video surveillance, behaviour analysis, and human-robot interaction. Most of the existing literature has addressed this field by analysing different sets of face features. However, in the last decade, several studies have shown how body movements can play a key role even in emotion recognition. The majority of these experiments on the body are performed by trained actors whose aim is to simulate emotional reactions. These unnatural expressions differ from the more challenging genuine emotions, thus invalidating the obtained results. In this paper, a solution for basic non-acted emotion recognition based on 3D skeleton and Deep Neural Networks (DNNs) is provided. The proposed work introduces three majors contributions. First, unlike the current state-of-the-art in non-acted body affect recognition, where only static or global body features are considered, in this work also temporal local movements performed by subjects in each frame are examined. Second, an original set of global and time-dependent features for body movement description is provided. Third, to the best of out knowledge, this is the first attempt to use deep learning methods for non-acted body affect recognition. Due to the novelty of the topic, only the UCLIC dataset is currently considered the benchmark for comparative tests. On the latter, the proposed method outperforms all the competitors.
- Published
- 2022
8. 2-D Skeleton-Based Action Recognition via Two-Branch Stacked LSTM-RNNs
- Author
-
Cristiano Massaroni, Luigi Cinque, Marco Cascio, Gian Luca Foresti, Danilo Avola, and Emanuele Rodolà
- Subjects
Artificial neural network ,Computer science ,business.industry ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,02 engineering and technology ,Computer Science Applications ,Recurrent neural network ,Robustness (computer science) ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,RGB color model ,Action recognition ,020201 artificial intelligence & image processing ,Artificial intelligence ,Electrical and Electronic Engineering ,business - Abstract
Action recognition in video sequences is an interesting field for many computer vision applications, including behavior analysis, event recognition, and video surveillance. In this article, a method based on 2D skeleton and two-branch stacked Recurrent Neural Networks (RNNs) with Long Short-Term Memory (LSTM) cells is proposed. Unlike 3D skeletons, usually generated by RGB-D cameras, the 2D skeletons adopted in this article are reconstructed starting from RGB video streams, therefore allowing the use of the proposed approach in both indoor and outdoor environments. Moreover, any case of missing skeletal data is managed by exploiting 3D-Convolutional Neural Networks (3D-CNNs). Comparative experiments with several key works on KTH and Weizmann datasets show that the method described in this paper outperforms the current state-of-the-art. Additional experiments on UCF Sports and IXMAS datasets demonstrate the effectiveness of our method in the presence of noisy data and perspective changes, respectively. Further investigations on UCF Sports, HMDB51, UCF101, and Kinetics400 highlight how the combination between the proposed two-branch stacked LSTM and the 3D-CNN-based network can manage missing skeleton information, greatly improving the overall accuracy. Moreover, additional tests on KTH and UCF Sports datasets also show the robustness of our approach in the presence of partial body occlusions. Finally, comparisons on UT-Kinect and NTU-RGB+D datasets show that the accuracy of the proposed method is fully comparable to that of works based on 3D skeletons.
- Published
- 2020
9. Fusing Self-Organized Neural Network and Keypoint Clustering for Localized Real-Time Background Subtraction
- Author
-
Marco Bernardi, Gian Luca Foresti, Danilo Avola, Cristiano Massaroni, and Luigi Cinque
- Subjects
Automated ,Neural Networks ,Computer Networks and Communications ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Datasets as Topic ,02 engineering and technology ,Pattern Recognition ,Pattern Recognition, Automated ,Computer ,Computer-Assisted ,background modeling ,Theoretical ,foreground detection ,Models ,Image Interpretation, Computer-Assisted ,0202 electrical engineering, electronic engineering, information engineering ,Cluster Analysis ,Humans ,Computer vision ,Cluster analysis ,Image Interpretation ,Foreground detection ,Background subtraction ,Artificial neural network ,business.industry ,clustering ,Self-organized neural network ,Models, Theoretical ,Neural Networks, Computer ,Subtraction Technique ,020207 software engineering ,General Medicine ,Object detection ,Key (cryptography) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business - Abstract
Moving object detection in video streams plays a key role in many computer vision applications. In particular, separation between background and foreground items represents a main prerequisite to carry out more complex tasks, such as object classification, vehicle tracking, and person re-identification. Despite the progress made in recent years, a main challenge of moving object detection still regards the management of dynamic aspects, including bootstrapping and illumination changes. In addition, the recent widespread of Pan–Tilt–Zoom (PTZ) cameras has made the management of these aspects even more complex in terms of performance due to their mixed movements (i.e. pan, tilt, and zoom). In this paper, a combined keypoint clustering and neural background subtraction method, based on Self-Organized Neural Network (SONN), for real-time moving object detection in video sequences acquired by PTZ cameras is proposed. Initially, the method performs a spatio-temporal tracking of the sets of moving keypoints to recognize the foreground areas and to establish the background. Then, it adopts a neural background subtraction, localized in these areas, to accomplish a foreground detection able to manage bootstrapping and gradual illumination changes. Experimental results on three well-known public datasets, and comparisons with different key works of the current literature, show the efficiency of the proposed method in terms of modeling and background subtraction.
- Published
- 2020
10. Adaptive bootstrapping management by keypoint clustering for background initialization
- Author
-
Gian Luca Foresti, Luigi Cinque, Marco Bernardi, Danilo Avola, and Cristiano Massaroni
- Subjects
DBSCAN ,Computer science ,Background initialization ,Background modeling ,Keypoint clustering ,Foreground detection ,Bootstrapping ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Initialization ,Software ,Signal Processing ,Artificial Intelligence ,02 engineering and technology ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Cluster analysis ,Background subtraction ,business.industry ,020207 software engineering ,Bootstrapping (linguistics) ,Pattern recognition ,Feature (computer vision) ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Noise (video) ,Artificial intelligence ,business - Abstract
The availability of a background model that describes the scene is a prerequisite for many computer vision applications. In several situations, the model cannot be easily generated when the background contains some foreground objects (i.e., bootstrapping problem). In this letter, an Adaptive Bootstrapping Management (ABM) method, based on keypoint clustering, is proposed to model the background on video sequences acquired by mobile and static cameras. First, keypoints are detected on each frame by the A-KAZE feature extractor, then Density-Based Spatial Clustering of Application with Noise (DBSCAN) is used to find keypoint clusters. These clusters represent the candidate regions of foreground elements inside the scene. The ABM method manages the scene changes generated by foreground elements, both in the background model initialization, managing the bootstrapping problem, and in the background model updating. Moreover, it achieves good results with both mobile and static cameras and it requires a small number of frames to initialize the background model.
- Published
- 2017
11. A keypoint-based method for background modeling and foreground detection using a PTZ camera
- Author
-
Luigi Cinque, Danilo Avola, Gian Luca Foresti, Daniele Pannone, and Cristiano Massaroni
- Subjects
Background reconstruction ,PTZ cameras ,Vehicle tracking system ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Initialization ,02 engineering and technology ,Tracking (particle physics) ,Background updating ,Grid strategy ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Background estimation ,Background modeling ,Foreground detection ,Spatio-temporal tracking of keypoints ,Software ,Signal Processing ,Computer vision ,Background subtraction ,background estimation ,background modeling ,background reconstruction ,background updating ,foreground detection ,grid strategy ,ptz cameras ,spatio-temporal tracking of keypoints ,business.industry ,Cognitive neuroscience of visual object recognition ,020207 software engineering ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Scale (map) - Abstract
Background modeling and foreground detection for PTZ cameras.Grid strategy for spatio-temporal tracking of keypoints.Background initialization and estimation.Background updating and reconstruction.Panoramic background reconstruction. Display Omitted The automatic scene analysis is still a topic of great interest in computer vision due to the growing possibilities provided by the increasingly sophisticated optical cameras. The background modeling, including its initialization and its updating, is a crucial aspect that can play a main role in a wide range of application domains, such as vehicle tracking, person re-identification and object recognition. In any case, many challenges still remain partially unsolved, including camera movements (i.e., pan/tilt), scale changes (i.e., zoom-in/zoom-out) and deletion of the initial foreground elements from the background model. This paper describes a method for background modeling and foreground detection able to address all the mentioned challenges. In particular, the proposed method uses a spatio-temporal tracking of sets of keypoints to distinguish the background from the foreground. It analyses these sets by a grid strategy to estimate both camera movements and scale changes. The same sets are also used to construct a panoramic background model and to delete the possible initial foreground elements from it. Experiments carried out on some challenging videos from three different datasets (i.e., PBI, VOT and Airport MotionSeg) demonstrate the effectiveness of the method on PTZ cameras. Other videos from a further dataset (i.e., FBMS) have been used to measure the accuracy of the proposed method with respect to some key works of the current state-of-the-art. Finally, some videos from another dataset (i.e., SBI) have been used to test the method on stationary cameras.
- Published
- 2017
12. Online separation of handwriting from freehand drawing using extreme learning machines
- Author
-
Cristiano Massaroni, Danilo Avola, Gian Luca Foresti, Marco Bernardi, and Luigi Cinque
- Subjects
Handwriting ,Computer science ,Extreme learning machines ,Freehand drawing ,Online separation algorithm ,Sketch-based interfaces ,Software ,Media Technology ,Hardware and Architecture ,Computer Networks and Communications ,02 engineering and technology ,Machine learning ,computer.software_genre ,Field (computer science) ,Task (project management) ,Set (abstract data type) ,0202 electrical engineering, electronic engineering, information engineering ,Extreme learning machine ,business.industry ,020207 software engineering ,Sketch ,Key (cryptography) ,Artificial intelligence ,business ,computer - Abstract
Online separation between handwriting and freehand drawing is still an active research area in the field of sketch-based interfaces. In the last years, most approaches in this area have been focused on the use of statistical separation methods, which have achieved significant results in terms of performance. More recently, Machine Learning (ML) techniques have proven to be even more effective by treating the separation problem like a classification task. Despite this, also in the use of these techniques several aspects can be still considered open problems, including: 1) the trade-off between separation performance and training time; 2) the separation of handwriting from different types of freehand drawings. To address the just reported drawbacks, in this paper a novel separation algorithm based on a set of original features and an Extreme Learning Machine (ELM) is proposed. Extensive experiments on a wide range of sketched schemes (i.e., text and graphical symbols), more numerous than those usually tested in any key work of the current literature, have highlighted the effectiveness of the proposed approach. Finally, measurements on accuracy and speed of computation, during both training and testing stages, have shown that the ELM can be considered, in this research area, the better choice even if compared with other popular ML techniques.
- Published
- 2019
13. Exploiting Recurrent Neural Networks and Leap Motion Controller for the Recognition of Sign Language and Semaphoric Hand Gestures
- Author
-
Cristiano Massaroni, Danilo Avola, Gian Luca Foresti, Luigi Cinque, and Marco Bernardi
- Subjects
Long Short Term Memory (LSTM) ,American Sign Language ,Computer science ,Feature vector ,Speech recognition ,Feature extraction ,Hand gesture recognition ,Leap Motion Controller (LMC) ,Recurrent Neural Network (RNN) ,semaphoric gestures ,sign language ,02 engineering and technology ,Sign language ,Gesture recognition ,assistive technology ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Electrical and Electronic Engineering ,Set (psychology) ,feature extraction ,thumb ,language.human_language ,Computer Science Applications ,Recurrent neural network ,three-dimensional displays ,Signal Processing ,language ,hand gesture recognition ,020201 artificial intelligence & image processing ,Gesture - Abstract
Hand gesture recognition is still a topic of great interest for the computer vision community. In particular, sign language and semaphoric hand gestures are two foremost areas of interest due to their importance in human–human communication and human–computer interaction, respectively. Any hand gesture can be represented by sets of feature vectors that change over time. Recurrent neural networks (RNNs) are suited to analyze this type of set thanks to their ability to model the long-term contextual information of temporal sequences. In this paper, an RNN is trained by using as features the angles formed by the finger bones of the human hands. The selected features, acquired by a leap motion controller sensor, are chosen because the majority of human hand gestures produce joint movements that generate truly characteristic corners. The proposed method, including the effectiveness of the selected angles, was initially tested by creating a very challenging dataset composed by a large number of gestures defined by the American sign language. On the latter, an accuracy of over 96% was achieved. Afterwards, by using the Shape Retrieval Contest (SHREC) dataset, a wide collection of semaphoric hand gestures, the method was also proven to outperform in accuracy competing approaches of the current literature.
- Published
- 2019
14. Master and Rookie Networks for Person Re-identification
- Author
-
Alessio Fagioli, Cristiano Massaroni, Danilo Avola, Gian Luca Foresti, Luigi Cinque, and Marco Cascio
- Subjects
Computer science ,business.industry ,Open problem ,Deep learning ,Feature extraction ,020207 software engineering ,02 engineering and technology ,Machine learning ,computer.software_genre ,Convolutional neural network ,Field (computer science) ,deep learning ,feature extraction ,person re-identification ,Person re-identification ,Discriminative model ,Convergence (routing) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Architecture ,business ,computer - Abstract
Recognizing different visual signatures of people across non-overlapping cameras is still an open problem of great interest for the computer vision community, especially due to its importance in automatic video surveillance on large-scale environments. A main aspect of this application field, known as person re-identification (re-id), is the feature extraction step used to define a robust appearance of a person. In this paper, a novel two-branch Convolutional Neural Network (CNN) architecture for person re-id in video sequences is proposed. A pre-trained branch, called Master, leads the learning phase of the other un-trained branch, called Rookie. Using this strategy, the Rookie network is able to learn complementary features with respect to those computed by the Master network, thus obtaining a more discriminative model. Extensive experiments on two popular challenging re-id datasets have shown increasing performance in terms of convergence speed as well as accuracy in comparison to standard models, thus providing an alternative and concrete contribution to the current re-id state-of-the-art.
- Published
- 2019
15. Feature-based SLAM algorithm for small scale UAV with nadir view
- Author
-
Alessio Fagioli, Daniele Pannone, Cristiano Massaroni, Luigi Cinque, Gian Luca Foresti, and Danilo Avola
- Subjects
0209 industrial biotechnology ,Computer science ,Computation ,GPSS ,ComputerApplications_COMPUTERSINOTHERSYSTEMS ,A-KAZE ,02 engineering and technology ,Simultaneous localization and mapping ,UAVs ,020901 industrial engineering & automation ,Data acquisition ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Search and rescue ,computer.programming_language ,business.industry ,Mosaicking ,SLAM algorithm ,Global Positioning System ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Change detection - Abstract
Small-scale Unmanned Aerial Vehicles (UAVs) have recently been used in several application areas, including search and rescue operations, precision agriculture, and environmental monitoring. Telemetry data, acquired by GPSs, plays a key role in supporting activities in areas like those just reported. In particular, this data is often used for the real-time computation of UAVs paths and heights, which are basic pre-requisites for many tasks. In some cases, however, the GPS sensors can lose their satellite connection, thus making the telemetry data acquisition impossible. This paper presents a feature-based Simultaneous Localisation and Mapping (SLAM) algorithm for small-scale UAVs with nadir view. The proposed algorithm allows to know the travelled route as well as the flight height by using both a calibration step and visual features extracted from the acquired images. Due to the novelty of the proposed algorithm no comparisons with other methods are reported. Anyway, extensive experiments on the recently released UAV Mosaicking and Change Detection (UMCD) dataset have shown the effectiveness and robustness of the proposed algorithm. The latter and the dataset can be used as baseline for future research in this application area.
- Published
- 2019
16. A new descriptor for Keypoint-Based background modeling
- Author
-
Marco Bernardi, Marco Cascio, Danilo Avola, Cristiano Massaroni, Luigi Cinque, and Gian Luca Foresti
- Subjects
Foreground detection ,keypoint clustering ,Computer science ,business.industry ,Pooling ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,dynamic background ,020207 software engineering ,02 engineering and technology ,Background modeling ,Dynamic background ,Keypoint clustering ,RootSIFT ,background modeling ,Robustness (computer science) ,foreground detection ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Invariant (mathematics) ,Cluster analysis ,business - Abstract
Background modeling is a preliminary task for many computer vision applications, describing static elements of a scene and isolating foreground ones. Defining a robust background model of uncontrolled environments is a current challenge since the model must manage many issues, e.g., moving cameras, dynamic background, bootstrapping, shadows, and illumination changes. Recently, methods based on keypoint clustering have shown remarkable robustness especially in bootstrapping and camera movements, highlighting however limitations in the analysis of dynamic background (i.e., trees blowing in the wind or gushing fountains). In this paper, an innovative combination between the RootSIFT descriptor and an average pooling is proposed in a keypoint clustering method for real-time background modeling and foreground detection. Compared to renowned descriptors, such as A-KAZE, this combination is invariant to small local changes in the scene, thus resulting more robust in dynamic background cases. Results, obtained on experiments carried out on two benchmark datasets, demonstrate how the proposed solution improves the previous keypoint-based models and overcomes several works of the current state-of-the-art.
- Published
- 2019
17. A multipurpose autonomous robot for target recognition in unknown environments
- Author
-
Cristiano Massaroni, Luca Lombardi, Danilo Avola, Gian Luca Foresti, Luigi Cinque, and Gabriele Vitale
- Subjects
0209 industrial biotechnology ,Robot kinematics ,Engineering ,Robot calibration ,business.industry ,Mobile robot ,02 engineering and technology ,Simultaneous localization and mapping ,Multipurpose Autonomous Robot ,Unknown Environments ,Autonomous robot ,Mobile robot navigation ,Robot control ,020901 industrial engineering & automation ,0202 electrical engineering, electronic engineering, information engineering ,Robot ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business - Abstract
In recent years, the technological improvements of consumer robots, in terms of processing capacity and sensors, are enabling an ever-increasing number of researchers to quickly develop both scale prototypes and alternative low cost solutions. In these contexts, a critical aspect is the design of ad-hoc algorithms according to the features of the available hardware. This paper proposes a prototype of an autonomous robot for mapping unknown environments and recognizing target objects. During the setup phase one or more target objects are shown to the RGB camera of the robot which, for each of them, extracts and stores a set of A-KAZE features. Afterwards, the robot adopts the ultrasonic distance measurement and the RGB stream to map the whole environment and search a set of A-KAZE features matchable with those previously acquired. The paper also reports both preliminary tests carried out on a reference indoor environment and a case study performed in an outdoor one that validate the proposed system.
- Published
- 2016
18. Combining keypoint clustering and neural background subtraction for real-time moving object detection by PTZ cameras
- Author
-
Cristiano Massaroni, Luigi Cinque, Danilo Avola, Marco Bernardi, and Gian Luca Foresti
- Subjects
Background subtraction ,Foreground Detection ,Keypoint Clustering ,Neural Background Subtraction ,Moving Objects ,PTZ Cameras ,Computer science ,business.industry ,020207 software engineering ,02 engineering and technology ,Object detection ,03 medical and health sciences ,0302 clinical medicine ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Artificial intelligence ,Cluster analysis ,business ,030217 neurology & neurosurgery
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.