123 results on '"Doulamis, Nikolaos"'
Search Results
2. Simultaneous Precise Localization And Classification of metal rust defects for robotic-driven maintenance and prefabrication using residual attention U-Net
- Author
-
Katsamenis, Iason, Doulamis, Nikolaos, Doulamis, Anastasios, Protopapadakis, Eftychios, and Voulodimos, Athanasios
- Published
- 2022
- Full Text
- View/download PDF
3. The Plegma dataset: Domestic appliance-level and aggregate electricity demand with metadata from Greece.
- Author
-
Athanasoulias, Sotirios, Guasselli, Fernanda, Doulamis, Nikolaos, Doulamis, Anastasios, Ipiotis, Nikolaos, Katsari, Athina, Stankovic, Lina, and Stankovic, Vladimir
- Subjects
ELECTRIC power consumption ,AGGREGATE demand ,SMART meters ,CONSUMPTION (Economics) ,SMART power grids ,MACHINE learning ,ENERGY consumption ,METADATA ,ACQUISITION of data - Abstract
The growing availability of smart meter data has facilitated the development of energy-saving services like demand response, personalized energy feedback, and non-intrusive-load-monitoring applications, all of which heavily rely on advanced machine learning algorithms trained on energy consumption datasets. To ensure the accuracy and reliability of these services, real-world smart meter data collection is crucial. The Plegma dataset described in this paper addresses this need bfy providing whole- house aggregate loads and appliance-level consumption measurements at 10-second intervals from 13 different households over a period of one year. It also includes environmental data such as humidity and temperature, building characteristics, demographic information, and user practice routines to enable quantitative as well as qualitative analysis. Plegma is the first high-frequency electricity measurements dataset in Greece, capturing the consumption behavior of people in the Mediterranean area who use devices not commonly included in other datasets, such as AC and electric-water boilers. The dataset comprises 218 million readings from 88 installed meters and sensors. The collected data are available in CSV format. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Spatio-temporal summarization of dance choreographies
- Author
-
Rallis, Ioannis, Doulamis, Nikolaos, Doulamis, Anastasios, Voulodimos, Athanasios, and Vescoukis, Vassilios
- Published
- 2018
- Full Text
- View/download PDF
5. Automatic crack detection for tunnel inspection using deep learning and heuristic image post-processing
- Author
-
Protopapadakis, Eftychios, Voulodimos, Athanasios, Doulamis, Anastasios, Doulamis, Nikolaos, and Stathaki, Tania
- Published
- 2019
- Full Text
- View/download PDF
6. Ensemble Learning for Blending Gridded Satellite and Gauge-Measured Precipitation Data.
- Author
-
Papacharalampous, Georgia, Tyralis, Hristos, Doulamis, Nikolaos, and Doulamis, Anastasios
- Subjects
BOOSTING algorithms ,MACHINE learning ,ARTIFICIAL neural networks ,INDEPENDENT variables ,RANDOM forest algorithms ,DATABASES - Abstract
Regression algorithms are regularly used for improving the accuracy of satellite precipitation products. In this context, satellite precipitation and topography data are the predictor variables, and gauged-measured precipitation data are the dependent variables. Alongside this, it is increasingly recognised in many fields that combinations of algorithms through ensemble learning can lead to substantial predictive performance improvements. Still, a sufficient number of ensemble learners for improving the accuracy of satellite precipitation products and their large-scale comparison are currently missing from the literature. In this study, we work towards filling in this specific gap by proposing 11 new ensemble learners in the field and by extensively comparing them. We apply the ensemble learners to monthly data from the PERSIANN (Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks) and IMERG (Integrated Multi-satellitE Retrievals for GPM) gridded datasets that span over a 15-year period and over the entire contiguous United States (CONUS). We also use gauge-measured precipitation data from the Global Historical Climatology Network monthly database, version 2 (GHCNm). The ensemble learners combine the predictions of six machine learning regression algorithms (base learners), namely the multivariate adaptive regression splines (MARS), multivariate adaptive polynomial splines (poly-MARS), random forests (RF), gradient boosting machines (GBM), extreme gradient boosting (XGBoost) and Bayesian regularized neural networks (BRNN), and each of them is based on a different combiner. The combiners include the equal-weight combiner, the median combiner, two best learners and seven variants of a sophisticated stacking method. The latter stacks a regression algorithm on top of the base learners to combine their independent predictions. Its seven variants are defined by seven different regression algorithms, specifically the linear regression (LR) algorithm and the six algorithms also used as base learners. The results suggest that sophisticated stacking performs significantly better than the base learners, especially when applied using the LR algorithm. It also beats the simpler combination methods. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
7. Comparison of Machine Learning Algorithms for Merging Gridded Satellite and Earth-Observed Precipitation Data.
- Author
-
Papacharalampous, Georgia, Tyralis, Hristos, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
MACHINE learning ,ARTIFICIAL neural networks ,INDEPENDENT variables ,RANDOM forest algorithms ,ERROR functions ,BOOSTING algorithms - Abstract
Gridded satellite precipitation datasets are useful in hydrological applications as they cover large regions with high density. However, they are not accurate in the sense that they do not agree with ground-based measurements. An established means for improving their accuracy is to correct them by adopting machine learning algorithms. This correction takes the form of a regression problem, in which the ground-based measurements have the role of the dependent variable and the satellite data are the predictor variables, together with topography factors (e.g., elevation). Most studies of this kind involve a limited number of machine learning algorithms and are conducted for a small region and for a limited time period. Thus, the results obtained through them are of local importance and do not provide more general guidance and best practices. To provide results that are generalizable and to contribute to the delivery of best practices, we here compare eight state-of-the-art machine learning algorithms in correcting satellite precipitation data for the entire contiguous United States and for a 15-year period. We use monthly data from the PERSIANN (Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks) gridded dataset, together with monthly earth-observed precipitation data from the Global Historical Climatology Network monthly database, version 2 (GHCNm). The results suggest that extreme gradient boosting (XGBoost) and random forests are the most accurate in terms of the squared error scoring function. The remaining algorithms can be ordered as follows, from the best to the worst: Bayesian regularized feed-forward neural networks, multivariate adaptive polynomial splines (poly-MARS), gradient boosting machines (gbm), multivariate adaptive regression splines (MARS), feed-forward neural networks and linear regression. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. A Low-Cost Gamified Urban Planning Methodology Enhanced with Co-Creation and Participatory Approaches.
- Author
-
Kavouras, Ioannis, Sardis, Emmanuel, Protopapadakis, Eftychios, Rallis, Ioannis, Doulamis, Anastasios, and Doulamis, Nikolaos
- Abstract
Targeted nature-based small-scale interventions is an approach commonly adopted by urban developers. The public acceptance of their implementation could be improved by participation, emphasizing residents or shopkeepers located close to the areas of interest. In this work, we propose a methodology that combines 3D technology, based on open data sources, user-generated content, 3D software and game engines for both minimizing the time and cost of the whole planning process and enhancing citizen participation. The proposed schemes are demonstrated in Piraeus (Greece) and Gladsaxe (Denmark). The core findings can be summarized as follows: (a) the time and cost are minimized by using online databases, (b) the gamification of the planning process enhances the decision making process and (c) the interactivity provided by the game engine inspired the participation of non-experts in the planning process (co-creation and co-evaluation), which decentralizes and democratizes the final planning solution. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. Comparison of Tree-Based Ensemble Algorithms for Merging Satellite and Earth-Observed Precipitation Data at the Daily Time Scale.
- Author
-
Papacharalampous, Georgia, Tyralis, Hristos, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
MACHINE learning ,STATISTICAL learning ,RANDOM forest algorithms ,ALGORITHMS ,DATABASES - Abstract
Merging satellite products and ground-based measurements is often required for obtaining precipitation datasets that simultaneously cover large regions with high density and are more accurate than pure satellite precipitation products. Machine and statistical learning regression algorithms are regularly utilized in this endeavor. At the same time, tree-based ensemble algorithms are adopted in various fields for solving regression problems with high accuracy and low computational costs. Still, information on which tree-based ensemble algorithm to select for correcting satellite precipitation products for the contiguous United States (US) at the daily time scale is missing from the literature. In this study, we worked towards filling this methodological gap by conducting an extensive comparison between three algorithms of the category of interest, specifically between random forests, gradient boosting machines (gbm) and extreme gradient boosting (XGBoost). We used daily data from the PERSIANN (Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks) and the IMERG (Integrated Multi-satellitE Retrievals for GPM) gridded datasets. We also used earth-observed precipitation data from the Global Historical Climatology Network daily (GHCNd) database. The experiments referred to the entire contiguous US and additionally included the application of the linear regression algorithm for benchmarking purposes. The results suggest that XGBoost is the best-performing tree-based ensemble algorithm among those compared. Indeed, the mean relative improvements that it provided with respect to linear regression (for the case that the latter algorithm was run with the same predictors as XGBoost) are equal to 52.66%, 56.26% and 64.55% (for three different predictor sets), while the respective values are 37.57%, 53.99% and 54.39% for random forests, and 34.72%, 47.99% and 62.61% for gbm. Lastly, the results suggest that IMERG is more useful than PERSIANN in the context investigated. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. A Prototype Machine Learning Tool Aiming to Support 3D Crowdsourced Cadastral Surveying of Self-Made Cities.
- Author
-
Potsiou, Chryssy, Doulamis, Nikolaos, Bakalos, Nikolaos, Gkeli, Maria, Ioannidis, Charalabos, and Markouizou, Selena
- Subjects
REAL property ,MACHINE learning ,INDOOR positioning systems ,DATA modeling ,MACHINE tools ,BLUETOOTH technology - Abstract
Land administration and management systems (LAMSs) have already made progress in the field of 3D Cadastre and the visualization of complex urban properties to support property markets and provide geospatial information for the sustainable management of smart cities. However, in less developed economies, with informally developed urban areas—the so-called self-made cities—the 2D LAMSs are left behind. Usually, they are less effective and mainly incomplete since a large number of informal constructions remain unregistered. This paper presents the latest results of an innovative on-going research aiming to structure, test and propose a low-cost but reliable enough methodology to support the simultaneous and fast implementation of both 2D land parcel and 3D property unit registration of informal, multi-story and unregistered constructions. An Indoor Positioning System (IPS) built upon low-cost Bluetooth technology combined with an innovative machine learning algorithm and connected with a 3D LADM-based cadastral mapping mobile application are the two key components of the technical solution under investigation. The proposed solution is tested for the first floor of a multi-room office building. The main conclusions concern the potential, usability and reliability of the method. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
11. Bayesian filter based behavior recognition in workflows allowing for user feedback
- Author
-
Kosmopoulos, Dimitrios I., Doulamis, Nikolaos D., and Voulodimos, Athanasios S.
- Published
- 2012
- Full Text
- View/download PDF
12. A service oriented architecture for decision support systems in environmental crisis management
- Author
-
Vescoukis, Vassilios, Doulamis, Nikolaos, and Karagiorgou, Sofia
- Published
- 2012
- Full Text
- View/download PDF
13. Active learning of user’s preferences estimation towards a personalized 3D navigation of geo-referenced scenes
- Author
-
Yiakoumettis, Christos, Doulamis, Nikolaos, Miaoulis, Georgios, and Ghazanfarpour, Djamchid
- Published
- 2014
- Full Text
- View/download PDF
14. A review of non-invasive sensors and artificial intelligence models for diabetic foot monitoring.
- Author
-
Kaselimi, Maria, Protopapadakis, Eftychios, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
DIABETIC foot ,ARTIFICIAL intelligence ,OPTICAL sensors ,DETECTORS ,TRUST - Abstract
Diabetic foot complications have multiple adverse effects in a person's quality of life. Yet, efficient monitoring schemes can mitigate or postpone any disorders, mainly by early detecting regions of interest. Nowadays, optical sensors and artificial intelligence (AI) tools can contribute efficiently to such monitoring processes. In this work, we provide information on the adopted imaging schemes and related optical sensors on this topic. The analysis considers both the physiology of the patients and the characteristics of the sensors. Currently, there are multiple approaches considering both visible and infrared bands (multiple ranges), most of them coupled with various AI tools. The source of the data (sensor type) can support different monitoring strategies and imposes restrictions on the AI tools that should be used with. This review provides a comprehensive literature review of AI-assisted DFU monitoring methods. The paper presents the outcomes of a large number of recently published scholarly articles. Furthermore, the paper discusses the highlights of these methods and the challenges for transferring these methods into a practical and trustworthy framework for sufficient remote management of the patients. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
15. Tensor-Based Learning for Detecting Abnormalities on Digital Mammograms.
- Author
-
Tzortzis, Ioannis N., Davradou, Agapi, Rallis, Ioannis, Kaselimi, Maria, Makantasis, Konstantinos, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
MAMMOGRAMS ,GENERAL Data Protection Regulation, 2016 ,ARTIFICIAL intelligence ,DEEP learning ,COMPUTER-aided diagnosis - Abstract
In this study, we propose a tensor-based learning model to efficiently detect abnormalities on digital mammograms. Due to the fact that the availability of medical data is limited and often restricted by GDPR (general data protection regulation) compliance, the need for more sophisticated and less data-hungry approaches is urgent. Accordingly, our proposed artificial intelligence framework utilizes the canonical polyadic decomposition to decrease the trainable parameters of the wrapped Rank-R FNN model, leading to efficient learning using small amounts of data. Our model was evaluated on the open source digital mammographic database INBreast and compared with state-of-the-art models in this domain. The experimental results show that the proposed solution performs well in comparison with the other deep learning models, such as AlexNet and SqueezeNet, achieving 90% ± 4% accuracy and an F1 score of 84% ± 5%. Additionally, our framework tends to attain more robust performance with small numbers of data and is computationally lighter for inference purposes, due to the small number of trainable parameters. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
16. Towards Trustworthy Energy Disaggregation: A Review of Challenges, Methods, and Perspectives for Non-Intrusive Load Monitoring.
- Author
-
Kaselimi, Maria, Protopapadakis, Eftychios, Voulodimos, Athanasios, Doulamis, Nikolaos, and Doulamis, Anastasios
- Subjects
TRUST ,SIGNAL processing ,MACHINE learning ,SCIENTIFIC community ,SCHOLARLY publishing - Abstract
Non-intrusive load monitoring (NILM) is the task of disaggregating the total power consumption into its individual sub-components. Over the years, signal processing and machine learning algorithms have been combined to achieve this. Many publications and extensive research works are performed on energy disaggregation or NILM for the state-of-the-art methods to reach the desired performance. The initial interest of the scientific community to formulate and describe mathematically the NILM problem using machine learning tools has now shifted into a more practical NILM. Currently, we are in the mature NILM period where there is an attempt for NILM to be applied in real-life application scenarios. Thus, the complexity of the algorithms, transferability, reliability, practicality, and, in general, trustworthiness are the main issues of interest. This review narrows the gap between the early immature NILM era and the mature one. In particular, the paper provides a comprehensive literature review of the NILM methods for residential appliances only. The paper analyzes, summarizes, and presents the outcomes of a large number of recently published scholarly articles. Furthermore, the paper discusses the highlights of these methods and introduces the research dilemmas that should be taken into consideration by researchers to apply NILM methods. Finally, we show the need for transferring the traditional disaggregation models into a practical and trustworthy framework. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
17. An evaluation study of clustering algorithms in the scope of user communities assessment
- Author
-
Karamolegkos, Pantelis N., Patrikakis, Charalampos Z., Doulamis, Nikolaos D., Vlacheas, Panagiotis T., and Nikolakopoulos, Ioannis G.
- Published
- 2009
- Full Text
- View/download PDF
18. Novel Insights in Spatial Epidemiology Utilizing Explainable AI (XAI) and Remote Sensing.
- Author
-
Temenos, Anastasios, Tzortzis, Ioannis N., Kaselimi, Maria, Rallis, Ioannis, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
REMOTE sensing ,EPIDEMIOLOGY ,VIRAL transmission ,REMOTE-sensing images ,COVID-19 pandemic ,MULTIDIMENSIONAL databases ,URBAN planning - Abstract
The COVID-19 pandemic has affected many aspects of human life around the world, due to its tremendous outcomes on public health and socio-economic activities. Policy makers have tried to develop efficient responses based on technologies and advanced pandemic control methodologies, to limit the wide spreading of the virus in urban areas. However, techniques such as social isolation and lockdown are short-term solutions that minimize the spread of the pandemic in cities and do not invert long-term issues that derive from climate change, air pollution and urban planning challenges that enhance the spreading ability. Thus, it seems crucial to understand what kind of factors assist or prevent the wide spreading of the virus. Although AI frameworks have a very efficient predictive ability as data-driven procedures, they often struggle to identify strong correlations among multidimensional data and provide robust explanations. In this paper, we propose the fusion of a heterogeneous, spatio-temporal dataset that combine data from eight European cities spanning from 1 January 2020 to 31 December 2021 and describe atmospheric, socio-economic, health, mobility and environmental factors all related to potential links with COVID-19. Remote sensing data are the key solution to monitor the availability on public green spaces between cities in the study period. So, we evaluate the benefits of NIR and RED bands of satellite images to calculate the NDVI and locate the percentage in vegetation cover on each city for each week of our 2-year study. This novel dataset is evaluated by a tree-based machine learning algorithm that utilizes ensemble learning and is trained to make robust predictions on daily cases and deaths. Comparisons with other machine learning techniques justify its robustness on the regression metrics RMSE and MAE. Furthermore, the explainable frameworks SHAP and LIME are utilized to locate potential positive or negative influence of the factors on global and local level, with respect to our model's predictive ability. A variation of SHAP, namely treeSHAP, is utilized for our tree-based algorithm to make fast and accurate explanations. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
19. STAMINA: Bioinformatics Platform for Monitoring and Mitigating Pandemic Outbreaks.
- Author
-
Bakalos, Nikolaos, Kaselimi, Maria, Doulamis, Nikolaos, Doulamis, Anastasios, Kalogeras, Dimitrios, Bimpas, Mathaios, Davradou, Agapi, Vlachostergiou, Aggeliki, Fotopoulos, Anaxagoras, Plakia, Maria, Karalis, Alexandros, Tsekeridou, Sofia, Anagnostopoulos, Themistoklis, Despotopoulou, Angela Maria, Bonavita, Ilaria, Petersen, Katrina, Pelepes, Leonidas, Voumvourakis, Lefteris, Anagnostou, Anastasia, and Groen, Derek
- Subjects
PANDEMICS ,DATA warehousing ,BIOINFORMATICS ,PREDICTION models ,CUSTOMIZATION - Abstract
This paper presents the components and integrated outcome of a system that aims to achieve early detection, monitoring and mitigation of pandemic outbreaks. The architecture of the platform aims at providing a number of pandemic-response-related services, on a modular basis, that allows for the easy customization of the platform to address user's needs per case. This customization is achieved through its ability to deploy only the necessary, loosely coupled services and tools for each case, and by providing a common authentication, data storage and data exchange infrastructure. This way, the platform can provide the necessary services without the burden of additional services that are not of use in the current deployment (e.g., predictive models for pathogens that are not endemic to the deployment area). All the decisions taken for the communication and integration of the tools that compose the platform adhere to this basic principle. The tools presented here as well as their integration is part of the project STAMINA. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
20. COVID-19 Spatio-Temporal Evolution Using Deep Learning at a European Level.
- Author
-
Kavouras, Ioannis, Kaselimi, Maria, Protopapadakis, Eftychios, Bakalos, Nikolaos, Doulamis, Nikolaos, and Doulamis, Anastasios
- Subjects
DEEP learning ,SPATIOTEMPORAL processes ,BOX-Jenkins forecasting ,COVID-19 ,COVID-19 pandemic ,VIRAL transmission - Abstract
COVID-19 evolution imposes significant challenges for the European healthcare system. The heterogeneous spread of the pandemic within EU regions elicited a wide range of policies, such as school closure, transport restrictions, etc. However, the implementation of these interventions is not accompanied by the implementation of quantitative methods, which would indicate their effectiveness. As a result, the efficacy of such policies on reducing the spread of the virus varies significantly. This paper investigates the effectiveness of using deep learning paradigms to accurately model the spread of COVID-19. The deep learning approaches proposed in this paper are able to effectively map the temporal evolution of a COVID-19 outbreak, while simultaneously taking into account policy interventions directly into the modelling process. Thus, our approach facilitates data-driven decision making by utilizing previous knowledge to train models that predict not only the spread of COVID-19, but also the effect of specific policy measures on minimizing this spread. Global models at the EU level are proposed, which can be successfully applied at the national level. These models use various inputs in order to successfully model the spatio-temporal variability of the phenomenon and obtain generalization abilities. The proposed models are compared against the traditional epidemiological and Autoregressive Integrated Moving Average (ARIMA) models. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
21. ELECTRIcity: An Efficient Transformer for Non-Intrusive Load Monitoring.
- Author
-
Sykiotis, Stavros, Kaselimi, Maria, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
DEEP learning ,VERNACULAR architecture ,ELECTRIC transformers ,HOUSEHOLD appliances ,CONSUMPTION (Economics) ,ELECTRICITY - Abstract
Non-Intrusive Load Monitoring (NILM) describes the process of inferring the consumption pattern of appliances by only having access to the aggregated household signal. Sequence-to-sequence deep learning models have been firmly established as state-of-the-art approaches for NILM, in an attempt to identify the pattern of the appliance power consumption signal into the aggregated power signal. Exceeding the limitations of recurrent models that have been widely used in sequential modeling, this paper proposes a transformer-based architecture for NILM. Our approach, called ELECTRIcity, utilizes transformer layers to accurately estimate the power signal of domestic appliances by relying entirely on attention mechanisms to extract global dependencies between the aggregate and the domestic appliance signals. Another additive value of the proposed model is that ELECTRIcity works with minimal dataset pre-processing and without requiring data balancing. Furthermore, ELECTRIcity introduces an efficient training routine compared to other traditional transformer-based architectures. According to this routine, ELECTRIcity splits model training into unsupervised pre-training and downstream task fine-tuning, which yields performance increases in both predictive accuracy and training time decrease. Experimental results indicate ELECTRIcity's superiority compared to several state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
22. Fall Detection Using Multi-Property Spatiotemporal Autoencoders in Maritime Environments.
- Author
-
Katsamenis, Iason, Bakalos, Nikolaos, Karolou, Eleni Eirini, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
DEEP learning ,VIDEO surveillance ,STREAMING video & television ,ANOMALY detection (Computer security) ,COMPUTER vision - Abstract
Man overboard is an emergency in which fast and efficient detection of the critical event is the key factor for the recovery of the victim. Its severity urges the utilization of intelligent video surveillance systems that monitor the ship's perimeter in real time and trigger the relative alarms that initiate the rescue mission. In terms of deep learning analysis, since man overboard incidents occur rarely, they present a severe class imbalance problem, and thus, supervised classification methods are not suitable. To tackle this obstacle, we follow an alternative philosophy and present a novel deep learning framework that formulates man overboard identification as an anomaly detection task. The proposed system, in the absence of training data, utilizes a multi-property spatiotemporal convolutional autoencoder that is trained only on the normal situation. We explore the use of RGB video sequences to extract specific properties of the scene, such as gradient and saliency, and utilize the autoencoders to detect anomalies. To the best of our knowledge, this is the first time that man overboard detection is made in a fully unsupervised manner while jointly learning the spatiotemporal features from RGB video streams. The algorithm achieved 97.30% accuracy and a 96.01% F1-score, surpassing the other state-of-the-art approaches significantly. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
23. Fair scheduling algorithms in grids
- Author
-
Doulamis, Nikolaos D., Doulamis, Anastasios D., Varvarigos, Emmanouel A., and Varvarigou, Theodora A.
- Subjects
Parallel processing -- Analysis ,Quality of service ,Parallel processing ,Business ,Computers ,Electronics ,Electronics and electrical industries - Abstract
Three new scheduling algorithms for the Grid environment are presented for implementation of scheduling in a fair way. The simple fair task order (SFTO), the adjusted fair task order (AFTO) and the max-min fair share (MMFS) are the three scheduling algorithms suitable for improving the task allocation performance.
- Published
- 2007
24. Optimal content-based video decomposition for interactive video navigation
- Author
-
Doulamis, Anastasios D. and Doulamis, Nikolaos D.
- Subjects
Image processing -- Research ,Business ,Computers ,Electronics ,Electronics and electrical industries - Abstract
In this paper, an interactive framework for navigating video sequences is presented using an optimal content-based video decomposition scheme. In particular, each video sequence is analyzed at different content resolution levels, creating a hierarchy from the lowest (coarse) to the highest (fine) resolution. This content hierarchy is represented as a tree structure, each level of which corresponds to a particular content resolution, while the tree nodes indicate the temporal video segments that the sequence content is partitioned at a given resolution. A criterion is introduced to measure the efficiency of the proposed scheme in organizing the video visual content and to compare it with other hierarchical video content representations and navigation schemes. The efficiency is measured as the difficulty for a user to locate a video segment of interest, while moving through different levels of hierarchy. In our case, video is decomposed so that the best efficiency is accomplished. However, the efficiency of a nonlinear video decomposition scheme depends on: 1) the number of paths required for a user to locate a relevant video segment and 2) the number of shot/frame classes (i.e., content representatives) extracted to represent the visual content. Both issues are addressed in this paper. In the first case, the probability of selecting a relevant video segment in the first path is maximized by extracting optimal content representatives through a minimization of a cross-correlation criterion. For the minimization, a genetic algorithm (GA) is adopted, since application of an exhaustive search to obtain the minimum value is too large to be implemented. The cross-correlation criterion is evaluated on the feature domain by extracting appropriate global and object-based descriptors for each video frame so that a better representation of the visual content is achieved. The second aspect (e.g., the number of content representatives) is addressed by minimizing the average transmitted information and simultaneously taking into consideration the temporal video segment complexity. More content representatives are extracted for video segments of high complexity, whereas a low number is required for low-complexity segments. In addition, a degree of interest is assigned to each video shot (or frame) to address the fact that, from the user's perception, the visual content of a set of shots (frames) satisfies his/her information needs. Finally, a computationally efficient algorithm is proposed to regulate the degree of detail (i.e., the number of shot/frames representatives) in case the visual content is not efficiently represented from the user's perceptive view. Experimental results on real-life video sequences indicate the performance of the proposed GA-based video decomposition scheme compared to other hierarchical video organization methods. Index Terms--Hierarchical summarization, MPEG-7, 21, video decomposition.
- Published
- 2004
25. Generalized nonlinear relevance feedback for interactive content-based retrieval and organization
- Author
-
Doulamis, Anastasios D. and Doulamis, Nikolaos D.
- Subjects
Image processing -- Research ,Business ,Computers ,Electronics ,Electronics and electrical industries - Abstract
In this paper, a novel relevance feedback algorithm is proposed for improving the performance of interactive content-based retrieval systems. The algorithm recursively estimates the similarity measure, which is used for data ranking in description environments where similarity-based queries are applied, using a set of relevant/irrelevant samples feedback by the user to the system so that the adjusted response is a better approximation of the current user's information needs and preferences. In particular, using concepts of functional analysis, the similarity measure is expressed as a parametric form of known monotone increasing functional components. Then, the contribution of each functional component to the similarity measure is estimated through a recursive and efficient on-line learning algorithm so that: 1) the current user's needs and preferences, as indicated by a set of selected relevant/irrelevant samples, are satisfied as much as possible, while simultaneously 2) a minimal modification of the already estimated similarity measure is accomplished. Experimental results on a large real-life database using objective evaluation criteria, such as the precision-recall curve and the average normalized modified retrieval rank (ANMRR), indicate that the proposed scheme outperforms the compared ones. In addition, the proposed algorithm requires low computational complexity and it can be implemented in a recursive way. Index Terms--Image retrieval, MPEG-7, relevance feedback.
- Published
- 2004
26. A combined fuzzy-neural network model for non-linear prediction of 3-D rendering workload in grid computing
- Author
-
Doulamis, Nikolaos D., Doulamis, Anastasios D., Panagakis, Athanasios, Dolkas, Konstantinos, Varvarigou, Theodora A., and Varvarigos, Emmanuel
- Subjects
Neural network ,Neural networks -- Research - Abstract
Implementation of a commercial application to a grid infrastructure introduces new challenges in managing the quality-of-service (QoS) requirements, most stem from the fact that negotiation on QoS between the user and the service provider should strictly be satisfied. An interesting commercial application with a wide impact on a variety of fields, which can benefit from the computational grid technologies, is three-dimensional (3-D) rendering. In order to implement, however, 3-D rendering to a grid infrastructure, we should develop appropriate scheduling and resource allocation mechanisms so that the negotiated (QoS) requirements are met. Efficient scheduling schemes require modeling and prediction of rendering workload. In this paper workload prediction is addressed based on a combined fuzzy classification and neural network model. Initially, appropriate descriptors are extracted to represent the synthetic world. The descriptors are obtained by parsing RIB formatted files, which provides a general structure for describing computer-generated images. Fuzzy classification is used for organizing rendering descriptor so that a reliable representation is accomplished which increases the prediction accuracy. Neural network performs workload prediction by modeling the nonlinear input-output relationship between rendering descriptors and the respective computational complexity. To increase prediction accuracy, a constructive algorithm is adopted in this paper to train the neural network so that network weights and size are simultaneously estimated. Then, a grid scheduler scheme is proposed to estimate the queuing order that the tasks should be executed and the most appopriate processor assignment so that the demanded QoS are satisfied as much as possible. A fair scheduling policy is considered as the most appropriate. Experimental results on a real grid infrastructure are presented to illustrate the efficiency of the proposed workload prediction--scheduling algorithm compared to other approaches presented in the literature. Index Terms--Grid computing, workload prediction, neural networks, three-dimensional (3-D) rendering.
- Published
- 2004
27. An efficient fully unsupervised video object segmentation scheme using an adaptive neural-network classifier architecture
- Author
-
Doulamis, Anastasios, Doulamis, Nikolaos, Ntalianis, Klimis, and Kollias, Stefanos
- Subjects
Neural networks -- Research ,Image coding -- Methods ,Image coding -- Evaluation ,Videoconferencing ,Object recognition (Computers) -- Evaluation ,Pattern recognition ,Neural network ,Videoconferencing ,Business ,Computers ,Electronics ,Electronics and electrical industries - Abstract
In this paper, an unsupervised video object (VO) segmentation and tracking algorithm is proposed based on an adaptable neural-network architecture. The proposed scheme comprises: 1) a VO tracking module and 2) an initial VO estimation module. Object tracking is handled as a classification problem and implemented through an adaptive network classifier, which provides better results compared to conventional motion-based tracking algorithms. Network adaptation is accomplished through an efficient and cost effective weight updating algorithm, providing a minimum degradation of the previous network knowledge and taking into account the current content conditions. A retraining set is constructed and used for this purpose based on initial VO estimation results. Two different scenarios are investigated. The first concerns extraction of human entities in video conferencing applications, while the second exploits depth information to identify generic VOs in stereoscopic video sequences. Human face/ body detection based on Gaussian distributions is accomplished in the first scenario, while segmentation fusion is obtained using color and depth information in the second scenario. A decision mechanism is also incorporated to detect time instances for weight updating. Experimental results and comparisons indicate the good performance of the proposed scheme even in sequences with complicated content (object bending, occlusion). Index Terms--Adaptive neural networks, MPEG-4, video object extraction.
- Published
- 2003
28. Deep Recurrent Neural Networks for Ionospheric Variations Estimation Using GNSS Measurements.
- Author
-
Kaselimi, Maria, Voulodimos, Athanasios, Doulamis, Nikolaos, Doulamis, Anastasios, and Delikaraoglou, Demitris
- Subjects
RECURRENT neural networks ,DEEP learning ,GLOBAL Positioning System ,FEATURE extraction - Abstract
Modeling ionospheric variability throughout a proper total electron content (TEC) parameter estimation is a demanding, however, crucial, process for achieving better accuracy and rapid convergence in precise point positioning (PPP). In particular, the single-frequency PPP (SF-PPP) method lacks accuracy due to the difficulty of dealing adequately with the ionospheric error sources. In order to apply ionosphere corrections in techniques, such as SF-PPP, external information of global ionosphere maps (GIMs) is crucial. In this article, we propose a deep learning model to efficiently predict TEC values and to replace the GIM-derived data that inherently have a global character, with equal or better in accuracy regional ones. The proposed model is suitable for predicting the ionosphere delay at different locations of receiver stations. The model is tested during different periods of time, under different solar and geomagnetic conditions and for stations in various latitudes, providing robust estimations of the ionospheric activity at the regional level. Our proposed model is a hybrid model comprising of a 1-D convolutional layer used for the optimal feature extraction and stacked recurrent layers used for temporal time series modeling. Thus, the model achieves good performance in TEC modeling compared to other state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
29. Unsupervised 3D Motion Summarization Using Stacked Auto-Encoders.
- Author
-
Protopapadakis, Eftychios, Rallis, Ioannis, Doulamis, Anastasios, Doulamis, Nikolaos, and Voulodimos, Athanasios
- Subjects
MOTION capture (Human mechanics) ,ALGORITHMS ,MOTION ,STANDARD deviations - Abstract
In this paper, a deep stacked auto-encoder (SAE) scheme followed by a hierarchical Sparse Modeling for Representative Selection (SMRS) algorithm is proposed to summarize dance video sequences, recorded using the VICON Motion capturing system. SAE's main task is to reduce the redundant information embedding in the raw data and, thus, to improve summarization performance. This becomes apparent when two dancers are performing simultaneously and severe errors are encountered in the humans' point joints, due to dancers' occlusions in the 3D space. Four summarization algorithms are applied to extract the key frames; density based, Kennard Stone, conventional SMRS and its hierarchical scheme called H-SMRS. Experimental results have been carried out on real-life dance sequences of Greek traditional dances while the results have been compared against ground truth data selected by dance experts. The results indicate that H-SMRS being applied after the SAE information reduction module extracts key frames which are deviated in time less than 0.3 s to the ones selected by the experts and with a standard deviation of 0.18 s. Thus, the proposed scheme can effectively represent the content of the dance sequence. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
30. A Stochastic Framework for Optimal Key Frame Extraction from MPEG Video Databases
- Author
-
Avrithis, Yannis S, Doulamis, Anastasios D, Doulamis, Nikolaos D, and Kollias, Stefanos D
- Published
- 1999
- Full Text
- View/download PDF
31. Gaussian Process Regression Tuned by Bayesian Optimization for Seawater Intrusion Prediction.
- Author
-
Kopsiaftis, George, Protopapadakis, Eftychios, Voulodimos, Athanasios, Doulamis, Nikolaos, and Mantoglou, Aristotelis
- Subjects
SALTWATER encroachment ,KRIGING ,WELLHEAD protection ,STANDARD deviations ,GROUNDWATER management ,LATIN hypercube sampling - Abstract
Accurate prediction of the seawater intrusion extent is necessary for many applications, such as groundwater management or protection of coastal aquifers from water quality deterioration. However, most applications require a large number of simulations usually at the expense of prediction accuracy. In this study, the Gaussian process regression method is investigated as a potential surrogate model for the computationally expensive variable density model. Gaussian process regression is a nonparametric kernel-based probabilistic model able to handle complex relations between input and output. In this study, the extent of seawater intrusion is represented by the location of the 0.5 kg/m3 iso-chlore at the bottom of the aquifer (seawater intrusion toe). The initial position of the toe, expressed as the distance of the specific line from a number of observation points across the coastline, along with the pumping rates are the surrogate model inputs, whereas the final position of the toe constitutes the output variable set. The training sample of the surrogate model consists of 4000 variable density simulations, which differ not only in the pumping rate pattern but also in the initial concentration distribution. The Latin hypercube sampling method is used to obtain the pumping rate patterns. For comparison purposes, a number of widely used regression methods are employed, specifically regression trees and Support Vector Machine regression (linear and nonlinear). A Bayesian optimization method is applied to all the regressors, to maximize their efficiency in the prediction of seawater intrusion. The final results indicate that the Gaussian process regression method, albeit more time consuming, proved to be more efficient in terms of the mean absolute error (MAE), the root mean square error (RMSE), and the coefficient of determination (R²). [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
32. Tensor-Based Classification Models for Hyperspectral Data Analysis.
- Author
-
Makantasis, Konstantinos, Doulamis, Anastasios D., Doulamis, Nikolaos D., and Nikitakis, Antonis
- Subjects
HYPERSPECTRAL imaging systems ,TENSOR algebra ,FEEDFORWARD neural networks ,SUPPORT vector machines ,LOGISTIC regression analysis ,MACHINE learning - Abstract
In this paper, we present tensor-based linear and nonlinear models for hyperspectral data classification and analysis. By exploiting the principles of tensor algebra, we introduce new classification architectures, the weight parameters of which satisfy the rank-1 canonical decomposition property. Then, we propose learning algorithms to train both linear and nonlinear classifiers. The advantages of the proposed classification approach are that: 1) it significantly reduces the number of weight parameters required to train the model (and thus the respective number of training samples); 2) it provides a physical interpretation of model coefficients on the classification output; and 3) it retains the spatial and spectral coherency of the input samples. The linear tensor-based model exploits the principles of logistic regression, assuming the rank-1 canonical decomposition property among its weights. For the nonlinear classifier, we propose a modification of a feedforward neural network (FNN), called rank-1 FNN, since its weights satisfy again the rank-1 canonical decomposition property. An appropriate learning algorithm is also proposed to train the network. Experimental results and comparisons with state-of-the-art classification methods, either linear (e.g., linear support vector machine) or nonlinear (e.g., deep learning), indicate the outperformance of the proposed scheme, especially in the cases where a small number of training samples is available. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
33. Data-Driven Background Subtraction Algorithm for In-Camera Acceleration in Thermal Imagery.
- Author
-
Makantasis, Konstantinos, Nikitakis, Antonios, Doulamis, Anastasios D., Doulamis, Nikolaos D., and Papaefstathiou, Ioannis
- Subjects
IMAGE processing ,MATHEMATICAL models ,INFRARED imaging ,ESTIMATION theory ,REAL-time computing ,HARDWARE - Abstract
Detection of moving objects in videos is a crucial step toward successful surveillance and monitoring applications. A key component for such tasks is called background subtraction and tries to extract regions of interest from the image background for further processing or action. For this reason, its accuracy and real-time performance are of great significance. Although effective background subtraction methods have been proposed, only a few of them take into consideration the special characteristics of thermal imagery. In this paper, we propose a background subtraction scheme, which models the thermal responses of each pixel as a mixture of Gaussians with unknown number of components. Following a Bayesian approach, our method automatically estimates the mixture structure, while simultaneously it avoids over-/underfitting. The pixel density estimate is followed by an efficient and highly accurate updating mechanism, which permits our system to be automatically adapted to dynamically changing operation conditions. We propose a reference implementation of our method in reconfigurable hardware achieving both adequate performance and low-power consumption. Adopting a high-level synthesis design and demanding floating point arithmetic operations are mapped in reconfigurable hardware, demonstrating fast prototyping and on-field customization at the same time. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
34. Dance Pose Identification from Motion Capture Data: A Comparison of Classifiers.
- Author
-
Protopapadakis, Eftychios, Voulodimos, Athanasios, Doulamis, Anastasios, Camarinopoulos, Stephanos, Doulamis, Nikolaos, and Miaoulis, Georgios
- Subjects
MOTION capture (Human mechanics) ,DANCE ,JOINTS (Anatomy) ,DETECTORS ,FOLK dancing - Abstract
In this paper, we scrutinize the effectiveness of classification techniques in recognizing dance types based on motion-captured human skeleton data. In particular, the goal is to identify poses which are characteristic for each dance performed, based on information on body joints, acquired by a Kinect sensor. The datasets used include sequences from six folk dances and their variations. Multiple pose identification schemes are applied using temporal constraints, spatial information, and feature space distributions for the creation of an adequate training dataset. The obtained results are evaluated and discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
35. Deep Learning for Computer Vision: A Brief Review.
- Author
-
Voulodimos, Athanasios, Doulamis, Nikolaos, Doulamis, Anastasios, and Protopapadakis, Eftychios
- Subjects
- *
DEEP learning , *ARTIFICIAL neural networks , *COMPUTER vision , *SIGNAL denoising , *FACE perception , *BOLTZMANN machine - Abstract
Over the last years deep learning methods have been shown to outperform previous state-of-the-art machine learning techniques in several fields, with computer vision being one of the most prominent cases. This review paper provides a brief overview of some of the most significant deep learning schemes used in computer vision problems, that is, Convolutional Neural Networks, Deep Boltzmann Machines and Deep Belief Networks, and Stacked Denoising Autoencoders. A brief account of their history, structure, advantages, and limitations is given, followed by a description of their applications in various computer vision tasks, such as object detection, face recognition, action and activity recognition, and human pose estimation. Finally, a brief overview is given of future directions in designing deep learning schemes for computer vision problems and the challenges involved therein. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
36. Stacked Autoencoders for Outlier Detection in Over-the-Horizon Radar Signals.
- Author
-
Protopapadakis, Eftychios, Voulodimos, Athanasios, Doulamis, Anastasios, Doulamis, Nikolaos, Dres, Dimitrios, and Bimpas, Matthaios
- Subjects
OUTLIER detection ,RADAR signal processing ,SURFACE waves (Seismic waves) ,DEEP learning ,CLUSTER analysis (Statistics) - Abstract
Detection of outliers in radar signals is a considerable challenge in maritime surveillance applications. High-Frequency Surface-Wave (HFSW) radars have attracted significant interest as potential tools for long-range target identification and outlier detection at over-the-horizon (OTH) distances. However, a number of disadvantages, such as their low spatial resolution and presence of clutter, have a negative impact on their accuracy. In this paper, we explore the applicability of deep learning techniques for detecting deviations from the norm in behavioral patterns of vessels (outliers) as they are tracked from an OTH radar. The proposed methodology exploits the nonlinear mapping capabilities of deep stacked autoencoders in combination with density-based clustering. A comparative experimental evaluation of the approach shows promising results in terms of the proposed methodology’s performance. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
37. Event Detection in Twitter Microblogging.
- Author
-
Doulamis, Nikolaos D., Doulamis, Anastasios D., Kokkinos, Panagiotis, and Varvarigos, Emmanouel Manos
- Abstract
The millions of tweets submitted daily overwhelm users who find it difficult to identify content of interest revealing the need for event detection algorithms in Twitter. Such algorithms are proposed in this paper covering both short (identifying what is currently happening) and long term periods (reviewing the most salient recently submitted events). For both scenarios, we propose fuzzy represented and timely evolved tweet-based theoretic information metrics to model Twitter dynamics. The Riemannian distance is also exploited with respect to words’ signatures to minimize temporal effects due to submission delays. Events are detected through a multiassignment graph partitioning algorithm that: 1) optimally retains maximum coherence within a cluster and 2) while allowing a word to belong to several clusters (events). Experimental results on real-life data demonstrate that our approach outperforms other methods. [ABSTRACT FROM PUBLISHER]
- Published
- 2016
- Full Text
- View/download PDF
38. TECHNICAL ASPECTS FOR THE CREATION OF A MULTI-DIMENSIONAL LAND INFORMATION SYSTEM.
- Author
-
Ioannidis, Charalabos, Potsiou, Chryssy, Soile, Sofia, Verykokou, Styliani, Mourafetis, George, and Doulamis, Nikolaos
- Subjects
URBAN land use ,DECISION making ,THREE-dimensional display systems - Abstract
The complexity of modern urban environments and civil demands for fast, reliable and affordable decision-making requires not only a 3D Land Information System, which tends to replace traditional 2D LIS architectures, but also the need to address the time and scale parameters, that is, the 3D geometry of buildings in various time instances (4
th dimension) at various levels of detail (LoDs - 5th dimension). This paper describes and proposes solutions for technical aspects that need to be addressed for the 5D modelling pipeline. Such solutions include the creation of a 3D model, the application of a selective modelling procedure between various time instances and at various LoDs, enriched with cadastral and other spatial data, and a procedural modelling approach for the representation of the inner parts of the buildings. The methodology is based on automatic change detection algorithms for spatialtemporal analysis of the changes that took place in subsequent time periods, using dense image matching and structure from motion algorithms. The selective modelling approach allows a detailed modelling only for the areas where spatial changes are detected. The procedural modelling techniques use programming languages for the textual semantic description of a building; they require the modeller to describe its part-to-whole relationships. Finally, a 5D viewer is developed, in order to tackle existing limitations that accompany the use of global systems, such as the Google Earth or the Google Maps, as visualization software. An application based on the proposed methodology in an urban area is presented and it provides satisfactory results. [ABSTRACT FROM AUTHOR]- Published
- 2016
- Full Text
- View/download PDF
39. 5D MODELLING: AN EFFICIENT APPROACH FOR CREATING SPATIOTEMPORAL PREDICTIVE 3D MAPS OF LARGE-SCALE CULTURAL RESOURCES.
- Author
-
Doulamis, Anastasios, Doulamis, Nikolaos, Ioannidis, Charalabos, Chrysouli, Christina, Grammalidis, Nikos, Dimitropoulos, Kosmas, Potsiou, Chryssy, Stathopoulou, Elisavet Konstantina, and Ioannides, Marinos
- Subjects
MODELING (Sculpture) ,GEOGRAPHY - Abstract
Outdoor large-scale cultural sites are mostly sensitive to environmental, natural and human made factors, implying an imminent need for a spatio-temporal assessment to identify regions of potential cultural interest (material degradation, structuring, conservation). On the other hand, in Cultural Heritage research quite different actors are involved (archaeologists, curators, conservators, simple users) each of diverse needs. All these statements advocate that a 5D modelling (3D geometry plus time plus levels of details) is ideally required for preservation and assessment of outdoor large scale cultural sites, which is currently implemented as a simple aggregation of 3D digital models at different time and levels of details. The main bottleneck of such an approach is its complexity, making 5D modelling impossible to be validated in real life conditions. In this paper, a cost effective and affordable framework for 5D modelling is proposed based on a spatial-temporal dependent aggregation of 3D digital models, by incorporating a predictive assessment procedure to indicate which regions (surfaces) of an object should be reconstructed at higher levels of details at next time instances and which at lower ones. In this way, dynamic change history maps are created, indicating spatial probabilities of regions needed further 3D modelling at forthcoming instances. Using these maps, predictive assessment can be made, that is, to localize surfaces within the objects where a high accuracy reconstruction process needs to be activated at the forthcoming time instances. The proposed 5D Digital Cultural Heritage Model (5D-DCHM) is implemented using open interoperable standards based on the CityGML framework, which also allows the description of additional semantic metadata information. Visualization aspects are also supported to allow easy manipulation, interaction and representation of the 5D-DCHM geometry and the respective semantic information. The open source 3DCityDB incorporating a PostgreSQL geo-database is used to manage and manipulate 3D data and their semantics. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
40. On the Exploration of Automatic Building Extraction from RGB Satellite Images Using Deep Learning Architectures Based on U-Net.
- Author
-
Temenos, Anastasios, Temenos, Nikos, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
DEEP learning ,REMOTE-sensing images ,CONVOLUTIONAL neural networks ,URBAN planning ,DATA mining ,FEATURE extraction - Abstract
Detecting and localizing buildings is of primary importance in urban planning tasks. Automating the building extraction process, however, has become attractive given the dominance of Convolutional Neural Networks (CNNs) in image classification tasks. In this work, we explore the effectiveness of the CNN-based architecture U-Net and its variations, namely, the Residual U-Net, the Attention U-Net, and the Attention Residual U-Net, in automatic building extraction. We showcase their robustness in feature extraction and information processing using exclusively RGB images, as they are a low-cost alternative to multi-spectral and LiDAR ones, selected from the SpaceNet 1 dataset. The experimental results show that U-Net achieves a 91.9 % accuracy, whereas introducing residual blocks, attention gates, or a combination of both improves the accuracy of the vanilla U-Net to 93.6 % , 94.0 % , and 93.7 % , respectively. Finally, the comparison between U-Net architectures and typical deep learning approaches from the literature highlights their increased performance in accurate building localization around corners and edges. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
41. Multiclass Confusion Matrix Reduction Method and Its Application on Net Promoter Score Classification Problem.
- Author
-
Markoulidakis, Ioannis, Rallis, Ioannis, Georgoulas, Ioannis, Kopsiaftis, George, Doulamis, Anastasios, and Doulamis, Nikolaos
- Subjects
RECEIVER operating characteristic curves ,CLASSIFICATION algorithms ,CURVES ,MACHINE learning ,KEY performance indicators (Management) ,MATRICES (Mathematics) ,CUSTOMER experience - Abstract
The current paper presents a novel method for reducing a multiclass confusion matrix into a 2 × 2 version enabling the exploitation of the relevant performance metrics and methods such as the receiver operating characteristic and area under the curve for the assessment of different classification algorithms. The reduction method is based on class grouping and leads to a special type of matrix called the reduced confusion matrix. The developed method is then exploited for the assessment of state of the art machine learning algorithms applied on the net promoter score classification problem in the field of customer experience analytics indicating the value of the proposed method in real world classification problems. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
42. Local Geometrically Enriched Mixtures for Stable and Robust Human Tracking in Detecting Falls.
- Author
-
Kokkinos, Michalis, Doulamis, Nikolaos D., and Doulamis, Anastasios D.
- Subjects
ROBUST control ,COMPUTER vision ,TRACKING algorithms ,CONSTRAINT satisfaction ,ROBOT motion ,ITERATIVE methods (Mathematics) - Abstract
Detecting a fall through visual cues is emerging as a hot research agenda for improving the independence of the elderly. However, the traditional motion-based algorithms are very sensitive to noise, reducing fall detection accuracy. Another approach is to efficiently localize and then track the foreground object followed by measurements that aid the identification of a fall. However, to perform robust and stable tracking over a long time is a challenging research aspect in computer vision society. In this paper, we introduce a stable human tracker able to efficiently cope with the trade-off between model stability (accurate tracking performance) and adaptability (model evolution to visual changes). In particular, we introduce local geometrically enriched mixture models for background modelling. Then, we incorporate iterative motion information methods, constrained by shape and time properties, to estimate high confidence image regions for background model updating. This way, we are able to detect and track the foreground objects even when visual conditions are dynamically changed over time (luminosity or background/foreground changes or active cameras). [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
43. ENABLING APPLICATIONS ON THE GRID: A GRIDLAB OVERVIEW.
- Author
-
Allen, Gabrielle, Davis, Kelly, Dolkas, Konstantinos N., Doulamis, Nikolaos D., Goodale, Tom, Goodale, Thilo, Merzky, André, Nabrzyski, Jarek, Pukacki, Juliusz, Radke, Thomas, Russell, Michael, Seidel, Ed, Shalf, John, and Taylor, Ian
- Subjects
GRID computing ,HIGH performance computing ,DISTRIBUTED computing ,COMPUTER systems ,COMPUTER scientists ,COMPUTER networks - Abstract
The primary aim of GridLab is to provide users and application developers with a simple and robust environment enabling them to produce applications that can exploit the full power and possibilities of the Grid. The GridLab project brings together computer scientists with computational scientists from various application areas to design and implement a Grid Application Toolkit (GAT), together with a set of Grid services, in a production grid environment. The GAT will provide functionality through a carefully constructed set of generic high-level APIs, through which an application will be able to call the underlying Grid services. The project will demonstrate the benefits of the GAT by developing and implementing real application scenarios, illustrating compelling new uses of the Grid. The development of the OAT is accompanied by the establishment of a pan-European testbed and by the development of Grid services of varying complexity, tailored to the needs of user community. These services are designed to complement and complete the existing Grid infrastructure, and to provide functionality needed by the GridLab applications in order to be usefully deployed in such environment.
- Published
- 2003
- Full Text
- View/download PDF
44. An Adaptable Neural-Network Model for Recursive Nonlinear Traffic Prediction and Modeling of MPEG Video Sources.
- Author
-
Doulamis, Anastasios D., Doulamis, Nikolaos D., and Kollias, Stefanos D.
- Subjects
- *
ARTIFICIAL neural networks , *MPEG (Video coding standard) - Abstract
Presents information on a study which described a neural network architecture for recursive nonlinear traffic prediction and modeling of MPEG video sources. Investigation of the problem of online traffic modeling; Implementation of the neural network architecture; Results and discussion; Conclusions.
- Published
- 2003
- Full Text
- View/download PDF
45. Efficient Content-Based Image Retrieval Using Fuzzy Organization and Optimal Relevance Feedback.
- Author
-
Doulamis, Anastasios, Doulamis, Nikolaos, and Varvarigou, Theodora
- Subjects
- *
IMAGE retrieval , *IMAGE processing , *FUZZY sets - Abstract
The performance of a Content-Based Image Retrieval System (CBIR) depends on (a) the system's adaptability to the user's information needs, which permits different types of indexing and simultaneously reduces the subjectivity of human perception for the interpretation of the image visual content and (b) the efficient organization of the extracted descriptors, which represent the rich visual information. Both issues are addressed in this paper. Descriptor organization is performed using a fuzzy classification scheme fragmented into multidimensional classes, instead of the previous works where fuzzy histograms were created in one dimension using, for example, the feature vector norm. Multidimensionality relates the descriptors with one another and thus allows a compact and meaningful visual representation by mapping the elements of the resulted feature vectors with a physical visual interpretation. Furthermore, fuzzy classification is applied for all visual content descriptors, in contrast to the previous approaches where only color information is exploited. Two kinds of content descriptors are extracted in our case; global-based and region-based. The first refers to the global image characteristics, while the second exploits the region-based properties. Regions are obtained by applying a multiresolution implementation of the Recursive Shortest Spanning Tree (RSST) algorithm, called M-RSST in this paper. The second issue is addressed by proposing a computationally efficient relevance feedback mechanism based on an optimal weight updating strategy. The scheme relies on the cross-correlation measure, instead of the Euclidean distance which is mainly used in most relevance feedback algorithms. Crosscorrelation is a normalized measure, which expresses how similar the two feature vectors are and thus it indicates a metric of their content similarity. The proposed scheme can be recursively implemented in the case of multiple feedback iterations, instead of the previous... [ABSTRACT FROM AUTHOR]
- Published
- 2003
- Full Text
- View/download PDF
46. NON-SEQUENTIAL VIDEO CONTENT REPRESENTATION USING TEMPORAL VARIATION OF FEATURE VECTORS.
- Author
-
Doulamis, Anastasios D. and Doulamis, Nikolaos
- Subjects
- *
IMAGE processing , *DIGITAL video , *DIGITAL electronics - Abstract
Deals with a study which proposed an algorithm for non-sequential video content representation. Depiction of the feature-based video representation; Description of the fuzzy classification of feature vectors; Presentation of experimental results using real life video sequences; Concluding remarks.
- Published
- 2000
- Full Text
- View/download PDF
47. Efficient Unsupervised Content-Based Segmentation in Stereoscopic Video Sequences.
- Author
-
Doulamis, Anastasios, Doulamis, Nikolaos, Ntalianis, Klimis, and Kollias, Stefanos
- Subjects
- *
THREE-dimensional imaging , *VIDEO recording - Abstract
This paper presents an efficient technique for unsupervised content-based segmentation in stereoscopic video sequences by appropriately combined different content descriptors in a hierarchical framework. Three main modules are involved in the proposed scheme; extraction of reliable depth information, image partition into color and depth regions and a constrained fusion algorithm of color segments using information derived from the depth map. In the first module, each stereo pair is analyzed and the disparity field and depth map are estimated. Occlusion detection and compensation are also applied for improving the depth map estimation. In the following phase, color and depth regions are created using a novel complexity-reducing multiresolution implementation of the Recursive Shortest Spanning Tree algorithm (M-RSST). While depth segments provide a coarse representation of the image content, color regions describe very accurately object boundaries. For this reason, in the final phase, a new segmentation fusion algorithm is employed which projects color segments onto depth segments. Experimental results are presented which exhibit the efficiency of the proposed scheme as content-based descriptor, even in case of images with complicated visual content. [ABSTRACT FROM AUTHOR]
- Published
- 2000
- Full Text
- View/download PDF
48. Efficient Summarization of Stereoscopic Video Sequences.
- Author
-
Doulamis, Nikolaos D. and Doulamis, Anastasios D.
- Subjects
- *
VIDEO compression , *DIGITAL video , *RECURSIVE functions , *FUZZY algorithms , *3-D television - Abstract
Features an efficient technique for summarization of stereoscopic video sequences which extracts a small but meaningful set of video frames using a content-based sampling algorithm. Capability of browsing digital stereoscopic video sequences; Efficient content-based queries and indexing; Recursive Shortest Spanning Tree algorithm; Multidimensional fuzzy classification of segment.
- Published
- 2000
- Full Text
- View/download PDF
49. Efficient modeling of VBR MPEG-1 coded video sources.
- Author
-
Doulamis, Nikolaos D. and Doulamis, Anastasios D.
- Subjects
- *
CODING theory , *MPEG (Video coding standard) , *VIDEO recording - Abstract
Proposes the layer modeling of Motion Picture Experts Group (MPEG)-1 coded video streams and statistical analysis of their characteristics at each layer. Basic characteristics of MPEG-1 encoders; Encoding algorithm; Study of autocorrelation function; Traffic model of frame layer; Intermediate layer modeling; Buffer configuration scheme; Prediction of video activity.
- Published
- 2000
- Full Text
- View/download PDF
50. On-Line Retrainable Neural Networks: Improving the Performance of Neural Networks in Image Analysis Problems.
- Author
-
Doulamis, Anastasios D. and Doulamis, Nikolaos D.
- Subjects
- *
ARTIFICIAL neural networks , *IMAGE reconstruction - Abstract
Presents information on a study which described an approach for improving the performance of neural-network classifiers in image recognition, segmentation or coding applications based on a retraining algorithm. Formulation of the problem; Optimal selection of the network retraining data; Decision mechanism for network retraining; Conclusions.
- Published
- 2000
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.