16 results on '"Gianniti, E"'
Search Results
2. An optimization framework for the capacity allocation and admission control of MapReduce jobs in cloud systems
- Author
-
Malekimajd, M., Ardagna, D., Ciavotta, M., Gianniti, E., Passacantando, M., and Rizzi, A. M.
- Published
- 2018
- Full Text
- View/download PDF
3. Optimizing Quality-Aware Big Data Applications in the Cloud
- Author
-
Gianniti, E, Ciavotta, M, Ardagna, D, Gianniti, E, Ciavotta, M, and Ardagna, D
- Abstract
The last years witnessed a steep rise in data generation worldwide and, consequently, the widespread adoption of software solutions able to support data-intensive applications. Competitiveness and innovation have strongly benefited from these new platforms and methodologies, and there is a great deal of interest around the new possibilities that Big Data analytics promise to make reality. Many companies currently engage in data-intensive processes as part of their core businesses; however, fully embracing the data-driven paradigm is still cumbersome, and establishing a production-ready, fine-tuned deployment is time-consuming, expensive, and resource-intensive. This situation calls for innovative models and techniques to streamline the process of deployment configuration for Big Data applications. In particular, the focus in this paper is on the rightsizing of Cloud deployed clusters, which represent a cost-eective alternative to installation on premises. This paper proposes a novel tool, integrated in a wider DevOps-inspired approach, implementing a parallel and distributed simulation-optimization technique that eciently and eectively explores the space of alternative Cloud configurations, seeking the minimum cost deployment that satisfies quality of service constraints. The soundness of the proposed solution has been thoroughly validated in a vast experimental campaign encompassing dierent applications and Big Data platforms.
- Published
- 2021
4. Predicting the performance of big data applications on the cloud
- Author
-
Ardagna, D., primary, Barbierato, E., additional, Gianniti, E., additional, Gribaudo, M., additional, Pinto, T. B. M., additional, da Silva, A. P. C., additional, and Almeida, J. M., additional
- Published
- 2020
- Full Text
- View/download PDF
5. Predicting the performance of big data applications on the cloud
- Author
-
Ardagna, D., Barbierato, Enrico, Gianniti, E., Gribaudo, M., Pinto, T. B. M., da Silva, A. P. C., Almeida, J. M., Barbierato, E. (ORCID:0000-0003-1466-0248), Ardagna, D., Barbierato, Enrico, Gianniti, E., Gribaudo, M., Pinto, T. B. M., da Silva, A. P. C., Almeida, J. M., and Barbierato, E. (ORCID:0000-0003-1466-0248)
- Abstract
Data science applications have become widespread as a means to extract knowledge from large datasets. Such applications are often characterized by highly heterogeneous and irregular data access patterns, thus often being referred to as big data applications. Such characteristics make the application execution quite challenging for existing software and hardware infrastructures to meet their resource demands. The cloud computing paradigm, in turn, ofers a natural hosting solution to such applications since its on-demand pricing model allows allocating efectively computing resources according to application’s needs. However, these properties impose extra challenge to the accurate performance prediction of cloud-based applications, which is a key step to adequate capacity planning and managing of the hosting infrastructure. In this article, we tackle this challenge by exploring three modeling approaches for predicting the performance of big data applications running on the cloud. We evaluate two queuing-based analytical models and dagSim, a fast ad-hoc simulator, in various scenarios based on diferent applications and infrastructure setups. The considered approaches are compared in terms of prediction accuracy and execution time. Our results indicate that our two best approaches, one analytical model and dagSim, can predict average application execution times with only up to a 7% relative error, on average. Moreover, a comparison with the widely used event-based simulator available with the Java Modeling Tool (JMT) suite demonstrates that both the analytical model and dagSim run very fast, requiring at least two orders of magnitude lower execution time than JMT while providing slightly better accuracy, being thus practical for online prediction.
- Published
- 2020
6. An optimization framework for the capacity allocation and admission control of MapReduce jobs in cloud systems
- Author
-
Malekimajd, M, Ardagna, D, Ciavotta, M, Gianniti, E, Passacantando, M, Rizzi, A, Rizzi, AM, Malekimajd, M, Ardagna, D, Ciavotta, M, Gianniti, E, Passacantando, M, Rizzi, A, and Rizzi, AM
- Abstract
Nowadays, we live in a Big Data world and many sectors of our economy are guided by data-driven decision processes. Big Data and Business Intelligence applications are facilitated by the MapReduce programming model, while, at infrastructural layer, cloud computing provides flexible and cost-effective solutions to provide on-demand large clusters. Capacity allocation in such systems, meant as the problem of providing computational power to support concurrent MapReduce applications in a cost-effective fashion, represents a challenge of paramount importance. In this paper we lay the foundation for a solution implementing admission control and capacity allocation for MapReduce jobs with a priori deadline guarantees. In particular, shared Hadoop 2.x clusters supporting batch and/or interactive jobs are targeted. We formulate a linear programming model able to minimize cloud resources costs and rejection penalties for the execution of jobs belonging to multiple classes with deadline guarantees. Scalability analyses demonstrated that the proposed method is able to determine the global optimal solution of the linear problem for systems including up to 10,000 classes in less than 1 s.
- Published
- 2018
7. Game-Theoretic Approach to Joint Admission Control and Capacity Allocation for MapReduce
- Author
-
Ciavotta, M, Gianniti, E, Ardagna, D, Passacantando, M, Ciavotta, M, Gianniti, E, Ardagna, D, and Passacantando, M
- Subjects
Admission Control ,Generalized Nash Equilibrium Problem ,Capacity Allocation - Published
- 2015
8. Fluid Petri nets for the performance evaluation of MapReduce applications
- Author
-
Gianniti, E., Rizzi, A. M., Barbierato, Enrico, Gribaudo, M., Ardagna, D., Barbierato E. (ORCID:0000-0003-1466-0248), Gianniti, E., Rizzi, A. M., Barbierato, Enrico, Gribaudo, M., Ardagna, D., and Barbierato E. (ORCID:0000-0003-1466-0248)
- Abstract
Big Data applications allow to successfully analyze large amounts of data not necessarily structured, though at the same time they present new challenges. For example, predicting the performance of frameworks such as Hadoop can be a costly task, hence the necessity to provide models that can be a valuable support for designers and developers. This paper provides a new contribution in studying a novel modeling approach based on fluid Petri nets to predict MapReduce jobs execution time. The experiments we performed at CINECA, the Italian supercomputing center, have shown that the achieved accuracy is within 16% of the actual measurements on average.
- Published
- 2017
9. A Game-Theoretic Approach for Runtime Capacity Allocation in MapReduce
- Author
-
Gianniti, E, Ardagna, D, Ciavotta, M, Passacantando, M, Gianniti, E, Ardagna, D, Ciavotta, M, and Passacantando, M
- Abstract
Nowadays many companies have available large amounts of raw, unstructured data. Among Big Data enabling technologies, a central place is held by the MapReduce framework and, in particular, by its open source implementation, Apache Hadoop. For cost effectiveness considerations, a common approach entails sharing server clusters among multiple users. The underlying infrastructure should provide every user with a fair share of computational resources, ensuring that service level agreements (SLAs) are met and avoiding wastes. In this paper we consider mathematical models for the optimal allocation of computational resources in a Hadoop 2.x cluster with the aim to develop new capacity allocation techniques that guarantee better performance in shared data centers. Our goal is to get a substantial reduction of power consumption while respecting the deadlines stated in the SLAs and avoiding penalties associated with job rejections. The core of this approach is a distributed algorithm for runtime capacity allocation, based on Game Theory models and techniques, that mimics the MapReduce dynamics by means of interacting players, namely the central Resource Manager and Class Managers.
- Published
- 2017
10. Capacity allocation for big data applications in the cloud
- Author
-
Ciavotta, M, Gianniti, E, Ardagna, D, Ciavotta, M, Gianniti, E, and Ardagna, D
- Abstract
The aim of this work is to present the problem of Capacity Allocation for multiple classes of Big Data applications running in the Cloud. The objective is the minimization of the renting out costs subject to the fulfillment of QoS requirements expressed in terms of application deadlines. We propose a preliminary version of a tool embedding a local- search-based algorithm exploiting also an integer nonlinear mathematical formulation and a queueing network simulation to solve the problem.
- Published
- 2017
11. D-SPACE4Cloud: A design tool for big data applications
- Author
-
Ciavotta, M, Gianniti, E, Ardagna, D, Ciavotta, M, Gianniti, E, and Ardagna, D
- Abstract
The last years have seen a steep rise in data generation worldwide, with the development and widespread adoption of several software projects targeting the Big Data paradigm. Many companies currently engage in Big Data analytics as part of their core business activities, nonetheless there are no tools or techniques to support the design of the underlying infrastructure configuration backing such systems. In particular, the focus in this paper is set on Cloud deployed clusters, which represent a cost-effective alternative to on premises installations. We propose a novel tool implementing a battery of optimization and prediction techniques integrated so as to efficiently assess several alternative resource configurations, in order to determine the minimum cost cluster deployment satisfying Quality of Service constraints. Further, the experimental campaign conducted on real systems shows the validity and relevance of the proposed method.
- Published
- 2016
12. Optimizing Quality-Aware Big Data Applications in the Cloud
- Author
-
Danilo Ardagna, Michele Ciavotta, Eugenio Gianniti, Gianniti, E, Ciavotta, M, and Ardagna, D
- Subjects
Big Data ,Optimization ,Computer Networks and Communications ,Computer science ,Test data generation ,Process (engineering) ,media_common.quotation_subject ,Cloud computing , Big Data , Tools , Optimization , Unified modeling language ,Big data ,Cloud computing ,02 engineering and technology ,ING-INF/05 - SISTEMI DI ELABORAZIONE DELLE INFORMAZIONI ,Data modeling ,Tools ,performance of system ,Nonlinear programming ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Quality (business) ,media_common ,business.industry ,Quality of service ,INF/01 - INFORMATICA ,020207 software engineering ,distributed system ,Data science ,Computer Science Applications ,Hardware and Architecture ,Software deployment ,MAT/09 - RICERCA OPERATIVA ,business ,Software ,Unified modeling language ,Information Systems - Abstract
The last years witnessed a steep rise in data generation worldwide and, consequently, the widespread adoption of software solutions able to support data-intensive application. Competitiveness and innovation have strongly benefited from these new platforms and methodologies, and there is a great deal of interest around the new possibilities that Big Data analytics promise to make reality. Many companies currently engage in data-intensive processes as part of their core businesses; however, fully embracing the data-driven paradigm is still cumbersome, and establishing a production-ready, fine-tuned deployment is time-consuming, expensive, and resource-intensive. This situation calls for innovative models and techniques to streamline the process of deployment configuration for Big Data applications. In particular, the focus in this paper is on the rightsizing of Cloud deployed clusters, which represent a cost-effective alternative to installation on premises. This paper proposes a novel tool, integrated in a wider DevOps-inspired approach, implementing a parallel and distributed simulation-optimization technique that efficiently and effectively explores the space of alternative Cloud configurations, seeking the minimum cost deployment that satisfies quality of service constraints. The soundness of the proposed solution has been thoroughly validated in a vast experimental campaign encompassing different applications and Big Data platforms.
- Published
- 2021
13. An optimization framework for the capacity allocation and admission control of MapReduce jobs in cloud systems
- Author
-
Danilo Ardagna, Marzieh Malekimajd, Alessandro Maria Rizzi, Mauro Passacantando, Eugenio Gianniti, Michele Ciavotta, Malekimajd, M, Ardagna, D, Ciavotta, M, Gianniti, E, Passacantando, M, and Rizzi, A
- Subjects
Optimization ,Linear programming ,Computer science ,Map Reduce ,Distributed computing ,Big data ,Cloud computing ,02 engineering and technology ,ING-INF/05 - SISTEMI DI ELABORAZIONE DELLE INFORMAZIONI ,Theoretical Computer Science ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Resource allocation ,business.industry ,INF/01 - INFORMATICA ,020206 networking & telecommunications ,Admission control ,Hardware and Architecture ,MAT/09 - RICERCA OPERATIVA ,Scalability ,Business intelligence ,Programming paradigm ,business ,Map Reduce, Resource allocation, Optimization ,Software ,Information Systems - Abstract
Nowadays, we live in a Big Data world and many sectors of our economy are guided by data-driven decision processes. Big Data and Business Intelligence applications are facilitated by the MapReduce programming model, while, at infrastructural layer, cloud computing provides flexible and cost-effective solutions to provide on-demand large clusters. Capacity allocation in such systems, meant as the problem of providing computational power to support concurrent MapReduce applications in a cost-effective fashion, represents a challenge of paramount importance. In this paper we lay the foundation for a solution implementing admission control and capacity allocation for MapReduce jobs with a priori deadline guarantees. In particular, shared Hadoop2.x clusters supporting batch and/or interactive jobs are targeted. We formulate a linear programming model able to minimize cloud resources costs and rejection penalties for the execution of jobs belonging to multiple classes with deadline guarantees. Scalability analyses demonstrated that the proposed method is able to determine the global optimal solution of the linear problem for systems including up to 10,000 classes in less than 1s.
- Published
- 2018
14. Capacity allocation for big data applications in the cloud
- Author
-
Danilo Ardagna, Michele Ciavotta, Eugenio Gianniti, Ciavotta, M, Gianniti, E, and Ardagna, D
- Subjects
0209 industrial biotechnology ,Mathematical optimization ,Computer science ,Real-time computing ,Big data ,QoS ,Cloud computing ,02 engineering and technology ,ING-INF/05 - SISTEMI DI ELABORAZIONE DELLE INFORMAZIONI ,Network simulation ,020901 industrial engineering & automation ,0202 electrical engineering, electronic engineering, information engineering ,Queueing theory ,business.industry ,Quality of service ,020208 electrical & electronic engineering ,INF/01 - INFORMATICA ,Capacity allocation ,Computer Science Applications1707 Computer Vision and Pattern Recognition ,Hardware and Architecture ,MAT/09 - RICERCA OPERATIVA ,Embedding ,Minification ,business ,Cloud ,Software ,Integer (computer science) - Abstract
The aim of this work is to present the problem of Capacity Allocation for multiple classes of Big Data applications running in the Cloud. The objective is the minimization of the renting out costs subject to the fulfillment of QoS requirements expressed in terms of application deadlines. We propose a preliminary version of a tool embedding a local- search-based algorithm exploiting also an integer nonlinear mathematical formulation and a queueing network simulation to solve the problem.
- Published
- 2017
15. A Game-Theoretic Approach for Runtime Capacity Allocation in MapReduce
- Author
-
Michele Ciavotta, Danilo Ardagna, Eugenio Gianniti, Mauro Passacantando, Gianniti, E, Ardagna, D, Ciavotta, M, and Passacantando, M
- Subjects
FOS: Computer and information sciences ,Computer science ,Cost effectiveness ,Computer Networks and Communications ,Distributed computing ,Admission Control ,Capacity Allocation ,Generalized Nash Equilibrium Problem ,Hadoop ,Hardware and Architecture ,Big data ,0211 other engineering and technologies ,02 engineering and technology ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Resource management ,021103 operations research ,business.industry ,Unstructured data ,Core (game theory) ,Computer Science - Distributed, Parallel, and Cluster Computing ,Distributed algorithm ,Service level ,Distributed, Parallel, and Cluster Computing (cs.DC) ,business ,Game theory - Abstract
Nowadays many companies have available large amounts of raw, unstructured data. Among Big Data enabling technologies, a central place is held by the MapReduce framework and, in particular, by its open source implementation, Apache Hadoop. For cost effectiveness considerations, a common approach entails sharing server clusters among multiple users. The underlying infrastructure should provide every user with a fair share of computational resources, ensuring that Service Level Agreements (SLAs) are met and avoiding wastes. In this paper we consider two mathematical programming problems that model the optimal allocation of computational resources in a Hadoop 2.x cluster with the aim to develop new capacity allocation techniques that guarantee better performance in shared data centers. Our goal is to get a substantial reduction of power consumption while respecting the deadlines stated in the SLAs and avoiding penalties associated with job rejections. The core of this approach is a distributed algorithm for runtime capacity allocation, based on Game Theory models and techniques, that mimics the MapReduce dynamics by means of interacting players, namely the central Resource Manager and Class Managers.
- Published
- 2017
16. D-SPACE4Cloud: A Design Tool for Big Data Applications
- Author
-
Danilo Ardagna, Eugenio Gianniti, Michele Ciavotta, Ciavotta, M, Gianniti, E, and Ardagna, D
- Subjects
FOS: Computer and information sciences ,Optimization ,Queueing network ,Computer science ,Test data generation ,Big data ,Cloud computing ,02 engineering and technology ,MapReduce ,Queueing networks ,Theoretical Computer Science ,Computer Science (all) ,Software ,Resource (project management) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Computer Science - Performance ,business.industry ,Quality of service ,Design tool ,020207 software engineering ,Data science ,Performance (cs.PF) ,Computer Science - Distributed, Parallel, and Cluster Computing ,Software deployment ,Distributed, Parallel, and Cluster Computing (cs.DC) ,business - Abstract
The last years have seen a steep rise in data generation worldwide, with the development and widespread adoption of several software projects targeting the Big Data paradigm. Many companies currently engage in Big Data analytics as part of their core business activities, nonetheless there are no tools or techniques to support the design of the underlying infrastructure configuration backing such systems. In particular, the focus in this paper is set on Cloud deployed clusters, which represent a cost-effective alternative to on premises installations. We propose a novel tool implementing a battery of optimization and prediction techniques integrated so as to efficiently assess several alternative resource configurations, in order to determine the minimum cost cluster deployment satisfying Quality of Service constraints. Further, the experimental campaign conducted on real systems shows the validity and relevance of the proposed method.
- Published
- 2016
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.