11 results on '"Wu, Yongwei"'
Search Results
2. DABGPM: A Double Auction Bayesian Game-Based Pricing Model in Cloud Market
- Author
-
Shang, Shifeng, Jiang, Jinlei, Wu, Yongwei, Huang, Zhenchun, Yang, Guangwen, Zheng, Weimin, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Ding, Chen, editor, Shao, Zhiyuan, editor, and Zheng, Ran, editor
- Published
- 2010
- Full Text
- View/download PDF
3. I/O-Conscious and Prediction-Enabled Virtual Machines Scheduling
- Author
-
Wu Yongwei, Weimin Zheng, Xun Zhao, and Jiang Jinlei
- Subjects
Rate-monotonic scheduling ,Earliest deadline first scheduling ,I/O scheduling ,Computer science ,Distributed computing ,Temporal isolation among virtual machines ,Processor scheduling ,050801 communication & media studies ,Cloud computing ,Dynamic priority scheduling ,computer.software_genre ,Fair-share scheduling ,Scheduling (computing) ,0508 media and communications ,Fixed-priority pre-emptive scheduling ,Input/output ,Job shop scheduling ,business.industry ,05 social sciences ,Flow shop scheduling ,Virtualization ,Round-robin scheduling ,Virtual machine ,Two-level scheduling ,0509 other social sciences ,050904 information & library sciences ,business ,computer - Abstract
Virtual machine (VM) technology plays an important role in modern cluster and cloud computing environments due to its advantages in application isolation, resource partition and load consolidation. Since a large number of VMs usually run simultaneously in these environments, VMs scheduling is key to achieve resources efficiency. Unfortunately, most existing scheduling strategies are complicated, with unacceptable overhead in a large-scale virtualization environment. Further more, they tend to produce biased results without considering the I/O characteristics of VMs and physical machines (PMs). To deal with the issue, this paper proposes a new strategy for scheduling VMs over physical clusters. Taking into consideration the I/O characteristics, the proposed strategy can obtain better scheduling results while with lower overhead. Moreover, a prediction algorithm is suggested, which, when combined with the proposed strategy, can further improve the scheduling results. The proposed strategy and prediction algorithm have been implemented and evaluated with the SPEC CPU 2006 and Netperf benchmarks. The experimental results show that the average completion time of SPEC reduces by about 23%, whereas the standard deviation of Netperf average bandwidth decreases from 10.9 to 4.2.
- Published
- 2016
4. Systematic Data Placement Optimization in Multi-Cloud Storage for Complex Requirements.
- Author
-
Su, Maomeng, Zhang, Lei, Wu, Yongwei, Chen, Kang, and Li, Keqin
- Subjects
CLOUD computing ,CLOUD storage ,PROGRAM transformation ,FAULT tolerance (Engineering) ,LINEAR programming - Abstract
Multi-cloud storage can provide better features such as availability and scalability. Current works use multiple cloud storage providers with erasure coding to achieve certain benefits including fault-tolerance improving or vendor lock-in avoiding. However, these works only use the multi-cloud storage in ad-hoc ways, and none of them considers the optimization issue in general. In fact, the key to optimize the multi-cloud storage is to effectively choose providers and erasure coding parameters. Meanwhile, the data placement should satisfy system or application developers’ requirements. As developers often demand various objectives to be optimized simultaneously, such complex requirement optimization cannot be easily fulfilled by ad-hoc ways. This paper presents Triones, a systematic model to formally formulate data placement in multi-cloud storage by using erasure coding. Firstly, Triones addresses the problem of data placement optimization by applying non-linear programming and geometric space abstraction. It could satisfy complex requirements involving multi-objective optimization. Secondly, Triones can effectively balance among different objectives in optimization and is scalable to incorporate new ones. The effectiveness of the model is proved by extensive experiments on multiple cloud storage providers in the real world. For simple requirements, Triones can achieve 50 percent access latency reduction, compared with the model in $\mu$
LibCloud. For complex requirements, Triones can improve fault-tolerance level by 2 $\times$- Published
- 2016
- Full Text
- View/download PDF
5. Cloud Performance Modeling with Benchmark Evaluation of Elastic Scaling Strategies.
- Author
-
Hwang, Kai, Bai, Xiaoying, Shi, Yue, Li, Muyang, Chen, Wen-Guang, and Wu, Yongwei
- Subjects
CLOUD computing ,SOFTWARE as a service ,DISTRIBUTED computing ,CLOUD storage ,WEB services - Abstract
In this paper, we present generic cloud performance models for evaluating Iaas, PaaS, SaaS, and mashup or hybrid clouds. We test clouds with real-life benchmark programs and propose some new performance metrics. Our benchmark experiments are conducted mainly on IaaS cloud platforms over scale-out and scale-up workloads. Cloud benchmarking results are analyzed with the efficiency, elasticity, QoS, productivity, and scalability of cloud performance. Five cloud benchmarks were tested on Amazon IaaS EC2 cloud: namely YCSB, CloudSuite, HiBench, BenchClouds, and TPC-W. To satisfy production services, the choice of scale-up or scale-out solutions should be made primarily by the workload patterns and resources utilization rates required. Scaling-out machine instances have much lower overhead than those experienced in scale-up experiments. However, scaling up is found more cost-effective in sustaining heavier workload. The cloud productivity is greatly attributed to system elasticity, efficiency, QoS and scalability. We find that auto-scaling is easy to implement but tends to over provision the resources. Lower resource utilization rate may result from auto-scaling, compared with using scale-out or scale-up strategies. We also demonstrate that the proposed cloud performance models are applicable to evaluate PaaS, SaaS and hybrid clouds as well. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
6. Improving the Effective IO Throughput by Adaptive Read-Ahead Strategy for Private Cloud Storage Service.
- Author
-
Wang, Qiuping, Chen, Kang, Wu, Yongwei, and Zheng, Weimin
- Abstract
We have employed the Linux on-demand read-ahead framework in our campus wide storage service system call MeePo. The appropriate read-ahead mechanism can significantly increase IO throughput and improve user experiences by hiding network latency which is critical for real-time applications. Our strategy is based on the data accessing characteristic of MeePo system. The read-ahead framework uses the strategy profile which is generated according to the analysis of access trace of a typical user in a storage community. Our test deployment environment involves more than 5000 registered users as well as 150+ communities. Based on our observation that most of files in our system have either sequential or interleaved accessing patterns. In such scenario, client IO throughout could increase 12\% for sequential stream and more than 180\% improvement for interleaved stream. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
7. EasyDeploy: Automatic Application Deployment in Virtual Clusters.
- Author
-
Gao, Tao, Xu, Yanjun, Wang, Xiaoying, Jiang, Jinlei, and Wu, Yongwei
- Abstract
Along with the fast development of Cloud computing, it has become a trend to use virtual clusters for scientific and business works. In spite of the fact, it is a big challenge to set up a virtual cluster to meet the user-specific requirement such as the applications to be used. In this paper we design and implement Easy Deploy, a system that can set up virtual clusters with user-specifying applications in Cloud computing environment automatically. Easy Deploy realizes its own automatic application deployment method in virtual clusters without the help of external tools for traditional clusters. It decouples application packages away from virtual machine images to save storage space. To reduce application package transfer time, cache and prefetching mechanism is provided. The experimental results show that in our settings we can create an eighteen nodes virtual cluster with Hadoop environment in less than 50 seconds. The cache and prefetching mechanism we designed can do reduce the transfer time of application packages. When we use both of them to create a virtual cluster, the transfer time will reduce by three times than that in the case without any optimization strategy. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
8. µLibCloud: Providing High Available and Uniform Accessing to Multiple Cloud Storages.
- Author
-
Mu, Shuai, Chen, Kang, Gao, Pin, Ye, Feng, Wu, Yongwei, and Zheng, Weimin
- Abstract
The increasing popularity of cloud storage services attracts large amounts of companies to store their data in cloud instead of building their own infrastructures. With large amounts of data stored in the cloud, it is expected to provide high availability and fine global access experiences. However, there are still major concerns of the availability of major cloud services, especially in a sparsely connected global network with complicated issues. In this paper, we introduce µLibCloud, a system based on Apache libCloud, aiming to improve the availability and global access experience of clouds, and to tolerate provider failures and outages. µLibCloud works as a library at client side, transparently spreading and collecting data smartly to/from different cloud providers through erasure code. In evaluation, we deployed the system into 7 major cloud providers and run a global benchmarks from 9 locations around the world. The results were compared to the original clouds and a content delivery network. We observed that µLibCloud achieved a higher and more uniformed read availability in most cases, with reasonable estimated extra costs. For example, the read latency of some original providers could be reduced by 50% -- 70% at different locations. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
9. NO2: Speeding up Parallel Processing of Massive Compute-Intensive Tasks.
- Author
-
Wu, Yongwei, Guo, Weichao, Ren, Jinglei, Zhao, Xun, and Zheng, Weimin
- Subjects
- *
PARALLEL processing , *CLOUD computing , *COMPUTER software , *CLUSTER analysis (Statistics) , *ELECTRONIC data processing - Abstract
Large-scale computing frameworks, either tenanted on the cloud or deployed in the high-end local cluster, have become an indispensable software infrastructure to support numerous enterprise and scientific applications. Tasks executed on these frameworks are generally classified into data-intensive and compute-intensive ones. However, most existing frameworks, led by MapReduce, are mainly suitable for data-intensive tasks. Their task schedulers assume that the proportion of data I/O reflects the task progress and state. Unfortunately, this assumption does not apply to most compute-intensive tasks. Due to biased estimation of task progress, traditional frameworks cannot timely cut off outliers and therefore largely prolong execution time when performing compute-intensive tasks. We propose a new framework designed for compute-intensive tasks. By using instrumentation and automatic instrument point selector, our framework estimates the compute-intensive task progress without resorting to data I/O. We employ a clustering method to identify outliers at runtime and perform speculative execution/aborting, speeding up task execution by up to 25%. Moreover, our improvement to bare instrumentation limits overhead within 0.1%, and the aborting-based execution only introduces 10% more average CPU usage. Low overhead and resource consumption make our framework practically usable in the production environment. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
10. Liquid: A Scalable Deduplication File System for Virtual Machine Images.
- Author
-
Zhao, Xun, Zhang, Yang, Wu, Yongwei, Chen, Kang, Jiang, Jinlei, and Li, Keqin
- Subjects
VIRTUAL machine systems ,IMAGE processing ,CLOUD computing ,COMPUTER input-output equipment ,STORAGE area networks (Computer networks) ,IMAGE storage & retrieval systems - Abstract
A virtual machine (VM) has been serving as a crucial component in cloud computing with its rich set of convenient features. The high overhead of a VM has been well addressed by hardware support such as Intel virtualization technology (VT), and by improvement in recent hypervisor implementation such as Xen, KVM, etc. However, the high demand on VM image storage remains a challenging problem. Existing systems have made efforts to reduce VM image storage consumption by means of deduplication within a storage area network (SAN) cluster. Nevertheless, an SAN cannot satisfy the increasing demand of large-scale VM hosting for cloud computing because of its cost limitation. In this paper, we propose Liquid, a scalable deduplication file system that has been particularly designed for large-scale VM deployment. Its design provides fast VM deployment with peer-to-peer (P2P) data transfer and low storage consumption by means of deduplication on VM images. It also provides a comprehensive set of storage features including instant cloning for VM images, on-demand fetching through a network, and caching with local disks by copy-on-read techniques. Experiments show that Liquid's features perform well and introduce minor performance overhead. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
11. Modeling of Distributed File Systems for Practical Performance Analysis.
- Author
-
Wu, Yongwei, Ye, Feng, Chen, Kang, and Zheng, Weimin
- Subjects
- *
CLOUD computing , *ELECTRONIC file management , *MIDDLEWARE , *ELECTRONIC data processing , *COMPUTER architecture , *DATA modeling , *MANAGEMENT - Abstract
Cloud computing has received significant attention recently. Delivering quality guaranteed services in clouds is highly desired. Distributed file systems (DFSs) are the key component of any cloud-scale data processing middleware. Evaluating the performance of DFSs is accordingly very important. To avoid cost for late life cycle performance fixes and architectural redesign, providing performance analysis before the deployment of DFSs is also particularly important. In this paper, we propose a systematic and practical performance analysis framework, driven by architecture and design models for defining the structure and behavior of typical master/slave DFSs. We put forward a configuration guideline for specifications of configuration alternatives of such DFSs, and a practical approach for both qualitatively and quantitatively performance analysis of DFSs with various configuration settings in a systematic way. What distinguish our approach from others is that 1) most of existing works rely on performance measurements under a variety of workloads/strategies, comparing with other DFSs or running application programs, but our approach is based on architecture and design level models and systematically derived performance models; 2) our approach is able to both qualitatively and quantitatively evaluate the performance of DFSs; and 3) our approach not only can evaluate the overall performance of a DFS but also its components and individual steps. We demonstrate the effectiveness of our approach by evaluating Hadoop distributed file system (HDFS). A series of real-world experiments on EC2 (Amazon Elastic Compute Cloud), Tansuo and Inspur Clusters, were conducted to qualitatively evaluate the effectiveness of our approach. We also performed a set of experiments of HDFS on EC2 to quantitatively analyze the performance and limitation of the metadata server of DFSs. Results show that our approach can achieve sufficient performance analysis. Similarly, the proposed approach could be also applied to evaluate other DFSs such as MooseFS, GFS, and zFS. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.