Haikun Liu

Huazhong University of Science and Technology, Wu-han-shih, Hubei, China

Are you Haikun Liu?

Claim your profile

Publications (15)5.94 Total impact

  • [Show abstract] [Hide abstract]
    ABSTRACT: Server consolidation using virtualization technologies allow large-scale datacenters to improve resource utilization and energy efficiency. However, most existing consolidation strategies solely focused on balancing the tradeoff between performance service-level-agreements (SLAs) desired by cloud applications and energy costs consumed by hosting servers. With the presence of fluctuating workloads in datacenters, the lifetime and reliability of servers under dynamic power-aware consolidation could be adversely impacted by repeated on–off thermal cycles, ware-and-tear and temperature rise. In this paper, we propose a Reliability-Aware server Consolidation stratEgy, named RACE, to address when and how to perform energy-efficient server consolidation in a reliability-friendly and profitable way. The focus is on the characterization and analysis of this problem as a multi-objective optimization, by developing a utility model that unifies multiple constraints on performance SLAs, reliability factors and energy costs in a holistic manner. An improved grouping genetic algorithm is proposed to search the global optimal solution, which takes advantage of a collection of reliability-aware resource buffering and virtual machines-to-servers re-mapping heuristics for generating good initial solutions and improving the convergence rate. Extensive simulations are conducted to validate the effectiveness, scalability and overhead of RACE—in improving the overall utility of datacenters while avoiding unprofitable consolidation in the long term—compared with pMapper and PADD strategies for server consolidation. Copyright © 2013 John Wiley & Sons, Ltd.
    International Journal of Communication Systems 10/2013; · 1.11 Impact Factor
  • [Show abstract] [Hide abstract]
    ABSTRACT: Server consolidation using virtualization technologies allow cloud-scale datacenters to improve resource utilization and energy efficiency. However, most existing consolidation strategies solely focused on balancing the tradeoff between service-level-agreements (SLAs) desired by cloud applications and energy costs consumed by hosting servers. With the presence of fluctuating workloads in datacenters, the lifetime and reliability of servers under dynamic power-aware consolidation could be adversely impacted by repeated on-off thermal cycles, wear-and-tear and temperature rise. In this paper, we propose a Reliability-Aware server Consolidation stratEgy, named RACE, to address when and how to perform energy-efficient server consolidation in a reliability-friendly and profitable way. The focus is on the characterization and analysis of this problem as a multi-objective optimization, by developing an utility model that unifies multiple constraints on performance SLAs, reliability factors, and energy costs in a holistic manner. An improved grouping genetic algorithm is proposed to search the global optimal solution, which takes advantage of a collection of reliability-aware resource buffering, and virtual machines-to-servers re-mapping heuristics for generating good initial solutions and improving the convergence rate. Extensive simulations are conducted to validate the effectiveness, scalability and overhead of RACE in improving the overall utility of datacenters while avoiding unprofitable consolidation in the long term — compared with pMapper and PADD strategies for server consolidation.
    Proceedings of the 2012 IEEE 4th International Conference on Cloud Computing Technology and Science (CloudCom); 12/2012
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Live migration of virtual machines (VM) across physical hosts provides a significant new benefit for administrators of data centers and clusters. Previous memory-to-memory approaches demonstrate the effectiveness of live VM migration in local area networks (LAN), but they would cause a long period of downtime in a wide area network (WAN) environment. This paper describes the design and implementation of a novel approach, namely, CR/TR-Motion, which adopts checkpointing/recovery and trace/replay technologies to provide fast, transparent VM migration for both LAN and WAN environments. With execution trace logged on the source host, a synchronization algorithm is performed to orchestrate the running source and target VMs until they reach a consistent state. CR/TR-Motion can greatly reduce the migration downtime and network bandwidth consumption. Experimental results show that the approach can drastically reduce migration overheads compared with memory-to-memory approach in a LAN: up to 72.4 percent on application observed downtime, up to 31.5 percent on total migration time, and up to 95.9 percent on the data to synchronize the VM state. The application performance overhead due to migration is kept within 8.54 percent on average. The results also show that for a variety of workloads migrated across WANs, the migration downtime is less than 300 milliseconds.
    IEEE Transactions on Parallel and Distributed Systems 01/2012; · 1.80 Impact Factor
  • Xiaofei Liao, Hai Jin, Haikun Liu
    [Show abstract] [Hide abstract]
    ABSTRACT: Since power is one of the major limiting factors for a data center or for large cluster growth, the objective of this study is to minimize the power consumption of the cluster without violating the performance constraints of the applications. We propose a runtime virtual machine (VM) mapping framework in a cluster or data center to save energy. The new framework can make reconfiguration decisions on time with the consideration of a low influence on the performance. In the GreenMap framework, one probabilistic, heuristic algorithm is designed for the optimization problem: mapping VMs onto a set of physical machines (PMs) under the constraint of multi-dimensional resource consumptions. Experimental measurements show that the new method can reduce the power consumption by up to 69.2% over base, with few performance penalties. The effectiveness and performance insights are also analytically verified.
    Future Generation Comp. Syst. 01/2012; 28:469-477.
  • [Show abstract] [Hide abstract]
    ABSTRACT: Desktop virtualization is a very hot concept in both industry and academic communities. Since virtualized desktop system is based on multiple virtual machines (VM), it is necessary to design a distributed storage system to manage the VM images. In this paper, we design a distributed storage system, VMStore, by taking into account three important characteristics: high performance VM snapshot, booting optimization from multiple images and redundancy removal of images data. We adopt a direct index structure of blocks for VM snapshots to speed up VM booting performance significantly; provide a distribute storage structure with good bandwidth scalability by dynamically changing the number of storage nodes; and propose a data preprocessing strategy with intelligent object partitioning techniques, which would eliminate duplication more effectively. Performance analysis for VMStore focuses on two metrics: the speedup of VM booting and the overhead of de-duplication. Experimental results show the efficiency and effectiveness of VMStore.
    Sciece China. Information Sciences 01/2011; 54:1104-1118. · 0.71 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Live migration of virtual machine (VM) provides a significant benefit for virtual server mobility without disrupting service. It is widely used for system management in virtualized data centers. However, migration costs may vary significantly for different workloads due to the variety of VM configurations and workload characteristics. To take into account the migration overhead in migration decision-making, we investigate design methodologies to quantitatively predict the migration performance and energy cost. We thoroughly analyze the key parameters that affect the migration cost from theory to practice. We construct two application-oblivious models for the cost prediction by using learned knowledge about the workloads at the hypervisor (also called VMM) level. This should be the first kind of work to estimate VM live migration cost in terms of both performance and energy in a quantitative approach. We evaluate the models using five representative workloads on a Xen virtualized environment. Experimental results show that the refined model yields higher than 90% prediction accuracy in comparison with measured cost. Model-guided decisions can significantly reduce the migration cost by more than 72.9% at an energy saving of 73.6%.
    Proceedings of the 20th ACM International Symposium on High Performance Distributed Computing, HPDC 2011, San Jose, CA, USA, June 8-11, 2011; 01/2011 · 0.78 Impact Factor
  • [Show abstract] [Hide abstract]
    ABSTRACT: Virtualization is being widely used now as an emerging trend. Rapid improvements in network bandwidth, ubiquitous security hazards and high total cost of ownership of personal computers have created a growing market for desktop virtualization. Much like server virtualization, virtualizing desktops involves separating the physical location of a client device from its logical interface. But, the performance and usability of some traditional desktop frameworks do not satisfy end-users. Other solutions, including WebOS, which needs to rebuild all daily-used applications into Client/Server mode, cannot be easily accepted by people in a short time. We present LVD, a system that combines the virtualization technology and inexpensive personal computers (PCs) to realize a lightweight virtual desktop system. Comparing to the previous desktop systems, LVD builds an integrated novel desktop environment, which can support the backup, mobility, suspending and resuming of per-user's working environment, and support synchronous using of incompatible applications on different platforms and achieves great saving in power consumption. We have implemented LVD in a cluster with Xen and compared its performance against widely used commercial approaches, including Microsoft RDP, Citrix MetaFrameXP and Sun Ray. Experimental results demonstrate that LVD is effective in performing the functions while imposing little overhead. Copyright © 2009 John Wiley & Sons, Ltd.
    Concurrency and Computation Practice and Experience 01/2010; 22:419-440. · 0.85 Impact Factor
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Virtual machine (VM) based logging-and-replay technology has attracted much attention for system security, fault tolerance, and debugging. However, so far there is not a replay system designed on virtual machine monitor Xen. In this paper, XenLR presents the design and implementation of a logging tool for full system replay on Xen. To reduce the design complexity, XenLR is achieved on a lightweight VM mini OS. Our preliminary work dedicates to identify architectural non-deterministic events and record them compactly. XenLR does not modify the guest OS and only need to record the external inputs (keyboard input and time update on mini OS). The log data are saved in the privileged domain¿s file system through transmission channels. The experiments indicate that XenLR results in low time and space overhead. Overhead due to logging is imperceptible for interactive use. The log files grow at 1.4 MByte/day, so it causes litter space consumption for today¿s large capacity disk storage.
    Frontier of Computer Science and Technology, 2008. FCST '08. Japan-China Joint Workshop on; 01/2009
  • Tao Cui, Hai Jin, Xiaofei Liao, Haikun Liu
    [Show abstract] [Hide abstract]
    ABSTRACT: Operating system debugging and system security are two important issues which are attracted more and more attentions. However, traditional solutions of software debugging can not make an integrated replay towards the status of operating system. Moreover, most intrusion detection methods all depend on the operating system excessively, but operating system ex-poses so many interfaces to outside that it can not ensure its own security and is attacked vulnerably. Therefore, a para-virtualized model based full system replay is developed in this paper. According to the para-virtualized device model pro-vided by Xen, this system captures the non-deterministic events in the target operating system through a real-time monitoring module, and saves the data and time point of these events occurrence as log file elaborately. Using these data, we can make a full-system replay to the target operating system accurately. Compared with the previous replay systems, this system is more efficient and has two prominent advantages: no need to modify the target operating system, and no need to reboot the target operating system before executing replay.
    NPC 2009, Sixth IFIP International Conference on Network and Parallel Computing, Gold Coast, Australia, October 19-21, 2009; 01/2009
  • [Show abstract] [Hide abstract]
    ABSTRACT: This paper describes the scheme of thread migration in a language virtual machine LAOVM - lightweight application-oriented virtual machine. To maintain high portability and flexibility, thread migration is implemented at language level. To migrate process or thread in runtime in LAOVM, the following issues are addressed: (1) supporting kinds of applications implemented by different programming language; (2) enabling transparent process migration indifferent operating system; (3) implementing fine-grained mobility. Experimental results show that the downtime of thread migration is reasonably small, even the application runtime image is required for occupying massive memory blocks.
    8th IEEE/ACIS International Conference on Computer and Information Science, IEEE/ACIS ICIS 2009, June 1-3, 2009, Shanghai, China; 01/2009
  • Source
    Haikun Liu, Hai Jin, Xiaofei Liao
    [Show abstract] [Hide abstract]
    ABSTRACT: Virtual Machine (VM) level checkpoints bring sever-al advantages which process-level checkpoint implementation can hardly provide: compatibility, transparence, flexibility, and simplicity. However, the size of VM-level checkpoint may be very large and even in the order of gigabytes. This disad-vantage causes the VM checkpointing and restart time become very long. To reduce the size of VM checkpoint, this paper proposes a memory exclusion scheme using ballooning me-chanism, which omits saving unnecessary free pages in the VM. We implement our prototype in Xen environment. Expe-rimental measurements show our approach can significantly reduce the size of VM checkpoint with minimal runtime over-head, thereby greatly improve the checkpoint performance.
    01/2009;
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Live migration of virtual machines (VM) across distinct physical hosts provides a significant new benefit for administrators of data centers and clusters. However, previous memory-to-memory ap- proaches only make the live VM migration be applied in local area networks (LAN), and their im- proved methods for wide area networks (WAN) environment bring the VM migration with a long period of downtime. This paper describes the design and implementation of a novel approach that adopts checkpointing/recovery and trace/replay technology to provide fast, transparent VM migra- tion for both LAN and WAN environments. With execution trace logged on the source host, a syn- chronization algorithm is performed to orchestrate the running source and target VM until they get a consistent state. This scheme can greatly reduce the migration downtime and network bandwidth consumption. Theoretic analysis indicates that the service downtime to migrate a VM across WANs is equivalent to the downtime cost while migrating a VM in a LAN. Experimental measurements show that for a variety of workloads migrated in a LAN or across WANs, the migration downtime is only 200 milliseconds with a little difference.
    Proceedings of the 18th ACM International Symposium on High Performance Distributed Computing, HPDC 2009, Garching, Germany, June 11-13, 2009; 01/2009
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: The concept of green computing has attracted much attention recently in cluster computing. However, previous local approaches focused on saving the energy cost of the components in a single workstation without a global vision on the whole cluster, so it achieved undesirable power reduction effect. Other cluster-wide energy saving techniques could only be applied to homogeneous workstations and specific applications. This paper describes the design and implementation of a novel approach that uses live migration of virtual machines to transfer load among the nodes on a multilayer ring-based overlay. This scheme can reduce the power consumption greatly by regarding all the cluster nodes as a whole. Plus, it can be applied to both the homogeneous and heterogeneous servers. Experimental measurements show that the new method can reduce the power consumption by 74.8% over base at most with certain adjustably acceptable overhead. The effectiveness and performance insights are also analytically verified.
    Cluster Computing, 2008 IEEE International Conference on; 11/2008
  • [Show abstract] [Hide abstract]
    ABSTRACT: Desktop portals provide convenience for end users to access heterogeneous resources in cloud. To meet the requirements of personalized customizing, simple managing, automatic deploying, suspending/resuming in multiple states, synchronizing incompatible resources from different platforms, we have designed, implemented and deployed a virtualized desktop, called DeskCloud, in cloud computing, to provide on-demand provisioning approaches based on system virtualizations in the last one and a half years in China education and research network. Experiments demonstrate and evaluate their approach's efficiency and effectiveness.
  • [Show abstract] [Hide abstract]
    ABSTRACT: Recent advance of virtualization technology provides a new approach to check-point/restart at the virtual machine (VM) level. In contrast to traditional process-level checkpointing, checkpointing at the virtualization layer brings up several advantages, such as compatibility, transparence, flexibility and simplicity. However, because the virtualization layer has little semantic knowledge about the operation system and the applications running atop, VM-layer checkpointing requires saving the entire operating system state rather than a single process. The overhead may render the approach impractical. To reduce the size of VM checkpoint, in this paper we propose a page eviction scheme and an incremental checkpointing mechanism to avoid saving unnecessary VM pages in the checkpoint. To keep the system online transparently, we propose a live checkpointing mechanism by saving the memory image in a copy-on-write (COW) manner. We implement the performance optimization mechanisms in a prototype system, called VMckpt. Experimental results with a group of representative applications show that our page eviction scheme and incremental checkpointing can significantly reduce the checkpoint file size by up to 87% and shorten the total checkpointing/restart time by a factor of up to 71%, in comparison with the Xens default checkpointing mechanism. The observed application downtimes due to checkpointing can be reduced to as small as 300 ms.
    Sciece China. Information Sciences 55(12). · 0.71 Impact Factor