Professional Documents
Culture Documents
International 5
International 5
Therefore, the time cost to transfer the In push phase of the proposed framework as
memory pages in the proposed framework is as shown in (1), the system first transfer the
follows. memory page except the Working Set list
that include the most recent pages during the
Execution time of ith iteration for push phase of migration time. In (5) for the push phase of the
the proposed framework: pre-copy based live migration, the system first
transfers the all of the memory pages. The
(1) proposed framework reduces the transfer
memory page using the working set list during
where i=1,2,3,…n and and the push phase of live migration. But the total
for first iteration and then modified memory migration of the proposed framework include not
page size is updated by: only the execution time for push phases and stop
and copy phase but also the overhead for pre-
(2) processing phase that apply the working set
prediction algorithm.
Execution time for stop phase of the proposed From the equations, we give a specific,
framework: numerical example to evaluate the total
migration time of the proposed framework.
Based on the experiment with Quake3 Server
(3)
workload, we make the assumption that before
migration. The virtual machine has the following
Total Migration Time of Proposed Framework states: M=131072 pages, B=5.625 MB/s,
= (4) =512 pages/s and the size of the page is
4KB. Then we have the results in Table 1. From
And the time cost to transfer the memory these results, the proposed framework can
pages in pre-copy based live migration is as migrate a virtual machine in less time than the
follows. pre-copy based live migration.
In Figure 4, we present the execution time
Execution time of ith iteration for push phase of for pre-processing phase that apply the working
pre-copy based live migration: set prediction algorithm of the proposed
framework with various working set size.
(5) According to the result, the larger size of
working set may cause the more execution time.
Figure 4. Execution Time for Pre-processing Figure 6. Total Migration Time for Proposed
Phase of the Proposed Framework Framework
In Figure 5, we present results comparing the Figure 7 shows the comparison of total
execution time for push phase of the proposed migration time for the proposed framework with
framework with pre-processing phase and pre-processing phase and pre-copy based live
without pre-processing phase using various migration using various working set size. From
working set size. Although we need less these calculation results, we can reduce the total
execution time for push phase with pre- migration time in the proposed framework than
processing phase, the system has the overhead pre-copy based live migration of virtual machine
for pre-processing phase. while the working set size is 1024 pages and
above. In 256 pages of working set, the
preprocessing phase can be overhead for total
migration time.
In Figure 6, the calculation results show the Figure 7. Total Migration Time of the
total migration time of the proposed framework proposed framework with pre-
with various working set size. According to the processing phase Vs pre-copy based
calculation results, 32 MB (8192 pages) is live migration
suitable for Working Set list of the proposed
framework. If the size of Working Set list 5. Conclusion
increased, the execution time for pre-processing
phase also increased. The system need to be In this paper, we have presented the
balance the effect and overhead using pre- framework which includes the pre-processing
processing phase. phase, push phase and stop phase for pre-copy
based live migration. Pre-copy based live
migration ensures that keep downtime small by International Conference on Cluster Computing
minimizing the amount of VM state. It provides (Cluster 2009), 2009.
to abort the migration should the destination VM [7] M. Nelson, B. Lim, and G. Hutchines, “Fast
ever crash during migration because the VM is transparent migration for virtual machines,” in
Proceedings of the USENIX Annual Technical
still running at the source. It needs to reduce the Conference (USENIX’05), 2005, pp. 391– 394.
total migration time and impact of migration on [8] M. R. Hines and K. Gopalan, “Post-copy based live
performance of source and destination VMs. We virtual machine migration using adaptive pre-paging
proposed the working set prediction algorithm and dynamic self-ballooning”, in Proceedings of the
which based on Least Recent Used (LRU) policy ACM/Usenix international conference on Virtual
collects the most recent used memory pages as execution environments (VEE’09), 2009, pp. 51–60
the Working Set list and collects the least recent [9] OPENVZ. Container-based Virtualization for
used memory pages as the inactive list in pre- Linux, http://www.openvz.com/.
processing phase. The proposed framework [10] P. Barham, B. Dragovic, K. Fraser, S. Hand, T.
Harris, A. Ho, R. Neugebauer, I. Pratt, and A.
migrate the inactive list in push phase and lastly Warfield, “Xen and the Art of Virtualization”,
it migrate the Working set list in stop phase to Proceedings of the Nineteenth ACM Symposium
reduce the amount of transferred data and Operating System Principles (SOSP19), pages
iteration times. From the calculation results, the 164.177. ACM Press, 2003.
proposed framework can reduce the total [11] P. Lu and K. Shen, “Virtual machine memory
migration time in live migration of virtual access tracing with hypervisor exclusive cache”. In
machines. ATC’07: 2007 USENIX Annual Technical Conference
on Proceedings of the USENIX Annual Technical
6. References Conference,pages 1–15, Berkeley, CA, USA, 2007.
USENIX Association. ISBN 999-8888-77-6.
[1] A. Kivity, Y. Kamay, and D. Laor, “kvm: the [12] R. Goldberg, “Survey of virtual machine
linux virtual machine monitor”. In Proc. of Ottawa research,” IEEE Computer, pp. 34–45, Jun. 1974.
Linux Symposium (2007). [13] S. Osman, D. Subhraveti, G. Su, and J. Nieh,
[2] C. P. Sapuntzakis, R. Chandra, B. Pfaff, J. Chow, “The Design and Implementation of Zap: A System
M. S. Lam, and M. Rosenblum, “Optimizing the for Migrating Computing Environments”, 5th
Migration of Virtual Computers”, Proceedings of the Symposium on Operating Systems Design and
5th Symposium on Operating Systems Design and Implementation (OSDI 2002), Boston, MA, December
Implementation, Boston, Massachusetts, USA, 2002.
December 9–11, 2002. [14] T. Yang, E. Berger, M. Hertz, S. Kaplan, and J.
[3] C. Clark, K. Fraser, S. Hand, J. G. Hansen, E. July, Moss. Automatic heap sizing: Taking real memory
C. Limpach, I. Pratt, and A. Warfield, “Live Migration into account, 2004. URL
of Virtual Machines”, Proceedings of the 2nd USENIX citeseer.ist.psu.edu/article/yang04automatic.html.
Symposium on Networked Systems Design and [15] T. Yang, E. Berger, S. Kaplan, and J. Moss.
Implementation, 2005. CRAMM: virtual memory support for garbage-
[4] D. Lee, J. Choi, J. Kim, S. Noh, S. Min, Y. Cho, collected applications. In OSDI ’06: Proceedings of
and C. Kim, “LRFU: A Spectrum of Policies that the 7th symposium on Operating systems design and
Subsumes the LRU and LFU Policies”, IEEE implementation, pages 103–116, Berkeley, CA, USA,
Transactions on Computers, vol. 50, no. 12, pp.1352- 2006. USENIX Association. ISBN 1-931971-47-1.
1361, Dec. 2001. [16] W. Huang, Q. Gao, J. Liu, and D. K. Panda,
[5] H. Liu, H. Jin, X. Liao, L. Hu, and C. Yu, “Live “High Performance Virtual Machine Migration with
migration of virtual machine based on full system RDMA over Modern Interconnects”, Proceedings of
trace and replay,” in Proceedings of the 18th IEEE Conference on Cluster Computing (Cluster
International Symposium on High Performance 2007), Austin, Texas. September, 2007
Distributed Computing (HPDC’09), 2009, pp.101–
110.
[6] H. Jin, L. Deng, S. Wu, X. Shi, and X. Pan. Live
virtual machine migration with adaptive memory
compression. In Proceedings of the 2009 IEEE