As the size of data increases and computation becomes complicated in fog computing environments, the need for highperformance computation is increasing. One of the most popular ways to improve the performance of a virtual machine (VM) is to allocate a graphic processing unit (GPU) to the VM for supporting general purpose computing on graphic processing unit (GPGPU) operations. The direct pass-through, often used for GPUs in VMs, is popular in the cloud because VMs can use the full functionality of the GPU and experience virtually no performance degradation owing to virtualization. Direct pass-through is very useful for improving the performance of VMs. However, since the GPU usage time is not considered in the VM scheduler that operates based on the central processing unit (CPU) usage time of the VM, the VM performing the GPGPU operation degrades the performance of other VMs. In this paper, we analyze the effect of the VM performing the GPGPU operation (GPGPU-intensive VM) on other VMs through experiments. Then, we propose a method to mitigate the performance degradation of other VMs by dynamically allocating the resource usage time of the VM and preventing the priority preemption of the GPGPU-intensive VM.
[1]
Kenli Li,et al.
vCUDA: GPU-Accelerated High-Performance Computing in Virtual Machines
,
2012,
IEEE Trans. Computers.
[2]
Gil Neiger,et al.
Intel ® Virtualization Technology for Directed I/O
,
2006
.
[3]
Yaozu Dong,et al.
A Full GPU Virtualization Solution with Mediated Pass-Through
,
2014,
USENIX Annual Technical Conference.
[4]
Ada Gavrilovska,et al.
GPUShare: Fair-Sharing Middleware for GPU Clouds
,
2016,
2016 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW).
[5]
Shinpei Kato,et al.
GPUvm: Why Not Virtualizing GPUs at the Hypervisor?
,
2014,
USENIX Annual Technical Conference.