Implications of Public Cloud Resource Heterogeneity for Inference Serving
暂无分享,去创建一个
Chita R. Das | Jashwant Raj Gunasekaran | Prashanth Thinakaran | Cyan Subhra Mishra | Mahmut Taylan Kandemir | M. Kandemir | C. Das | J. Gunasekaran | P. Thinakaran
[1] Gregory R. Ganger,et al. Tributary: spot-dancing for elastic services with latency SLOs , 2018, USENIX ATC.
[2] Mengyuan Li,et al. Peeking Behind the Curtains of Serverless Platforms , 2018, USENIX Annual Technical Conference.
[3] Minlan Yu,et al. CherryPick: Adaptively Unearthing the Best Cloud Configurations for Big Data Analytics , 2017, NSDI.
[4] Gregory R. Ganger,et al. Stratus: cost-aware container scheduling in the public cloud , 2018, SoCC.
[5] Qian Li,et al. A Case for Managed and Model-less Inference Serving , 2019, HotOS.
[6] George Kesidis,et al. Spock: Exploiting Serverless Functions for SLO and Cost Aware Resource Procurement in Public Cloud , 2019, 2019 IEEE 12th International Conference on Cloud Computing (CLOUD).
[7] Wei Wang,et al. MArk: Exploiting Cloud Services for Cost-Effective, SLO-Aware Machine Learning Inference Serving , 2019, USENIX Annual Technical Conference.
[8] Sameh Elnikety,et al. Swayam: distributed autoscaling to meet SLAs of machine learning inference services with resource efficiency , 2017, Middleware.
[9] Thomas F. Wenisch,et al. SoftSKU: Optimizing Server Architectures for Microservice Diversity @Scale , 2019, 2019 ACM/IEEE 46th Annual International Symposium on Computer Architecture (ISCA).
[10] Alexandru Agache,et al. Firecracker: Lightweight Virtualization for Serverless Applications , 2020, NSDI.