Empirical Analysis and Modeling of Compute Times of CNN Operations on AWS Cloud

2020 IEEE International Symposium on Workload Characterization (IISWC)(2020)

引用 5|浏览6
暂无评分
摘要
Given the widespread use of Convolutional Neural Networks (CNNs) in image classification applications, cloud providers now routinely offer several GPU-equipped instances with varying price points and hardware specifications. From a practitioner's perspective, given an arbitrary CNN, it is not obvious which GPU instance should be employed to minimize the model training time and/or rental cost. This paper presents Ceer, a model-driven approach to determine the optimal GPU instance(s) for any given CNN. Based on an operation-level empirical analysis of various CNNs, we develop regression models for heavy GPU operations (where input size is a key feature) and employ the sample median estimator for light GPU and CPU operations. To estimate the communication overhead between CPU and GPU(s), especially in the case of multi-GPU training, we develop a model that relates this communication overhead to the number of model parameters in the CNN. Evaluation results on AWS Cloud show that Ceer can accurately predict training time and cost (less than 5% average prediction error) across CNNs, enabling 36% -44% cost savings over simpler strategies that employ the cheapest or the latest generation GPU instances.
更多
查看译文
关键词
AWS Cloud,model parameters,multiGPU training,communication overhead,sample median estimator,input size,heavy GPU operations,regression models,operation-level empirical analysis,optimal GPU instance,model-driven approach,Ceer,rental cost,model training time,hardware specifications,price points,GPU-equipped instances,cloud providers,image classification applications,convolutional neural networks,CNN operations,compute times
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要