|Deep Learning Models on CPUs: A Methodology for Efficient Training|
GPUs have been favored for training deep learning models due to their highly parallelized architecture. As a result, most studies on training optimization focus on GPUs. There is often a trade-off, however, between cost and efficiency when deciding how to choose the proper hardware for training. In particular, CPU servers can be beneficial if training on CPUs was more efficient, as they incur fewer hardware update costs and better utilize existing infrastructure.
|Year of Publication||
Journal of Machine Learning Theory, Applications and Practice
|Google Scholar | BibTeX | XML | DOI|