Authors
Daniel Justus, John Brennan, Stephen Bonner, Andrew Stephen McGough
Publication date
2018/12/10
Conference
2018 IEEE international conference on big data (Big Data)
Pages
3873-3882
Publisher
IEEE
Description
Deep learning is rapidly becoming a go-to tool for many artificial intelligence problems due to its ability to outperform other approaches and even humans at many problems. Despite its popularity we are still unable to accurately predict the time it will take to train a deep learning network to solve a given problem. This training time can be seen as the product of the training time per epoch and the number of epochs which need to be performed to reach the desired level of accuracy. Some work has been carried out to predict the training time for an epoch - most have been based around the assumption that the training time is linearly related to the number of floating point operations required. However, this relationship is not true and becomes exacerbated in cases where other activities start to dominate the execution time. Such as the time to load data from memory or loss of performance due to non-optimal parallel …
Total citations
20192020202120222023202473748747846
Scholar articles
D Justus, J Brennan, S Bonner, AS McGough - 2018 IEEE international conference on big data (Big …, 2018