In the last decade, Machine Learning has experienced a dramatic increase in performance on a wide variety of tasks, including computer vision, speech recognition, text parsing, and language translation, just to name a few. This has corresponded to an understandable hype especially for the remarkable results achieved in some cases. Therefore, practitioners of Scientific Disciplines have become interested in utilizing new Machine Learning techniques, and have sometimes started doing so with mixed success. The purpose of this paper is to describe some of the common Traps, Pitfalls and Misconceptions of Machine Learning as relevant to the Scientific Discipline, and how to avoid them. In fact, Machine Learning and Deep Learning are fast evolving fields, and some of the astonishing results achieved recently sit on small but important details which have become the state of the art. Some of these details are not broadly known by the scientific community. No new scientific result is presented in this paper, which is a survey and a summary of the best of the field, for the benefit of researchers with limited experience. It is not the intention of the authors to provide any criticism to the work of experienced practitioners, particularly not to the ones working on the cutting edge of what is currently possible: in these cases expert researchers may well be doing exactly what we recommend here to avoid, and for a good reason. However we believe that the advice provided here will be useful, and perhaps even a reference, for the newcomers of the field.
[1]
Jian Sun,et al.
Deep Residual Learning for Image Recognition
,
2015,
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2]
Guigang Zhang,et al.
Deep Learning
,
2016,
Int. J. Semantic Comput..
[3]
Judea Pearl,et al.
Robustness of Causal Claims
,
2004,
UAI.
[4]
Lukasz Kaiser,et al.
Attention is All you Need
,
2017,
NIPS.
[5]
Jürgen Schmidhuber,et al.
Long Short-Term Memory
,
1997,
Neural Computation.
[6]
Geoffrey E. Hinton,et al.
ImageNet classification with deep convolutional neural networks
,
2012,
Commun. ACM.
[7]
Yoshua Bengio,et al.
Gradient-based learning applied to document recognition
,
1998,
Proc. IEEE.
[8]
Zenghui Wang,et al.
Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review
,
2017,
Neural Computation.
[9]
Prabhat,et al.
Exascale Deep Learning for Climate Analytics
,
2018,
SC18: International Conference for High Performance Computing, Networking, Storage and Analysis.
[10]
Toru Baji,et al.
Evolution of the GPU Device widely used in AI and Massive Parallel Processing
,
2018,
2018 IEEE 2nd Electron Devices Technology and Manufacturing Conference (EDTM).
[11]
Simon Haykin,et al.
GradientBased Learning Applied to Document Recognition
,
2001
.