Hard to say. But should check these out.
https://www.alexirpan.com/2017/04/26/perils-batch-norm.html
https://blog.paperspace.com/busting-the-myths-about-batch-normalization/
Pretty neat idea
https://vpr-norman.ou.edu/FY18-Federal-Research-Budgets
Interviews of deep learning icons by Andrew Ng (he himself of course is an icon also). Highly recommend. Especially interviews of the big three (Hinton, Bengio, LeCun).
Some interesting excerpt of this
…
As glimpses of meta-learning, I was especially fascinated with Ng’s lectures and labs for:
My fascination has motivated me to learn about various meta-learning approaches, such as:
The insight is that one should eagerly explore the new secret sauce for DL — meta-learning.