Reflecting on 2020, I realized I spent a lot of time reviewing. I reviewed 34 conference papers, 3 journal papers and 4 workshop papers. To my surprise, Lihua Lei reviewed 48 papers, 20 of which were for journals and probably took extra time due to revisions.
Reviewing is an important part of doing a PhD and it actually helps when writing papers. However, it does not help anymore once reviewing takes more time than writing. Of course, one solution that seems to be a common choice is to spend less time on each reviewed paper but this has led to the popular joke of the notorious Reviewer #2. I plan, nevertheless, to keep spending as much time as I need for every paper. My only hope is that the overall system at ML conferences would improve in near future, lowering the load on the people involved, for example by adopting the reviewing system of ICLR.
Peter Richtárik, Filip Hanzely and I are organizing a session on Optimization for Deep Learning at the SIAM Conference on Mathematics of Datascience (MDS20). See the link below for more info: https://www.siam.org/conferences/cm/conference/mds20
Our presenters are:
Simon Shaolei Du from the Institute for Advanced Study of Princeton (http://simonshaoleidu.com/)
Haihao Lu from Google Research NYC (http://web.mit.edu/haihao/www/)
Sai Praneeth Karimireddy from EPFL (https://people.epfl.ch/sai.karimireddy)
Filip Hanzely from KAUST (https://fhanzely.github.io/)