Very cool book. I think a reason why ML has seen so much progress despite benchmark overfitting/abuse is that results are "regularized" by real world applications and the Lindy effect. Methods, or research, that abuse benchmarks aren't adopted by follow-up research so they tend not to survive. And they aren't adopted because people try them but then find out that they don't generalize to other/newer benchmarks. So the system works not because of specific benchmarks, but because of how the community as a whole deals with benchmarks.
I'm a director at the Max Planck Institute for Intelligent Systems. Prior to joining the institute, I was Associate Professor for Electrical Engineering and Computer Sciences at the University of California, Berkeley. My research contributes to the scientific foundations of machine learning and algorithmic decision making with a focus on social questions.[0]
Also simply knowing of him doesn't answer the question.