This paper by He et al demonstrates an improvement on the Xavier/Glorot initialiaser. The newer He initializer works better with deep ReLU networks, the Xavier was initially developed to work with the sigmoid activation.
Is there already support for alternative initializers? (I haven’t looked
(NB with TF2 I believe the default initializer is Glorot_Uniform)