On non-approximability of zero loss global ${\mathcal L}^2$ minimizers by gradient descent in Deep Learning