We analyze the classical EM algorithm for parameter estimation in the symmetric two-component Gaussian mixtures in d dimensions. We show that, even in the absence of any separation between components, provided that the sample size satisfies n=\Omega(d \log^4 d) , the randomly initialized EM algorithm converges to an estimate in at most O(\sqrt{n}) iterations with high probability, which is at most O((d / n)^{1/4} \log n) in Euclidean distance from the true parameter and within logarithmic factors of the minimax rate of (d / n)^{1/4} . Both the nonparametric statistical rate and the sublinear convergence rate are direct consequences of the zero Fisher information in the worst case. Refined pointwise guarantees beyond worst-case analysis and convergence to the MLE are also shown under mild conditions. This improves the previous result of Balakrishnan, Wainwright, and Yu (2017), which requires strong conditions on both the separation of the components and the quality of the initialization, and that of Daskalakis, Tzamos, and Zampetakis (2017), which requires sample splitting and restarting the EM iteration.