The growing cyber space with the developments in cyber network technologies in smart grid (SG) systems has necessitated questioning the reliability of networks and taking precautions against possible cyber threats. For this reason, defensive strategies and approaches against cyber attacks must be improved to sustain secure information flow of the network connections used in electricity generation, transmission, distribution, and consumption. This paper proposes a multi-agent multi environment deep reinforcement learning (MM-DRL) based defender response against cyber epidemics consisting coordinated cyber-attacks (multi-CAs) in the same time frame scheme to sustain security for SG networks. In this regard, the PMU-connected 123-bus system is integrated as a Markov game. MM-DRL approach is implemented for subenvironments of a typical SG system. Multi-CAs game aims to coordinate PMU signals across intersections to improve the network efficiency of a SG. DRL has been applied to data control recently and demonstrated promising performance where each data signal is regarded as an agent. Conversely, multi-CAs are self-renewing emerging causative agent of electricity theft, network disturbances, and data manipulation in SG systems characterized with wide characteristic diversity and rapid evolution. The game results show that the presented request response algorithm is able to minimize system attack damage and maintain protection duties when compared to a benchmark without request response. In addition, the performance of the MM-DRL approach compared to other developed methods is examined.