The ability to sculpt complex reference waves and probe diverse radiation field patterns have facilitated the rise of metasurface antennas, while there is still a compromise between the required wide operation band and the non-overlapping characteristic of radiation field patterns. Specifically, the current computational image formation process with a classic matched filter and other sparsity-driven algorithms would inevitably face the challenge of a relatively confined scene information sampling ratio and high computational complexity. In this paper, we marry the concepts of a deep convolutional neural network with computational imaging literature. Compared with the current matched filter and compressed sensing reconstruction technique, our proposal could handle a relatively high correlation of measurement modes and low scene sampling ratio. With the delicately trained reconstruction network, point-size objects and more complicated targets can both be quickly and accurately reconstructed. In addition, the unavoidable heavy computation burden and essential large operation frequency band can be effectively mitigated. The simulated experiments with measured radiation field data verify the effectiveness of the proposed method.