It is shown that the discrete-time disturbance rejection problem, formulated in finite and infinite horizons, and under perfect state measurements, can be solved by making direct use of some results on linear-quadratic zero-sum dynamic games. For the finite-horizon problem an optimal (minimax) controller exists, and can be expressed in terms of a generalized (time-varying) discrete-time Riccati equation. The existence of an optimum also holds in the infinite-horizon case, under an appropriate observability condition, with the optimal control, given in terms of a generalized algebraic Riccati equation, also being stabilizing. In both cases, the corresponding worst-case disturbances turn out to be correlated random sequences with discrete distributions, which means that the problem (viewed as a dynamic game between the controller and the disturbance) does not admit a pure-strategy saddle point. Results for the delayed state measurement and the nonzero initial state cases are presented.< <ETX xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">></ETX>