This paper investigates the problem of Hankel norm model reduction for linear systems with time-varying delay in the state. For a given stable system, our attention is focused on the construction of reduced-order model, which guarantees the corresponding error system to be asymptotically stable and has a specified Hankel norm error performance. Two different approaches are proposed to solve this problem. One casts the model reduction into a convex optimization problem by using a linearization procedure, and the other is based on the cone complementarity linearization idea, which casts the model reduction into a sequential minimization problem subject to linear matrix inequality constraints. Both continuous and discrete time cases are considered. A numerical example is provided to show the effectiveness of the proposed theory.