AbstractLifespan face age transformation aims to generate facial images that accurately depict an individual's appearance at different age stages. This task is highly challenging due to the need for reasonable changes in facial features while preserving identity characteristics. Existing methods tend to synthesize unsatisfactory results, such as entangled facial attributes and low identity preservation, especially when dealing with large age gaps. Furthermore, over‐manipulating the style vector may deviate it from the latent space and damage image quality. To address these issues, this paper introduces a novel nonlinear regression model‐Disentangled Lifespan face Aging (DL‐Aging) to achieve high‐quality age transformation images. Specifically, we propose an age modulation encoder to extract age‐related multi‐scale facial features as key and value, and use the reconstructed style vector of the image as the query. The multi‐head cross‐attention in the W+ space is utilized to update the query for aging image reconstruction iteratively. This nonlinear transformation enables the model to learn a more disentangled mode of transformation, which is crucial for alleviating facial attribute entanglement. Additionally, we introduce a W+ space age regularization term to prevent excessive manipulation of the style vector and ensure it remains within the W+ space during transformation, thereby improving generation quality and aging accuracy. Extensive qualitative and quantitative experiments demonstrate that the proposed DL‐Aging outperforms state‐of‐the‐art methods regarding aging accuracy, image quality, attribute disentanglement, and identity preservation, especially for large age gaps.
Read full abstract