The COVID-19 pandemic has made us all understand that wearing a face mask protects us from the spread of respiratory viruses. Face authentication systems, which are trained on the basis of facial key points such as the eyes, nose, and mouth, found it difficult to identify a person when the majority of the face is covered by a face mask. Removing the mask for authentication will cause the infection to spread. The possible solutions are (a) to train face recognition systems to identify the person with the upper face features, (b) reconstruct the complete face of the person with a generative model, and (c) train the model with a dataset of the masked faces of the people. In this article, we explore the scope of generative models for image synthesis. We used stable diffusion to generate masked face images of popular celebrities on various text prompts. A realistic dataset of 15K masked face images of 100 celebrities was generated and is called the Realistic Synthetic Masked Face Dataset (RSMFD). The model and the generated dataset will be made public so that researchers can augment the dataset. To the best of our knowledge, this is the largest masked face recognition dataset with realistic images. The generated images were tested on popular deep face recognition models and achieved significant results. The dataset is also trained and tested on some of the famous image classification models, and the results are competitive. The dataset is available at https://drive.google.com/drive/folders/1yetcgUOL1TOP4rod1geGsOkIrIJHtcEw?usp=sharing .
Read full abstract