The development of the Internet of Things has led to a significant increase in the number of devices, consequently generating a vast amount of data and resulting in an influx of unlabeled data. Collecting these data enables the training of robust models to support a broader range of applications. However, labeling these data can be costly, and the models dependent on labeled data are often unsuitable for rapidly evolving fields like vehicular networks and mobile Internet of Things, where new data continuously emerge. To address this challenge, Self-Supervised Learning (SSL) offers a way to train models without the need for labels. Nevertheless, the data stored locally in vehicles are considered private, and vehicles are reluctant to share data with others. Federated Learning (FL) is an advanced distributed machine learning approach that protects each vehicle’s privacy by allowing models to be trained locally and the model parameters to be exchanged across multiple devices simultaneously. Additionally, vehicles capture images while driving through cameras mounted on their rooftops. If a vehicle’s velocity is too high, the captured images, donated as local data, may be blurred. Simple aggregation of such data can negatively impact the accuracy of the aggregated model and slow down the convergence speed of FL. This paper proposes a FL algorithm for aggregation based on image blur levels, which is called FLSimCo. This algorithm does not require labels and serves as a pre-training stage for SSL in vehicular networks. Simulation results demonstrate that the proposed algorithm achieves fast and stable convergence.