Field-road trajectory classification is a crucial task for agricultural machinery behavior mode recognition, aiming to distinguish field operation mode and road driving mode automatically. However, the imbalanced distribution of agricultural machine trajectories brings challenges for the field-road trajectory classification task. Additionally, most existing field-road trajectory classification methods have certain shortcomings. For instance, they encounter difficulties in accurately representing the state of agricultural machinery movement using the current features. The data transformation process often leads to information loss, and the model’s generalization capabilities are limited. The performance of the models is constrained by each of these elements. To address these shortcomings, this paper introduces a general image classification model for agricultural machinery trajectory mode recognition named ATRNet. First, to address the issue of imbalanced field-road proportions in agricultural machinery trajectory data, a Conditional Tabular Generative Adversarial Network (CTGAN) is employed to generate quasi trajectories, balancing the distribution of positive and negative samples in the data. This step aims to eliminate biases during the model training process. Second, to accurately characterize the motion status of agricultural machinery, we propose a multiangle feature enhancement method to extract rich spatiotemporal features from trajectory data. Finally, different from conventional field-road trajectory classification models that primarily rely on spatial and temporal information for identifying trajectories, we present a lossless trajectory data representation paradigm. This paradigm maps each trajectory point into a “feature map” and uses an image classification model to capture latent feature representations of trajectory points for the recognition of different behavior modes of agricultural machinery. This paradigm can generalize image classification networks to the field-road trajectory classification task, providing a general vision model solution for agricultural machinery trajectory mode recognition. To validate the effectiveness of the ATRNet model, experiments were conducted on real corn and wheat harvester trajectory datasets. The results demonstrate that the proposed model achieves remarkable performance improvements over the state-of-the-art (SOTA) models. In the corn harvester trajectory dataset, ATRNet achieves an accuracy of 92.36% and an F1-score of 92.34%, surpassing existing SOTA models by 3.12% and 12.46%, respectively. Similarly, in the wheat harvester trajectory dataset, ATRNet achieves an accuracy of 92.36% and an F1-score of 92.33%, outperforming the existing optimal algorithm by 4.76% and 18.18%, respectively.
Read full abstract