In the field of hyperspectral image classification, deep learning technology, especially convolutional neural networks, has achieved remarkable progress. However, convolutional neural network models encounter challenges in hyperspectral image classification due to limitations in their receptive fields. Conversely, the global modeling capability of Transformers has garnered attention in hyperspectral image classification. Nevertheless, the high computational cost and inadequate local feature extraction hinder its widespread application. In this study, we propose a novel fusion model of convolutional neural networks and Transformers to enhance performance in hyperspectral image classification, namely the dual-branch multi-granularity convolutional cross-substitution Transformer (DMCCT). The proposed model adopts a dual-branch structure to separately extract spatial and spectral features, thereby mitigating mutual interference and information loss between spectral and spatial data during feature extraction. Moreover, a multi-granularity embedding module is introduced to facilitate multi-scale and multi-level local feature extraction for spatial and spectral information. In particular, the improved convolutional cross-substitution Transformer module effectively integrates convolution and Transformer, reducing the complexity of attention operations and enhancing the accuracy of hyperspectral image classification tasks. Subsequently, the proposed method is evaluated against existing approaches using three classical datasets, namely Pavia University, Kennedy Space Center, and Indian Pines. Experimental results demonstrate the efficacy of the proposed method, achieving significant classification results on these datasets with overall classification accuracies of 98.57%, 97.96%, and 96.59%, respectively. These results establish the superiority of the proposed method in the context of hyperspectral image classification under similar experimental conditions.