Alzheimer’s disease (AD) is a prevalent and widespread neurodegenerative disorder among the older adult population worldwide. Among the numerous cognitive screening methods, neuropsychological scale assessments (NSA) are the widely utilized screening tools in clinical practice. The NSA places significant emphasis on speech-related questions, and thus the role of Automatic Speech Recognition (ASR) technology in AI-based NSA evaluations becomes particularly critical. However, the majority of ASR research pays limited attention to the study of speech recognition for various dialects. One of the primary reasons and main challenges for the scarcity of research in dialectal speech recognition is the limited availability of annotated dialectal data. Furthermore, the older adults’ unclear pronunciation leads to erroneous recognition of similar phonetic words by ASR. To overcome the challenges of limited annotated dialect data in various regions and the issues related to confusable pronunciation of older adults, we propose a dialectal collaborative encoder mechanism and a machine reading comprehension augmented model (MRCAM) that automatically corrects the texts in the NSA. The dialectal collaborative encoder mechanism utilizes layer frozenness to transfer recognition ability to the dialectal model. Meanwhile, the MRCAM model fills the gap by combining speech recognition algorithms with natural language processing (NLP) technology to recognize the dialects spoken by older adults. The experimental evaluation results indicate that the proposed augmented model achieves over 90% consistency with the physician’s judgment, providing the premise of the proposed methods for the task of automated screening for AD based on ASR.