The recommended readability of online health education materials is at or below the sixth- to eighth-grade level. Nevertheless, more than a decade of research has demonstrated that most online education materials pertaining to orthopaedic surgery do not meet these recommendations. The repeated evidence of this limited progress underscores that unaddressed barriers exist to improving readability, such as the added time and cost associated with writing easily readable materials that cover complex topics. Freely available artificial intelligence (AI) platforms might facilitate the conversion of patient-education materials at scale, but to our knowledge, this has not been evaluated in orthopaedic surgery. (1) Can a freely available AI dialogue platform rewrite orthopaedic patient education materials to reduce the required reading skill level from the high-school level to the sixth-grade level (which is approximately the median reading level in the United States)? (2) Were the converted materials accurate, and did they retain sufficient content detail to be informative as education materials for patients? Descriptions of lumbar disc herniation, scoliosis, and spinal stenosis, as well as TKA and THA, were identified from educational materials published online by orthopaedic surgery specialty organizations and leading orthopaedic institutions. The descriptions were entered into an AI dialogue platform with the prompt "translate to fifth-grade reading level" to convert each group of text at or below the sixth-grade reading level. The fifth-grade reading level was selected to account for potential variation in how readability is defined by the AI platform, given that there are several widely used preexisting methods for defining readability levels. The Flesch Reading Ease score and Flesch-Kincaid grade level were determined for each description before and after AI conversion. The time to convert was also recorded. Each education material and its respective conversion was reviewed for factual inaccuracies, and each conversion was reviewed for its retention of sufficient detail for intended use as a patient education document. As presented to the public, the current descriptions of herniated lumbar disc, scoliosis, and stenosis had median (range) Flesch-Kincaid grade levels of 9.5 (9.1 to 10.5), 12.6 (10.8 to 15), and 10.9 (8 to 13.6), respectively. After conversion by the AI dialogue platform, the median Flesch-Kincaid grade level scores for herniated lumbar disc, scoliosis, and stenosis were 5.0 (3.3 to 8.2), 5.6 (4.1 to 7.3), and 6.9 (5 to 7.8), respectively. Similarly, descriptions of TKA and THA improved from 12.0 (11.2 to 13.5) to 6.3 (5.8 to 7.6) and from 11.6 (9.5 to 12.6) to 6.1 (5.4 to 7.1), respectively. The Flesch Reading Ease scores followed a similar trend. Seconds per sentence conversion was median 4.5 (3.3 to 4.9) and 4.5 (3.5 to 4.8) for spine conditions and arthroplasty, respectively. Evaluation of the materials that were converted for ease of reading still provided a sufficient level of nuance for patient education, and no factual errors or inaccuracies were identified. We found that a freely available AI dialogue platform can improve the reading accessibility of orthopaedic surgery online patient education materials to recommended levels quickly and effectively. Professional organizations and practices should determine whether their patient education materials exceed current recommended reading levels by using widely available measurement tools, and then apply an AI dialogue platform to facilitate converting their materials to more accessible levels if needed. Additional research is needed to determine whether this technology can be applied to additional materials meant to inform patients, such as surgical consent documents or postoperative instructions, and whether the methods presented here are applicable to non-English language materials.