Abstract
AbstractThe Gene Ontology (GO) describes molecular functions, biological processes, and cellular components of gene products using controlled-vocabulary terms that are related to each other in a structure that facilitates computing on GO annotations within and across species. Experimentally-based GO annotations that are manually curated from the literature are often used to predict the functions of related uncharacterized proteins. The accuracy of such annotations is thus critically important, particularly for a well-studied model organism such as Saccharomyces cerevisiae which is frequently used as the source of the experimental data.Comparison of experimentally-based annotations with those predicted by computational methods for the same gene products may reveal inaccuracies in curation of the experimental data, and could additionally be used to evaluate and improve the computational methods. We will present the results of an analysis at SGD that identified four major reasons for discrepancies between the two kinds of annotation. Some discrepancies revealed cases in which human error led to errors or omissions in the manual curation, prompting prioritization for review and correction. In another category, the computational annotations were not supported or were refuted by the literature, thereby suggesting ways in which the accuracy of the prediction methods could be improved. Yet another type of discrepancy resulted from issues with the GO structure, such as missing parentage for certain terms, leading to reexamination and improvement of the ontology. Finally, some discrepancies arose because the computational predictions were entirely novel, and no relevant experimental evidence was available. These cases highlight potential interesting new avenues for experimentation.
Highlights
Algorithm uses a protein‐protein linkage map derived from diverse genomic data to predict a process‐specific network
ConclusionsThis type of analysis can result in improvements to manual annotations, computational methods, and the GO ontology we hope to use this method to target and prioritize manual annotations that need review Still to do: comparison of manual annotations to computational predictions other than InterPro
Summary
1. To improve manual annotation quality, finding: ‐ errors ‐ omissions ‐ “shallow” annotations (i.e., not as granular as possible). 2. To improve computational prediction methods: ‐ are certain domains incorrectly mapped to GO terms? ‐ are prediction algorithms consistently generating incorrect predictions in any particular area of biology?. 3. To improve the Gene Ontology content and structure: ‐ do inconsistencies between manual annotations and predictions reveal issues with GO structure, such as incorrect or missing parentage, or true path violations?
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.