Abstract

AbstractThe Gene Ontology (GO) describes molecular functions, biological processes, and cellular components of gene products using controlled-vocabulary terms that are related to each other in a structure that facilitates computing on GO annotations within and across species. Experimentally-based GO annotations that are manually curated from the literature are often used to predict the functions of related uncharacterized proteins. The accuracy of such annotations is thus critically important, particularly for a well-studied model organism such as Saccharomyces cerevisiae which is frequently used as the source of the experimental data.Comparison of experimentally-based annotations with those predicted by computational methods for the same gene products may reveal inaccuracies in curation of the experimental data, and could additionally be used to evaluate and improve the computational methods. We will present the results of an analysis at SGD that identified four major reasons for discrepancies between the two kinds of annotation. Some discrepancies revealed cases in which human error led to errors or omissions in the manual curation, prompting prioritization for review and correction. In another category, the computational annotations were not supported or were refuted by the literature, thereby suggesting ways in which the accuracy of the prediction methods could be improved. Yet another type of discrepancy resulted from issues with the GO structure, such as missing parentage for certain terms, leading to reexamination and improvement of the ontology. Finally, some discrepancies arose because the computational predictions were entirely novel, and no relevant experimental evidence was available. These cases highlight potential interesting new avenues for experimentation.

Highlights

  • Algorithm uses a protein‐protein linkage map derived from diverse genomic data to predict a process‐specific network

  • ConclusionsThis
type
of
analysis
can
result
in
improvements
to
manual
annotations, computational
methods,
and
the
GO
ontology 
we
hope
to
use
this
method
to
target
and
prioritize
manual
annotations
that need
review 
Still
to
do:
comparison
of
manual
annotations
to
computational
predictions other
than
InterPro

Read more

Summary

Why compare manual and computational annotations?

1. To
improve
manual
annotation
quality,
finding: ‐
errors ‐
omissions ‐
“shallow”
annotations
(i.e.,
not
as
granular
as
possible). 2. To
improve
computational
prediction
methods: 






‐
are
certain
domains
incorrectly
mapped
to
GO
terms? ‐
are
prediction
algorithms
consistently
generating
incorrect predictions
in
any
particular
area
of
biology?. 3. To
improve
the
Gene
Ontology
content
and
structure: 






‐
do
inconsistencies
between
manual
annotations
and
predictions reveal
issues
with
GO
structure,
such
as
incorrect
or
missing parentage,
or
true
path
violations?

Manual annotations reviewed
Sometimes the GO structure needs to be changed
Conclusions

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.