fgga 1.2.0
As volume of genomic data grows, computational methods become essential for providing a first glimpse onto gene annotations. Automated Gene Ontology (GO) annotation methods based on hierarchical ensemble classification techniques are particularly interesting when interpretability of annotation results is a main concern. In these methods, raw GO-term predictions computed by base binary classifiers are leveraged by checking the consistency of predefined GO relationships. FGGA is a factor graph approach to the hierarchical ensemble formulation of the automated GO annotation problem. In this formal method, a core factor graph is first built based on the GO structure and then enriched to take into account the noisy nature of GO-term predictions. Hence, starting from raw GO-term predictions, an iterative message passing algorithm between nodes of the factor graph is used to compute marginal probabilities of target GO-terms.
The method is detailed in the original publications [1, 2], but this brief vignette explains how to use FGGA to predict a set of GO-terms (GO node labels) from gene-product sequences of a given organism. The aim is to improve the quality of existing electronic annotations and provide a new annotation for those unknown sequences that can not be classified by traditional methods such as Blast.
Thus, FGGA is a tool to the automated annotation of protein sequences, however, the annotation of other types of striking functional gene products is also possible, e.g., long non-coding RNAs. FGGA may bring an opportunity for improving the annotation of long non-coding RNA sequences through boosting the confidence of base binary classifiers by the characterization of their structures primary, secondary and tertiary. Along the vignette, we used protein coding genes data from Cannis familiaris organism obtained with UniProt.
The fgga R source package can be downloaded from Bioconductor repository or GitHub repository. This R package contains a experimental dataset as example, one pre-run R object and all functions needed to run FGGA.
## From Bioconductor repository
if (!requireNamespace("BiocManager", quietly = TRUE)) {
install.packages("BiocManager")
}
BiocManager::install("fgga")
## Or from GitHub repository using devtools
BiocManager::install("devtools")
devtools::install_github("fspetale/fgga")
After the package is installed, it can be loaded into R workspace by
library(fgga)
At present, the method directly supports data characterized from gene product sequences. This characterization can be generated according to the expert’s criteria. For example, possible characterizations can be: PFAM motifs, physico-chemical properties, signal peptides, among others. The admitted values can be of the numeric, boolean or character type. However, for greater efficiency of the binary classification algorithm it is recommended that the use of normalized numerical values. These datasets must have at least 50 annotations per GO-term to train the FGGA model.
In this section, we apply FGGA to predict the biological functions, GO-terms, of Canis lupus familiaris proteins. We download 6962 Canis lupus familiaris, Cf, protein sequences with their GO annotations from UniProt and then were characterized through physico-chemical properties from R package Peptides.
Let us import the toy dataset in the workspace:
# Loading Canis lupus familiaris dataset and example R objects
data(CfData)
# To see the summarized experiment object
summary(CfData)
#> Length Class Mode
#> dxCf 501264 -none- numeric
#> graphCfGO 1296 -none- numeric
#> indexGO 2 -none- list
#> tableCfGO 250632 -none- numeric
#> nodesGO 36 -none- character
#> varianceGOs 36 -none- numeric
# To see the information of characterized data
dim(CfData$dxCf)
#> [1] 6962 72
colnames(CfData$dxCf)[1:20]
#> [1] "Numb_Amino" "A" "C" "D" "E" "F" "G"
#> [8] "H" "I" "K" "L" "M" "N" "P"
#> [15] "Q" "R" "S" "T" "V" "W"
rownames(CfData$dxCf)[1:10]
#> [1] "E2QTB2" "F1PQ93" "E2RQU4" "E2RN67" "E2RGI4" "E2RFJ5" "J9P0Z9" "E2RL61" "F1PAF1" "J9P1G2"
head.matrix(CfData$dxCf[, 51:61], n = 10)
#> pK.C Surrounding.hydrop Mol.weight pI Pos_charged_res2 Pos_changed_res3
#> E2QTB2 -0.0790 0.0474 58225.55 4.9492 0.1175 0.1368
#> F1PQ93 -0.0953 0.0494 27833.13 4.6956 0.1250 0.1371
#> E2RQU4 -0.0117 0.1234 83933.57 9.0031 0.1344 0.1694
#> E2RN67 0.0185 0.0424 73821.11 7.9705 0.0848 0.1161
#> E2RGI4 0.0715 0.0368 28713.98 5.3223 0.1304 0.1423
#> E2RFJ5 -0.0194 0.0066 30101.85 4.6895 0.1049 0.1199
#> J9P0Z9 0.0010 0.0619 67510.79 6.3768 0.1267 0.1467
#> E2RL61 -0.0559 0.0226 163422.56 5.6757 0.1291 0.1463
#> F1PAF1 -0.0407 -0.0550 107808.45 5.4839 0.0932 0.1087
#> J9P1G2 -0.1119 0.0330 26295.98 9.1588 0.1040 0.1200
#> Neg_changed_res Carbon Hydrogen Nitrogen Oxygen
#> E2QTB2 0.1676 2518 3935 739 816
#> F1PQ93 0.1976 1209 1922 330 402
#> E2RQU4 0.1183 3779 5963 1033 1083
#> E2RN67 0.0818 3374 5180 910 914
#> E2RGI4 0.1542 1289 2035 331 387
#> E2RFJ5 0.1723 1327 2072 348 426
#> J9P0Z9 0.1333 3003 4694 808 908
#> E2RL61 0.1470 7269 11378 2012 2209
#> F1PAF1 0.1118 4755 7406 1294 1451
#> J9P1G2 0.0800 1166 1842 328 347
# to see the information of GO data
dim(CfData$tableCfGO)
#> [1] 6962 36
colnames(CfData$tableCfGO)[1:10]
#> [1] "GO:0000166" "GO:0003674" "GO:0003676" "GO:0003677" "GO:0003824" "GO:0004888" "GO:0005102"
#> [8] "GO:0005215" "GO:0005488" "GO:0005515"
rownames(CfData$tableCfGO)[1:10]
#> [1] "E2QTB2" "F1PQ93" "E2RQU4" "E2RN67" "E2RGI4" "E2RFJ5" "J9P0Z9" "E2RL61" "F1PAF1" "J9P1G2"
head(CfData$tableCfGO)[, 1:8]
#> GO:0000166 GO:0003674 GO:0003676 GO:0003677 GO:0003824 GO:0004888 GO:0005102 GO:0005215
#> E2QTB2 0 1 1 0 0 0 0 0
#> F1PQ93 0 1 0 0 0 0 0 0
#> E2RQU4 1 1 0 0 1 0 0 0
#> E2RN67 0 1 0 0 1 0 0 0
#> E2RGI4 0 1 0 0 0 0 0 1
#> E2RFJ5 0 1 1 0 0 0 0 0
tableCfGO is a binary matrix whose \((i, j)_{th}\) component is 1 if protein \(i\) is annotated with \(GO-term_j\), 0 otherwise. Note that the row names of both dxCf and tableCfGO are identical. Our binary classifiers require at least 50 annotations per GO-terms. Therefore, this condition is checked and those GO-terms that do not comply are eliminated.
# Checking the amount of annotations by GO-term
apply(CfData$tableCfGO, MARGIN=2, sum)
#> GO:0000166 GO:0003674 GO:0003676 GO:0003677 GO:0003824 GO:0004888 GO:0005102 GO:0005215 GO:0005488
#> 995 6962 1558 857 2679 516 520 565 5183
#> GO:0005515 GO:0005524 GO:0008144 GO:0016740 GO:0016787 GO:0017076 GO:0019899 GO:0030554 GO:0032553
#> 2738 682 764 1040 1097 878 870 707 882
#> GO:0032555 GO:0032559 GO:0035639 GO:0036094 GO:0038023 GO:0042802 GO:0043167 GO:0043168 GO:0043169
#> 871 702 847 1147 593 656 2273 1232 1307
#> GO:0046872 GO:0060089 GO:0097159 GO:0097367 GO:0098772 GO:0140096 GO:0140110 GO:1901265 GO:1901363
#> 1274 610 2523 992 650 978 615 995 2498
If we want to predict GO-terms in a single subdomain, BP, MF or CC , we must build the GO-DAG associated with these GO-terms.
library(GO.db)
library(GOstats)
mygraph <- GOGraph(CfData$nodesGO, GOMFPARENTS)
# Delete root node called all
mygraph <- subGraph(CfData$nodesGO, mygraph)
# We adapt the graph to the format used by FGGA
mygraph <- t(as(mygraph, "matrix"))
mygraphGO <- as(mygraph, "graphNEL")
# We search the root GO-term
rootGO <- leaves(mygraphGO, "in")
rootGO
#> [1] "GO:0003674" "GO:0008144"
plot(mygraphGO)
On the other hand, if you want to predict in two or three subdomains you should use the preCoreFG function. This function builds a graph respecting the GO constraints of inference and also links the GO-terms of the selected subdomains.
# We add GO-terms corresponding to Cellular Component subdomain
myGOs <- c(CfData[['nodesGO']], "GO:1902494", "GO:0032991", "GO:1990234",
"GO:0005575")
# We build a graph respecting the GO constraints of inference to MF and CC subdomains
mygraphGO <- preCoreFG(myGOs, domains="MF-CC")
plot(mygraphGO)
Let’s a GO subgraph, GO-terms GO:i are mapped to binary-valued variable nodes \(x_i\) of a factor graph while relationships between GO-terms are mapped to logical factor nodes \(f_k\) describing valid GO:i configurations under the True Path Graph constraint.
modelFGGA <- fgga2bipartite(mygraphGO)
Now, let’s use the MF subgraph to build our model of binary SVM classifiers with a training set of the Cf data.
# We take a subset of Cf data to train our model
idsTrain <- CfData$indexGO[["indexTrain"]][1:750]
# We build our model of binary SVM classifiers
modelSVMs <- lapply(CfData[["nodesGO"]], FUN = svmTrain,
tableGOs = CfData[["tableCfGO"]][idsTrain, ],
dxCharacterized = CfData[["dxCf"]][idsTrain, ],
graphGO = mygraphGO, kernelSVM = "radial")
# We calculate the reliability of each GO-term
varianceGOs <- varianceGO(tableGOs = CfData[["tableCfGO"]][idsTrain, ],
dxCharacterized = CfData[["dxCf"]][idsTrain, ],
kFold = 5, graphGO = mygraphGO, rootNode = rootGO,
kernelSVM = "radial")
varianceGOs
#> GO:0000166 GO:0003674 GO:0003676 GO:0003677 GO:0003824 GO:0004888 GO:0005102 GO:0005215 GO:0005488
#> 1.401 0.001 2.591 1.545 3.344 1.434 1.476 1.785 3.896
#> GO:0005515 GO:0005524 GO:0008144 GO:0016740 GO:0016787 GO:0017076 GO:0019899 GO:0030554 GO:0032553
#> 3.153 0.911 1.039 2.026 2.428 1.128 2.171 0.891 0.960
#> GO:0032555 GO:0032559 GO:0035639 GO:0036094 GO:0038023 GO:0042802 GO:0043167 GO:0043168 GO:0043169
#> 1.014 0.949 0.975 1.243 1.770 1.383 2.982 1.654 2.973
#> GO:0046872 GO:0060089 GO:0097159 GO:0097367 GO:0098772 GO:0140096 GO:0140110 GO:1901265 GO:1901363
#> 3.070 1.548 3.112 1.431 1.839 2.205 1.745 1.429 3.135
Next, we predict each GO-terms from a test set using our model of binary SVM classifiers.
dxTestCharacterized <- CfData[["dxCf"]][CfData$indexGO[["indexTest"]][1:50], ]
matrixGOTest <- svmGO(svmMoldel = modelSVMs,
dxCharacterized = dxTestCharacterized,
rootNode = rootGO,
varianceSVM = varianceGOs)
head(matrixGOTest)[,1:8]
#> GO:0000166 GO:0003674 GO:0003676 GO:0003677 GO:0003824 GO:0004888 GO:0005102 GO:0005215
#> J9P8X1 0.58526122 0.9999 0.5673337 0.30253745 0.3213095 0.1898543 0.3502084 0.5466443
#> F1PVD9 0.36945833 0.9999 0.6810226 0.76184402 0.1875907 0.1401090 0.4604712 0.2453967
#> F1PTQ9 0.09777731 0.9999 0.5674606 0.42287327 0.3993549 0.5838850 0.2890876 0.2802917
#> E2R8P1 0.48728672 0.9999 0.2439744 0.12009956 0.5969643 0.4153217 0.3781195 0.2488501
#> F6UWV4 0.02660784 0.9999 0.1485072 0.09705293 0.5342228 0.1674889 0.3285540 0.3734480
#> E2RQR5 0.27287685 0.9999 0.1871167 0.04250878 0.5083126 0.1650280 0.1253899 0.6402917
Once a factor graph model FG for the automated GO annotation problem has been defined, an iterative message passing algorithm between nodes of FG can be used to compute maximum a posteriori (MAP) estimates of variable nodes \(x_i\) modeling actual GO:i annotations.
The function msgFGGA returns a matrix with k rows and m columns corresponding to the Cf proteins and MF GO-terms respectively.
matrixFGGATest <- t(apply(matrixGOTest, MARGIN = 1, FUN = msgFGGA,
matrixFGGA = modelFGGA, graphGO = mygraphGO,
tmax = 50, epsilon = 0.001))
head(matrixFGGATest)[,1:8]
#> GO:0000166 GO:0003674 GO:0003676 GO:0003677 GO:0003824 GO:0004888 GO:0005102
#> J9P8X1 4.575186e-01 1.0000000 0.511057236 0.1545014018 0.5849553 0.071376105 0.27608105
#> F1PVD9 5.009218e-02 1.0000000 0.830522494 0.6327655048 0.2641987 0.007589162 0.42190161
#> F1PTQ9 1.469920e-03 1.0000000 0.451043778 0.1906380924 0.5795067 0.212323780 0.24219123
#> E2R8P1 1.000000e+00 1.0000000 0.268339107 0.0322274075 0.9335443 0.063682709 0.26862341
#> F6UWV4 1.744926e-06 0.9999999 0.003387351 0.0003283196 0.9751549 0.017807844 0.11977568
#> E2RQR5 9.425340e-01 1.0000000 0.188026131 0.0079917903 0.8846683 0.007549413 0.09432469
#> GO:0005215
#> J9P8X1 0.5466443
#> F1PVD9 0.2453967
#> F1PTQ9 0.2802917
#> E2R8P1 0.2488501
#> F6UWV4 0.3734480
#> E2RQR5 0.6402917
We now evaluate the performance of FGGA in terms of hierarchical F-score. The hierarchical classification performance metrics like the hierarchical precision (HP), the hierarchical recall (HR), and the hierarchical F-score (HF) measures publications [3] properly recognize partially correct classifications and correspondingly penalize more distant or more superficial errors. The formulas of the hierarchical metrics are shown below:
\[ HP(s) = \frac{1}{\mid l(P_{G}(s)) \mid} \hspace{1.25mm} \sum_{q \hspace{0.5mm} \in \hspace{0.5mm} l(P_{G}(s))} \hspace{1.5mm} \max_{c \hspace{0.5mm} \in \hspace{0.5mm} l(C_{G}(s))} \frac{\mid \uparrow c \hspace{1mm} \cap \uparrow q \mid}{\uparrow q} \]
\[ HR(s) = \frac{1}{\mid l(C_{G}(s)) \mid} \hspace{1.25mm} \sum_{c \hspace{0.5mm} \in \hspace{0.5mm} (C_{G}(s))} \hspace{1.5mm} \max_{q \hspace{0.5mm} \in \hspace{0.5mm} l(P_{G}(s))} \frac{\mid \uparrow c \hspace{1mm} \cap \uparrow q \mid}{\uparrow c} \]
\[ HF(s) = \frac{2 \cdot HP \cdot HR}{HP + HR} \]
HP HR HF samples
0.5949843 0.7047748 0.6178593 2399
Finally, we evaluate the performance of FGGA in terms of Area under the ROC Curve (AUC) and in terms of Precision at x Recall level (PxR). We use the R package PerfMeas, which provides functions to compute the performance measures we need.
library(PerfMeas)
for (i in 1:dim(matrixFGGATest)[1]){
matrixFGGATest[which(matrixFGGATest[i, ] >= 0.5),] <- 1
matrixFGGATest[which(matrixFGGATest[i, ] < 0.5),] <- 0
}
# Computing F-score
Fs <- F.measure.single.over.classes(
CfData$tableCfGO[rownames(matrixFGGATest), ], matrixFGGATest)
# Average F-score
Fs$average[4]
#> F
#> 0.3076908
# Computing AUC
auc <- AUC.single.over.classes(CfData$tableCfGO[rownames(matrixFGGATest), ],
matrixFGGATest);
# Average AUC
auc$average
#> [1] 0.4406522
# Computing precision at different recall levels
PXR <- precision.at.multiple.recall.level.over.classes(
CfData$tableCfGO[rownames(matrixFGGATest), ], matrixFGGATest,
seq(from = 0.1, to = 1, by = 0.1));
# Average PxR
PXR$avgPXR
#> 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
#> 0.3551945 0.3496750 0.3637629 0.3566131 0.3297087 0.3184688 0.3181759 0.3188827 0.3023128 0.3038879
1: Spetale F.E., Tapia E., Krsticevic F., Roda F. and Bulacio P. “A Factor Graph Approach to Automated GO Annotation”. PLoS ONE 11(1): e0146986, 2016.
2: Spetale Flavio E., Arce D., Krsticevic F., Bulacio P. and Tapia E. “Consistent prediction of GO protein localization”. Scientific Report 7787(8), 2018
3: Verspoor K., Cohn J., Mnizewski S., Joslyn C. “A categorization approach to automated ontological function annotation”. Protein Science. 2006;15:1544–1549