DDintensity Supplementary Material

Limitation: Requirement for Large Datasets

Table S1. Specific number of positive and negative samples for CVs

Dataset (remain)Y_train0Y_train1Y_test0Y_test1 
80%5698115618142453905 
64%4558412495113963124 
51.2%36478998591202496 
20%1432038363580959 
4%2869762718190 
DDI2013*2792510699127 

* The number of remaining relationship pairs in the DDI2013 dataset after filtering and alignment.

The performance of the small dataset of DDI2013 is relatively poor compared to the large datasets of Ddinter and Drugbank. Therefore, we hope to continue to explore whether our method has high requirements for the dataset. We have added the following experiments

We randomly deleted relationship pairs from Ddinter based on the ratio of positive and negative samples to ensure sufficient samples for each training session. Afterwards, we conducted a 5-fold cross-validation on the new dataset and recorded the AUC and AUPR values. The specific number of positive and negative samples for cross-validations is shown in Table S1. The results obtained are presented in Figure S1 and Figure S2.‌

As illustrated in the figure, when we randomly decrease the number of positive and negative samples involved in training, regardless of the specific reduction amount, it is evident that our prediction metrics, AUC and AUPR, continuously decline. Notably, when more than 80% of the data is removed, there is a steep drop in both AUC and AUPR.‌

When we delete 96% of the data, we are left with 2869 entries labeled as 0 and 762 entries labeled as 1. The remaining test set contains 718 entries labeled as 0 and 190 entries labeled as 1. It can be observed that when the original dataset is reduced to only 4% (roughly the same number of relationship pairs as the DDI2013 dataset), the model’s performance deteriorates, with the AUC dropping to 0.742 and the AUPR decreasing to 0.452. This indicates that our method requires a large and accurate dataset to support its performance, which means, the quantity and quality of training samples have a significant impact on model performance. Reducing the dataset size, especially to a large extent, severely affects the model’s ability to generalize and make accurate predictions

Figure S1. Trend of AUC After Deleting a Proportion of the Dataset. When we randomly decrease the number of positive and negative samples involved in training, regardless of the specific reduction amount, it is evident that AUC continuously declines.
Figure S2. Trend of AUPR After Deleting a Proportion of the Dataset. When we randomly decrease the number of positive and negative samples involved in training, regardless of the specific reduction amount, it is evident that AUPR continuously declines.

Limitation 2: The storage space and computing power requirements of the dataset constructed using the splicing method will be further increased

Table S2. Details of each dataset

DatasetBalancedSamplingDD-TypePositiveNegativeFile_size
DDI2013NoNoDDI63734910.12GB
DrugBankYesYesDDI19180819180811.57 GB
DDinterNoNoDD intensity (Major? or not)24647903113.43GB

If there are high requirements for the size of the dataset, what we need to further explore is whether further expansion will lead to the need for particularly large storage space and computing power.‌

Our method involves concatenating features, so as the dataset expands, more storage and computing resources will be required to scale up the training.

For instance, Table S2 shows that DDI2013 contains 4,128 relationship pairs and occupies 0.12GB, while DrugBank has 383,616 pairs and occupies 11.57GB. The dataset expanded by 92.93 times, but the storage increased by 96.42 times.

Therefore, it can be inferred that as the dataset grows, the storage demand will further increase. Currently, our algorithm runs on a GTX4080 platform with 12GB memory, an i9-13900HX processor, and 64GB DDR4 RAM. For larger datasets, more advanced hardware will be needed.‌

Imbalanced issue

To address the issue of the imbalanced dataset, we designed another experiment. Currently, in our dataset, the number of positive samples (major) is much smaller than the number of negative samples (moderate, minor). Since it is difficult to introduce relationships using oversampling methods (no other datasets contain information on drug-drug interactions), we opted for a down sampling method (i.e., randomly deleting some negative samples). This resulted in a balanced dataset, DS1, and two other imbalanced datasets, DS05 and DS2. In DS1, the number of negative samples is equal to the number of positive samples; in DS05, the number of negative samples is half that of the positive samples; and in DS2, the number of negative samples is double that of the positive samples.

Table S3. Down-sampling results

DatasetAUC(mean)Std (AUC.)AUPR(mean)Std (AUC.)
DS050.94550.00260.9670.0026
DS10.95770.00150.95430.0023
DS20.96570.00110.93710.0024
DS_original0.97040.00110.91980.003

According to the results, our model performs well on both the balanced (DS1) and imbalanced (DS0.5, DS2, DS_original) drug interaction datasets.

Set label “major” as label ‘One’ issue

 Firstly, most existing methods solely aim to predict whether there are adverse effects among proved drug pairs, ignoring the classification of risk levels within different drug combinations.

Secondly, our case study on the drugs in the dataset that were not involved in training shows that using this method for modeling and prediction yields excellent results.

Thirdly, our approach is more inclined to identify stronger relationships in drug-drug interactions. An intuitive hypothesis is that if we set major as 1 and the others as 0 during dataset design, it could better help us discover “strong” relationships.

To further validate our hypothesis, we designed a new experiment. First, we randomly selected 5,000 major relationship pairs from the dataset (and removed them from the training set) to create an independent dataset. Next, we trained the model using Major as label1, and then using both major and moderate as label1, with the rest set as label0.

Subsequently, we randomly selected 5,000 drug-drug relationship pairs as a control group of negative samples. We then compared the average scores of the 5,000 major independent dataset with the 5,000 randomly selected negative samples. The results are shown in Table S4. When using the Major label as label1, the average score was lower, but it was more likely to classify the randomly selected negative samples as negative (resulting in a lower score). In contrast, when using the moderate label along with Major as label1, the average score was higher, but it also tended to assign higher scores to negative samples (leaning towards classifying randomly selected negative samples as positive).

We compared the difference in averages between the two approaches and found that using the major label as label1 resulted in a greater difference in average scores between the independent positive sample dataset and the randomly selected negative sample dataset. This indicates that this method allows the model to better identify strong DDI relationships.

Table S4. Independent dataset/ random selected negative samples results

Label typeIndependent (Positive mean)Random selected (Negative mean)Variance
Major(1)2.6593-5.40558.0647
Ma+Moderate(1)5.81713.58362.23353

Other metrics

As DDintensity is a predict and rank model, our framework focuses on more threshold-less metrics, such as AUC and AUPR.

It usually needs a threshold to calculate the metric for accuracy, precision, and recall. Moreover, the threshold is critical as it may influence performance. However, to further evaluate our models, we calculated F1, Recall, and Precision. For reference, we set the threshold as the mean value of all predicted scores for each modality. As shown in table, S4, which records F1, Recall, and Precision. We can see that the embedding we used ranks first in F1, second in Recall, and first in Precision among all the embeddings involved in the comparison.

Table S5. Results comparisons with different embeddings

EmbeddingsF1F1_stdRecallRecall_stdPrecisionPre_std
3DResnet500.73120.00470.94220.00420.70690.0042
DinoVitb160.7270.00360.95040.00270.70880.0031
GAT(DDI)0.67310.00470.9440.00430.68340.0037
Bart0.76330.0050.97120.00250.73180.004
SapBert0.760.0040.97480.00230.72680.0035
SMILEFIN0.70450.00460.92880.00460.69390.0038
RDF2VEC0.760.00420.98010.00280.72540.0036
SciBert0.75270.00360.97150.00230.72230.003
GADNN0.66840.00690.91180.00740.66720.0059
ChemBERTa-20.72930.00290.95930.00310.71240.0029
bioGPT0.77040.00370.97610.00240.73670.0032