Table 1 Performance comparison between the trained transformers and random-forest models across several metrics
From: A chemical language model for molecular taste prediction
Unweighted average | ||||||
|---|---|---|---|---|---|---|
Model | Accuracy | Precision | Recall | F1 Score | AUROC | Support |
XGBoost: fingerprints (fp) | 0.8988 | 0.8169 | 0.7400 | 0.7661 | 0.8520 | 100% |
XGBoost: fp+descriptors | 0.8962 | 0.8842 | 0.7402 | 0.7779 | 0.8513 | 100% |
Balanced Random Forest: fp | 0.7972 | 0.5845 | 0.7322 | 0.6014 | 0.8391 | 100% |
Chemprop | 0.8851 | 0.8037 | 0.7224 | 0.7499 | 0.8392 | 100% |
FART | 0.8860 | 0.7720 | 0.6873 | 0.7118 | 0.9639 | 100% |
FART augmented | 0.8940 | 0.8789 | 0.7388 | 0.7737 | 0.9744 | 100% |
FART augmented + confidence | 0.9155 | 0.9000 | 0.7617 | 0.7956 | 0.9806 | 94% |