Advancements in neural machinery have led to a wide range of algorithmic solutions for molecular property prediction. Two classes of models in particular have yielded promising results: neural networks applied to computed molecular fingerprints or expert-crafted descriptors and graph convolutional neural networks that construct a learned molecular representation by operating on the graph structure of the molecule. However, recent literature has yet to clearly determine which of these two methods is superior when generalizing to new chemical space. Furthermore, prior research has rarely examined these new models in industry research settings in comparison to existing employed models. In this paper, we benchmark models extensively on 19 public and 16 proprietary industrial data sets spanning a wide variety of chemical end points. In addition, we introduce a graph convolutional model that consistently matches or outperforms models using fixed molecular descriptors as well as previous graph neural architectures on both public and proprietary data sets. Our empirical findings indicate that while approaches based on these representations have yet to reach the level of experimental reproducibility, our proposed model nevertheless offers significant improvements over models currently used in industrial workflows.
Highlights d A deep learning model is trained to predict antibiotics based on structure d Halicin is predicted as an antibacterial molecule from the Drug Repurposing Hub d Halicin shows broad-spectrum antibiotic activities in mice d More antibiotics with distinct structures are predicted from the ZINC15 database
Uncertainty quantification (UQ) is an important component of molecular property prediction, particularly for drug discovery applications where model predictions direct experimental design and where unanticipated imprecision wastes valuable time and resources. The need for UQ is especially acute for neural models, which are becoming increasingly standard yet are challenging to interpret. While several approaches to UQ have been proposed in the literature, there is no clear consensus on the comparative performance of these models. In this paper, we study this question in the context of regression tasks. We systematically evaluate several methods on five regression data sets using multiple complementary performance metrics. Our experiments show that none of the methods we tested is unequivocally superior to all others, and none produces a particularly reliable ranking of errors across multiple data sets. While we believe that these results show that existing UQ methods are not sufficient for all common use cases and further research is needed, we conclude with a practical recommendation as to which existing techniques seem to perform well relative to others.
ammographic breast density can mask cancers at mammography and is an independent risk factor for breast cancer (1-3). Legislation mandating patients be notified of mammographic breast density has passed in more than 30 states, and a federal bill is under consideration. Details of state legislation vary, but most states require direct reporting to the patient that breast density can mask cancers at mammography and that the patient may benefit from additional testing. Qualitative assessment of mammographic breast density is subjective and varies widely between radiologists (4-10). In a study of 83 radiologists who assessed breast density, Sprague et al (4) found extreme variation in qualitative density assessment per the Breast Imaging Reporting and Data System (BI-RADS), with 6%-85% of mammograms assessed as either heterogeneously or extremely dense depending on radiologist interpretation. In a study of 34 radiologists, the intraradiologist agreement of density assessments among women who underwent two examinations varied from 62% to 87% (6). Commercially available methods for automated assessment of breast density do exist; however, they yield mixed results in agreement with expert qualitative density assessments, with k scores of 0.32-0.61 (11,12). These methods tend to result in over-or underreporting of breast density when compared with qualitative assessment by radiologists (11,13). A recent study found significant differences in density assessments in the same 4170 women with two software programs (Volpara, Volpara Solutions, Wellington, New Zealand; Quantra, Hologic, Bedford, Mass), with the software programs showing 37% and 51%, respectively, of women had dense breast tissue. In the same set of mammograms, radiologists determined 43% of the women had dense breast tissue (13). Deep learning (DL) has been gaining traction in radiology (12,14-17). Specifically, there has been preliminary work with DL methods to assess breast density (12,18); however, none of these techniques have been implemented in clinical practice, raising questions about clinical acceptance by practicing radiologists and the effect on patient care. In contrast, our purpose was to develop a DL algorithm we could use to reliably assess breast density and to measure the acceptance of its predictions in real-time clinical practice. We hypothesize that DL models can be applied to assess breast density at the same level as experienced breast imagers and that they can be accepted into routine clinical practice.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.