Objectives To perform a systematic review of design and reporting of imaging studies applying convolutional neural network models for radiological cancer diagnosis. Methods A comprehensive search of PUBMED, EMBASE, MEDLINE and SCOPUS was performed for published studies applying convolutional neural network models to radiological cancer diagnosis from January 1, 2016, to August 1, 2020. Two independent reviewers measured compliance with the Checklist for Artificial Intelligence in Medical Imaging (CLAIM). Compliance was defined as the proportion of applicable CLAIM items satisfied. Results One hundred eighty-six of 655 screened studies were included. Many studies did not meet the criteria for current design and reporting guidelines. Twenty-seven percent of studies documented eligibility criteria for their data (50/186, 95% CI 21–34%), 31% reported demographics for their study population (58/186, 95% CI 25–39%) and 49% of studies assessed model performance on test data partitions (91/186, 95% CI 42–57%). Median CLAIM compliance was 0.40 (IQR 0.33–0.49). Compliance correlated positively with publication year (ρ = 0.15, p = .04) and journal H-index (ρ = 0.27, p < .001). Clinical journals demonstrated higher mean compliance than technical journals (0.44 vs. 0.37, p < .001). Conclusions Our findings highlight opportunities for improved design and reporting of convolutional neural network research for radiological cancer diagnosis. Key Points • Imaging studies applying convolutional neural networks (CNNs) for cancer diagnosis frequently omit key clinical information including eligibility criteria and population demographics. • Fewer than half of imaging studies assessed model performance on explicitly unobserved test data partitions. • Design and reporting standards have improved in CNN research for radiological cancer diagnosis, though many opportunities remain for further progress.
Polynomial time solutions to the graph isomorphism problem have remained elusive to date, presenting one of the most important issues in computer science. This study describes a spectral algorithm to infer quadratic permutations mapping tuples of isomorphic graphs in O(n^4) time. Robustness to degeneracy and multiple isomorphisms are achieved through low dimensional eigenspace projections and iterative perturbations respectively. An analytic proof is provided and experimental performance is demonstrated in synthetic and biological data.
Objectives Radiomic models present an avenue to improve oesophageal adenocarcinoma assessment through quantitative medical image analysis. However, model selection is complicated by the abundance of available predictors and the uncertainty of their relevance and reproducibility. This analysis reviews recent research to facilitate precedent-based model selection for prospective validation studies. Methods This analysis reviews research on 18F-FDG PET/CT, PET/MRI and CT radiomics in oesophageal adenocarcinoma between 2016 and 2021. Model design, testing and reporting are evaluated according to the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) score and Radiomics Quality Score (RQS). Key results and limitations are analysed to identify opportunities for future research in the area. Results Radiomic models of stage and therapeutic response demonstrated discriminative capacity, though clinical applications require greater sensitivity. Although radiomic models predict survival within institutions, generalisability is limited. Few radiomic features have been recommended independently by multiple studies. Conclusions Future research must prioritise prospective validation of previously proposed models to further clinical translation.
Background: Evaluation of gene interaction models in cancer genomics is challenging, as the true distribution is uncertain. Previous analyses have benchmarked models using synthetic data or databases of experimentally verified interactions – approaches which are susceptible to misrepresentation and incompleteness, respectively. The objectives of this analysis are to (1) provide a real-world data-driven approach for comparing performance of genomic model inference algorithms, (2) compare the performance of LASSO, elastic net, best-subset selection, [Formula: see text] penalisation and [Formula: see text] penalisation in real genomic data and (3) compare algorithmic preselection according to performance in our benchmark datasets to algorithmic selection by internal cross-validation. Methods: Five large [Formula: see text] genomic datasets were extracted from Gene Expression Omnibus. ‘Gold-standard’ regression models were trained on subspaces of these datasets ([Formula: see text], [Formula: see text]). Penalised regression models were trained on small samples from these subspaces ([Formula: see text]) and validated against the gold-standard models. Variable selection performance and out-of-sample prediction were assessed. Penalty ‘preselection’ according to test performance in the other 4 datasets was compared to selection internal cross-validation error minimisation. Results: [Formula: see text]-penalisation achieved the highest cosine similarity between estimated coefficients and those of gold-standard models. [Formula: see text]-penalised models explained the greatest proportion of variance in test responses, though performance was unreliable in low signal:noise conditions. [Formula: see text] also attained the highest overall median variable selection F1 score. Penalty preselection significantly outperformed selection by internal cross-validation in each of 3 examined metrics. Conclusions: This analysis explores a novel approach for comparisons of model selection approaches in real genomic data from 5 cancers. Our benchmarking datasets have been made publicly available for use in future research. Our findings support the use of [Formula: see text] penalisation for structural selection and [Formula: see text] penalisation for coefficient recovery in genomic data. Evaluation of learning algorithms according to observed test performance in external genomic datasets yields valuable insights into actual test performance, providing a data-driven complement to internal cross-validation in genomic regression tasks.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.