We hypothesize that a flexible, configurable suite of automated informatics tools can reduce significantly the effort needed to generate systematic reviews while maintaining or even improving their quality. To test this hypothesis, we propose:
Aim 1. To extend our research on automated RCT tagging to include additional study types and provide public resources. A) Machine learning models will be created that automatically assign probability estimates to three types of observational studies that are widely examined by systematic reviewers. B) The RCT and other taggers will be evaluated prospectively for newly published PubMed articles. C) All PubMed articles will be automatically tagged for RCT, cohort, case-control and cross-sectional studies and annotated in a public dataset linked to a public query interface. Users will also receive tags for articles from non-PubMed data sources on demand.
Aim 2. To evaluate the performance and usability of our tools when used by systematic reviewers under field conditions. A) The tools will be customized and integrated to facilitate field evaluation. B) A three-stage evaluation: 1. Retrospective evaluation of Metta and RCT Tagger performance. 2. Real-time ?shadowing?. 3. Prospective controlled study.
Aim 3. To identify additional clinical trial articles, appearing after a published systematic review was completed, that are relevant to the review topic.
Aim 4. To identify publications related to specific ClinicalTrials.gov registered trials.
Aim 5. To develop and evaluate new machine learning methods and tools that will facilitate rapid evidence scoping for new systematic review topics. A) Methods will be developed for ranking articles with respect to their relevance to a proposed new systematic review topic. B) A scoping tool will be created that displays articles ranked by predicted relevance, tagged with study design attributes, sample sizes, and Cochrane risk of bias estimates. The proposed studies will advance the automation of early steps in the process of writing systematic reviews, and thereby enhance evidence-based medicine and the incorporation of best practices into clinical care.

Public Health Relevance

Systematic reviews are essential for determining which treatments and interventions are safe and effective. At present, systematic reviews are written largely by laborious manual methods. The proposed studies will reduce the time and effort needed to write systematic reviews, and thereby enhance evidence-based medicine and the incorporation of best practices into clinical care.

Agency
National Institute of Health (NIH)
Institute
National Library of Medicine (NLM)
Type
Research Project (R01)
Project #
5R01LM010817-08
Application #
9731665
Study Section
Biomedical Library and Informatics Review Committee (BLR)
Program Officer
Ye, Jane
Project Start
2010-09-30
Project End
2021-06-30
Budget Start
2019-07-01
Budget End
2021-06-30
Support Year
8
Fiscal Year
2019
Total Cost
Indirect Cost
Name
University of Illinois at Chicago
Department
Psychiatry
Type
Schools of Medicine
DUNS #
098987217
City
Chicago
State
IL
Country
United States
Zip Code
60612
Smalheiser, Neil R (2017) Rediscovering Don Swanson: the Past, Present and Future of Literature-Based Discovery. J Data Inf Sci 2:43-64
Peng, Yufang; Bonifield, Gary; Smalheiser, Neil R (2017) Gaps within the Biomedical Literature: Initial Characterization and Assessment of Strategies for Discovery. Front Res Metr Anal 2:
Wallace, Byron C; Noel-Storr, Anna; Marshall, Iain J et al. (2017) Identifying reports of randomized controlled trials (RCTs) via a hybrid machine learning and crowdsourcing approach. J Am Med Inform Assoc 24:1165-1168
Smalheiser, Neil R; Bonifield, Gary (2016) Two Similarity Metrics for Medical Subject Headings (MeSH): An Aid to Biomedical Text Mining and Author Name Disambiguation. J Biomed Discov Collab 7:e1
Smalheiser, Neil R; Shao, Weixiang; Yu, Philip S (2015) Nuggets: findings shared in multiple clinical case reports. J Med Libr Assoc 103:171-6
Cohen, Aaron M; Smalheiser, Neil R; McDonagh, Marian S et al. (2015) Automated confidence ranked classification of randomized controlled trial articles: an aid to evidence-based medicine. J Am Med Inform Assoc 22:707-17
Shao, Weixiang; Adams, Clive E; Cohen, Aaron M et al. (2015) Aggregator: a machine learning approach to identifying MEDLINE articles that derive from the same underlying clinical trial. Methods 74:65-70
D'Souza, Jennifer L; Smalheiser, Neil R (2014) Three journal similarity metrics and their application to biomedical journals. PLoS One 9:e115681
Jiang, Yu; Lin, Can; Meng, Weiyi et al. (2014) Rule-based deduplication of article records from bibliographic databases. Database (Oxford) 2014:bat086
Edinger, Tracy; Cohen, Aaron M (2013) A large-scale analysis of the reasons given for excluding articles that are retrieved by literature search during systematic review. AMIA Annu Symp Proc 2013:379-87

Showing the most recent 10 out of 11 publications