2012
DOI: 10.1136/amiajnl-2012-000862
|View full text |Cite
|
Sign up to set email alerts
|

Grid Binary LOgistic REgression (GLORE): building shared models without sharing data

Abstract: ObjectiveThe classification of complex or rare patterns in clinical and genomic data requires the availability of a large, labeled patient set. While methods that operate on large, centralized data sources have been extensively used, little attention has been paid to understanding whether models such as binary logistic regression (LR) can be developed in a distributed manner, allowing researchers to share models without necessarily sharing patient data.Material and methodsInstead of bringing data to a central … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
162
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
7
2
1

Relationship

3
7

Authors

Journals

citations
Cited by 155 publications
(165 citation statements)
references
References 43 publications
0
162
0
Order By: Relevance
“…This is possible as the central 'master' sends possible prediction models rather than fetching the data from remote nodes. Only statistical indexes totally unrelated to specific patients are exchanged between nodes and their master [47,[53][54]. Although this protocol does not require an intervention and the data is fully deidentified, internal review board ethics approval is recommended before implementing a local node.…”
Section: Privacy Protection Of Patientsmentioning
confidence: 99%
“…This is possible as the central 'master' sends possible prediction models rather than fetching the data from remote nodes. Only statistical indexes totally unrelated to specific patients are exchanged between nodes and their master [47,[53][54]. Although this protocol does not require an intervention and the data is fully deidentified, internal review board ethics approval is recommended before implementing a local node.…”
Section: Privacy Protection Of Patientsmentioning
confidence: 99%
“…In this setting, the amount of data per transfer diminishes; however, the data transfer frequency increases. A thorough explanation how distributed machine learning algorithms work is given by Boyd et al [5] and Wu et al [34]. From this work of Boyd et al, we reused the MapReduce concept, developed by Dean and Ghemawat [8], to implement the distributed machine learning concept.…”
Section: Distributed Machine Learningmentioning
confidence: 99%
“…T e VM import model can also enable distributed computation, with each party installing the same VM and contributing results of its local computation to a coordinating center. For example, we have shown that it is possible to create an accurate predictive model by exporting the computation to dif erent centers and aggregating results only, without any individual patient data ever being transferred (21).…”
Section: " "mentioning
confidence: 99%