2021
DOI: 10.48550/arxiv.2106.11360
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Hi-BEHRT: Hierarchical Transformer-based model for accurate prediction of clinical events using multimodal longitudinal electronic health records

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…In this study, we compared the performance of PCB with standard DL models for HF risk prediction, to investigate the trade-off between counterfactual reasoning and predic-tion accuracy. For PCB models, we constructed the initial representation learning architecture m(•) by adopting a previous high-performing Transformer (Devlin et al, 2019) model architecture, Hi-BEHRT, and its parameters (Li et al, 2021) (see Supplementary, Method S2 for more). Next, a two-layer multi-layer perceptron g(•) and a vector quantization component h(•) were trained to map m(x) to c and l, respectively.…”
Section: Baseline Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…In this study, we compared the performance of PCB with standard DL models for HF risk prediction, to investigate the trade-off between counterfactual reasoning and predic-tion accuracy. For PCB models, we constructed the initial representation learning architecture m(•) by adopting a previous high-performing Transformer (Devlin et al, 2019) model architecture, Hi-BEHRT, and its parameters (Li et al, 2021) (see Supplementary, Method S2 for more). Next, a two-layer multi-layer perceptron g(•) and a vector quantization component h(•) were trained to map m(x) to c and l, respectively.…”
Section: Baseline Modelsmentioning
confidence: 99%
“…Hi-BEHRT (Li et al, 2021) We used the identical Hi-BEHRT as mentioned above for latent representation learning. However, instead of pooling the first-time step for classification, we used a two-layer multi-layer perceptron to map the representation to the highlevel concepts with 64 units for the first layer.…”
Section: S22 Hi-behrtmentioning
confidence: 99%
“…Recently, more studies focused on the use of time-series information. Methods for such an approach include autoencoders, convolutional neural networks [3], or sequential models like recurrent neural networks (RNN) [4] or transformer-based models [5][6][7][8][9]. Transformer-based models originate from natural language processing (NLP) and have recently gained much attention since they have achieved excellent results in many areas [10][11][12][13][14].…”
Section: Introductionmentioning
confidence: 99%
“…Later, Li et al developed BERT for EHR (BEHRT), which generated a patient embedding based on the history of diagnoses and used it for disease prediction in different time windows [5]. Since BEHRT -like most transformer-based models -is limited with respect to the maximum sequence length, the authors later developed a hierarchical BEHRT variant (HI-BEHRT), which can process longer medical histories [6]. Another model, called the Bidirectional Representation Learning model with a Transformer architecture on Multimodal EHR (BRLTM), was published by Meng et al in 2021[8].…”
Section: Introductionmentioning
confidence: 99%