2021
DOI: 10.1109/access.2021.3077350
|View full text |Cite
|
Sign up to set email alerts
|

The NLP Cookbook: Modern Recipes for Transformer Based Deep Learning Architectures

Abstract: In recent years, Natural Language Processing (NLP) models have achieved phenomenal success in linguistic and semantic tasks like text classification, machine translation, cognitive dialogue systems, information retrieval via Natural Language Understanding (NLU), and Natural Language Generation (NLG). This feat is primarily attributed due to the seminal Transformer architecture, leading to designs such as BERT, GPT (I, II, III), etc. Although these large-size models have achieved unprecedented performances, the… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
50
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 82 publications
(51 citation statements)
references
References 44 publications
0
50
0
1
Order By: Relevance
“…This includes but is not limited to transfer learning [ 76 ] or adversarial learning [ 77 ], which include a variety of neural networks structures or knowledge graphs that have been at the core of NLP research. This also includes Transformer-based methods [ 78 ] (e.g., large pre-trained language models such as BERT [ 79 ]), which have made a significant impact on the field of NLP over recent years and could prove to be useful in NLP for occupational exposure research. This type of deep learning method is based on attention [ 80 ], which has been shown to improve results in a variety of other domains that have utilised NLP (e.g., healthcare).…”
Section: Discussionmentioning
confidence: 99%
“…This includes but is not limited to transfer learning [ 76 ] or adversarial learning [ 77 ], which include a variety of neural networks structures or knowledge graphs that have been at the core of NLP research. This also includes Transformer-based methods [ 78 ] (e.g., large pre-trained language models such as BERT [ 79 ]), which have made a significant impact on the field of NLP over recent years and could prove to be useful in NLP for occupational exposure research. This type of deep learning method is based on attention [ 80 ], which has been shown to improve results in a variety of other domains that have utilised NLP (e.g., healthcare).…”
Section: Discussionmentioning
confidence: 99%
“…However, computational algorithms have difficulty in recognizing different text sentences as the same [47], [48]. Natural Language Processing approaches [49]- [51] have been proposed over the years. However, they depend on the quality and completeness of the information presented in the text.…”
Section: A Structured Task Modelmentioning
confidence: 99%
“…Self-attention enables the look-up of remaining input words at various positions to determine the significance of the currently processed word. This is done for all input words to improve the encoding and context understanding of all words [ 96 ]. We present an illustration of this architecture in Fig.…”
Section: Introductionmentioning
confidence: 99%