2021
DOI: 10.1109/access.2021.3115659
|View full text |Cite
|
Sign up to set email alerts
|

SpaceTransformers: Language Modeling for Space Systems

Abstract: The transformers architecture and transfer learning have radically modified the Natural Language Processing (NLP) landscape, enabling new applications in fields where open source labelled datasets are scarce. Space systems engineering is a field with limited access to large labelled corpora and a need for enhanced knowledge reuse of accumulated design data. Transformers models such as the Bidirectional Encoder Representations from Transformers (BERT) and the Robustly Optimised BERT Pretraining Approach (RoBERT… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 23 publications
(33 reference statements)
0
1
0
Order By: Relevance
“…We need large models that contain representations of vast amounts of knowledge in core and progressively capilar scientific disciplines, with a grounding in world knowledge and commonsense understanding, as well as the ability to continuously acquire and update the model's beliefs as the state of the art evolves. Until now, scientific language models like SciBert (Beltagy et al, 2019), BioBert (Lee et al, 2019) or SpaceRoBERTa (Berquand et al, 2021) have tried to address the challenge of domain-specifty through additional pre-training on large amounts of scientific documents that leverage largescale open access scientific resources like OpenAire 3 , arXiv 4 , Web of Science 5 or Semantic Scholar. 6 However, only such knowledge which is statistically significant in the training data is effectively captured, limiting the usefulness of pre-trained language models as knowledge bases or reasoning engines.…”
Section: Modularity Adaptation and Autonomous Learningmentioning
confidence: 99%
“…We need large models that contain representations of vast amounts of knowledge in core and progressively capilar scientific disciplines, with a grounding in world knowledge and commonsense understanding, as well as the ability to continuously acquire and update the model's beliefs as the state of the art evolves. Until now, scientific language models like SciBert (Beltagy et al, 2019), BioBert (Lee et al, 2019) or SpaceRoBERTa (Berquand et al, 2021) have tried to address the challenge of domain-specifty through additional pre-training on large amounts of scientific documents that leverage largescale open access scientific resources like OpenAire 3 , arXiv 4 , Web of Science 5 or Semantic Scholar. 6 However, only such knowledge which is statistically significant in the training data is effectively captured, limiting the usefulness of pre-trained language models as knowledge bases or reasoning engines.…”
Section: Modularity Adaptation and Autonomous Learningmentioning
confidence: 99%