2022
DOI: 10.48550/arxiv.2211.09085
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Galactica: A Large Language Model for Science

Abstract: Information overload is a major obstacle to scientific progress. The explosive growth in scientific literature and data has made it ever harder to discover useful insights in a large mass of information. Today scientific knowledge is accessed through search engines, but they are unable to organize scientific knowledge alone. In this paper we introduce Galactica: a large language model that can store, combine and reason about scientific knowledge. We train on a large scientific corpus of papers, reference mater… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
58
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 67 publications
(87 citation statements)
references
References 28 publications
0
58
0
Order By: Relevance
“…This is an advancement not seen in previous smaller language models and likely reflects the model's ability to learn and extract more knowledge from its training data. Transformer-based LLMs have demonstrated close to human-level performances in medical question and answering benchmarks and summarisation tasks ( 7 9 ), and with techniques like self-consistency ( 9 ), chain of thought prompting ( 10 ), and reinforcement learning from human feedback ( 11 ) the model performance can be further enhanced. Given their rapid rate of advancement, it is probable that LLM based conversational AI (chatbots) will soon be developed for healthcare use.…”
Section: Introductionmentioning
confidence: 99%
“…This is an advancement not seen in previous smaller language models and likely reflects the model's ability to learn and extract more knowledge from its training data. Transformer-based LLMs have demonstrated close to human-level performances in medical question and answering benchmarks and summarisation tasks ( 7 9 ), and with techniques like self-consistency ( 9 ), chain of thought prompting ( 10 ), and reinforcement learning from human feedback ( 11 ) the model performance can be further enhanced. Given their rapid rate of advancement, it is probable that LLM based conversational AI (chatbots) will soon be developed for healthcare use.…”
Section: Introductionmentioning
confidence: 99%
“…Such weighting is more "pliable" than a rule-based generation and mimics human-like statistical learning to some extent by enabling associations in the representation space. However, this same blending contributes to falsehoods (45).…”
Section: Factors Contributing To Nlg Errorsmentioning
confidence: 99%
“…As the number of new publications over time grew and methodical and linguistic diversity increased, the task has become constantly more challenging. Similar developments in science in general 63 have inspired a series of innovative attempts to solve the problem of scientific knowledge discovery and education 36 by using of large language models to support both research 64 and writing of scientific publications 65 albeit with limited success and apparent upper bounds despite large numbers of parameters at present. 66 Given current technical limitations and employing an established design mechanism in machine learning, 67,68 we suggest a hybrid, human-in-the-loop approach to information management, combining state-of-the-art natural language processing tools to find and curate publications for review with subject matter expert human control.…”
Section: An Interactive Living Surveymentioning
confidence: 99%