2023
DOI: 10.48550/arxiv.2301.12586
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unifying Molecular and Textual Representations via Multi-task Language Modelling

Abstract: The recent advances in neural language models have also been successfully applied to the field of chemistry, offering generative solutions for classical problems in molecular design and synthesis planning. These new methods have the potential to optimize laboratory operations and fuel a new era of data-driven automation in scientific discovery. However, specialized models are still typically required for each task, leading to the need for problem-specific fine-tuning and neglecting task interrelations. The mai… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 31 publications
0
1
0
Order By: Relevance
“…CLAMP [29] introduced a fusion approach, combining a molecule encoder and a text encoder for property prediction tasks. Christofidellis et al [30] presented a unified model capable of handling various text-to-text, text-tomolecule, molecule-to-text, and molecule-to-molecule tasks. MolReGPT [31] implemented tasks such as molecule captioning and text-based molecule generation by assigning ChatGPT a role as a biochemist, facilitating in-context learning.…”
Section: Data Generation Frameworkmentioning
confidence: 99%
“…CLAMP [29] introduced a fusion approach, combining a molecule encoder and a text encoder for property prediction tasks. Christofidellis et al [30] presented a unified model capable of handling various text-to-text, text-tomolecule, molecule-to-text, and molecule-to-molecule tasks. MolReGPT [31] implemented tasks such as molecule captioning and text-based molecule generation by assigning ChatGPT a role as a biochemist, facilitating in-context learning.…”
Section: Data Generation Frameworkmentioning
confidence: 99%
“…To enable higher-level control over molecular design, multi-modal models (Edwards et al, 2021;Vall et al, 2021;Zeng et al, 2022;Xu and Wang, 2022;Su et al, 2022;Seidl et al, 2023;Xu et al, 2023;Zhao et al, 2023;Liu et al, 2023b) have been proposed. Existing work focuses on cross-modal retrieval (Edwards et al, 2021;Zeng et al, 2022), translation (Edwards et al, 2022;Liu et al, 2023c;Christofidellis et al, 2023), and editing (Liu et al, 2022).…”
Section: B1 Multi-modal Models For Chemistrymentioning
confidence: 99%