Feb. 7, 2024, 5:44 a.m. | Pengfei Liu Yiming Ren Jun Tao Zhixiang Ren

cs.LG updates on arXiv.org arxiv.org

Large language models have made significant strides in natural language processing, enabling innovative applications in molecular science by processing textual representations of molecules. However, most existing language models cannot capture the rich information with complex molecular structures or images. In this paper, we introduce GIT-Mol, a multi-modal large language model that integrates the Graph, Image, and Text information. To facilitate the integration of multi-modal molecular data, we propose GIT-Former, a novel architecture that is capable of aligning all modalities into …

applications cs.cl cs.lg enabling git graph image images information language language model language models language processing large language large language model large language models modal molecular science molecules multi-modal natural natural language natural language processing paper processing q-bio.bm science text textual

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne