中文 |

Newsroom

Advancing Chemical Text Mining: Fine-Tuning Large Language Models for Knowledge Extraction

Jun 17, 2024

Chemical literature contains a wealth of data, and extracting key chemical information through text mining techniques to build extensive databases can provide experimental chemists with detailed physical and chemical properties and synthesis routes. It can also offer computational chemists rich data and insights for model building and prediction. However, due to the complexity of chemical language and the diversity of writing styles, extracting standardized and structured data from chemical literature is a highly challenging task.
In a study published in Chemical Science, a research team led by ZHENG Mingyue from the Shanghai Institute of Materia Medica (SIMM) of the Chinese Academy of Sciences (CAS), conducted a comprehensive study to explore the capabilities of various large language models (LLMs) on five key chemical text mining tasks. This work demonstrated the versatility and efficiency of fine-tuning LLMs for extracting chemical information from literature and provides valuable insights into the practical applications of LLMs.
In this study, researchers evaluated various LLMs on five tasks: compound entity recognition, reaction role labeling, metal-organic framework (MOF) synthesis information extraction, nuclear magnetic resonance (NMR) data extraction, and reaction synthesis paragraph-to-action sequence conversion. They explored multiple strategies using different LLMs, including guiding GPT-4 and GPT-3.5-turbo with zero-shot and few-shot prompt engineering, as well as parameter efficient fine-tuning or full parameter fine-tuning models like GPT-3.5-turbo, Llama3, Mistral, T5, and BART. 
The research revealed that fully fine-tuned LLMs exhibited impressive performance, and meanwhile, reduced reliance on prompting techniques. The fine-tuned ChatGPT models excelled at all tasks, achieving exact match accuracies between 69% and 95% and outperforming specific models pre-trained on extensive domain-specific data. Additionally, open-source models like Mistral and Llama3, once fine-tuned, demonstrated competitive performance and showcased the potential of LLMs in mining private data locally. 
This study underscored the flexibility, robustness, efficiency, and low-code nature of fine-tuned LLMs in chemical text mining. In light of these properties, fine-tuning LLMs can emerge as a powerful method for information extraction and accelerate data collection and scientific discovery across various fields.
Schematics of cheminformatics insights extracted from paragraphs with LLMs. (Image by ZHENG Mingyue's Laboratory)
Contact

JIANG Qingling

Shanghai Institute of Materia Medica

E-mail:

Fine-tuning large language models for chemical text mining

Related Articles
Contact Us
  • 86-10-68597521 (day)

    86-10-68597289 (night)

  • 86-10-68511095 (day)

    86-10-68512458 (night)

  • cas_en@cas.cn

  • 52 Sanlihe Rd., Xicheng District,

    Beijing, China (100864)

Copyright © 2002 - Chinese Academy of Sciences