Fierens, Amaury
[UCL]
Jodogne, Sébastien
[UCL]
Many medical applications are envisioned for Large Language Models (LLMs), such as the automated summary of the health condition of a patient, or the automated codification of electronic health records. Even though the training of LLMs directly inside hospitals is highly desirable to exploit the local clinical data while avoiding data privacy concerns, this process requires a costly, complex computing infrastructure. This paper explores the recent Cramming approach as a cost-effective way to train LLMs within medical institutions in one day using one GPU. We show that the Cramming approach that was originally designed for English can be transposed to French, and that the resulting models can be successfully fine-tuned to healthcare-related tasks in the French language. This research opens the path to the creation of LLMs that are tailored to the specific needs of institutions that handle sensitive textual data in another language than English.
Bibliographic reference |
Fierens, Amaury ; Jodogne, Sébastien. BERTinchamps: Cost-Effective Training of Large Language Models for Medical Tasks in French.Workshop on Natural Language for Artificial Intelligence (NL4AI, 7th edition) (Rome, Italy, du 06/11/2023 au 09/11/2023). In: CEUR Workshop Proceedings, Vol. 3551 (2023) |
Permanent URL |
http://hdl.handle.net/2078.1/279237 |