Open Access iconOpen Access

ARTICLE

Multilingual Text Summarization in Healthcare Using Pre-Trained Transformer-Based Language Models

Josua Käser1, Thomas Nagy1, Patrick Stirnemann1, Thomas Hanne2,*

1 School of Business, University of Applied Sciences and Arts Northwestern Switzerland, Olten, 4600, Switzerland
2 Institute for Information Systems, University of Applied Sciences and Arts Northwestern Switzerland, Olten, 4600, Switzerland

* Corresponding Author: Thomas Hanne. Email: email

Computers, Materials & Continua 2025, 83(1), 201-217. https://doi.org/10.32604/cmc.2025.061527

Abstract

We analyze the suitability of existing pre-trained transformer-based language models (PLMs) for abstractive text summarization on German technical healthcare texts. The study focuses on the multilingual capabilities of these models and their ability to perform the task of abstractive text summarization in the healthcare field. The research hypothesis was that large language models could perform high-quality abstractive text summarization on German technical healthcare texts, even if the model is not specifically trained in that language. Through experiments, the research questions explore the performance of transformer language models in dealing with complex syntax constructs, the difference in performance between models trained in English and German, and the impact of translating the source text to English before conducting the summarization. We conducted an evaluation of four PLMs (GPT-3, a translation-based approach also utilizing GPT-3, a German language Model, and a domain-specific bio-medical model approach). The evaluation considered the informativeness using 3 types of metrics based on Recall-Oriented Understudy for Gisting Evaluation (ROUGE) and the quality of results which is manually evaluated considering 5 aspects. The results show that text summarization models could be used in the German healthcare domain and that domain-independent language models achieved the best results. The study proves that text summarization models can simplify the search for pre-existing German knowledge in various domains.

Keywords

Text summarization; pre-trained transformer-based language models; large language models; technical healthcare texts; natural language processing

Cite This Article

APA Style
Käser, J., Nagy, T., Stirnemann, P., Hanne, T. (2025). Multilingual text summarization in healthcare using pre-trained transformer-based language models. Computers, Materials & Continua, 83(1), 201–217. https://doi.org/10.32604/cmc.2025.061527
Vancouver Style
Käser J, Nagy T, Stirnemann P, Hanne T. Multilingual text summarization in healthcare using pre-trained transformer-based language models. Comput Mater Contin. 2025;83(1):201–217. https://doi.org/10.32604/cmc.2025.061527
IEEE Style
J. Käser, T. Nagy, P. Stirnemann, and T. Hanne, “Multilingual Text Summarization in Healthcare Using Pre-Trained Transformer-Based Language Models,” Comput. Mater. Contin., vol. 83, no. 1, pp. 201–217, 2025. https://doi.org/10.32604/cmc.2025.061527



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 154

    View

  • 82

    Download

  • 0

    Like

Share Link