Digitala Vetenskapliga Arkivet

Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Impact of Language-Model Size on the Quality of Automatic Text Summarization Comparing Larger Models against Smaller Open-Source Alternatives
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.
2024 (English)Independent thesis Advanced level (degree of Master (One Year)), 10 credits / 15 HE creditsStudent thesis
Abstract [en]

Automatic text summarization condenses large amounts of text into concise representations, making information retrieval and comprehension more efficient. This research investigates the impact of language-model size on the quality of automatic summarization. We compare the performance of large models against smaller open-source alternatives.

Background research highlights the increasing prevalence of large language models and the significant role they play in Natural Language Processing (NLP) tasks. However, the relationship between model size and summarization quality remains an open question, this research gap is addressed in our study by comparing the summarization quality of large and small size models, understanding the trade-offs between model size, cost and performance along with answering the practical implication of using smaller models interms of cost, deployment and data privacy.

Our methodology employs an experimental design together with a correlational design as an alternative research design methodology. We utilize a benchmark dataset encompassing diverse news text types to ensure generalizability. 10 large and small LMs are evaluated on CNN-Daily Mail Dataset, using Recall-Oriented Understudy for Gisting Evaluation (ROUGE) scores (ROUGE-1, ROUGE-2, ROUGE-L) to further showcase the quality of summarization results at model level sample generated summaries from 8 models are presented. In the analysis factors beyond ROUGE scores, such as fluency, coherence, and factuality are discussed. The findings of this study with it’s limitations underlines that using small open-source models for abstractive news summarization offers advantages in terms of cost, maintenance, deployment, and potential data privacy. However, it's important to weigh these benefits against the potential trade-off in summarization quality and functionality.

Place, publisher, year, edition, pages
2024.
Keywords [en]
Abstractive Text Summarization, Automatic Text Summarization, Large Language Models, Model Size, Summarization Quality, ROUGE
National Category
Other Computer and Information Science
Identifiers
URN: urn:nbn:se:su:diva-242747OAI: oai:DiVA.org:su-242747DiVA, id: diva2:1955679
Available from: 2025-04-30 Created: 2025-04-30

Open Access in DiVA

fulltext(1155 kB)12 downloads
File information
File name FULLTEXT01.pdfFile size 1155 kBChecksum SHA-512
cb354198e15883d9f8c8e8418a4041b204a4b9e7f2cdfc98bad17c85b370cea8d7c5295993ca4d6d090c6e522af981d9e865d78c0b47637515cf2f784effda9e
Type fulltextMimetype application/pdf

Search in DiVA

By author/editor
Gebreyesus, Yasmin Tesfaldet
By organisation
Department of Computer and Systems Sciences
Other Computer and Information Science

Search outside of DiVA

GoogleGoogle Scholar
Total: 12 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 13 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf