Efficient adaptation of Large Language Models for digital media and government applications
dc.contributor.advisor | Minghim, Rosane | |
dc.contributor.advisor | Zahran, Ahmed | |
dc.contributor.author | Trust, Paul | |
dc.contributor.funder | Science Foundation Ireland | en |
dc.date.accessioned | 2025-10-01T11:45:28Z | |
dc.date.available | 2025-10-01T11:45:28Z | |
dc.date.issued | 2024 | |
dc.date.submitted | 2024 | |
dc.description.abstract | The digital transformation has greatly increased the amount of data, particularly text generated across various fields, including the public sector and digital media. Information such as political campaigns, media reports, citizen feedback, and press releases is now commonly shared on digital platforms. Politicians, government agencies, businesses, and citizens use these platforms to express their views, strategies, goals, and policies. Analyzing this data can provide valuable insights into public opinion, ongoing policy discussions, business strategies, and socio-political dynamics. However, the sheer volume of data makes traditional manual analysis or early computational methods impractical, highlighting the need for more efficient automated approaches from Natural Language Processing (NLP), particularly Large Language Models (LLMs), to manage, analyze, and summarize this information. In the course of this work, document-based learning has seen a landmark advance with the progress of generative Artificial Intelligence and the availability of engines and models that are revolutionizing NLP. Despite significant adoption and investment in the private sector, academia, and high-resource fields, LLMs are less utilized in low-resource fields and the public sector due to constraints such as lack of labeled data, and insufficient budget allocation for machine learning infrastructure and training. In this Thesis, which was developed during a transition time of fast development of LLMs, we study ways of adapting the most up to date models to novel scenarios in order to both achieve efficiency and understanding of how to adapt LLMs to applications. To achieve this, this Thesis adapts LLMs to applications in the public sector and digital media. The main approaches developed include: applying weak supervision by leveraging synthetic labels generated by other LLMs to fine-tune models for classifying news articles related to Economic Policy Uncertainty; proposing LLM-based methods for classifying citizen feedback into different categories, as well as for summarization and question answering of citizen feedback; adapting LLMs for automating the handling and navigation of public documents by incorporating strategies such as Retrieval Augmented Generation (RAG), LLM agents; and developing techniques for hallucination detection in these domains. | en |
dc.description.status | Not peer reviewed | en |
dc.description.version | Accepted Version | en |
dc.format.mimetype | application/pdf | en |
dc.identifier.citation | Trust, P. 2024. Efficient adaptation of Large Language Models for digital media and government applications. PhD Thesis, University College Cork. | |
dc.identifier.endpage | 160 | |
dc.identifier.uri | https://hdl.handle.net/10468/17932 | |
dc.language.iso | en | en |
dc.publisher | University College Cork | en |
dc.relation.project | info:eu-repo/grantAgreement/SFI/NSF Student Mobility Programme/18/CRT/6222 (S5)/IE/18/CRT/6222 Supplement/ | |
dc.rights | © 2024, Paul Trust. | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.subject | Large Language Models | en |
dc.subject | Generative AI | en |
dc.subject | Citzen feedback | en |
dc.subject | RAG | en |
dc.title | Efficient adaptation of Large Language Models for digital media and government applications | |
dc.type | Doctoral thesis | en |
dc.type.qualificationlevel | Doctoral | en |
dc.type.qualificationname | PhD - Doctor of Philosophy | en |
Files
Original bundle
1 - 2 of 2
Loading...
- Name:
- TrustP_ PhD2024.pdf
- Size:
- 2.11 MB
- Format:
- Adobe Portable Document Format
- Description:
- Full Text E-thesis
Loading...
- Name:
- 3. 120222601 - Paul Trust - Submission for examination form.pdf
- Size:
- 793.46 KB
- Format:
- Adobe Portable Document Format
License bundle
1 - 1 of 1
Loading...
- Name:
- license.txt
- Size:
- 5.2 KB
- Format:
- Item-specific license agreed upon to submission
- Description: