Abstract
Information accessibility has been transformed in the field of Artificial Intelligence (AI), particularly in natural language processing (NLP), owing to the widespread use of technologies such as ChatGPT. This paper explores the field of human-directed AI solutions with particular emphasis on newspaper summarization, a useful tool in today's busy world. Utilizing comprehensive models (LLMs), we explore extractive and abstractive summarization methods. To maximize the LLM performance, our strategy entails creating a customized news dataset enhanced with human-centric summaries and using cutting-edge data preparation techniques. To improve the accuracy assessment, we present a modified evaluation metric called Sem-rouge, which augments established units of measurement. In a comparative analysis, it was noticed that the proposed metric can highlight both syntactic and semantic similarities; hence, the metric is suitable for both extractive and abstractive summarization methods. We highlight the significance of dataset selection, data processing methods, and assessment criteria in fine-tuning auto-generated summaries using rigorous comparison analysis. Further studies will focus on improving semantic similarity techniques, integrating advanced models such as The BERT algorithm or Generative Pre-trained Transformer algorithm, and overcoming challenges such as overfitting. Finally, our study emphasizes the importance of meticulously training models and modifying them frequently to enhance automated summarization skills.
Keywords
Get full access to this article
View all access options for this article.
