This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.

This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.

Hierarchical Bidirectional Transformers for Text Summarization

FOROOZANDE NEJAD, PARISA
2024/2025

Abstract

This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.
2024
Hierarchical Bidirectional Transformers for Text Summarization
This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.
LLMs
NLP
Transformers
Text Summarization
Deep Learning
File in questo prodotto:
File Dimensione Formato  
Foroozande-Nejad.pdf

accesso aperto

Dimensione 944.24 kB
Formato Adobe PDF
944.24 kB Adobe PDF Visualizza/Apri

The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12608/91828