This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.
This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.
Hierarchical Bidirectional Transformers for Text Summarization
FOROOZANDE NEJAD, PARISA
2024/2025
Abstract
This thesis presents a novel approach to unsupervised extractive text summarization using hierarchical transformer architectures. With the exponential growth of digital content, the need for efficient text summarization methods has become increasingly critical. Our research addresses this challenge by combining the hierarchical bidirectional transformer architecture of HIBERT with the unsupervised ranking criteria of STAS, enhanced by a Pointwise Mutual Information (PMI)-based redundancy control mechanism. The proposed method employs a two-level processing structure that captures both local sentence semantics and global document structure. Unlike traditional approaches that rely on surface-level features, our hierarchical transformer architecture enables effective sentence-level attention mechanisms for ranking sentences in unsupervised extractive summarization. The model processes documents by generating contextual sentence representations through HIBERT's pre-trained hierarchical encoder, then applies STAS ranking criteria combined with PMI-based redundancy measures to select the most salient sentences. We evaluate our approach on the CNN/DailyMail dataset, a standard benchmark for summarization tasks. The experimental results demonstrate that our method achieves highly competitive performance, with significant improvements attributed to the redundancy control mechanism.| File | Dimensione | Formato | |
|---|---|---|---|
|
Foroozande-Nejad.pdf
accesso aperto
Dimensione
944.24 kB
Formato
Adobe PDF
|
944.24 kB | Adobe PDF | Visualizza/Apri |
The text of this website © Università degli studi di Padova. Full Text are published under a non-exclusive license. Metadata are under a CC0 License
https://hdl.handle.net/20.500.12608/91828