Arabic abstractive text summarization via supervised transfer and deep reinforcement learning
Loading...
Date
2024
Journal Title
Journal ISSN
Volume Title
Publisher
Faculty of sciences
Abstract
The rapid growth of digital content has created a pressing need for efficient text summarization techniques, particularly for languages with complex features like Arabic. Although progress in Arabic abstractive text summarization has been limited
compared to languages like English, recent advancements leveraging transfer
learning with pretrained transformer-based language models (PTLMs) have shown promise. These approaches have primarily relied on supervised learning techniques for finetuning, which typically focus on maximizing next-word prediction like lihood rather than summary quality. They also suffer from in consistencies between
training and inference conditions, as well as exposure bias. Notably, reinforcement learning (RL), which offers potential solutions to thes issues, especially its ability to directly optimize non-differentiable objectives, remains largely unexplored in Arabic ATS. This thesis proposes a novel approach combining transfer learning and RL to address these limitations and advance Arabic ATS. We introduce a novel two- phase framework: (1) supervised finetuning (SFT) of a PTLM, followed by (2) RL-based
optimization via Proximal Policy Optimization. Our approach is evaluated on the XL-Sum and AraSum datasets, using reward functions derived from different automatic evaluation metrics and textual entailment. Experimental results demonstrate that our RL-finetuned models out perform our supervised baseline across multiple metrics, while demonstrating enhanced factual consistency. Comparison with prior work shows that our models achieve new state-of-the-art performance in terms of BERT Score, while achieving competitive and more balanced ROUGE scores on both datasets.
Moreover, a test-only transfer evaluation on a new dataset reveals that our
RL-optimized models exhibit superior generalization capabilities compared to the supervised baseline. We also conduct ablation studies to analyze the
contributions of some critical components in our approach