首页 | 本学科首页   官方微博 | 高级检索  
     检索      


Arabic abstractive text summarization using RNN-based and transformer-based architectures
Institution:1. Syrian Virtual University, Damascus, Syria;2. ESC Rennes School of Business in Rennes, France;3. Burgundy School of Business in Dijon, France;1. National Authority for Remote Sensing and Space Sciences, Cairo, Egypt;2. Faculty of Computer Engineering, Al-Azhar University, Cairo, Egypt;1. Department of Management, Beijing Electronic Science & Technology Institute, Beijing, 100070, China;2. Department of Cybersecurity, Beijing Electronic Science & Technology Institute, Beijing, 100070, China;1. School of Computer Science and Technology, Guangdong University of Technology, China;2. Department of Computing, Hong Kong Polytechnic University, Hong Kong, China
Abstract:Recently, the Transformer model architecture and the pre-trained Transformer-based language models have shown impressive performance when used in solving both natural language understanding and text generation tasks. Nevertheless, there is little research done on using these models for text generation in Arabic. This research aims at leveraging and comparing the performance of different model architectures, including RNN-based and Transformer-based ones, and different pre-trained language models, including mBERT, AraBERT, AraGPT2, and AraT5 for Arabic abstractive summarization. We first built an Arabic summarization dataset of 84,764 high-quality text-summary pairs. To use mBERT and AraBERT in the context of text summarization, we employed a BERT2BERT-based encoder-decoder model where we initialized both the encoder and decoder with the respective model weights. The proposed models have been tested using ROUGE metrics and manual human evaluation. We also compared their performance on out-of-domain data. Our pre-trained Transformer-based models give a large improvement in performance with ~79% less data. We found that AraT5 scores ~3 ROUGE higher than a BERT2BERT-based model that is initialized with AraBERT, indicating that an encoder-decoder pre-trained Transformer is more suitable for summarizing Arabic text. Also, both of these two models perform better than AraGPT2 by a clear margin, which we found to produce summaries with high readability but with relatively lesser quality. On the other hand, we found that both AraT5 and AraGPT2 are better at summarizing out-of-domain text. We released our models and dataset publicly1,.2
Keywords:
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号