Vui lòng dùng định danh này để trích dẫn hoặc liên kết đến tài liệu này:
https://elib.vku.udn.vn/handle/123456789/4019
Nhan đề: | THASUM: Transformer for High-Performance Abstractive Summarizing Vietnamese Large-scale Dataset |
Tác giả: | Nguyen, Ti Hon Do, Thanh Nghi |
Từ khoá: | Abstractive Text Summarization Transformer Vietnamese Large-scale Dataset |
Năm xuất bản: | thá-2024 |
Nhà xuất bản: | Vietnam-Korea University of Information and Communication Technology |
Tùng thư/Số báo cáo: | CITA; |
Tóm tắt: | Our investigation aims to propose a high-performance abstractive text summarization model for Vietnamese languages. We based on the transformer network with a full encoder-decoder to study the high-quality features of the training data. Next, we scaled down the network size to increase the number of documents the model can summarize in a time frame. We trained the model with a large-scale dataset, including 880,895 documents in the training set and 110, 103 in the testing set. The summarizing speed for the testing set significantly improves with 5.93 hours when using a multiple-core CPU and 0.31 hours on a small GPU. The numerical test results of F1 are also close to the state-of-the-art with 51.03% in ROUGE-1, 18.17% in ROUGE-2, and 31.60% in ROUGE-L. |
Mô tả: | Proceedings of the 13th International Conference on Information Technology and Its Applications (CITA 2024); pp: 100-111. |
Định danh: | https://elib.vku.udn.vn/handle/123456789/4019 |
ISBN: | 978-604-80-9774-5 |
Bộ sưu tập: | CITA 2024 (Proceeding - Vol 2) |
Khi sử dụng các tài liệu trong Thư viện số phải tuân thủ Luật bản quyền.