LSPGS Wikipedia Long Sentences Summarization Task
A LSPGS Wikipedia Long Sentences Summarization Task is a Multi-Document Text Summarization Task that is based on a Transformer-based Neural Network architecture that generates Wikipedia articles.
- AKA: Text Summarization via Transformer Neural Network, Liu-Saleh-Pot-Goodrich-Sepassi Wikipedia Long Sentences Summarization Task, Liu's Wikipedia Long Sentences Summarization Task.
- Context:
- Task Input(s): Wikipedia Topic.
- Task Output(s): Wikipedia generated article.
- Task Requirement(s):
- Benchmark Datasets:
- Performance Metrics:
- Baseline Models:
- It can be solved LSPGS Wikipedia Long Sentences Summarization System that implements LSPGS Wikipedia Long Sentences Summarization Algorithms.
- Example(s):
- Extractive methods comparison (Liu et al., 2018):
- Models Performance (Liu et al., 2018):
{{class="wikitable" style="border:1px; text-align:left; solid black; border-spacing:1px; margin: 1em auto; width: 80%"
- Models Performance (Liu et al., 2018):
|- ! Model !!Test perplexity!! ROUGE-L. |- |seq2seq-attention, $L = 500 || 5.04952|| 12.7 |- |Transformer-ED, $L = 500 || 2.46645|| 34.2 |- |Transformer-D, $L = 4000$|| 2.22216 ||33.6 |- | Transformer-DMCA, no MoE-layer, $L = 11000$ ||2.05159 ||36.2 |- |Transformer-DMCA, MoE-128, $L = 11000 || 1.92871|| 37.9 |- |Transformer-DMCA, MoE-256, $L = 7500$|| 1.90325|| 38.8 |- |}
- Counter-Example(s):
- See: Abstractive Text Summarization Task, Neural Abstractive Summarization Task, Self-Attention Mechanism, Sequence-to-Sequence (seq2seq) Neural Network, Natural Language Generation Task, Natural Language Understanding Task, Natural Language Translation Task, Natural Language Processing Task.
References
2018
- (Liu et al., 2018) ⇒ Peter J. Liu, Mohammad Saleh, Etienne Pot, Ben Goodrich, Ryan Sepassi, Lukasz Kaiser, and Noam Shazeer. (2018). “Generating Wikipedia by Summarizing Long Sequences.” In: Proceedings of the Sixth International Conference on Learning Representations (ICLR-2018).
2017
- (Vaswani et al., 2017) ⇒ Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. (2017). “Attention is all You Need.” In: Advances in Neural Information Processing Systems.
2016a
- (Nallapati et al., 2016) ⇒ Ramesh Nallapati, Bowen Zhou, Cicero Nogueira dos Santos, Caglar Gulcehre, and Bing Xiang. (2016). “Abstractive Text Summarization Using Sequence-to-sequence RNNs and Beyond.” In: Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning (CoNLL 2016). DOI:10.18653/v1/K16-1028.
2016b
- (Wu et al., 2016) ⇒ Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, and 27 co-authors. (2016). "Google's Neural Machine Translation System: Bridging the Gap Between Human and Machine Translation". In: arXiv:1609.08144.
2015
- (Bahdanau et al., 2015) ⇒ Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. (2015). “Neural Machine Translation by Jointly Learning to Align and Translate.” In: Proceedings of the Third International Conference on Learning Representations, (ICLR-2015).
2005
- (Nenkova & Vanderwende, 2005) ⇒ Ani Nenkova, and Lucy Vanderwende (2005). “The Impact of Frequency on Summarization". Microsoft Research, Redmond, Washington, Tech. Rep. MSR-TR-2005, 101.
2004
- (Mihalcea & Tarau, 2004) ⇒ Rada Mihalcea, and Paul Tarau. (2004). “TextRank: Bringing Order into Texts.” In: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (EMNLP 2004)
2003a
- (Graff & Cieri, 2003) ⇒ David Graff, and Christopher Cieri (2003). "English Gigaword". Linguistic Data Consortium, Philadelphia, 2003.
- QUOTE: English Gigaword was produced by Linguistic Data Consortium (LDC) catalog number LDC2003T05 and ISBN 1-58563-260-0, and is distributed on DVD. This is a comprehensive archive of newswire text data in English that has been acquired over several years by the LDC.
Four distinct international sources of English newswire are represented here:
- QUOTE: English Gigaword was produced by Linguistic Data Consortium (LDC) catalog number LDC2003T05 and ISBN 1-58563-260-0, and is distributed on DVD. This is a comprehensive archive of newswire text data in English that has been acquired over several years by the LDC.
2003b
- (Ramos, 2003) ⇒ Juan Ramos (2003). “Using TF-IDF to Determine Word Relevance in Document Queries". In: Proceedings of the first instructional conference on Machine Learning, volume 242, pp. 133–142.