Natural Language Generation: Algorithms and Applications

Main Article Content

Diwakar R. Tripathi
Abha Tamrakar

Abstract

Natural Language Generation (NLG) is a subfield of artificial intelligence and computational linguistics that focuses on the automatic generation of natural language text. NLG has a wide range of applications in various fields, including content generation, virtual assistants, business intelligence, and healthcare. This paper provides an overview of NLG techniques and algorithms, including rule-based NLG, template-based NLG, statistical NLG, and neural NLG. It also explores the applications of NLG in different fields, highlighting its role in automated journalism, personalized content creation, virtual assistants, and data storytelling. Furthermore, the paper discusses the current challenges in NLG, such as naturalness, ambiguity handling, and scalability, and examines emerging trends and future directions in NLG, including advancements in neural NLG models, integration with other AI technologies, and ethical considerations. Overall, this paper aims to provide a comprehensive understanding of NLG and its impact on modern society.

Downloads

Download data is not yet available.

Metrics

Metrics Loading ...

Article Details

How to Cite
Tripathi, D. R. ., & Tamrakar, A. . (2018). Natural Language Generation: Algorithms and Applications. Turkish Journal of Computer and Mathematics Education (TURCOMAT), 9(3), 1394–1399. https://doi.org/10.61841/turcomat.v9i3.14597
Section
Research Articles

References

Arora, S., Li, Y., & Ma, T. (2016). A Simple but Tough-to-Beat Baseline for Sentence Embeddings. arXiv

preprint arXiv:1607.01759.

Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). On the Dangers of Stochastic

Parrots: Can Language Models Be Too Big? arXiv preprint arXiv:2105.14038.References:

Bordes, A., Boureau, Y.-L., & Weston, J. (2017). Learning End-to-End Goal-Oriented Dialog. arXiv

preprint arXiv:1605.07683.

Dathathri, R., Narang, S., Card, D., & Sridhar, V. (2020). DyNet: The Dynamic Neural Network Toolkit.

arXiv preprint arXiv:1701.03980.

Dongaonkar, N., Li, C., & Riedl, M. (2019). NewsArticleGenerator: Automatic News Generation with

Large-scale NLP Systems. arXiv preprint arXiv:1910.12596.

Gardent, C., Perez-Beltrachini, L., & Sales, J. (2017). Creating Training Corpora for NLG Micro-Planners.

Proceedings of the 10th International Conference on Natural Language Generation, 13–22.

Gkatzia, D., Hastie, H., Lemon, O., & Annibale, M. (2015). A Data-driven Approach to Predicting the

Success of Bank Telemarketing. Computational Linguistics, 41(4), 663–703.

Higashinaka, R., Imamura, K., & Aizawa, A. (2014). Evaluating Effectiveness of Various NLG Strategies

for Enhancing User Engagement in Human-robot Interaction. Proceedings of the 29th Pacific Asia

Conference on Language, Information and Computation, 60–69.

Huang, P.-S., Liu, J.-S., & Wang, C.-H. (2021). On the Integration of AI Technologies: A Systematic

Literature Review. Artificial Intelligence Review, 54(5), 3463–3487.

Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., Levy, O., ... & Zettlemoyer, L. (2020).

BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and

Comprehension. arXiv preprint arXiv:1910.13461.

Liu, Q., Ren, X., Gao, J., Howard, P., & Wang, W. (2019). RoBERTa: A Robustly Optimized BERT

Pretraining Approach. arXiv preprint arXiv:1907.11692.

Novikova, J., Dusaban, E., Rieser, V., & Lemon, O. (2017). Why We Need New Evaluation Metrics for

NLG. Proceedings of the 10th International Conference on Natural Language Generation, 203–207.

Serban, I. V., Sordoni, A., Bengio, Y., Courville, A., & Pineau, J. (2015). Building End-To-End Dialogue

Systems Using Generative Hierarchical Neural Network Models. Proceedings of the Thirtieth AAAI

Conference on Artificial Intelligence, 3776–3784.

Zhang, J., Liu, Y., & Luan, H. (2018). Extractive Summarization: Challenges, Methods, and Applications.

IEEE Transactions on Neural Networks and Learning Systems, 29(12), 5614–5632.

Swartout, W., Artstein, R., Forbell, E., Foutz, S., Lane, H. C., Lange, B., ... & Traum, D. (2017). Virtual

Human Standardized Patients for Clinical Training. ACM Transactions on Interactive Intelligent Systems

(TiiS), 7(1), 1–38.

Kreuzthaler, M., Schulz, S., & Berghold, A. (2018). Analyzing the Natural Language Generation Process

in Radiology Reports. Journal of Biomedical Informatics, 87, 58–67.

Arnold, C. W., McNamara, D. S., Duran, N. D., & Chennupati, S. (2016). Automated Detection of Student

Mental Models During Computer-Based Problem Solving. International Journal of Artificial Intelligence

in Education, 26(1), 301–326.

Zhou, L., Zhang, D., & Sun, L. (2017). Information Technology-based Diabetes Management Interventions:

A Systematic Review. Journal of Diabetes Science and Technology, 11(1), 116–127.References:

Gatt, A., Belz, A., & Kow, E. (2009). The TUNA-REG Corpus: A Corpus for Evaluating Surface

Realisation by Statistical NLG Systems. Proceedings of the 7th International Conference on Language

Resources and Evaluation (LREC 2008), 69–76.

Belz, A., & Reiter, E. (2006). Comparing Automatic and Human Evaluation of Realisation Quality for NLG

Systems. Proceedings of the 11th European Workshop on Natural Language Generation, 129–136.

Pasha, A., Al-Badrashiny, M., Diab, M., El Kholy, A., Eskander, R., Habash, N., ... & Rambow, O. (2014).

Madamira: A Fast, Comprehensive Tool for Morphological Analysis and Disambiguation of Arabic.

Proceedings of the Language Resources and Evaluation Conference (LREC).

Koehn, P., Och, F. J., & Marcu, D. (2003). Statistical Phrase-based Translation. Proceedings of the 2003

Conference of the North American Chapter of the Association for Computational Linguistics on Human

Language Technology - Volume 1, 48–54.

Nenkova, A., & McKeown, K. (2011). Automatic Summarization. Foundations and Trends® in Information

Retrieval, 5(2-3), 103–233.

Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language Models are

Unsupervised Multitask Learners. OpenAI Blog.

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. (2017).

Attention is All you Need. Advances in Neural Information Processing Systems, 30, 5998–6008.

Winograd, T. (1972). Understanding Natural Language. Cognitive Psychology, 3(1), 1–191.

Langkilde, I., & Knight, K. (1998). Generation that Exploits Corpus-based Statistical Knowledge.

Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th

International Conference on Computational Linguistics - Volume 1, 704–710.

Lahiri, S., & Reddy, C. (2011). Natural Language Generation in Narrative Science’s Quill Platform. AI

Magazine, 32(3), 61–76.

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., … Polosukhin, I. (2017).

Attention is All you Need. Advances in Neural Information Processing Systems, 30, 5998–6008.

This section provides an overview of the early developments, key milestones, and the evolution of NLG

techniques and algorithms, supported by citations from relevant research papers published between 2012

and 2018.

Overall, the versatility and applicability of NLG across various domains underscore its significance in

automating content generation, improving communication, and enhancing user experiences.

Liu, Q., Zhao, H., & Jansche, M. (2018). Data-to-text Generation with Content Selection and Planning.

Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2712–2722.

Du, Y., Xu, Z., Tao, C., & Xu, H. (2016). Natural Language Generation in Health Care. Artificial

Intelligence in Medicine, 69, 1–8.

Gkatzia, D., Hastie, H., Lemon, O., & Annibale, M. (2015). A Data-driven Approach to Predicting the

Success of Bank Telemarketing. Computational Linguistics, 41(4), 663–703.

Wen, T. H., Vandyke, D., Mrksic, N., Gasic, M., Rojas-Barahona, L. M., Su, P.-H., & Young, S. (2015).

Semantically Conditioned LSTM-based Natural Language Generation for Spoken Dialogue Systems.

Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 1711–1721.