Transforming Text Summarization with Deep Neural Networks

Authors

  • Chadha Chawla India Author

Keywords:

deep learning algorithms, convolutional neural networks , long short-term memory

Abstract

The advent of deep neural networks has revolutionized the field of text summarization, offering unprecedented capabilities for extracting meaningful summaries from large text corpora. This paper explores the transformation of text summarization processes through the application of advanced deep learning algorithms. We investigate various neural network architectures, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), and more specifically, long short-term memory (LSTM) networks, to assess their efficacy in generating coherent and contextually relevant summaries. By analyzing the performance metrics across different datasets, we demonstrate the superiority of deep learning methods over traditional approaches. Our findings suggest that deep neural networks not only enhance the quality of text summaries but also provide robust mechanisms for handling diverse and complex linguistic structures. This study lays the groundwork for further research into optimizing neural network parameters and architectures to achieve even more effective summarization outcomes.

References

Bengio, Y., Ducharme, R., Vincent, P., & Jauvin, C. (2003). A neural probabilistic language model. Journal of Machine Learning Research, 3, 1137-1155.

Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., & Kuksa, P. (2011). Natural language processing (almost) from scratch. Journal of Machine Learning Research, 12, 2493-2537.

Graves, A., & Schmidhuber, J. (2005). Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Networks, 18(5-6), 602-610.

Hinton, G. E., Osindero, S., & Teh, Y. W. (2006). A fast learning algorithm for deep belief nets. Neural Computation, 18(7), 1527-1554.

Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735-1780.

LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278-2324.

Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.

Nallapati, R., Zhai, F., & Zhou, B. (2012). Summarunner: A recurrent neural network based sequence model for extractive summarization of documents. arXiv preprint arXiv:1602.03644.

Salakhutdinov, R., & Hinton, G. (2009). Deep Boltzmann machines. Proceedings of the International Conference on Artificial Intelligence and Statistics, 448-455.

Socher, R., Huval, B., Manning, C. D., & Ng, A. Y. (2012). Semantic compositionality through recursive matrix-vector spaces. Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 1201-1211.

Published

2014-05-12

How to Cite

Transforming Text Summarization with Deep Neural Networks. (2014). JOURNAL OF RECENT TRENDS IN COMPUTER SCIENCE AND ENGINEERING ( JRTCSE), 2(1), 32-41. https://jrtcse.com/index.php/home/article/view/JRTCSE.2014.1.4