Evolution of Natural Language Understanding Using Transformer-Based AI Models
Keywords:
Natural Language Understanding, Transformer Models, Language Models, AI Evolution, NLP Architectures, Foundation Models, Attention MechanismSynopsis
References
(1) Vaswani, Ashish, et al. “Attention Is All You Need.” Advances in Neural Information Processing Systems, vol. 30, 2017, pp. 5998–6008.
(2) Devlin, Jacob, et al. “BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding.” Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-HLT), 2019, pp. 4171–4186.
(3) Brown, Tom B., et al. “Language Models Are Few-Shot Learners.” Advances in Neural Information Processing Systems, vol. 33, 2020, pp. 1877–1901.
(4) Raffel, Colin, et al. “Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.” Journal of Machine Learning Research, vol. 21, no. 140, 2020, pp. 1–67.
(5) Sirimalla, A. (2022). End-to-end automation for cross-database DevOps deployments: CI/CD pipelines, schema drift detection, and performance regression testing in the cloud. World Journal of Advanced Research and Reviews, 14(3), 871–889. https://doi.org/10.30574/wjarr.2022.14.3.0555
(6) Liu, Yinhan, et al. “RoBERTa: A Robustly Optimized BERT Pretraining Approach.” arXiv, arXiv:1907.11692.
(7) Yang, Zhilin, et al. “XLNet: Generalized Autoregressive Pretraining for Language Understanding.” Advances in Neural Information Processing Systems, vol. 32, 2019, pp. 5753–5763.
(8) Lan, Zhenzhong, et al. “ALBERT: A Lite BERT for Self-Supervised Learning of Language Representations.” Proceedings of the International Conference on Learning Representations, 2020.
(9) Peters, Matthew E., et al. “Deep Contextualized Word Representations.” Proceedings of NAACL-HLT, 2018, pp. 2227–2237.
(10) Radford, Alec, et al. “Improving Language Understanding by Generative Pre-Training.” OpenAI Technical Report, 2018.
(11) Radford, Alec, et al. “Language Models Are Unsupervised Multitask Learners.” OpenAI Technical Report, 2019.
(12) Sirimalla A. Autonomous Performance Tuning Framework for Databases Using Python and Machine Learning. J Artif Intell Mach Learn & Data Sci 2023 1(4), 3139-3147. DOI: doi.org/10.51219/JAIMLD/adithya-sirimalla/642
(13) Sanh, Victor, et al. “DistilBERT, a Distilled Version of BERT: Smaller, Faster, Cheaper and Lighter.” Proceedings of the NeurIPS EMC² Workshop, 2019.
(14) Hendrycks, Dan, et al. “Measuring Massive Multitask Language Understanding.” Proceedings of the International Conference on Learning Representations, 2021.
Published
Series
License

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.