Skip to main content
Top

2024 | OriginalPaper | Chapter

Transformer Models in Natural Language Processing

Authors : László Kovács, László Csépányi-Fürjes, Walelign Tewabe

Published in: The 17th International Conference Interdisciplinarity in Engineering

Publisher: Springer Nature Switzerland

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

The development of transformer-based language models brings a paradigm shift in the world of smart applications. The ChatGPT model opened new horizons in the field of natural language understanding and generation. This paper presents a survey on the history of transformer models, on the basic architecture and application areas. The last section is devoted to two use cases experiments on the application of ChatGPT. The first domain relates to Human-Level Programming and the second focuses on the semantic functional parsing of text sentences. The performed analysis demonstrates the big potential in the transformer language models.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
3.
go back to reference Heyman, G., Huysegems, R., Justen, P., van Cutsem, T.: Natural language-guided programming. In: Onward! 2021 - Proceedings of the 2021 ACM SIGPLAN International Symposium on New Ideas, New Paradigms, and Reflections on Programming and Software, Co-Located with SPLASH 2021, pp. 39–55 (2021). https://doi.org/10.1145/3486607.3486749 Heyman, G., Huysegems, R., Justen, P., van Cutsem, T.: Natural language-guided programming. In: Onward! 2021 - Proceedings of the 2021 ACM SIGPLAN International Symposium on New Ideas, New Paradigms, and Reflections on Programming and Software, Co-Located with SPLASH 2021, pp. 39–55 (2021). https://​doi.​org/​10.​1145/​3486607.​3486749
4.
go back to reference Kovács, L.: Elemző és robotizált folyamatautomatizálási rendszer fejlesztése nagy terhelésű ügyfélszolgálatok számára: Kutatási jelentések 2022/1 Miskolc-Egyetemváros, Magyarország: Miskolci Egyetemi Kiadó, p. 167 (2022). ISBN: 9789633582626 Kovács, L.: Elemző és robotizált folyamatautomatizálási rendszer fejlesztése nagy terhelésű ügyfélszolgálatok számára: Kutatási jelentések 2022/1 Miskolc-Egyetemváros, Magyarország: Miskolci Egyetemi Kiadó, p. 167 (2022). ISBN: 9789633582626
8.
go back to reference Haque, S., Eberhart, Z., Bansal, A., McMillan, C.: Semantic similarity metrics for evaluating source code summarization. In: IEEE International Conference on Program Comprehension, IEEE Computer Society, pp. 36–47 (2022) Haque, S., Eberhart, Z., Bansal, A., McMillan, C.: Semantic similarity metrics for evaluating source code summarization. In: IEEE International Conference on Program Comprehension, IEEE Computer Society, pp. 36–47 (2022)
12.
go back to reference Pascanu, R., Mikolov, T., Bengio, Y.: On the difficulty of training recurrent neural networks. In: International Conference on Machine Learning. PMLR (2013) Pascanu, R., Mikolov, T., Bengio, Y.: On the difficulty of training recurrent neural networks. In: International Conference on Machine Learning. PMLR (2013)
13.
go back to reference Zhou, C., et al.: A comprehensive survey on pretrained foundation models: A history from BERT to ChatGPT (2023). arXiv preprint arXiv:2302.09419. x0x0 Zhou, C., et al.: A comprehensive survey on pretrained foundation models: A history from BERT to ChatGPT (2023). arXiv preprint arXiv:​2302.​09419. x0x0
14.
go back to reference Liu, Y., et al.: RoBERTa: A robustly optimized BERT pretraining approach (2019). CoRR Liu, Y., et al.: RoBERTa: A robustly optimized BERT pretraining approach (2019). CoRR
15.
go back to reference Li, J., et al.: Pretrained language models for text generation: A survey (2022). arXiv preprint arXiv: 2201.05273 Li, J., et al.: Pretrained language models for text generation: A survey (2022). arXiv preprint arXiv: 2201.05273
17.
go back to reference Lewis, M., et al.: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension (2019). arXiv preprint arXiv: 1910.13461 Lewis, M., et al.: BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension (2019). arXiv preprint arXiv: 1910.13461
18.
go back to reference Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020) Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)
19.
go back to reference Biswas, S.S.: Role of chat GPT in public health. Ann. Biomed. Eng. 51(5), 868–869 (2023)CrossRef Biswas, S.S.: Role of chat GPT in public health. Ann. Biomed. Eng. 51(5), 868–869 (2023)CrossRef
20.
go back to reference Li, Jianning, et al. “ChatGPT in Healthcare: A Taxonomy and Systematic Review.” medRxiv (2023): 2023–03 Li, Jianning, et al. “ChatGPT in Healthcare: A Taxonomy and Systematic Review.” medRxiv (2023): 2023–03
21.
go back to reference Biswas, S.: ChatGPT and the future of medical writing. Radiology 307(2), e223312 (2023)CrossRef Biswas, S.: ChatGPT and the future of medical writing. Radiology 307(2), e223312 (2023)CrossRef
22.
go back to reference Wang, C., et al.: Ethical Considerations of Using ChatGPT in Health Care. J. Med. Internet Res. 25, e48009 (2023) Wang, C., et al.: Ethical Considerations of Using ChatGPT in Health Care. J. Med. Internet Res. 25, e48009 (2023)
23.
go back to reference Hill-Yardin, E.L., et al.: A Chat (GPT) about the future of scientific publishing. Brain Behav. Immun. 110, 152–154 (2023) Hill-Yardin, E.L., et al.: A Chat (GPT) about the future of scientific publishing. Brain Behav. Immun. 110, 152–154 (2023)
24.
go back to reference Rahimi, F., Abadi, A.T.B.: ChatGPT and publication ethics. Arch. Med. Res. 54(3), 272–274 (2023)CrossRef Rahimi, F., Abadi, A.T.B.: ChatGPT and publication ethics. Arch. Med. Res. 54(3), 272–274 (2023)CrossRef
25.
go back to reference Zhu, J.-J., et al.: ChatGPT and environmental research. Environ. Sci. Technol. 57(46), 17667–17670 (2023) Zhu, J.-J., et al.: ChatGPT and environmental research. Environ. Sci. Technol. 57(46), 17667–17670 (2023)
26.
go back to reference Surameery, N.M.S., Shakor, M.Y.: Use chat GPT to solve programming bugs. Int. J. Inf. Technol. Comput. Eng. (IJITC) 3(01), 17–22 (2023). ISSN: 2455–5290 Surameery, N.M.S., Shakor, M.Y.: Use chat GPT to solve programming bugs. Int. J. Inf. Technol. Comput. Eng. (IJITC) 3(01), 17–22 (2023). ISSN: 2455–5290
27.
go back to reference Lo, C.K.: What is the impact of ChatGPT on education? A rapid review of the literature. Educ. Sci. 13(4), 410 (2023)CrossRef Lo, C.K.: What is the impact of ChatGPT on education? A rapid review of the literature. Educ. Sci. 13(4), 410 (2023)CrossRef
Metadata
Title
Transformer Models in Natural Language Processing
Authors
László Kovács
László Csépányi-Fürjes
Walelign Tewabe
Copyright Year
2024
DOI
https://doi.org/10.1007/978-3-031-54674-7_14

Premium Partners