Skip to main content

2024 | OriginalPaper | Buchkapitel

Analyzing the Impact of Carbon Emission in Training Neural Machine Translation Models: A Case Study

verfasst von : Goutam Datta, Nisheeth Joshi, Kusum Gupta

Erschienen in: ICT: Innovation and Computing

Verlag: Springer Nature Singapore

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

As the field of machine learning grows rapidly, a lot of attention has been paid to how training complex models affects the environment. Carbon emissions caused by the computing needs of machine learning algorithms are becoming a big concern. This is because these models need a lot of computing power and energy. The goal of this paper is to find out how training Neural Machine Translation models affects the environment in terms of carbon footprint and to look into ways to reduce that effect. Machine translation, which automatically translates from source to target, is an area of natural language processing where researchers have been actively working for a long time. The performance of Neural Machine Translation (NMT) is enhanced by exploiting Artificial Neural Networks (ANN) in its model implementation. However, NMT is highly data-hungry and it requires longer training time. In this paper, an attempt has been made to estimate carbon emission when the different NMT models are trained in low-resource language pairs such as English to Hindi and English to Bengali language pairs on different hardware configurations. Finally, different alternatives have also been suggested to reduce this carbon emission and thereby its adverse impact on the environment can be minimized.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Koehn P, Zens R, Dyer C et al (2007) Moses: open source toolkit for statistical machine translation. In: Proceedings of the 45th annual meeting of the ACL on interactive poster and demonstration sessions—ACL’07, June 2007, p 177. https://doi.org/10.3115/1557769.1557821 Koehn P, Zens R, Dyer C et al (2007) Moses: open source toolkit for statistical machine translation. In: Proceedings of the 45th annual meeting of the ACL on interactive poster and demonstration sessions—ACL’07, June 2007, p 177. https://​doi.​org/​10.​3115/​1557769.​1557821
6.
Zurück zum Zitat Raffel C, Shazeer N, Roberts A et al (2020) Exploring the limits of transfer learning with a unified text-to-text transformer. J Mach Learn Res 21:1–67MathSciNet Raffel C, Shazeer N, Roberts A et al (2020) Exploring the limits of transfer learning with a unified text-to-text transformer. J Mach Learn Res 21:1–67MathSciNet
8.
Zurück zum Zitat Henderson P, Romoff J, Brunskill E (2020) Footprints of machine learning 21:1–44 Henderson P, Romoff J, Brunskill E (2020) Footprints of machine learning 21:1–44
9.
Zurück zum Zitat Strubell E, Ganesh A, McCallum A (2020) Energy and policy considerations for deep learning in NLP. In: ACL 2019—Proceedings of the conference on 57th annual meeting of the Association for Computational Linguistics, vol 1, pp 3645–3650 Strubell E, Ganesh A, McCallum A (2020) Energy and policy considerations for deep learning in NLP. In: ACL 2019—Proceedings of the conference on 57th annual meeting of the Association for Computational Linguistics, vol 1, pp 3645–3650
10.
Zurück zum Zitat Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. In: Advances in neural information processing system, pp 2377–2385 Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. In: Advances in neural information processing system, pp 2377–2385
11.
Zurück zum Zitat Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. Adv Neural Inf Process Syst 4(January):3104–3112 Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. Adv Neural Inf Process Syst 4(January):3104–3112
12.
Zurück zum Zitat Bahdanau D, Cho KH, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3rd International conference on learning representations, ICLR 2015—conference track proceedings, pp 1–15. Published online Bahdanau D, Cho KH, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: 3rd International conference on learning representations, ICLR 2015—conference track proceedings, pp 1–15. Published online
13.
Zurück zum Zitat Luong MT, Pham H, Manning CD (2015) Effective approaches to attention-based neural machine translation. In: Conference proceedings—EMNLP 2015 conference on empirical methods in natural language processing, pp 1412–1421. Published online. https://doi.org/10.18653/v1/d15-1166 Luong MT, Pham H, Manning CD (2015) Effective approaches to attention-based neural machine translation. In: Conference proceedings—EMNLP 2015 conference on empirical methods in natural language processing, pp 1412–1421. Published online. https://​doi.​org/​10.​18653/​v1/​d15-1166
21.
Zurück zum Zitat Bergstra J, Bengio Y (2012) Random search for hyper-parameter optimization. J Mach Learn Res 13:281–305MathSciNet Bergstra J, Bengio Y (2012) Random search for hyper-parameter optimization. J Mach Learn Res 13:281–305MathSciNet
22.
Zurück zum Zitat Papineni K, Roukos S, Ward T, Zhu W-J (2002) BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the Association for Computational Linguistics. Association for Computational Linguistics, pp 311–318. https://doi.org/10.3115/1073083.1073135 Papineni K, Roukos S, Ward T, Zhu W-J (2002) BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the Association for Computational Linguistics. Association for Computational Linguistics, pp 311–318. https://​doi.​org/​10.​3115/​1073083.​1073135
Metadaten
Titel
Analyzing the Impact of Carbon Emission in Training Neural Machine Translation Models: A Case Study
verfasst von
Goutam Datta
Nisheeth Joshi
Kusum Gupta
Copyright-Jahr
2024
Verlag
Springer Nature Singapore
DOI
https://doi.org/10.1007/978-981-99-9486-1_7

Neuer Inhalt