Measuring the Effectiveness of Carbon-Aware AI Training Strategies in Cloud Instances: A Confirmation Study
Abstract
:1. Introduction
- RQ1: How effective are the Flexible Start and Pause & Resume carbon-aware AI training strategies in reducing carbon emissions when applied to different Anomaly Detection algorithms?
- RQ2: What impact do geographical variations, specifically the carbon intensity of different regions, have on the carbon savings achieved by the Flexible Start and Pause & Resume strategies?
- RQ3: How do the carbon savings from Flexible Start and Pause & Resume strategies differ when applied to Machine Learning (ML) and Deep Learning (DL) Anomaly Detection algorithms, and what factors influence these variations?
2. Relevant Literature
3. Carbon-Aware Training Strategies
- GPUs’ energy consumption tracked each 5 min;
- Historical data of local marginal emissions supplied by WattTime with a 5 min granularity.
3.1. Flexible Start
- is the row vector of 5 min energy consumption, where n is the number of intervals that compose the workload.
- is the 1 column vector that represents the local marginal emissions at 5 min intervals for the starting time.
- is the operational emissions for Flexible Start strategy.
- i is in the range of s to e, where s is the start-time of the time window, and e is the end time.
3.2. Pause & Resume
3.3. Algorithms Comparison and Application
- {6, 12, 18, 24} h for Flexible Start.
- {25%, 50%, 75%, 100%} for Pause & Resume.
- Flexible Start. This strategy has been shown to be particularly efficient for short-duration workloads, such as in the case of DenseNet, where the best percentage of emissions reduction reaches in the West US region. The optimal time window was found to be 24 h. In contrast, for jobs longer than one day, the reductions are less significant. For instance, applying this strategy to a 6.1B parameter transformer led to reductions of less than . This may be explained by the fact that shorter jobs are less affected by the variability of marginal emissions during the time window. Therefore, this strategy is applicable to workloads that need to be run regularly but have some flexibility regarding their start times.
- Pause & Resume. This strategy yields considerable reductions in regions with wide variability of marginal emissions throughout the day and for workloads longer than a day. When applied to DenseNet, it brought a small reduction of less than . Conversely, when applied to the 6.1B parameter transformer, the reduction reaches almost .
4. Materials and Methods
4.1. Research Methodology
4.2. Anomaly Detection AI Algorithms
- Environmental impact: While there are AI sectors with more significant emissions, it is important to also pay attention to other sectors, such as AD, which will be demonstrated to have a non-negligible environmental impact.
- Data accessibility: This study uses a dataset provided by an Italian bank containing both fraudulent and non-fraudulent banking transactions. Access to a real dataset, rather than a synthetic one, allows for more realistic evaluations and assessments.
- Isolation Forest (IF) [33]. This algorithm is used in AD to identify anomalies within a dataset. It is based on an unsupervised learning approach and operates on the principle that anomalies are rarer and therefore more “isolated” than normal instances. The algorithm works by creating a set of random decision trees, known as isolation trees. Each tree is constructed by randomly selecting a subset of the data and recursively partitioning it based on random attribute values. The splitting process continues until each data instance is isolated in a single terminal node of the tree. To identify anomalies, the algorithm evaluates the separation path within the trees. Instances that require fewer splits to be isolated are considered more anomalous. Therefore, anomalies will have a shorter average path length compared with normal instances. The Isolation Forest has several advantageous features, including its ability to handle large-scale datasets efficiently, robustness to variations in data dimensionality, and capability to detect different types of anomalies, such as point anomalies and collective anomalies. However, it is important to note that the Isolation Forest may not be effective in detecting certain complex or subtle anomalies that require more contextual analysis. Thus, it is often used in combination with other AD algorithms to improve overall system performance.
- Support Vector Machine (SVM) [34]. This is a ML algorithm used for classification and regression problems. It is a supervised learning model that can separate data examples into different classes. The basic idea of an SVM is to find the optimal hyperplane that divides the feature space to maximise the separation between different data classes. The hyperplane is selected to be the maximum distance from a set of training examples of one class (called support vectors) while maximising the distance from the support vectors of the other classes. In practice, an SVM uses a set of labelled training examples to train a model that can subsequently classify new unlabelled examples. During the training phase, the SVM algorithm optimises an objective function that balances the maximum separation between classes with the reduction of classification error. The use of kernel functions is an important aspect of SVMs. A kernel allows data to be mapped into a high-dimensional feature space, where it is easier to find a linearly separating hyperplane. This enables SVMs to address nonlinear classification problems. SVMs have proven effective in a variety of applications, including image recognition, text analysis, bioinformatics, and more. They are known for their ability to handle large datasets and generalise well even on unseen data during training.
- HF-SCA [35]. This algorithm is a U-Net, a specific neural network whose architecture resembles the letter “U”. This structure is designed such that information, starting from the top left of the “U”, is progressively compressed through downsampling using convolutional layers and pooling. Once it reaches the bottom point, the information travels back up towards the top right of the “U”, reconstructing the data through upsampling. In the case of HF-SCA, depending on the reconstruction error of the input data, the algorithm determines whether a transaction is fraudulent or not. Additionally, the uniqueness of this algorithm lies in replacing the convolutional layers of the U-Net with squeeze-and-excitation blocks, which have improved performance for the specific case being examined.
- Autoencoder (AE) [36]. This is a neural network architecture commonly used in AD. It falls under the category of unsupervised learning, as it learns to reconstruct the input data without relying on labelled examples. The main idea behind the autoencoder is to use an encoder network to compress the input data into a lower-dimensional representation, also known as the latent space. This compressed representation contains the most essential features of the input data. Then, a decoder network reconstructs the original input data from the latent space representation. During the training phase, the autoencoder aims to minimise the reconstruction error, which is the difference between the original input data and the reconstructed output. By learning to reconstruct normal instances accurately, the autoencoder becomes less capable of reconstructing anomalous instances, leading to higher reconstruction errors for anomalies.
4.3. Benchmarking Software
- Workload Runner. The runner script is deployed on a remote machine with SSH access provided by CINECA, the largest computing centre in Italy. The purpose of the script is to execute ML and DL trainings while monitoring their energy consumption. Specifically, this runner script handles four different workloads and launches the selected workload. Each workload trains its corresponding model using a dataset in .csv format and, upon completion, generates an additional .csv file containing the energy consumption data for that workload. This file is required for the execution of the subsequent script.
- Strategy Launcher. The launcher script is executed on the local machine and simulates the application of one of the carbon-aware training strategies to the chosen workload. This simulation uses the .csv file of energy consumption produced by the Workload Runner script. Once this file is transferred to the local machine via SCP, the launcher script calculates the CO2eq emissions by mapping the obtained data against marginal emissions for a specific region and day in the year 2021. These marginal emissions data are also obtained from .csv datasets provided by WattTime. After calculating the emissions, the script computes the percentage reductions and visually presents a comparative analysis of the results for all the carbon-aware strategies.
- NumPy [38]: a library for scientific computing used for scalar product calculations, as well as determining maximum, minimum, and average values.
- Pandas [39]: a library for easy handling of structured data such as CSV files. It was primarily used for reading datasets and performing preprocessing tasks, including handling incomplete or dirty data.
- CodeCarbon [43]: The previously mentioned library used for tracking energy consumption. It even offers the ability to directly calculate carbon emissions for a given function using APIs. In this work, we used CodeCarbon only to acquire energy consumption, while emissions were calculated separately using carbon-aware strategies.
- Matplotlib [44]: a library capable of generating static, interactive, and animated graphs in Python. It was used for generating various visualisations, including timesheets for strategies and bar plots for annual percentage reductions.
4.4. Computing Environment
- Compliance with privacy policies: The General Data Protection Regulation (GDPR) is a regulation of the European Union regarding the processing of personal data and privacy. It stipulates that personal data can only be transferred to countries or organisations that provide an adequate level of data protection.
- Access to regions that use both renewable and non-renewable energy sources: this allows for a diverse analysis of emissions across different energy profiles.
- Availability of specific cities from the list of AWS Cloud instances: these regions provide a credible testbed for experiments.
- Timestamp: This represents the date and time of each collected sample. As previously mentioned, the data was recorded with a granularity of 5 min.
- MOER: Marginal Operating Emission Rate, which refers to the rate of carbon emissions emitted per unit of consumed energy, expressed in lbs/kWh. This value has been converted because the power consumption was tracked in kWh, and the carbon emissions were provided in gCO2eq (grammes of carbon dioxide equivalent).
- MOER version: refers to the version of the provided MOER data.
- Frequency: the duration in seconds for which the data is valid from point_time, where point_time is the date/time format indicating when this data became valid.
5. Results
5.1. Baseline
5.2. Emission Reduction
- The time window will be calculated by adding each value from the set , which we will refer to as the hours-set, to the length of the workload.
- The size of the time window will be calculated by increasing the workload length by each percentage value in the set , which we will refer to as the percentage set for simplicity.
- The checking time will be set to values in the set .
5.2.1. Isolation Forest
5.2.2. SVM
5.2.3. HF-SCA
5.2.4. Autoencoder
5.3. Average across All Regions
6. Discussion
6.1. Emission Reduction by Region
6.1.1. Flexible Start
6.1.2. Pause & Resume
6.2. Comparable Duration Increases
6.3. Final Considerations
6.4. Comparison with Existing Literature
6.5. Threat to Validity
7. Conclusions
- We reimplemented the code for Flexible Start and Pause & Resume, which is available upon request.
- We confirmed the effectiveness of two existing carbon-aware AI training strategies, providing additional useful insights.
- We contributed to the field of Green AI, hopefully encouraging further research efforts in this specific domain.
- Industry: our findings provide more evidence on the effectiveness of the proposed strategies, enabling industry players to decide if, when, and how to apply these strategies to make their AI usage greener and more aligned with the ecological transition, particularly with regard to ESG requirements.
- AI Community: the results can be used to develop AI libraries and frameworks that natively support carbon-aware scheduling of training workloads.
- Researchers in Green AI and Green Software: researchers can use our results to assess and compare the effectiveness of the proposed strategies in new scenarios and to develop and benchmark new strategies against those analysed in this work.
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
ML | Machine Learning |
DL | Deep Learning |
AD | Anomaly Detection |
AWS | Amazon Web Services |
NLP | Natural Language Processing |
MOER | Marginal Operating Emission Rate |
IF | Isolation Forest |
SVM | Support Vectors Machine |
HF-SCA | Hands-Free Strong Customer Authentication |
AE | Autoencoder |
AUC | Area Under the ROC Curve |
GDPR | General Data Protection Regulation |
ESG | Environment Social Governance |
References
- Aslan, J.; Mayers, K.; Koomey, J.G.; France, C. Electricity intensity of internet data transmission: Untangling the estimates. J. Ind. Ecol. 2018, 22, 785–798. [Google Scholar] [CrossRef]
- Mainetti, L.; Aprile, M.; Mele, E.; Vergallo, R. A sustainable approach to delivering programmable peer-to-peer offline payments. Sensors 2023, 23, 1336. [Google Scholar] [CrossRef] [PubMed]
- Masanet, E.; Shehabi, A.; Lei, N.; Smith, S.; Koomey, J. Recalibrating global data center energy-use estimates. Science 2020, 367, 984–986. [Google Scholar] [CrossRef] [PubMed]
- Vergallo, R.; D’Alò, T.; Mainetti, L.; Paiano, R.; Matino, S. Evaluating Sustainable Digitalization: A Carbon-Aware Framework for Enhancing Eco-Friendly Business Process Reengineering. Sustainability 2024, 16, 7789. [Google Scholar] [CrossRef]
- Corcoran, P.; Andrae, A. Emerging Trends in Electricity Consumption for Consumer ICT; National University of Ireland: Galway, Ireland, 2013. [Google Scholar]
- Tamburrini, G. The AI carbon footprint and responsibilities of AI scientists. Philosophies 2022, 7, 4. [Google Scholar] [CrossRef]
- Strubell, E.; Ganesh, A.; McCallum, A. Energy and policy considerations for modern deep learning research. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13693–13696. [Google Scholar]
- Taylor, J.A.; Callaway, D.S.; Poolla, K. Competitive energy storage in the presence of renewables. IEEE Trans. Power Syst. 2012, 28, 985–996. [Google Scholar] [CrossRef]
- Dodge, J.; Prewitt, T.; des Combes, R.T.; Odmark, E.; Schwartz, R.; Strubell, E.; Buchanan, W. Measuring the carbon intensity of AI in cloud instances. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, 21–24 June 2022; pp. 1877–1894. [Google Scholar]
- Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. Acm Comput. Surv. (CSUR) 2009, 41, 1–58. [Google Scholar] [CrossRef]
- Verdecchia, R.; Sallou, J.; Cruz, L. A systematic review of Green AI. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2023, 13, e1507. [Google Scholar] [CrossRef]
- Verdecchia, R.; Lago, P.; Ebert, C.; De Vries, C. Green IT and green software. IEEE Softw. 2021, 38, 7–15. [Google Scholar] [CrossRef]
- Wiedmann, T.; Minx, J. A definition of “carbon footprint”. Ecol. Econ. Res. Trends 2008, 1, 1–11. [Google Scholar]
- Matuštík, J.; Kočí, V. What is a footprint? A conceptual analysis of environmental footprint indicators. J. Clean. Prod. 2021, 285, 124833. [Google Scholar] [CrossRef]
- Luccioni, A.S.; Hernandez-Garcia, A. Counting carbon: A survey of factors influencing the emissions of machine learning. arXiv 2023, arXiv:2302.08476. [Google Scholar]
- Fraga-Lamas, P.; Lopes, S.I.; Fernández-Caramés, T.M. Green IoT and edge AI as key technological enablers for a sustainable digital transition towards a smart circular economy: An industry 5.0 use case. Sensors 2021, 21, 5745. [Google Scholar] [CrossRef] [PubMed]
- Wu, C.J.; Raghavendra, R.; Gupta, U.; Acun, B.; Ardalani, N.; Maeng, K.; Hazelwood, K. Sustainable ai: Environmental implications, challenges and opportunities. Proc. Mach. Learn. Syst. 2022, 4, 795–813. [Google Scholar]
- Ferro, M.; Silva, G.D.; de Paula, F.B.; Vieira, V.; Schulze, B. Towards a sustainable artificial intelligence: A case study of energy efficiency in decision tree algorithms. Concurr. Comput. Pract. Exp. 2023, 35, e6815. [Google Scholar] [CrossRef]
- Asperti, A.; Evangelista, D.; Loli Piccolomini, E. A survey on variational autoencoders from a green AI perspective. SN Comput. Sci. 2021, 2, 301. [Google Scholar] [CrossRef]
- Jooste, W.; Haque, R.; Way, A. Knowledge distillation: A method for making neural machine translation more efficient. Information 2022, 13, 88. [Google Scholar] [CrossRef]
- Bannour, N.; Ghannay, S.; Névéol, A.; Ligozat, A.L. Evaluating the carbon footprint of NLP methods: A survey and analysis of existing tools. In Proceedings of the Second Workshop on Simple and Efficient Natural Language Processing, Onilne, 10 November 2021; pp. 11–21. [Google Scholar]
- Kaack, L.H.; Donti, P.L.; Strubell, E.; Kamiya, G.; Creutzig, F.; Rolnick, D. Aligning artificial intelligence with climate change mitigation. Nat. Clim. Chang. 2022, 12, 518–527. [Google Scholar] [CrossRef]
- Willenbacher, M.; Hornauer, T.; Wohlgemuth, V. Rebound effects in methods of artificial intelligence. In Environmental Informatics; Springer International Publishing: Cham, Switzerland, 2021; pp. 73–85. [Google Scholar]
- Patterson, D.; Gonzalez, J.; Hölzle, U.; Le, Q.; Liang, C.; Munguia, L.M.; Dean, J. The carbon footprint of machine learning training will plateau, then shrink. Computer 2022, 55, 18–28. [Google Scholar] [CrossRef]
- Ligozat, A.L.; Lefevre, J.; Bugeau, A.; Combaz, J. Unraveling the hidden environmental impacts of AI solutions for environment life cycle assessment of AI solutions. Sustainability 2022, 14, 5172. [Google Scholar] [CrossRef]
- Dhar, P. The carbon impact of artificial intelligence. Nat. Mach. Intell. 2020, 2, 423–425. [Google Scholar] [CrossRef]
- Rohde, F.; Gossen, M.; Wagner, J.; Santarius, T. Sustainability challenges of artificial intelligence and policy implications. öKologisches-Wirtsch.-Fachz. 2021, 36, 36–40. [Google Scholar] [CrossRef]
- Wenninger, S.; Kaymakci, C.; Wiethe, C.; Römmelt, J.; Baur, L.; Häckel, B.; Sauer, A. How sustainable is machine learning in energy applications?–The sustainable machine learning balance sheet. In Proceedings of the International Conference on Wirtschaftsinformatik (WI) 2022, Online, 21–23 February 2022. [Google Scholar]
- Henderson, P.; Hu, J.; Romoff, J.; Brunskill, E.; Jurafsky, D.; Pineau, J. Towards the systematic reporting of the energy and carbon footprints of machine learning. J. Mach. Learn. Res. 2020, 21, 1–43. [Google Scholar]
- Anthony, L.F.W.; Kanding, B.; Selvan, R. Carbontracker: Tracking and predicting the carbon footprint of training deep learning models. arXiv 2020, arXiv:2007.03051. [Google Scholar]
- Lannelongue, L.; Grealey, J.; Inouye, M. Green algorithms: Quantifying the carbon footprint of computation. Adv. Sci. 2021, 8, 2100707. [Google Scholar] [CrossRef]
- Jurj, S.L.; Opritoiu, F.; Vladutiu, M. Environmentally-friendly metrics for evaluating the performance of deep learning models and systems. In Neural Information, Processing: 27th International Conference, ICONIP 2020; Bangkok, Thailand, 23–27 November 2020, Proceedings, Part III 27; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 232–244. [Google Scholar]
- Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar]
- Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef]
- Distante, C.; Fineo, L.; Mainetti, L.; Manco, L.; Taccardi, B.; Vergallo, R. HF-SCA: Hands-Free Strong Customer Authentication Based on a Memory-Guided Attention Mechanisms. J. Risk Financ. Manag. 2022, 15, 342. [Google Scholar] [CrossRef]
- Bank, D.; Koenigstein, N.; Giryes, R. Autoencoders. arXiv 2020, arXiv:2003.05991. [Google Scholar]
- Sandbox Regolamentare; Banca d’Italia. Retrieved June 2024. Available online: https://www.bancaditalia.it/focus/sandbox (accessed on 11 September 2024).
- NumPy; Hugunin, J. Retrieved June 2024. Available online: https://numpy.org/doc/stable/ (accessed on 11 September 2024).
- Pandas; McKinney, W.; Mendel, J.B.; Van den Bossche, J.; Reback, J. Retrieved June 2024. Available online: https://pandas.pydata.org (accessed on 11 September 2024).
- Cournapeau, D.; Grisel, O.; Varoquaux, G.; Mueller, A.G.E.A. Scikit-Learn Machine Learning in Python. Retrieved June 2024. Available online: https://scikit-learn.org (accessed on 11 September 2024).
- Tensorflow; Google Brain. Retrieved June 2024. Available online: https://www.tensorflow.org (accessed on 11 September 2024).
- PyTorch, Meta AI. Retrieved June 2024. Available online: https://pytorch.org (accessed on 11 September 2024).
- Track and Reduce CO2 Emissions from Your Computing, Clever Cloud. Retrieved June 2024. Available online: https://codecarbon.io/ (accessed on 11 September 2024).
- Hunter, J.D. Matplotlib: Visualization with Python. Retrieved June 2024. Available online: https://matplotlib.org (accessed on 11 September 2024).
- Jagannadharao, A.; Beckage, N.; Nafus, D.; Chamberlin, S. Timeshifting strategies for carbon-efficient long-running large language model training. Innov. Syst. Softw. Eng. 2023, 1–15. [Google Scholar] [CrossRef]
- Wiesner, P.; Behnke, I.; Kilian, P.; Steinke, M.; Kao, O. Vessim: A Testbed for Carbon-Aware Applications and Systems; Technische Universität Berlin: Berlin, Germany, 2024. [Google Scholar]
- Ross Fairbanks. Retrieved June 2024. Available online: https://rossfairbanks.com/2023/07/12/carbon-aware-spatial-shifting-with-karmada/ (accessed on 11 September 2024).
- Retrieved June 2024. Available online: https://www.wonderingchimp.com/exploring-the-green-apis/ (accessed on 11 September 2024).
- Gama, J.; Žliobaitė, I.; Bifet, A.; Pechenizkiy, M.; Bouchachia, A. A survey on concept drift adaptation. ACM Comput. Surv. (CSUR) 2014, 46, 1–37. [Google Scholar] [CrossRef]
- Poenaru-Olaru, L.; Karpova, N.; Cruz, L.; Rellermeyer, J.; van Deursen, A. Maintenance Techniques for Anomaly Detection AIOps Solutions. arXiv 2023, arXiv:2311.10421. [Google Scholar]
- Järvenpää, H.; Lago, P.; Bogner, J.; Lewis, G.; Muccini, H.; Ozkaya, I. A synthesis of green architectural tactics for ml-enabled systems. In Proceedings of the 46th International Conference on Software Engineering: Software Engineering in Societ, Lisbon, Portugal, 14–20 April 2024; pp. 130–141. [Google Scholar]
- Poenaru-Olaru, L.; Sallou, J.; Cruz, L.; Rellermeyer, J.S.; Van Deursen, A. Retrain AI Systems Responsibly! Use Sustainable Concept Drift Adaptation Techniques. In Proceedings of the 7th International Workshop on Green and Sustainable Software (GREENS), Melbourne, Australia, 14 May 2023; pp. 17–18. [Google Scholar]
Model | BERT Finetune | BERT LM | 6.1B Transformer | Dense 121 | Dense 169 | Dense 201 | ViT Tiny | ViT Small | ViT Base | ViT Large | ViT Huge | |
---|---|---|---|---|---|---|---|---|---|---|---|---|
Hours | 6 | 36 | 192 | 0.3 | 0.3 | 0.4 | 19 | 19 | 21 | 90 | 216 | |
kWh | 3.1 | 37.3 | 13,812.4 | 0.02 | 0.03 | 0.04 | 1.3 | 2.2 | 4.7 | 93.3 | 237.6 | |
24 h increase | Flexible Start | 14.5% | 3.4% | 0.5% | 26.8% | 26.4% | 25.9% | 5.6% | 5.3% | 4.2% | 1.3% | 0.5% |
Pause & Resume | 19.0% | 8.5% | 2.5% | 27.7% | 27.3% | 27.1% | 12.5% | 12.3% | 11.7% | 4.7% | 2.4% | |
100% increase | Flexible Start | 7.0% | 4.1% | 2.6% | 1.8% | 2.5% | 2.7% | 5.0% | 4.8% | 3.9% | 3.3% | 3.0% |
Pause & Resume | 9.5% | 11.0% | 11.4% | 2.0% | 2.8% | 3.1% | 11.0% | 11.0% | 10.8% | 11.4% | 11.3% |
Isolation Forest | SVM | HF-SCA | Autoencoder | |
---|---|---|---|---|
AUC score | 0.56 | 0.51 | 0.97 | 0.73 |
Energy consumption (kWh) | 0.825 | 0.493 | 3.310 | 0.615 |
Training time (h) | 4:15 | 2:30 | 16:00 | 3:30 |
Algorithm | Strategy | Hours-Set | |||
---|---|---|---|---|---|
6 | 12 | 18 | 24 | ||
IF | Flexible Start | 3.91% | 4.50% | 5.20% | 6.05% |
Pause & Resume | 3.23% | 5.09% | 5.71% | 7.01% | |
SVM | Flexible Start | 3.07% | 5.17% | 6.29% | 7.51% |
Pause & Resume | 3.81% | 5.44% | 5.59% | 6.74% | |
HF-SCA | Flexible Start | 2.27% | 2.49% | 2.63% | 2.80% |
Pause & Resume | 1.48% | 3.20% | 4.88% | 4.99% | |
AE | Flexible Start | 3.97% | 4.90% | 5.39% | 6.56% |
Pause & Resume | 3.20% | 5.31% | 6.04% | 7.31% |
Strategy | Avg Time Dilatation | Std Time Dilatation | Avg Carbon Reduction |
---|---|---|---|
Flexible Start | 7:54 h | 2:46 h | 5.72% |
Pause & Resume | 11:15 h | 3:26 h | 6.51% |
Paper | Year | Green AI Definition | Venue Type | Study Type | Topic | Domain | Type of Data | Artifact Considered | Considered Phase | Validation Type | Considered System Size | Saving | Industry Involvement | Intended Reader | Providing Tool |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
This work | 2024 | carbon footprint | journal | observational | deployment, monitoring | cloud | text | all | training | field experiment | 635K | % | Mix | Academic | Yes |
Strubell et al. [7] | 2019 | ecological footprint, carbon footprint, energy efficiency | conference | observational | monitoring | general | NLP | algorithm - deep neural network | training | - | - | Academic | Academic | No | |
Dodge et al. [9] | 2022 | carbon footprint | conference | observational | deployment, monitoring | cloud | NLP, image | all | all | field experiment | up to 6B | % | Mix | Academic | No |
Fraga-Lamas et al. [16] | 2021 | energy efficiency, carbon footprint | journal | observational | deployment, emissions, network-architecture | edge | image | all | all | laboratory | - | - | Academic | Academic | No |
Wu et al. [17] | 2022 | ecological footprint, carbon footprint | conference | observational | monitoring | general | - | algorithm - general | all | field experiment | - | - | Academic | Academic | No |
Ferro et al. [18] | 2021 | carbon footprint, energy efficiency | journal | observational | monitoring, model-comparison | general | number, text | algorithm - decision tree | all | laboratory | 5M | - | Academic | Academic | No |
Asperti et al. [19] | 2021 | energy efficiency, carbon footprint | journal | observational | model-comparison | general | image | algorithm - deep neural network | training | laboratory | 10K | - | Academic | Academic | No |
Jooste et al. [20] | 2022 | carbon footprint | journal | observational | hyperparameter-tuning | general | text | algorithm - deep neural network | training | laboratory | 2M | 50% | Academic | Academic | No |
Bannour et al. [21] | 2021 | carbon footprint | workshop | observational | model-comparison, monitoring | general | NLP | algorithm | all | laboratory | 1.2K | - | Academic | Mix | No |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Vergallo, R.; Mainetti, L. Measuring the Effectiveness of Carbon-Aware AI Training Strategies in Cloud Instances: A Confirmation Study. Future Internet 2024, 16, 334. https://doi.org/10.3390/fi16090334
Vergallo R, Mainetti L. Measuring the Effectiveness of Carbon-Aware AI Training Strategies in Cloud Instances: A Confirmation Study. Future Internet. 2024; 16(9):334. https://doi.org/10.3390/fi16090334
Chicago/Turabian StyleVergallo, Roberto, and Luca Mainetti. 2024. "Measuring the Effectiveness of Carbon-Aware AI Training Strategies in Cloud Instances: A Confirmation Study" Future Internet 16, no. 9: 334. https://doi.org/10.3390/fi16090334
APA StyleVergallo, R., & Mainetti, L. (2024). Measuring the Effectiveness of Carbon-Aware AI Training Strategies in Cloud Instances: A Confirmation Study. Future Internet, 16(9), 334. https://doi.org/10.3390/fi16090334