New method forecasts computation, energy costs for sustainable AI models

Forecasting Renewable Energy Generation with Machine Learning and ...

The process of updating deep learning/AI models when they face new tasks or must accommodate changes in data can have significant costs in terms of computational resources and energy consumption. Researchers have developed a novel method that predicts those costs, allowing users to make informed decisions about when to update AI models to improve AI sustainability. The study is published on the arXiv preprint server.

“There have been studies that focused on making deep learning model training more efficient,” says Jung-Eun Kim, corresponding author of a paper on the work and an assistant professor of computer science at North Carolina State University. “However, over a model’s life cycle, it will likely need to be updated many times. One reason is that, as our work here shows, retraining an existing model is much more cost effective than training a new model from scratch.

“If we want to address sustainability issues related to deep learning AI, we must look at computational and energy costs across a model’s entire life cycle—including the costs associated with updates. If you cannot predict what the costs will be ahead of time, it is impossible to engage in the type of planning that makes sustainability efforts possible. That makes our work here particularly valuable.”

Training a deep learning model is a computationally intensive process, and users want to go as long as possible without having to update the AI. However, two types of shifts can happen that make these updates inevitable. First, the task that the AI is performing may need to be modified. For example, if a model was initially tasked with only classifying digits and traffic symbols, you may need to modify the task to identify vehicles and humans as well. This is called a task shift.

Second, the data users provide to the model may change. For example, you may need to make use of a new kind of data, or perhaps the data you are working with is being coded in a different way. Either way, the AI needs to be updated to accommodate the change. This is called a distribution shift.

“Regardless of what is driving the need for an update, it is extremely useful for AI practitioners to have a realistic estimate of the computational demand that will be required for the update,” Kim says. “This can help them make informed decisions about when to conduct the update, as well as how much computational demand they will need to budget for the update.”

To forecast what the computational and energy costs will be, the researchers developed a new technique they call the REpresentation Shift QUantifying Estimator (RESQUE).

Essentially, RESQUE allows users to compare the dataset that a deep learning model was initially trained on to the new dataset that will be used to update the model. This comparison is done in a way that estimates the computational and energy costs associated with conducting the update.

Those costs are presented as a single index value, which can then be compared with five metrics: epochs, parameter change, gradient norm, carbon and energy. Epochs, parameter change and gradient norm are all ways of measuring the amount of computational effort necessary to retrain the model.

“However, to provide insight regarding what this means in a broader sustainability context, we also tell users how much energy, in kilowatt hours, will be needed to retrain the model,” Kim says. “And we predict how much carbon, in kilograms, will be released into the atmosphere in order to provide that energy.”

The researchers conducted extensive experiments involving multiple data sets, many different distribution shifts, and many different task shifts to validate RESQUE’s performance.

“We found that the RESQUE predictions aligned very closely with the real-world costs of conducting deep learning model updates,” Kim says. “Also, as I noted earlier, all of our experimental findings tell us that training a new model from scratch demands far more computational power and energy than retraining an existing model.”

In the short term, RESQUE is a useful methodology for anyone who needs to update a deep learning model.

“RESQUE can be used to help users budget computational resources for updates, allow them to predict how long the update will take, and so on,” Kim says.

“In the bigger picture, this work offers a deeper understanding of the costs associated with deep learning models across their entire life cycle, which can help us make informed decisions related to the sustainability of the models and how they are used. Because if we want AI to be viable and useful, these models must be not only dynamic but sustainable.”

The paper, “RESQUE: Quantifying Estimator to Task and Distribution Shift for Sustainable Model Reusability,” will be presented at The 39th Association for the Advancement of Artificial Intelligence (AAAI) Conference on Artificial Intelligence, which will be held Feb. 25–Mar. 4 in Philadelphia, Penn. The first author of the paper is Vishwesh Sangarya, a graduate student at NC State.

More information:
Vishwesh Sangarya et al, RESQUE: Quantifying Estimator to Task and Distribution Shift for Sustainable Model Reusability, arXiv (2024). DOI: 10.48550/arxiv.2412.15511

Journal information:
arXiv

Provided by
North Carolina State University

Citation:
New method forecasts computation, energy costs for sustainable AI models (2025, January 13)

Subscribe
Don't miss the best news ! Subscribe to our free newsletter :