skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Geographical Server Relocation: Opportunities and Challenges
The enormous growth of AI computing has led to a surging demand for electricity. To stem the resulting energy cost and environmental impact, this paper explores opportunities enabled by the increasing hardware heterogeneity and introduces the concept of Geographical Server Relocation (GSR). Specifically, GSR physically balances the available AI servers across geographically distributed data centers subject to AI computing demand and power capacity constraints in each location. The key idea of GSR is to relocate older and less energy-efficient servers to regions with more renewables, better water efficiencies and/or lower electricity prices. Our case study demonstrates that, even with modest flexibility of relocation, GSR can substantially reduce the total operational environmental footprints and operation costs of AI computing. We conclude this paper by discussing major challenges of GSR, including service migration, software management, and algorithms.  more » « less
Award ID(s):
2324941 2007115
PAR ID:
10544904
Author(s) / Creator(s):
; ; ;
Publisher / Repository:
2024 HotCarbon Workshop on Sustainable Computer Systems
Date Published:
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. The enormous growth of AI computing has led to a surging demand for electricity. To stem the resulting energy cost and environmental impact, this paper explores opportunities enabled by the increasing hardware heterogeneity and introduces the concept of Geographical Server Relocation (GSR). Specifically, GSRphysicallybalances the available AI servers across geographically distributed data centers subject to AI computing demand and power capacity constraints in each location. The key idea of GSR is to relocate older and less energy-efficient servers to regions with more renewables, better water efficiencies and/or lower electricity prices. Our case study demonstrates that, even with modest flexibility of relocation, GSR can substantially reduce the total operational environmental footprints and operation costs of AI computing. We conclude this paper by discussing major challenges of GSR, including service migration, software management, and algorithms. 
    more » « less
  2. Since the advent of widely accessible AI tools, AI technology has been in high demand by businesses, academic researchers and individuals. Technology companies are building AI infrastructure at a rapid pace, and these facilities consume vast and growing resources, particularly electricity and water, with significant real and projected climate impacts. There is a need for new research initiatives to support long time horizon efforts to develop energy efficient computing capabilities to support the continued growth of AI infrastructure in a sustainable fashion. Such efficiency is required at both the hardware and software levels. Where can industry turn for examples of ultra-low power, energy efficient computing? We argue here that neurobiological principles offer rich and under-exploited sources of inspiration for energy efficient NeuroAI, and that new partnerships between industry and academia should be developed in this direction 
    more » « less
  3. Since the advent of widely accessible AI tools, AI technology has been in high demand by businesses, academic researchers and individuals. Technology companies are building AI infrastructure at a rapid pace, and these facilities consume vast and growing resources, particularly electricity and water, with significant real and projected climate impacts. There is a need for new research initiatives to support long time horizon efforts to develop energy efficient computing capabilities to support the continued growth of AI infrastructure in a sustainable fashion. Such efficiency is required at both the hardware and software levels. Where can industry turn for examples of ultra-low power, energy efficient computing? We argue here that neurobiological principles offer rich and under-exploited sources of inspiration for energy efficient NeuroAI, and that new partnerships between industry and academia should be developed in this direction 
    more » « less
  4. The remarkable progress in artificial intelligence (AI) has ushered in a new era characterized by models with billions of parameters, enabling extraordinary capabilities across diverse domains. However, these achievements come at a significant cost in terms of memory and energy consumption. The growing demand for computational resources raises grand challenges for the sustainable development of energy-efficient AI systems. This paper delves into the paradigm of memory-based computing as a promising avenue to address these challenges. By capitalizing on the inherent characteristics of memory and its efficient utilization, memory-based computing offers a novel approach to enhance AI performance while reducing the associated energy costs. Our paper systematically analyzes the multifaceted aspects of this paradigm, highlighting its potential benefits and outlining the challenges it poses. Through an exploration of various methodologies, architectures, and algorithms, we elucidate the intricate interplay between memory utilization, computational efficiency, and AI model complexity. Furthermore, we review the evolving area of hardware and software solutions for memory-based computing, underscoring their implications for achieving energy-efficient AI systems. As AI continues its rapid evolution, identifying the key challenges and insights presented in this paper serve as a foundational guide for researchers striving to navigate the complex field of memory-based computing and its pivotal role in shaping the future of energy-efficient AI. 
    more » « less
  5. Smart Servers, Smarter Speed Scaling: A Decentralized Algorithm for Data Center Efficiency A team of researchers from Georgia Tech and the University of Minnesota has introduced a cutting-edge algorithm designed to optimize energy use in large-scale data centers. As detailed in their paper “Distributed Rate Scaling in Large-Scale Service Systems,” the team developed a decentralized method allowing each server to adjust its processing speed autonomously without the need for communication or knowledge of system-wide traffic. The algorithm uses idle time as a local signal to guide processing speed, ensuring that all servers converge toward a globally optimal performance rate. This innovation addresses a critical issue in modern computing infrastructure: balancing energy efficiency with performance under uncertainty and scale. The authors demonstrate that their approach not only stabilizes the system but achieves asymptotic optimality as the number of servers increases. The work is poised to significantly reduce energy consumption in data centers, which are projected to account for up to 8% of U.S. electricity use by 2030. 
    more » « less