Award Abstract # 1837924
CSR: Small: Cross-Layer Design of Power Delivery and Load Balancing for Green Data Centers

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: REGENTS OF THE UNIVERSITY OF CALIFORNIA, THE
Initial Amendment Date: June 28, 2018
Latest Amendment Date: June 28, 2018
Award Number: 1837924
Award Instrument: Standard Grant
Program Manager: Marilyn McClure
mmcclure@nsf.gov
 (703)292-5197
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: January 1, 2018
End Date: August 31, 2019 (Estimated)
Total Intended Award Amount: $212,488.00
Total Awarded Amount to Date: $228,488.00
Funds Obligated to Date: FY 2015 = $212,488.00
FY 2017 = $16,000.00
History of Investigator:
  • Robert Pilawa-Podgurski (Principal Investigator)
    pilawa@berkeley.edu
Recipient Sponsored Research Office: University of California-Berkeley
1608 4TH ST STE 201
BERKELEY
CA  US  94710-1749
(510)643-3891
Sponsor Congressional District: 12
Primary Place of Performance: University of California-Berkeley
Berkeley
CA  US  94704-5940
Primary Place of Performance
Congressional District:
12
Unique Entity Identifier (UEI): GS3YEVSS12N6
Parent UEI:
NSF Program(s): Special Projects - CNS,
CSR-Computer Systems Research
Primary Program Source: 01001516DB NSF RESEARCH & RELATED ACTIVIT
01001718DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 9251
Program Element Code(s): 171400, 735400
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

As ever more computing is moved to the cloud, the energy consumption of data centers becomes increasingly important, both from an environmental and cost viewpoint. As a result, there is an increasing trend towards reducing the energy and carbon foot- print of data centers. While there has been considerable efforts to reduce the energy consumption, relatively little attention has been paid towards the power delivery in data centers. The objective of this work is to reduce the high voltage conversion losses (currently contributing 10-15% power loss) to almost zero by designing a joint software and hardware power delivery architecture specifically for a multi-server environment. This research, which could lead to drastic reduction of power conversion losses in data centers, has far-reaching impact on the design of sustainable and green data centers. Participation of underrepresented groups is encouraged, and portions of the research is incorporated into cloud computing courses, as hardware projects into a laboratory power electronics course, and as a case study of data center power delivery in an advanced graduate level power electronics course. An interactive online power usage portal, that visualizes in real-time the power usage of each individual server in the test-cluster provides opportunities for public interaction with the research. These open-source software and hardware demonstrations enable practitioners from around the world to learn more about sustainable computing.

This research explores a cross-layer design approach to data centers, where the power delivery architecture and software load balancing algorithms work together to achieve the highest possible power delivery efficiency. The research explores electrically series-connected racks of servers, to minimize overall power conversion and attendant losses. A key challenge in series voltage stacking is the variation in input voltage of each server due to imbalance of computational load in a series-stack. In this research, the challenge is addressed both in hardware and software. In software, scalable load balancing algorithms that ensure uniform power consumption in each server in the rack are developed. The load balancing algorithms simultaneously optimize for response time and power loss. Moreover, hardware power converters and distributed energy storage (e.g., capacitors, batteries) provide filtering and power balance in cases when software alone does not suffice. A key question being addressed is the suitable size of energy storage, and the required control bandwidth of the power converters to ensure proper operation for realistic workloads. In addition, high speed sensing and communication of electrical measurements of voltage and currents are employed in combination with operation of servers at asymmetric input voltages for static power consumption mismatch mitigation. The load balancing algorithms is tested with two types of workloads: (1) Interactive web workloads with short turnaround time and homogeneous servers; and (2) Map-reduce type workloads with long turnaround time and servers with data locality constraint.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Dimitrios Skarlatos, Renji Thomas, Aditya Agrawal, Shibin Qin, Robert Pilawa-Podgurski, Ulya R. Karpuzcu , Radu Teodorescu, Nam Sung Kim, and Josep Torrellas. "Snatch: Opportunistically Reassigning Power Allocation between Processor and Memory in 3D Stacks" nternational Symposium On Microarchitecture (MICRO) , 2016
A. Stillwell, R.C.N. Pilawa-Podgurski "A Resonant Switched-Capacitor Converter with GaN Transistors for High Efficiency Power Delivery to Series-Stacked Processors" IEEE Journal of Emerging and Selected Topics in Power Electronics , 2019 10.1109/JESTPE.2019.2917658

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

We have developed hardware and software solutions to reduce the high voltage conversion losses present in today's data centers to almost zero by designing a power delivery architecture specifically for a multi-server environment. We have demonstrated, in hardware and software, the idea of series-stacked servers, which, with the assistance of well-designed hardware and software, can drastically reduce the power delivery loss. Specifically, we have developed custom differential power processing (DPP) power converters, capable of powering a stack of series-connected ARM Cortex-based embedded computers, running Linux operating systems and distributed data processing software. We have demonstrated a peak efficiency of 99% power delivery, using computer workloads similar to that which occurs in datacenters. Moreover, we have successfully demonstrated startup, shutdown, and hot-swapping of individual computers, with un-interrupted continued operation. The high efficiency operation of this system relies on two primary research efforts, based in hardware and software innovations:

 

1. Differential power conversion. We addressed the challenge of differing server voltage caused by load imbalance by designing and employing differential power converters. The differential power converters provide voltage balancing by processing the difference in power between imbalanced loads, hence replacing inefficient high-voltage bulk power conversion with differential power conversion between servers. By processing only the difference in power between servers, substantial efficiency improvements can be realized. This was demonstrated using a resonant switched-capacitor power converter architectures with ultra-high efficiency.

 

2. Randomized load balancing. The conversion loss in differential power converters depends on the amount of load imbalance in a series-connected stack. This makes effective load balancing critical to

reducing the conversion loss. In addition, the load balancing algorithm needs to be scalable in a large

data center and must optimize for both response time and power at the same time. Computational load

needs to be balanced within each series stack to minimize the mismatch and processed power. This was accomplished through custom software to ensure distributed computing and load balancing across the series-stacked computing domains

 


Last Modified: 01/11/2021
Modified by: Robert Pilawa-Podgurski

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page