
NSF Org: |
CNS Division Of Computer and Network Systems |
Recipient: |
|
Initial Amendment Date: | June 28, 2018 |
Latest Amendment Date: | June 28, 2018 |
Award Number: | 1837924 |
Award Instrument: | Standard Grant |
Program Manager: |
Marilyn McClure
mmcclure@nsf.gov (703)292-5197 CNS Division Of Computer and Network Systems CSE Directorate for Computer and Information Science and Engineering |
Start Date: | January 1, 2018 |
End Date: | August 31, 2019 (Estimated) |
Total Intended Award Amount: | $212,488.00 |
Total Awarded Amount to Date: | $228,488.00 |
Funds Obligated to Date: |
FY 2017 = $16,000.00 |
History of Investigator: |
|
Recipient Sponsored Research Office: |
1608 4TH ST STE 201 BERKELEY CA US 94710-1749 (510)643-3891 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
Berkeley CA US 94704-5940 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): |
Special Projects - CNS, CSR-Computer Systems Research |
Primary Program Source: |
01001718DB NSF RESEARCH & RELATED ACTIVIT |
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
As ever more computing is moved to the cloud, the energy consumption of data centers becomes increasingly important, both from an environmental and cost viewpoint. As a result, there is an increasing trend towards reducing the energy and carbon foot- print of data centers. While there has been considerable efforts to reduce the energy consumption, relatively little attention has been paid towards the power delivery in data centers. The objective of this work is to reduce the high voltage conversion losses (currently contributing 10-15% power loss) to almost zero by designing a joint software and hardware power delivery architecture specifically for a multi-server environment. This research, which could lead to drastic reduction of power conversion losses in data centers, has far-reaching impact on the design of sustainable and green data centers. Participation of underrepresented groups is encouraged, and portions of the research is incorporated into cloud computing courses, as hardware projects into a laboratory power electronics course, and as a case study of data center power delivery in an advanced graduate level power electronics course. An interactive online power usage portal, that visualizes in real-time the power usage of each individual server in the test-cluster provides opportunities for public interaction with the research. These open-source software and hardware demonstrations enable practitioners from around the world to learn more about sustainable computing.
This research explores a cross-layer design approach to data centers, where the power delivery architecture and software load balancing algorithms work together to achieve the highest possible power delivery efficiency. The research explores electrically series-connected racks of servers, to minimize overall power conversion and attendant losses. A key challenge in series voltage stacking is the variation in input voltage of each server due to imbalance of computational load in a series-stack. In this research, the challenge is addressed both in hardware and software. In software, scalable load balancing algorithms that ensure uniform power consumption in each server in the rack are developed. The load balancing algorithms simultaneously optimize for response time and power loss. Moreover, hardware power converters and distributed energy storage (e.g., capacitors, batteries) provide filtering and power balance in cases when software alone does not suffice. A key question being addressed is the suitable size of energy storage, and the required control bandwidth of the power converters to ensure proper operation for realistic workloads. In addition, high speed sensing and communication of electrical measurements of voltage and currents are employed in combination with operation of servers at asymmetric input voltages for static power consumption mismatch mitigation. The load balancing algorithms is tested with two types of workloads: (1) Interactive web workloads with short turnaround time and homogeneous servers; and (2) Map-reduce type workloads with long turnaround time and servers with data locality constraint.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
We have developed hardware and software solutions to reduce the high voltage conversion losses present in today's data centers to almost zero by designing a power delivery architecture specifically for a multi-server environment. We have demonstrated, in hardware and software, the idea of series-stacked servers, which, with the assistance of well-designed hardware and software, can drastically reduce the power delivery loss. Specifically, we have developed custom differential power processing (DPP) power converters, capable of powering a stack of series-connected ARM Cortex-based embedded computers, running Linux operating systems and distributed data processing software. We have demonstrated a peak efficiency of 99% power delivery, using computer workloads similar to that which occurs in datacenters. Moreover, we have successfully demonstrated startup, shutdown, and hot-swapping of individual computers, with un-interrupted continued operation. The high efficiency operation of this system relies on two primary research efforts, based in hardware and software innovations:
1. Differential power conversion. We addressed the challenge of differing server voltage caused by load imbalance by designing and employing differential power converters. The differential power converters provide voltage balancing by processing the difference in power between imbalanced loads, hence replacing inefficient high-voltage bulk power conversion with differential power conversion between servers. By processing only the difference in power between servers, substantial efficiency improvements can be realized. This was demonstrated using a resonant switched-capacitor power converter architectures with ultra-high efficiency.
2. Randomized load balancing. The conversion loss in differential power converters depends on the amount of load imbalance in a series-connected stack. This makes effective load balancing critical to
reducing the conversion loss. In addition, the load balancing algorithm needs to be scalable in a large
data center and must optimize for both response time and power at the same time. Computational load
needs to be balanced within each series stack to minimize the mismatch and processed power. This was accomplished through custom software to ensure distributed computing and load balancing across the series-stacked computing domains
Last Modified: 01/11/2021
Modified by: Robert Pilawa-Podgurski
Please report errors in award information by writing to: awardsearch@nsf.gov.