Award Abstract # 2203167
NeTS: Small: New Abstractions for First-hop Networking in Cloud Data Centers

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: UNIVERSITY OF TEXAS AT AUSTIN
Initial Amendment Date: February 5, 2022
Latest Amendment Date: February 5, 2022
Award Number: 2203167
Award Instrument: Standard Grant
Program Manager: Darleen Fisher
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: October 1, 2021
End Date: August 31, 2023 (Estimated)
Total Intended Award Amount: $500,000.00
Total Awarded Amount to Date: $167,504.00
Funds Obligated to Date: FY 2017 = $167,504.00
History of Investigator:
  • Aditya Akella (Principal Investigator)
    akella@cs.utexas.edu
Recipient Sponsored Research Office: University of Texas at Austin
110 INNER CAMPUS DR
AUSTIN
TX  US  78712-1139
(512)471-6424
Sponsor Congressional District: 25
Primary Place of Performance: University of Texas at Austin
TX  US  78759-5316
Primary Place of Performance
Congressional District:
37
Unique Entity Identifier (UEI): V6AFQPN18437
Parent UEI:
NSF Program(s): Networking Technology and Syst
Primary Program Source: 01001718DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7923
Program Element Code(s): 736300
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

This project seeks to improve the performance of future data center compute servers and networks, enabling a growing number of performance-sensitive applications to co-exist on each server. To allow multiple applications on a server to drive the Network Interface Card (NIC) at high speed, this research aims to rethink the mechanisms employed in various layers of the server networking stack, and create new interfaces and abstractions between the layers and for applications. This research will lead to improved scalability, utilization, and efficiency of cloud computing infrastructure, allowing providers to support many more applications and tenants while maintaining their current infrastructure footprint.

This project seeks to develop new algorithms and programming abstractions for server operating systems and Network Interface Cards (NICs) to improve scheduling of outbound network traffic. These algorithms will overcome current performance bottlenecks, and will improve the ability for multiple applications to drive a NIC at high speed, with the ambitious goal of sustaining line rates while meeting application latency/throughput/fairness objectives as well as infrastructure-wide utilization and efficiency objectives. The PIs plan to integrate the proposed research into courses at UW-Madison. The PIs will organize various `network programming bootcamps' aimed at network engineers, undergraduate students, and high school students from under-represented groups.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Ji, Tao and Saxena, Divyanshu and Stephens, Brent E. and Akella, Aditya "Yama: Providing Performance Isolation for Black-Box Offloads" SoCC , 2023 https://doi.org/10.1145/3620678.3624792 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Modern data centers run various applications with different performance needs, ranging from highly latency-sensitive applications to throughput-sensitive batch (big data and machine learning) workloads. To meet the growing performance demands of these applications, data center operators are scaling up network speeds. However, ensuring that the applications can drive server network interfaces (NICs) at full line rates while not impacting each other’s performance objectives, and while maintaining optimal levels of server/network utilization and efficiency, continues to remain a challenge. The main reason is that existing solutions—such as hardware/software parallelism to aid application multiplexing, and offloading application computation onto network hardware—serve as specialized point optimizations. Because of loose coordination across layers of the network stacks where these techniques apply, they can often lead to detrimental outcomes such as high and variable latency, unreasonable and wasteful CPU utilization, and severe network congestion.

To allow multiple applications on a server to drive the NIC at high speed, our project aims to rethink the mechanisms employed in various layers of the server networking stack and create new interfaces and abstractions between the layers and applications. The goal is to sustain high line rates while meeting application latency/throughput/fairness objectives and infrastructure-wide utilization and efficiency goals.

Intellectual merit: Toward this goal, our project made several successful breakthroughs. We developed a new NIC that allows application computation to be offloaded from CPUs to network hardware, enabling optimal CPU utilization and high application and network performance. We created hardware modules that can be hosted on this NIC to host other performance- and resource-intensive functionalities such as load balancing, core allocation, and application scheduling. Finally, we showed how the NICs can host data-intensive applications spanning big data processing and machine learning domains. Our impacts and advances not only networking, but also operating systems, computer architecture, and programming languages.

Broader impacts: Results from this project have been successfully incorporated into graduate and undergraduate courses at UW Madison and UT Austin. The outcomes of this work are now being considered for production deployments at hyperscaler networks including Google and Microsoft. The project contributed to the PhD thesis of four students across UW and UT, including two female PhD students. It also led to the honors thesis of two undergraduate students at UT Austin.


 


 


Last Modified: 12/15/2023
Modified by: Aditya Akella

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page