Award Abstract # 1553248
CAREER: An Information Theoretic Perspective of Consistent Distributed Storage Systems

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: THE PENNSYLVANIA STATE UNIVERSITY
Initial Amendment Date: January 29, 2016
Latest Amendment Date: February 26, 2020
Award Number: 1553248
Award Instrument: Continuing Grant
Program Manager: Phillip Regalia
pregalia@nsf.gov
 (703)292-2981
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: February 15, 2016
End Date: January 31, 2023 (Estimated)
Total Intended Award Amount: $497,847.00
Total Awarded Amount to Date: $497,847.00
Funds Obligated to Date: FY 2016 = $192,086.00
FY 2018 = $201,554.00

FY 2020 = $104,207.00
History of Investigator:
  • Viveck Cadambe (Principal Investigator)
Recipient Sponsored Research Office: Pennsylvania State Univ University Park
201 OLD MAIN
UNIVERSITY PARK
PA  US  16802-1503
(814)865-1372
Sponsor Congressional District: 15
Primary Place of Performance: Pennsylvania State Univ University Park
PA  US  16802-7000
Primary Place of Performance
Congressional District:
Unique Entity Identifier (UEI): NPM2J7MSCF61
Parent UEI:
NSF Program(s): Comm & Information Foundations
Primary Program Source: 01001617DB NSF RESEARCH & RELATED ACTIVIT
01001819DB NSF RESEARCH & RELATED ACTIVIT

01002021DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 7935
Program Element Code(s): 779700
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Broader Significance:

Key-value stores form an integral infrastructural component of numerous modern web-based applications including retail stores, multi-player games, reservation systems, news feeds, and social and professional networks. Cloud computing service providers commonly implement key-value stores over large scale distributed data storage systems. At the heart of key-value store implementations in distributed storage systems, there are carefully crafted algorithms that expose a consistent, current view of the stored data to a user who reads the data. The main purpose of this project is to undertake a formal study of the storage costs incurred in distributed storage systems which aspire to present a consistent, current view of the stored data. The project has the long-term potential to aid the development of new data storage techniques that can benefit key-value store implementations by reducing their storage cost and energy consumption.

Technical Description:

An important requirement of a distributed data storage system is fault tolerance, that is, the data must be accessible even if the system components fail. In applications of distributed storage to distributed computing and implementation of key-value stores, the following property known as consistency is also critical: when the data is being constantly updated, a user that reads from the system should obtain the latest version of the data. Algorithms that ensure consistency and fault tolerance in storage systems have been extensively studied in distributed systems theory and practice. The goal of this project is to obtain, for the first time, an information theoretic understanding of the storage costs incurred in consistent, fault-tolerant distributed storage systems.

Building on preliminary work by the investigator, the project will develop and study several new information theoretic formulations inspired by distributed systems theory and practice. The proposed formulations naturally expose trade-offs between the degrees of redundancy and consistency, and other physical parameters of storage systems. New coding schemes and information theoretic converses for the proposed formulations will be developed using tools from algebra, combinatorics and network information theory. The project will also pursue the development of new bounds for network coding, which naturally apply to the newly proposed formulations and to other families of codes including locally repairable codes and regenerating codes. The project will likewise develop an education plan that eyes the long term goal of developing interdisciplinary researchers and engineers who are trained in information theory, coding theory, and the theory and design of distributed systems.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 19)
Cadambe, Viveck R. and Lyu, Shihang "Brief Announcement: CausalEC: A Causally Consistent Data Storage Algorithm based on Cross-Object Erasure Coding" PODC '23: Proceedings of the 2023 ACM Symposium on Principles of Distributed Computing , 2023 https://doi.org/10.1145/3583668.3594603 Citation Details
Dutta, Sanghamitra and Fahim, Mohammad and Haddadpour, Farzin and Jeong, Haewon and Cadambe, Viveck and Grover, Pulkit "On the optimal recovery threshold of coded matrix multiplication" IEEE Transactions on Information Theory , v.66 , 2020 10.1109/TIT.2019.2929328
Haddadpour, Farzin and Cadambe, Viveck R "Codes for distributed finite alphabet matrix-vector multiplication" 2018 IEEE International Symposium on Information Theory (ISIT) , 2018 10.1109/ISIT.2018.8437542
Haddadpour, Farzin and Yang, Yaoqing and Cadambe, Viveck and Grover, Pulkit "Cross-Iteration Coded Computing" 2018 56th Annual Allerton Conference on Communication, Control, and Computing , 2018 10.1109/ALLERTON.2018.8635933
Hamidreza Zare, Viveck R. Cadambe, Bhuvan Urgaonkar, Praneet Soni, Nader Alfares, Chetan Sharma, Arif Merchant "LEGOStore: A Linearizable Geo-Distributed Store Combining Replication and Erasure Coding" Proceedings of the VLDB Endowment , v.5 , 2022
Mohammad Fahim, Haewon Jeong, Farzin Haddadpou, Sanghamitra Dutta, Viveck Cadambe, and Pulkit Grover "On the Optimal Recovery Threshold of CodedMatrix Multiplication" 2017 Annual Allerton Conference on Communication, Control and Computing , 2017
Nicolas Nicolaou, Kishori M. Konwar, Viveck Cadambe, N. Prakash, Nancy Lynch, Muriel Medard "ARES: Adaptive, Reconfigurable, Erasure coded, atomic Storage" 2019 IEEE International Conference on Distributed Computing Systems (ICDCS) , 2019 10.1109/ICDCS.2019.00216
Nicolas Nicolaou, Viveck Cadambe, N. Prakash, Andrea Trigeorgi, Kishori M. Konwar, Muriel Medard and Nancy Lynch "ARES: Adaptive, Reconfigurable, Erasure coded, Atomic Storage" ACM Transactions on Storage , v.4 , 2022 https://doi.org/10.1145/3510613
Ramy Ali, Viveck R. Cadambe, Antonia Tulino, Jaime Llorca "Multi-version Coding with Side Information" IEEE International Symposium on Information Theory (ISIT) , 2018 10.1109/ISIT.2018.8437658
Ramy E. Ali and Viveck R. Cadambe "Consistent Distributed Storage of Correlated Data Updates Via Multi-version Coding" IEEE Information Theory Workshop (ITW) , v.Sep , 2016 10.1109/ITW.2016.7606819
Ramy E. Ali and Viveck R. Cadambe "Harnessing Correlations in Distributed Erasure Coded Key-Value Stores" IEEE Transactions on Communications , v.67 , 2019 10.1109/TCOMM.2019.2923616
(Showing: 1 - 10 of 19)

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Modern cloud data storage services are designed to ensure that the data is available to users even if some system components (such as servers) fail. This failure tolerance property is commonly ensured by replicating the data on multiple servers. Several applications such as reservation systems, financial transactions, and multiplayer gaming require storage systems to satisfy a property called consistency in addition to failure tolerance. In simple terms, consistency means that the latest update of the data must be available to users accessing the data. Consistent data storage services employ several carefully designed distributed algorithms that simultaneously ensure both failure tolerance and consistency.

Notably, data replication comes with a redundancy overhead to ensure failure tolerance. In the absence of consistency requirements, a technique called erasure coding from the area of information theory is well-known to incur significantly lower redundancy overhead as compared to data replication without lowering the degree of failure tolerance. This project studied the use of erasure coding for storage systems with consistency requirements. Specifically, the project developed information-theoretic trade-offs between failure tolerance and redundancy overhead for systems that can use erasure coding. It identified new factors that emerge in this trade-off due to the requirement of maintaining consistency. New erasure codes that are nearly optimal with respect to failure tolerance redundancy trade-offs were developed. The developed erasure coding techniques were incorporated into new distributed algorithms for consistent data storage. The cost benefits of these algorithms were analyzed and demonstrated through a prototype data storage system.

The project trained master's and Ph.D. level students in the areas of information theory, error-correcting codes, and distributed algorithms.


Last Modified: 05/01/2023
Modified by: Viveck R Cadambe

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page