Award Abstract # 2007714
CIF: Small: Compression for Learning over networks

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: UNIVERSITY OF CALIFORNIA, LOS ANGELES
Initial Amendment Date: July 30, 2020
Latest Amendment Date: July 30, 2020
Award Number: 2007714
Award Instrument: Standard Grant
Program Manager: Phillip Regalia
pregalia@nsf.gov
 (703)292-2981
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: August 1, 2020
End Date: July 31, 2024 (Estimated)
Total Intended Award Amount: $523,925.00
Total Awarded Amount to Date: $523,925.00
Funds Obligated to Date: FY 2020 = $523,925.00
History of Investigator:
  • Christina Fragouli (Principal Investigator)
    christina.fragouli@ucla.edu
  • Suhas Diggavi (Co-Principal Investigator)
Recipient Sponsored Research Office: University of California-Los Angeles
10889 WILSHIRE BLVD STE 700
LOS ANGELES
CA  US  90024-4200
(310)794-0102
Sponsor Congressional District: 36
Primary Place of Performance: University of California-Los Angeles
Los Angeles
CA  US  90095-1406
Primary Place of Performance
Congressional District:
36
Unique Entity Identifier (UEI): RN64EPNH8JC6
Parent UEI:
NSF Program(s): Special Projects - CCF,
Comm & Information Foundations
Primary Program Source: 01002021DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 075Z, 2878, 7923, 7936, 9178, 9251
Program Element Code(s): 287800, 779700
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Data compression is a core component of all communication protocols, as it can translate to bandwidth savings, energy efficiency and low delay operations. In the traditional setup, an information source compresses its messages so that they can be communicated efficiently with the goal of ensuring accurate reconstruction at the destination. This project seeks to design compression schemes that are specifically tailored to Machine Learning applications: If the transmitted messages support a given learning task (e.g., classification or learning), the desired compression schemes should provide better support for the learning task instead of focusing on reconstruction accuracy. This approach to compression could potentially yield significant benefits in terms of communication efficiency, while simultaneously promoting the successful implementation of Machine Learning algorithms. By improving communication efficiency, such schemes are expected to contribute to the successful implementation of distributed machine learning algorithms over networks.

Traditionally, compression schemes are evaluated using rate-distortion trade-offs; this project is interested in rate-accuracy trade-offs, where accuracy captures the effect that quantization may have on a specific machine learning task. There is particular interest in information-theoretic lower bounds and trade-offs, and in explicit compression for the following two questions: (1) How to compress for model training, when we need to use distributed communication constrained nodes to learn a model, fast and efficiently; and (2) How to compress for communication during inference. The project will derive bounds and algorithms for distributed compression of features coming from composite distributions that will be used for a machine learning task, such as classification. This work will advance the state of the art, and build new connections between the areas of data compression and distributed machine learning.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 40)
Cao, Xuanyu and Baar, Tamer and Diggavi, Suhas and Eldar, Yonina C. and Letaief, Khaled B. and Poor, H. Vincent and Zhang, Junshan "Communication-Efficient Distributed Learning: An Overview" IEEE Journal on Selected Areas in Communications , v.41 , 2023 https://doi.org/10.1109/JSAC.2023.3242710 Citation Details
Cao, Xuanyu and Baar, Tamer and Diggavi, Suhas and Eldar, Yonina C. and Letaief, Khaled B. and Poor, H. Vincent and Zhang, Junshan "Guest Editorial Communication-Efficient Distributed Learning Over Networks" IEEE Journal on Selected Areas in Communications , v.41 , 2023 https://doi.org/10.1109/JSAC.2023.3241848 Citation Details
Data, Deepesh and Diggavi, Suhas N. "Byzantine-Resilient High-Dimensional Federated Learning" IEEE Transactions on Information Theory , v.69 , 2023 https://doi.org/10.1109/TIT.2023.3284427 Citation Details
Data, Deepesh and Diggavi, Suhas N. "Byzantine-Resilient High-Dimensional SGD with Local Iterations on Heterogeneous Data" International Conference on Machine Learning (ICML) , 2021 Citation Details
Data, Deepesh and Diggavi, Suhas N. "Byzantine-Resilient SGD in High Dimensions on Heterogeneous Data" IEEE International Symposium on Information Theory (ISIT) , 2021 Citation Details
Girgis, A. M and Data, D and Diggavi, S N. "Renyi differential privacy of the subsampled shuffle model in distributed learning" Advances in Neural Information Processing Systems (NeurIPS) , 2021 Citation Details
Girgis, Antonious M and Data, Deepesh and and Diggavi, Suhas. "Differentially Private Federated Learning with Shuffling and Client Self-Sampling" IEEE International Symposium on Information Theory (ISIT) , 2021 Citation Details
Girgis, Antonious M. and Data, Deepesh and Diggavi, Suhas "Distributed User-Level Private Mean Estimation" IEEE International Symposium on Information Theory (ISIT) , 2022 https://doi.org/10.1109/ISIT50566.2022.9834713 Citation Details
Girgis, Antonious M. and Data, Deepesh and Diggavi, Suhas and Suresh, Ananda Theertha and Kairouz, Peter "On the Rรฉnyi Differential Privacy of the Shuffle Model" ACM Symposium on Computer and Communication Security (CCS) , 2021 https://doi.org/10.1145/3460120.3484794 Citation Details
Girgis, Antonious M and Diggavi, Suhas "Multi-Message Shuffled Privacy in Federated Learning" IEEE Journal on Selected Areas in Information Theory , v.5 , 2024 https://doi.org/10.1109/JSAIT.2024.3366225 Citation Details
Hanna, Osama A. and Ezzeldin, Yahya H. and Fragouli, Christina and Diggavi, Suhas "Quantization of Distributed Data for Learning" IEEE Journal on Selected Areas in Information Theory , v.2 , 2021 https://doi.org/10.1109/JSAIT.2021.3105359 Citation Details
(Showing: 1 - 10 of 40)

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Traditionally, compression schemes are evaluated using rate-distortion 
trade-offs; this project is interested in rate-accuracy trade-offs, 
where accuracy captures the effect that quantization may have on a 
specific machine learning task. There is particular interest in 
information-theoretic lower bounds and trade-offs, and in explicit 
compression for the following two questions: (1) How to compress for 
model training, when we need to use distributed communication 
constrained nodes to learn a model, fast and efficiently; and (2) How to 
compress for communication during inference. The project will derive 
bounds and algorithms for distributed compression of features coming 
from composite distributions that will be used for a machine learning 
task, such as classification. The goals of this project is to create new 
foundational theory and algorithms which will advance the state of the 
art, and build new connections between the areas of data compression and 
distributed machine learning. If successful, the project will not only 
build new theory, but also could have impact in practical applications 
of learning over communication constrained networks. The project was 
arranged in three thrusts, with Thrust 1 focused on compression for 
distributed model training, Thrust 2 exploring compression for inference 
time  and Thrust 3 that combined both training and inference time 
compression. We believe that we had significant results in the project 
towards these goals.

One measure of academic impact of the project is the number of 
publications during the duration of the project; 45 publications 
resulted from this project (that acknowledged support of this grant) in 
top tier conference and journals.  Papers acknowledging support from 
this project have garnered nearly 500 citations according to Google 
Scholar. It also supported (in part) the PhD work of 7 graduate 
students, one postdoctoral researchers and several undergraduate (REU) 
students. The four graduated PhD students have gone on to careers in 
industrial research labs. Another measure of impact is the recognition 
of the papers and PIs through this project. For example, one of papers 
supported (in part) by this project was recognized through  the 
prestigious 2021 ACM Conference on Computer and Communications Security 
(CCS) best paper award. One of the students supported by this award 
received the UCLA PhD dissertation award. It also resulted in faculty 
research awards on related topics for the PIs from Amazon and Meta. The 
research in this project also fostered a collaboration with industrial 
research labs (Google) through joint publications. The work was also 
disseminated through several invited lectures in various industrial and 
academics settings by the PIs, and through organization of workshops and 
conferences.

This project developed several fundamental ideas compression for 
distributed learning systems, which were discovered and published during 
the course of this project. The details are in the publications and were 
also summarized in annual reports of the project. Among the several 
works (including over 25 journal and journal-equivalent conference 
papers) we would like to highlight the following: (i) Gradient 
compression for decentralized networks: we proposed new algorithms that 
used error-compensation, event-triggering and sparsification to enable 
schemes whose convergence was equivalent to vanilla stochastic 
optimization with orders of magnitude reduction in communications; this 
work was published in several communities including Control (Trans. Aut. 
Control, Automatica, CDC), information theory (JSAIT, ISIT) etc. (ii) 
Compression for contextual multi-arm bandits: In this line of work we 
initated new compression methods for multi-arm bandits and also showed 
fundamental connections between contextual MAB and linear bandit 
problems. These works were published both in learning venues (NeurIPS, 
AISTATS, COLT) as well as information theory venues (JSAIT, ISIT)  (iii) 
Compression for interactive distributed inference: we developed new 
schemes and information theoretic bounds by introducing interaction in 
distributed inference, demonstrating a new way to break dependencies in 
observations (iv) Information theory and algorithms for personalized 
learning: we developed a new statistical framework for personalized 
learning and through that developed both information theoretic bounds as 
well as algorithms. These were published in learning venues (e.g., 
NeurIPS, ICLR) as well as information theory venues (ISIT). These are 
only a few of the many results from the project, which also included a 
review paper on compression for distributed learning and co-editing a 
special issue on the topic in IEEE Journal of Selected Areas in 
Communications. The project also involved out reach and training of 
undergraduate students, including several female students.



Last Modified: 01/19/2025
Modified by: Christina Fragouli

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page