Award Abstract # 2154191
Collaborative Research: CNS CORE: Small: RUI: Hierarchical Deep Reinforcement Learning for Routing in Mobile Wireless Networks

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: UNIVERSITY OF CONNECTICUT
Initial Amendment Date: April 19, 2022
Latest Amendment Date: May 8, 2023
Award Number: 2154191
Award Instrument: Standard Grant
Program Manager: Alhussein Abouzeid
aabouzei@nsf.gov
 (703)292-7855
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: April 15, 2022
End Date: March 31, 2026 (Estimated)
Total Intended Award Amount: $273,015.00
Total Awarded Amount to Date: $287,415.00
Funds Obligated to Date: FY 2022 = $273,015.00
FY 2023 = $14,400.00
History of Investigator:
  • Bing Wang (Principal Investigator)
    bing@uconn.edu
  • Dongjin Song (Co-Principal Investigator)
Recipient Sponsored Research Office: University of Connecticut
438 WHITNEY RD EXTENSION UNIT 1133
STORRS
CT  US  06269-9018
(860)486-3622
Sponsor Congressional District: 02
Primary Place of Performance: University of Connecticut
438 Whitney Road Ext.
Storrs
CT  US  06269-1133
Primary Place of Performance
Congressional District:
02
Unique Entity Identifier (UEI): WNTPS995QBM7
Parent UEI:
NSF Program(s): Special Projects - CNS,
Networking Technology and Syst
Primary Program Source: 01002223DB NSF RESEARCH & RELATED ACTIVIT
01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 9251, 7923, 9102, 9178
Program Element Code(s): 171400, 736300
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

The use of multi-hop routing in mobile wireless networks is becoming more prevalent, just as these networks are becoming more dense, dynamic, and heterogeneous. Designing a universal multi-hop routing strategy for mobile wireless networks is challenging, however, due to the need to seamlessly adapt routing behavior to spatially diverse and temporally changing network conditions. An alternative to using hand-crafted routing strategies is to use Reinforcement Learning (RL) to learn adaptive multi-hop routing strategies automatically. RL focuses on the design of intelligent agents: an RL agent interacts with its environment to learn a policy, i.e., which actions to take in different environmental states. By using function approximation like deep neural networks (DNNs) as in deep reinforcement learning (DeepRL) to approximate the policy, the RL agent can learn to generalize from its training experience to unseen network conditions and scale the learned routing strategy to larger networks. The PIs will continue their current practice of involving under-represented groups in research, and will use the project research to promote teaching and training through postdoctoral mentoring, course development, and outreach activities.

The goal of this project is to use DeepRL to develop a universal multi-hop routing strategy for mobile wireless networks that is scalable, generalizable, and adaptive. Specifically, this project will build a novel routing framework that uses hierarchical DeepRL to design an option hierarchy, comprised of multiple layers of routing decisions working together to achieve the overall goals of the network. To enable the same routing strategy to be used at different devices and in unseen network scenarios, the framework will use relational features combined with novel neural network models to handle mobility and perform feature estimation. To further enhance generalizability, the framework will use continual learning to ensure that the routing behaviors learned for more recently seen network scenarios do not dominate the learned routing policy. The developed routing strategies will be thoroughly evaluated using both simulation and experimental testbeds. Through the use of hierarchical DeepRL, this project will provide a significant step forward in developing RL-based routing strategies, and will facilitate development of adaptive strategies for a wide range of mobile wireless networks.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Zhang, Xikun and Song, Dongjin and Tao, Dacheng "Ricci Curvature-Based Graph Sparsification for Continual Graph Representation Learning" IEEE Transactions on Neural Networks and Learning Systems , 2024 https://doi.org/10.1109/TNNLS.2023.3303454 Citation Details
Zhang, Xikun and Song, Dongjin and Tao, Dacheng "Hierarchical Prototype Networks for Continual Graph Representation Learning" IEEE Transactions on Pattern Analysis and Machine Intelligence , v.45 , 2023 https://doi.org/10.1109/TPAMI.2022.3186909 Citation Details
Zhang, Xikun and Song, Dongjin and Tao, Dacheng "Sparsified Subgraph Memory for Continual Graph Representation Learning" IEEE International Conference on Data Mining (ICDM) 2022 , 2022 https://doi.org/10.1109/ICDM54844.2022.00177 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page