Cutting Edge '25

Sparse-MDI: Adaptive Sparsification for Communication-Efficient Model Distributed Inference in The Edge

By

Catagories

Play Video

The rapid growth of edge computing applications, such as autonomous vehicles and IoT systems, necessitates real-time, low-latency inference of deep neural networks in environments with limited computational resources. Most existing distributed inference methods face significant communication overheads when handling large, complex models, especially in resource-constrained edge networks. This research explores an adaptive sparsification technique by introducing a novel peer- 2-peer (P2P) model distributed inference (MDI) framework to address these communication challenges. Adjusting the sparsity of activations based on network and device conditions can be adopted by any peer-2-peer MDI framework to improve communication overhead. By dynamically adjusting data loads, adaptive sparsification enhances performance, reduces latency, and preserves model accuracy.

Vision Quest

Check out the visionary projects our students have brought up in this year
VisuaLit

VisuaLit is an AI-powered eBook reader that redefines traditional reading by merging visual storytelling, audio narration, and contextual learning into…

VenDoor

The VenDoor application is a fully functional mobile application designed to create a bridge between mobile vendors and their customers…

UniGuide

UniGuide is a student-focused platform that helps individuals make smart educational and career decisions. It offers a comprehensive database of…