Cited By
View all- Tuli SJha N(2023)EdgeTran: Device-Aware Co-Search of Transformers for Efficient Inference on Mobile Edge PlatformsIEEE Transactions on Mobile Computing10.1109/TMC.2023.332828723:6(7012-7029)Online publication date: 30-Oct-2023
With the recent trend of on-device deep learning, inference latency has become a crucial metric in running Deep Neural Network (DNN) models on various mobile and edge devices. To this end, latency prediction of DNN model inference is highly desirable ...
The asynchronous transfer mode (ATM) is the choice of transport mode for broadband integrated service digital networks (B-ISDNs). We propose a window-based contention resolution algorithm to achieve higher throughput for nonblocking switches in ATM ...
Model compression is an important technique to facilitate efficient embedded and hardware implementations of deep neural networks (DNNs), a number of prior works are dedicated to model compression techniques. The target is to simultaneously reduce the ...
Association for Computing Machinery
New York, NY, United States
Check if you have access through your login credentials or your institution to get full access on this article.
Sign in