Efficient acceleration of deep learning inference on resource-constrained edge devices: A review
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …
in breakthroughs in many areas. However, deploying these highly accurate models for data …
Deep learning in mobile and wireless networking: A survey
The rapid uptake of mobile devices and the rising popularity of mobile applications and
services pose unprecedented demands on mobile and wireless networking infrastructure …
services pose unprecedented demands on mobile and wireless networking infrastructure …
Edge intelligence: Empowering intelligence to the edge of network
Edge intelligence refers to a set of connected systems and devices for data collection,
caching, processing, and analysis proximity to where data are captured based on artificial …
caching, processing, and analysis proximity to where data are captured based on artificial …
A survey on optimized implementation of deep learning models on the nvidia jetson platform
S Mittal - Journal of Systems Architecture, 2019 - Elsevier
Abstract Design of hardware accelerators for neural network (NN) applications involves
walking a tight rope amidst the constraints of low-power, high accuracy and throughput …
walking a tight rope amidst the constraints of low-power, high accuracy and throughput …
Edge intelligence: Architectures, challenges, and applications
Edge intelligence refers to a set of connected systems and devices for data collection,
caching, processing, and analysis in locations close to where data is captured based on …
caching, processing, and analysis in locations close to where data is captured based on …
Tensorrt-based framework and optimization methodology for deep learning inference on jetson boards
As deep learning inference applications are increasing in embedded devices, an embedded
device tends to equip neural processing units (NPUs) in addition to a multi-core CPU and a …
device tends to equip neural processing units (NPUs) in addition to a multi-core CPU and a …
Adaptive deep learning model selection on embedded systems
The recent ground-breaking advances in deep learning networks (DNNs) make them
attractive for embedded systems. However, it can take a long time for DNNs to make an …
attractive for embedded systems. However, it can take a long time for DNNs to make an …
Scope of machine learning applications for addressing the challenges in next‐generation wireless networks
The convenience of availing quality services at affordable costs anytime and anywhere
makes mobile technology very popular among users. Due to this popularity, there has been …
makes mobile technology very popular among users. Due to this popularity, there has been …
A survey of deep learning on cpus: opportunities and co-optimizations
CPU is a powerful, pervasive, and indispensable platform for running deep learning (DL)
workloads in systems ranging from mobile to extreme-end servers. In this article, we present …
workloads in systems ranging from mobile to extreme-end servers. In this article, we present …
Optimizing deep learning inference on embedded systems through adaptive model selection
Deep neural networks (DNNs) are becoming a key enabling technique for many application
domains. However, on-device inference on battery-powered, resource-constrained …
domains. However, on-device inference on battery-powered, resource-constrained …