Abstract

Due to limited resources of the Internet of Things (IoT) edge devices, deep neural network (DNN) inference requires collaboration with cloud server platforms, where DNN inference is partitioned and offloaded to high-performance servers to reduce end-to-end latency. As data-intensive intermediate feature space at the partitioned layer should be transmitted to the servers, efficient compression of the feature space is imperative for high-throughput inference. However, the feature space at deeper layers has different characteristics than natural images, limiting the compression performance by conventional preprocessing and encoding techniques. To tackle this limitation, we introduce a new method for compressing DNN intermediate feature space using a specialized autoencoder, called auto-tiler. The proposed auto-tiler is designed to include the tiling process and provide multiple input/output dimensions to support various partitioned layers and compression ratios. The results show that auto-tiler achieves 18% to 67% higher percent point accuracy compared to the existing methods at the same bitrate while reducing the process latency by 73% to 81%. The dimension variability of an auto-tiler also reduces the storage overhead by 62% with negligible accuracy loss.

Highlights

  • Artificial Intelligence (AI) is rapidly spreading into Internet of Things (IoT) devices, including face recognition for smart security systems [1,2,3], voice assistant with AI speakers [4,5,6], and smart cars [7,8]

  • Autoencoders are a special type of a neural network which is typically used to reduce the dimension of the input feature space using a bottleneck layer

  • The result indicates that the proposed method retains noticeably higher structural similarity after compression than the existing methods shown in Figure 2b,c, allowing for a more efficient compression

Read more

Summary

Introduction

Artificial Intelligence (AI) is rapidly spreading into Internet of Things (IoT) devices, including face recognition for smart security systems [1,2,3], voice assistant with AI speakers [4,5,6], and smart cars [7,8]. There is a rich literature on compressing input image/video, including Joint Photographic Experts Group (JPEG) and High Efficiency Video Coding (HEVC) These codecs have evolved over decades and can achieve a very high compression ratio on vision data. The others [15,16] tile multiple maps to build a large frame, which introduces blockiness in the combined frame and degrades the efficiency of natural-image-based codecs To address these limitations, this paper presents a new preprocessing technique for intermediate compression, called ‘auto-tiler’ (Figure 1). Some suggest preprocessing methods to make feature space compressed by conventional image/video codecs such as JPEG and HEVC They utilize state-of-the-art codecs to achieve high compression efficiency and can be classified into two categories based on how they process multi-channel features

Existing Feature Preprocessing Methods and Their Limitations
Frame per Channel Method
Tiling Method
Autoencoder as a Preprocessor
Variable Output Dimensions
DNN Model Training
Video Encoder Settings
Compression Performance Comparison
Memory Overhead Comparison
Structural Similarity Comparison
Effective Compression Ratio Comparison
Latency Comparison
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call