TUHH Open Research
Help
  • Log In
    New user? Click here to register.Have you forgotten your password?
  • English
  • Deutsch
  • Communities & Collections
  • Publications
  • Research Data
  • People
  • Institutions
  • Projects
  • Statistics
  1. Home
  2. TUHH
  3. Publication References
  4. Activation sparsity and dynamic pruning for split computing in edge AI
 
Options

Activation sparsity and dynamic pruning for split computing in edge AI

Publikationstyp
Conference Paper
Date Issued
2022-12
Sprache
English
Author(s)
Haberer, Janek  
Landsiedel, Olaf  
TORE-URI
https://hdl.handle.net/11420/53866
Start Page
30
End Page
36
Citation
DistributedML 2022 - Proceedings of the 3rd International Workshop on Distributed Machine Learning, Part of CoNEXT 2022: 30-36
Contribution to Conference
3rd International Workshop on Distributed Machine Learning, DistributedML 2022  
18th International Conference on emerging Networking EXperiments and Technologies, CoNEXT 2022  
Publisher DOI
10.1145/3565010.3569066
Scopus ID
2-s2.0-85144825965
Publisher
Association for Computing Machinery
ISBN
978-1-4503-9922-7
Deep neural networks are getting larger and, therefore, harder to deploy on constrained IoT devices. Split computing provides a solution by splitting a network and placing the first few layers on the IoT device. The output of these layers is transmitted to the cloud where inference continues. Earlier works indicate a degree of high sparsity in intermediate activation outputs, this paper analyzes and exploits activation sparsity to reduce the network communication overhead when transmitting intermediate data to the cloud. Specifically, we analyze the intermediate activations of two early layers in ResNet-50 on CIFAR-10 and ImageNet, focusing on sparsity to guide the process of choosing a splitting point. We employ dynamic pruning of activations and feature maps and find that sparsity is very dependent on the size of a layer, and weights do not correlate with activation sparsity in convolutional layers. Additionally, we show that sparse intermediate outputs can be compressed by a factor of 3.3X at an accuracy loss of 1.1% without any fine-tuning. When adding fine-tuning, the compression factor increases up to 14X at a total accuracy loss of 1%.
Subjects
Activation Sparsity | Deep Learning | Edge Computing | Feature Map Pruning | Offloading
MLE@TUHH
DDC Class
600: Technology
TUHH
Weiterführende Links
  • Contact
  • Send Feedback
  • Cookie settings
  • Privacy policy
  • Impress
DSpace Software

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science
Design by effective webwork GmbH

  • Deutsche NationalbibliothekDeutsche Nationalbibliothek
  • ORCiD Member OrganizationORCiD Member Organization
  • DataCiteDataCite
  • Re3DataRe3Data
  • OpenDOAROpenDOAR
  • OpenAireOpenAire
  • BASE Bielefeld Academic Search EngineBASE Bielefeld Academic Search Engine
Feedback