Chrome Extension
WeChat Mini Program
Use on ChatGLM

SPARK: Scalable and Precision-Aware Acceleration of Neural Networks Via Efficient Encoding

2024 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE, HPCA 2024(2024)

Cited 0|Views12
No score
Key words
Neural Network,Efficient Encoding,Accuracy Of Model,Deep Neural Network,Source Code,Model Size,Values Of Quantities,Compression Method,Hardware Accelerators,Bit Length,Hardware Overhead,Convolutional Neural Network,Artificial Neural Network,High Precision,Distribution Of Parameters,Original Value,Convolutional Neural Network Model,Low Precision,Accuracy Loss,Deep Neural Network Model,High Precision Values,Input Bits,Uniform Quantization,Vision Transformer,Most Significant Bit,Quantization Parameter,High Sparsity,Decimal Value,Efficient Integration,Quantization Scheme
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined