Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Transform quantization for CNN compression
In this paper, we compress convolutional neural network (CNN) weights post-training via
transform quantization. Previous CNN quantization techniques tend to ignore the joint …
transform quantization. Previous CNN quantization techniques tend to ignore the joint …
Optimal gradient compression for distributed and federated learning
Communicating information, like gradient vectors, between computing nodes in distributed
and federated learning is typically an unavoidable burden, resulting in scalability issues …
and federated learning is typically an unavoidable burden, resulting in scalability issues …
A gradient flow framework for analyzing network pruning
Recent network pruning methods focus on pruning models early-on in training. To estimate
the impact of removing a parameter, these methods use importance measures that were …
the impact of removing a parameter, these methods use importance measures that were …
Finite blocklength lossy source coding for discrete memoryless sources
Shannon propounded a theoretical framework (collectively called information theory) that
uses mathematical tools to understand, model and analyze modern mobile wireless …
uses mathematical tools to understand, model and analyze modern mobile wireless …
An information-theoretic justification for model pruning
We study the neural network (NN) compression problem, viewing the tension between the
compression ratio and NN performance through the lens of rate-distortion theory. We choose …
compression ratio and NN performance through the lens of rate-distortion theory. We choose …
Fundamental limitation of semantic communications: Neural estimation for rate-distortion
This paper studies the fundamental limit of semantic communications over the discrete
memoryless channel. We consider the scenario to send a semantic source consisting of an …
memoryless channel. We consider the scenario to send a semantic source consisting of an …
Rdo-q: Extremely fine-grained channel-wise quantization via rate-distortion optimization
Allocating different bit widths to different channels and quantizing them independently bring
higher quantization precision and accuracy. Most of prior works use equal bit width to …
higher quantization precision and accuracy. Most of prior works use equal bit width to …
On distributed quantization for classification
We consider the problem of distributed feature quantization, where the goal is to enable a
pretrained classifier at a central node to carry out its classification on features that are …
pretrained classifier at a central node to carry out its classification on features that are …
[HTML][HTML] Population risk improvement with model compression: An information-theoretic approach
It has been reported in many recent works on deep model compression that the population
risk of a compressed model can be even better than that of the original model. In this paper …
risk of a compressed model can be even better than that of the original model. In this paper …
Taxonomy and evaluation of structured compression of convolutional neural networks
The success of deep neural networks in many real-world applications is leading to new
challenges in building more efficient architectures. One effective way of making networks …
challenges in building more efficient architectures. One effective way of making networks …