ParameterNet: Parameters Are All You Need for Large-scale Visual Pretraining of Mobile Networks.
Computing Research Repository (CoRR)(2024)
Huawei Noah's Ark Lab | University of Sydney | University of Macau
Abstract
The large-scale visual pretraining has significantly improve the performance of large vision models. However, we observe the low FLOPs pitfall that the existing low-FLOPs models cannot benefit from large-scale pretraining. In this paper, we introduce a novel design principle, termed ParameterNet, aimed at augmenting the number of parame-ters in large-scale visual pretraining models while minimizing the increase in FLOPs. We leverage dynamic convolutions to incorporate additional parameters into the networks with only a marginal rise in FLOPs. The ParameterNet approach allows low-FLOPs networks to take advantage of large-scale visual pretraining. Furthermore, we extend the ParameterNet concept to the language domain to enhance inference results while preserving inference speed. Experiments on the large-scale ImageNet-22K have shown the superiority of our ParameterNet scheme. For example, ParameterNet-600M can achieve higher accuracy than the widely-used Swin Transformer ( 81.6% vs. 80.9%) and has much lower FLOPs (0.6G vs. 4.5G). The code will be released at https://parameternet.github.io/.
MoreTranslated text
Key words
Large-scale Pretraining,Rational Design,Visual Model,Language Domains,Large Datasets,Learning Rate,Convolutional Layers,Computer Vision,Object Detection,Large-scale Datasets,Visual Task,Weight Decay,ImageNet,Semantic Segmentation,Fully-connected Layer,Language Model,Number Of Experts,Top-1 Accuracy,Transformer Architecture,Standard Convolution,Vision Transformer,Base Learning Rate,AdamW Optimizer,Hidden Size,Training Loss,Training Dataset,Neural Network,Trainable Parameters
PDF
View via Publisher
AI Read Science
Must-Reading Tree
Example

Generate MRT to find the research sequence of this paper
Related Papers
2019
被引用6425 | 浏览
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper
去 AI 文献库 对话