MobyDeep: A lightweight CNN architecture to configure models for text classification

Knowledge-Based Systems(2022)

引用 5|浏览20
暂无评分
摘要
Nowadays, trends in deep learning for text classification are addressed to create complex models to deal with huge datasets. Deeper models are usually based on cutting edge neural network architectures, achieving good results in general but demanding better hardware than shallow ones. In this work, a new Convolutional Neural Network (CNN) architecture (MobyDeep) for text classification tasks is proposed. Designed as a configurable tool, resultant models (MobyNets) are able to manage big corpora sizes under low computational costs. To achieve those milestones, the architecture was conceived to produce lightweight models, having their internal layers based on a new proposed convolutional block. That block was designed and customized by adapting ideas from image to text processing, helping to squeezing model sizes and to reduce computational costs. The architecture was also designed as a residual network, covering complex functions by extending models up to 28 layers. Moreover, middle layers were optimized by residual connections, helping to remove fully connected layers on top and resulting in Fully CNN. Corpus were chosen from the recent literature, aiming to define real scenarios when comparing configured MobyDeep models with other state-of the-art works. Thus, three models were configured in 8, 16 and 28 layers respectively, offering competitive accuracy results.
更多
查看译文
关键词
Convolutional lightweight architecture,Network layer optimization,Text classification
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要