Distributed optimization for deep learning with gossip exchange.

Neurocomputing(2019)

引用 41|浏览33
暂无评分
摘要
We address the issue of speeding up the training of convolutional neural networks by studying a distributed method adapted to stochastic gradient descent. Our parallel optimization setup uses several threads, each applying individual gradient descents on a local variable. We propose a new way of sharing information between different threads based on gossip algorithms that show good consensus convergence properties. Our method called GoSGD has the advantage to be fully asynchronous and decentralized.
更多
查看译文
关键词
Optimization,Distributed gradient descent,Gossip,Deep,Learning,Neural networks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要