由买买提看人间百态

boards

本页内容为未名空间相应帖子的节选和存档,一周内的贴子最多显示50字,超过一周显示500字 访问原贴
Programming版 - IBM 弄出个DDL,16天缩短到7小时,很暴力
相关主题
hinton的胶囊本版什么评价卷积这东西真神了
Big Data’s Next Big Battle: IBM Versus… GE?CNN transfer learning 为啥这么牛逼?
IBM小型机+Oracle在什么情况下就不再适用了?Deepmind 的星际二挑战
Resnet没有max-pooling,为啥能比cnn好?deep learning现在还是冰山一角
从数据压缩角度理解CNN请教数据存储问题
tf还是很混乱再来推广下picpac
feifei li 也去google创建数据库问题 (转载)
CNN网络之后一般还要加FNN?请教有关java spring MVC的问题
相关话题的讨论汇总
话题: ddl话题: accuracy话题: our话题: gpus话题: microsoft
进入Programming版参与讨论
1 (共1页)
m*****n
发帖数: 1631
1
Our software does deep learning training fully synchronously with very low
communication overhead. As a result, when we scaled to a large cluster with
100s of NVIDAI GPUs, it yielded record image recognition accuracy of 33.8%
on 7.5M images from the ImageNet-22k dataset vs the previous best published
result of 29.8% by Microsoft. A 4% increase in accuracy is a big leap
forward; typical improvements in the past have been less than 1%. Our
innovative distributed deep learning (DDL) approach enabled us to not just
improve accuracy, but also to train a ResNet-101 neural network model in
just 7 hours, by leveraging the power of 10s of servers, equipped with 100s
of NVIDIA GPUs; Microsoft took 10 days to train the same model. This
achievement required we create the DDL code and algorithms to overcome
issues inherent to scaling these otherwise powerful deep learning frameworks.
https://www.ibm.com/blogs/research/2017/08/distributed-deep-learning/
a*****g
发帖数: 19398
2
不错啊。拿来练习围棋。

with
%
published
100s

【在 m*****n 的大作中提到】
: Our software does deep learning training fully synchronously with very low
: communication overhead. As a result, when we scaled to a large cluster with
: 100s of NVIDAI GPUs, it yielded record image recognition accuracy of 33.8%
: on 7.5M images from the ImageNet-22k dataset vs the previous best published
: result of 29.8% by Microsoft. A 4% increase in accuracy is a big leap
: forward; typical improvements in the past have been less than 1%. Our
: innovative distributed deep learning (DDL) approach enabled us to not just
: improve accuracy, but also to train a ResNet-101 neural network model in
: just 7 hours, by leveraging the power of 10s of servers, equipped with 100s
: of NVIDIA GPUs; Microsoft took 10 days to train the same model. This

1 (共1页)
进入Programming版参与讨论
相关主题
请教有关java spring MVC的问题从数据压缩角度理解CNN
wdong帮忙几个alphaGo的问题tf还是很混乱
这个机器怎么样?feifei li 也去google
ubuntu 16.04上1070不workCNN网络之后一般还要加FNN?
hinton的胶囊本版什么评价卷积这东西真神了
Big Data’s Next Big Battle: IBM Versus… GE?CNN transfer learning 为啥这么牛逼?
IBM小型机+Oracle在什么情况下就不再适用了?Deepmind 的星际二挑战
Resnet没有max-pooling,为啥能比cnn好?deep learning现在还是冰山一角
相关话题的讨论汇总
话题: ddl话题: accuracy话题: our话题: gpus话题: microsoft