Letian Zhao

Follow

Generating author description...

Common Coauthors
Coauthor Papers Together
Teng Tian 1
Wei Wu 1
Xi Jin 1
Chio-in Ieong 1
Xiaotian Wang 1
Tianqi Wang 1
Rui Xu 1
Commonly Cited References
Action Title Year Authors # of times referenced
+ Very Deep Convolutional Networks for Large-Scale Image Recognition 2014 Karen Simonyan
Andrew Zisserman
1
+ PDF Chat Deep Residual Learning for Image Recognition 2016 Kaiming He
Xiangyu Zhang
Shaoqing Ren
Jian Sun
1
+ PDF Chat Identity Mappings in Deep Residual Networks 2016 Kaiming He
Xiangyu Zhang
Shaoqing Ren
Jian Sun
1
+ Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation 2016 Yonghui Wu
Mike Schuster
Zhifeng Chen
Quoc V. Le
Mohammad Norouzi
Wolfgang Macherey
Maxim Krikun
Yuan Cao
Qin Gao
Klaus Macherey
1
+ Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour 2017 Priya Goyal
Piotr Dollár
Ross Girshick
Pieter Noordhuis
Lukasz Wesolowski
Aapo Kyrola
Andrew Tulloch
Yangqing Jia
Kaiming He
1
+ Exploring Hidden Dimensions in Parallelizing Convolutional Neural Networks 2018 Zhihao Jia
Sina Lin
Charles R. Qi
Alex Aiken
1
+ Attention is All you Need 2017 Ashish Vaswani
Noam Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan N. Gomez
Łukasz Kaiser
Illia Polosukhin
1
+ Dynamic Scheduling of MPI-based Distributed Deep Learning Training Jobs. 2019 Tim Capes
Vishal Raheja
Mete Kemertas
Iqbal Mohomed
1
+ PDF Chat Optimizing Multi-GPU Parallelization Strategies for Deep Learning Training 2019 Saptadeep Pal
Eiman Ebrahimi
Arslan Zulfiqar
Yaosheng Fu
Victor Zhang
Szymon Migacz
David Nellans
Puneet Gupta
1
+ PyTorch: An Imperative Style, High-Performance Deep Learning Library 2019 Adam Paszke
Sam Gross
Francisco Massa
Adam Lerer
James Bradbury
Gregory Chanan
Trevor Killeen
Zeming Lin
Natalia Gimelshein
Luca Antiga
1
+ PDF Chat Speeding up Deep Learning with Transient Servers 2019 Shijian Li
Robert J. Walls
Lijie Xu
Tian Guo
1
+ Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism 2019 Mohammad Shoeybi
Mostofa Patwary
Raul Puri
Patrick LeGresley
Jared Casper
Bryan Catanzaro
1
+ PDF Chat Dynamic Stale Synchronous Parallel Distributed Training for Deep Learning 2019 Xing Zhao
Aijun An
Junfeng Liu
Bao Xin Chen
1
+ GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism 2018 Yanping Huang
Youlong Cheng
Ankur Bapna
Orhan Fırat
Mia Xu Chen
Dehao Chen
HyoukJoong Lee
Jiquan Ngiam
Quoc V. Le
Yonghui Wu
1
+ Characterizing and Modeling Distributed Training with Transient Cloud GPU Servers 2020 Shijian Li
Robert J. Walls
Tian Guo
1
+ Beyond Data and Model Parallelism for Deep Neural Networks 2018 Zhihao Jia
Matei Zaharia
Alex Aiken
1
+ FPDeep: Scalable Acceleration of CNN Training on Deeply-Pipelined FPGA Clusters 2020 Tianqi Wang
Tong Geng
Ang Li
Xi Jin
Martin C. Herbordt
1
+ Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity. 2021 William Fedus
Barret Zoph
Noam Shazeer
1