基本信息
views: 193
Career Trajectory
Bio
I'm working to make machines more helpful through unsupervised learning that scales.
I developed the Sparse Transformer with Scott Gray, and also coauthored work showing the emergent capabilities of large language models in a variety of settings (GPT-2, GPT-3, Image GPT, and more).
More recently I've worked on reducing the limitations of those techniques (very deep VAEs) while continuing to apply them on larger supercomputers (MT-NLG and PaLM).
I developed the Sparse Transformer with Scott Gray, and also coauthored work showing the emergent capabilities of large language models in a variety of settings (GPT-2, GPT-3, Image GPT, and more).
More recently I've worked on reducing the limitations of those techniques (very deep VAEs) while continuing to apply them on larger supercomputers (MT-NLG and PaLM).
Research Interests
Papers共 14 篇Author StatisticsCo-AuthorSimilar Experts
By YearBy Citation主题筛选期刊级别筛选合作者筛选合作机构筛选
时间
引用量
主题
期刊级别
合作者
合作机构
JOURNAL OF MACHINE LEARNING RESEARCH (2023): 240:1-240:113
arXiv (Cornell University) (2022)
Cited300Views0EIBibtex
300
0
CoRR (2020)
Cited1774Views0EIBibtex
1774
0
NIPS 2020 (2020): 1877-1901
ICML (2020): 1691-1703
Cited1571Views0EIBibtex
1571
0
ICML 2020 (2020): 5006-5019
Cited56Views0EIBibtex
56
0
user-5f1692da4c775ed682f59262(2019)
Cited12103Views0Bibtex
12103
0
arXiv: Learning (2019)
Cited1383Views0EIBibtex
1383
0
Load More
Author Statistics
Co-Author
Co-Institution
D-Core
- 合作者
- 学生
- 导师
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn