Sledovat
Sam Shleifer
Sam Shleifer
Facebook AI Research
E-mailová adresa ověřena na: fb.com
Název
Citace
Citace
Rok
Transformers: State-of-the-art natural language processing
T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ...
Proceedings of the 2020 conference on empirical methods in natural language …, 2020
61892020
Huggingface's transformers: State-of-the-art natural language processing
T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ...
arXiv preprint arXiv:1910.03771, 2019
37032019
Opt: Open pre-trained transformer language models
S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ...
arXiv preprint arXiv:2205.01068, 2022
29902022
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ...
arXiv preprint arXiv:2206.04615, 2022
14202022
Pytorch fsdp: experiences on scaling fully sharded data parallel
Y Zhao, A Gu, R Varma, L Luo, CC Huang, M Xu, L Wright, H Shojanazeri, ...
arXiv preprint arXiv:2304.11277, 2023
2892023
8-bit optimizers via block-wise quantization
T Dettmers, M Lewis, S Shleifer, L Zettlemoyer
arXiv preprint arXiv:2110.02861, 2021
2702021
Efficient large scale language modeling with mixtures of experts
M Artetxe, S Bhosale, N Goyal, T Mihaylov, M Ott, S Shleifer, XV Lin, J Du, ...
arXiv preprint arXiv:2112.10684, 2021
1352021
Automated coronary calcium scoring using deep learning with multicenter external validation
D Eng, C Chute, N Khandwala, P Rajpurkar, J Long, S Shleifer, ...
NPJ digital medicine 4 (1), 88, 2021
1202021
Pre-trained summarization distillation
S Shleifer, AM Rush
arXiv preprint arXiv:2010.13002, 2020
1142020
Huggingface’s transformers: State-of-the-art natural language processing. arXiv
T Wolf, L Debut, V Sanh, J Chaumond, C Delangue, A Moi, P Cistac, ...
arXiv preprint arXiv:1910.03771 10, 2019
992019
Few-shot learning with multilingual generative language models
XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen, D Simig, M Ott, N Goyal, ...
Proceedings of the 2022 conference on empirical methods in natural language …, 2022
812022
Low resource text classification with ulmfit and backtranslation
S Shleifer
arXiv preprint arXiv:1903.09244, 2019
782019
Opt: Open pre-trained transformer language models. arXiv 2022
S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ...
arXiv preprint arXiv:2205.01068, 2023
652023
Normformer: Improved transformer pretraining with extra normalization
S Shleifer, J Weston, M Ott
arXiv preprint arXiv:2110.09456, 2021
652021
Few-shot learning with multilingual language models
XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen, D Simig, M Ott, N Goyal, ...
arXiv preprint arXiv:2112.10668, 2021
622021
Incrementally improving graph wavenet performance on traffic prediction
S Shleifer, C McCreery, V Chitters
arXiv preprint arXiv:1912.07390, 2019
312019
Using small proxy datasets to accelerate hyperparameter search
S Shleifer, E Prokop
arXiv preprint arXiv:1906.04887, 2019
242019
Shen Li
Y Zhao, A Gu, R Varma, L Luo, CC Huang, M Xu, L Wright, H Shojanazeri, ...
Pytorch fsdp: Experiences on scaling fully sharded data parallel, 2023
142023
Efficient language modeling with sparse all-mlp
P Yu, M Artetxe, M Ott, S Shleifer, H Gong, V Stoyanov, X Li
arXiv preprint arXiv:2203.06850, 2022
132022
Automated coronary calcium scoring using deep learning with multicenter external validation. npj Digit Med. 2021; 4: 88
D Eng, C Chute, N Khandwala, P Rajpurkar, J Long, S Shleifer
Přejít k původnímu zdroji... Přejít na PubMed, 0
5
Systém momentálně nemůže danou operaci provést. Zkuste to znovu později.
Články 1–20