Follow
Suchin Gururangan
Suchin Gururangan
Verified email at cs.washington.edu - Homepage
Title
Cited by
Cited by
Year
Don't stop pretraining: Adapt language models to domains and tasks
S Gururangan, A Marasović, S Swayamdipta, K Lo, I Beltagy, D Downey, ...
arXiv preprint arXiv:2004.10964, 2020
19302020
Annotation artifacts in natural language inference data
S Gururangan, S Swayamdipta, O Levy, R Schwartz, SR Bowman, ...
arXiv preprint arXiv:1803.02324, 2018
11222018
Realtoxicityprompts: Evaluating neural toxic degeneration in language models
S Gehman, S Gururangan, M Sap, Y Choi, NA Smith
arXiv preprint arXiv:2009.11462, 2020
6922020
All that's' human'is not gold: Evaluating human evaluation of generated text
E Clark, T August, S Serrano, N Haduong, S Gururangan, NA Smith
arXiv preprint arXiv:2107.00061, 2021
2562021
Show your work: Improved reporting of experimental results
J Dodge, S Gururangan, D Card, R Schwartz, NA Smith
arXiv preprint arXiv:1909.03004, 2019
2432019
Editing models with task arithmetic
G Ilharco, MT Ribeiro, M Wortsman, S Gururangan, L Schmidt, ...
arXiv preprint arXiv:2212.04089, 2022
1242022
Variational pretraining for semi-supervised text classification
S Gururangan, T Dang, D Card, NA Smith
arXiv preprint arXiv:1906.02242, 2019
1232019
Detoxifying language models risks marginalizing minority voices
A Xu, E Pathak, E Wallace, S Gururangan, M Sap, D Klein
arXiv preprint arXiv:2104.06390, 2021
912021
Demix layers: Disentangling domains for modular language modeling
S Gururangan, M Lewis, A Holtzman, NA Smith, L Zettlemoyer
arXiv preprint arXiv:2108.05036, 2021
832021
Branch-train-merge: Embarrassingly parallel training of expert language models
M Li, S Gururangan, T Dettmers, M Lewis, T Althoff, NA Smith, ...
arXiv preprint arXiv:2208.03306, 2022
782022
Time waits for no one! analysis and challenges of temporal misalignment
K Luu, D Khashabi, S Gururangan, K Mandyam, NA Smith
arXiv preprint arXiv:2111.07408, 2021
502021
Nearest neighbor zero-shot inference
W Shi, J Michael, S Gururangan, L Zettlemoyer
Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022
312022
Analysis of graph invariants in functional neocortical circuitry reveals generalized features common to three areas of sensory cortex
SS Gururangan, AJ Sadovsky, JN MacLean
PLoS computational biology 10 (7), e1003710, 2014
172014
Silo language models: Isolating legal risk in a nonparametric datastore
S Min, S Gururangan, E Wallace, H Hajishirzi, NA Smith, L Zettlemoyer
arXiv preprint arXiv:2308.04430, 2023
132023
Scaling expert language models with unsupervised domain discovery
S Gururangan, M Li, M Lewis, W Shi, T Althoff, NA Smith, L Zettlemoyer
arXiv preprint arXiv:2303.14177, 2023
132023
M2D2: A massively multi-domain language modeling dataset
M Reid, V Zhong, S Gururangan, L Zettlemoyer
arXiv preprint arXiv:2210.07370, 2022
122022
Emergent coordination underlying learning to reach to grasp with a brain-machine interface
M Vaidya, K Balasubramanian, J Southerland, I Badreldin, A Eleryan, ...
Journal of neurophysiology 119 (4), 1291-1304, 2018
112018
Whose language counts as high quality? measuring language ideologies in text data selection
S Gururangan, D Card, SK Dreier, EK Gade, LZ Wang, Z Wang, ...
arXiv preprint arXiv:2201.10474, 2022
102022
Less: Selecting influential data for targeted instruction tuning
M Xia, S Malladi, S Gururangan, S Arora, D Chen
arXiv preprint arXiv:2402.04333, 2024
42024
lo-fi: distributed fine-tuning without communication
M Wortsman, S Gururangan, S Li, A Farhadi, L Schmidt, M Rabbat, ...
arXiv preprint arXiv:2210.11948, 2022
32022
The system can't perform the operation now. Try again later.
Articles 1–20