Back to Search Start Over

On the Opportunities and Risks of Foundation Models

Authors :
Bommasani, Rishi
Hudson, Drew A.
Adeli, Ehsan
Altman, Russ
Arora, Simran
von Arx, Sydney
Bernstein, Michael S.
Bohg, Jeannette
Bosselut, Antoine
Brunskill, Emma
Brynjolfsson, Erik
Buch, Shyamal
Card, Dallas
Castellon, Rodrigo
Chatterji, Niladri
Chen, Annie
Creel, Kathleen
Davis, Jared Quincy
Demszky, Dora
Donahue, Chris
Doumbouya, Moussa
Durmus, Esin
Ermon, Stefano
Etchemendy, John
Ethayarajh, Kawin
Fei-Fei, Li
Finn, Chelsea
Gale, Trevor
Gillespie, Lauren
Goel, Karan
Goodman, Noah
Grossman, Shelby
Guha, Neel
Hashimoto, Tatsunori
Henderson, Peter
Hewitt, John
Ho, Daniel E.
Hong, Jenny
Hsu, Kyle
Huang, Jing
Icard, Thomas
Jain, Saahil
Jurafsky, Dan
Kalluri, Pratyusha
Karamcheti, Siddharth
Keeling, Geoff
Khani, Fereshte
Khattab, Omar
Koh, Pang Wei
Krass, Mark
Krishna, Ranjay
Kuditipudi, Rohith
Kumar, Ananya
Ladhak, Faisal
Lee, Mina
Lee, Tony
Leskovec, Jure
Levent, Isabelle
Li, Xiang Lisa
Li, Xuechen
Ma, Tengyu
Malik, Ali
Manning, Christopher D.
Mirchandani, Suvir
Mitchell, Eric
Munyikwa, Zanele
Nair, Suraj
Narayan, Avanika
Narayanan, Deepak
Newman, Ben
Nie, Allen
Niebles, Juan Carlos
Nilforoshan, Hamed
Nyarko, Julian
Ogut, Giray
Orr, Laurel
Papadimitriou, Isabel
Park, Joon Sung
Piech, Chris
Portelance, Eva
Potts, Christopher
Raghunathan, Aditi
Reich, Rob
Ren, Hongyu
Rong, Frieda
Roohani, Yusuf
Ruiz, Camilo
Ryan, Jack
Ré, Christopher
Sadigh, Dorsa
Sagawa, Shiori
Santhanam, Keshav
Shih, Andy
Srinivasan, Krishnan
Tamkin, Alex
Taori, Rohan
Thomas, Armin W.
Tramèr, Florian
Wang, Rose E.
Wang, William
Wu, Bohan
Wu, Jiajun
Wu, Yuhuai
Xie, Sang Michael
Yasunaga, Michihiro
You, Jiaxuan
Zaharia, Matei
Zhang, Michael
Zhang, Tianyi
Zhang, Xikun
Zhang, Yuhui
Zheng, Lucia
Zhou, Kaitlyn
Liang, Percy
Publication Year :
2021

Abstract

AI is undergoing a paradigm shift with the rise of models (e.g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks. We call these models foundation models to underscore their critically central yet incomplete character. This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (e.g., language, vision, robotics, reasoning, human interaction) and technical principles(e.g., model architectures, training procedures, data, systems, security, evaluation, theory) to their applications (e.g., law, healthcare, education) and societal impact (e.g., inequity, misuse, economic and environmental impact, legal and ethical considerations). Though foundation models are based on standard deep learning and transfer learning, their scale results in new emergent capabilities,and their effectiveness across so many tasks incentivizes homogenization. Homogenization provides powerful leverage but demands caution, as the defects of the foundation model are inherited by all the adapted models downstream. Despite the impending widespread deployment of foundation models, we currently lack a clear understanding of how they work, when they fail, and what they are even capable of due to their emergent properties. To tackle these questions, we believe much of the critical research on foundation models will require deep interdisciplinary collaboration commensurate with their fundamentally sociotechnical nature.<br />Comment: Authored by the Center for Research on Foundation Models (CRFM) at the Stanford Institute for Human-Centered Artificial Intelligence (HAI). Report page with citation guidelines: https://crfm.stanford.edu/report.html

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2108.07258
Document Type :
Working Paper