Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Superintelligence
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
=== Transformer LLMs and ASI === The rapid advancement of transformer-based LLMs has led to speculation about their potential path to ASI. Some researchers argue that scaled-up versions of these models could exhibit ASI-like capabilities:<ref>{{Cite web |date=2021 |title=On the Opportunities and Risks of Foundation Models |url=https://crfm.stanford.edu/report.html |website=Stanford University |arxiv=2108.07258 |last1=Bommasani |first1=Rishi |last2=Hudson |first2=Drew A. |last3=Adeli |first3=Ehsan |last4=Altman |first4=Russ |last5=Arora |first5=Simran |author6=Sydney von Arx |last7=Bernstein |first7=Michael S. |last8=Bohg |first8=Jeannette |last9=Bosselut |first9=Antoine |last10=Brunskill |first10=Emma |last11=Brynjolfsson |first11=Erik |last12=Buch |first12=Shyamal |last13=Card |first13=Dallas |last14=Castellon |first14=Rodrigo |last15=Chatterji |first15=Niladri |last16=Chen |first16=Annie |last17=Creel |first17=Kathleen |author18=Jared Quincy Davis |last19=Demszky |first19=Dora |last20=Donahue |first20=Chris |last21=Doumbouya |first21=Moussa |last22=Durmus |first22=Esin |last23=Ermon |first23=Stefano |last24=Etchemendy |first24=John |last25=Ethayarajh |first25=Kawin |last26=Fei-Fei |first26=Li |last27=Finn |first27=Chelsea |last28=Gale |first28=Trevor |last29=Gillespie |first29=Lauren |last30=Goel |first30=Karan |display-authors=1 }}</ref> * Emergent abilities β As LLMs increase in size and complexity, they demonstrate unexpected capabilities not present in smaller models.<ref name=":0">{{Cite journal |date=2022-06-26 |title=Emergent Abilities of Large Language Models |journal=Transactions on Machine Learning Research |language=en |arxiv=2206.07682 |issn=2835-8856 |last1=Wei |first1=Jason |last2=Tay |first2=Yi |last3=Bommasani |first3=Rishi |last4=Raffel |first4=Colin |last5=Zoph |first5=Barret |last6=Borgeaud |first6=Sebastian |last7=Yogatama |first7=Dani |last8=Bosma |first8=Maarten |last9=Zhou |first9=Denny |last10=Metzler |first10=Donald |last11=Chi |first11=Ed H. |last12=Hashimoto |first12=Tatsunori |last13=Vinyals |first13=Oriol |last14=Liang |first14=Percy |last15=Dean |first15=Jeff |last16=Fedus |first16=William }}</ref> * In-context learning β LLMs show the ability to adapt to new tasks without fine-tuning, potentially mimicking general intelligence.<ref>{{Cite journal |date=2020 |title=Language Models are Few-Shot Learners |journal=NeurIPS |arxiv=2005.14165 |last1=Brown |first1=Tom B. |last2=Mann |first2=Benjamin |last3=Ryder |first3=Nick |last4=Subbiah |first4=Melanie |last5=Kaplan |first5=Jared |last6=Dhariwal |first6=Prafulla |last7=Neelakantan |first7=Arvind |last8=Shyam |first8=Pranav |last9=Sastry |first9=Girish |last10=Askell |first10=Amanda |last11=Agarwal |first11=Sandhini |last12=Herbert-Voss |first12=Ariel |last13=Krueger |first13=Gretchen |last14=Henighan |first14=Tom |last15=Child |first15=Rewon |last16=Ramesh |first16=Aditya |last17=Ziegler |first17=Daniel M. |last18=Wu |first18=Jeffrey |last19=Winter |first19=Clemens |last20=Hesse |first20=Christopher |last21=Chen |first21=Mark |last22=Sigler |first22=Eric |last23=Litwin |first23=Mateusz |last24=Gray |first24=Scott |last25=Chess |first25=Benjamin |last26=Clark |first26=Jack |last27=Berner |first27=Christopher |last28=McCandlish |first28=Sam |last29=Radford |first29=Alec |last30=Sutskever |first30=Ilya |display-authors=1 }}</ref> * Multi-modal integration β Recent models can process and generate various types of data, including text, images, and audio.<ref>{{Cite journal |date=2022 |title=Flamingo: a Visual Language Model for Few-Shot Learning |journal=NeurIPS|arxiv=2204.14198 |last1=Alayrac |first1=Jean-Baptiste |last2=Donahue |first2=Jeff |last3=Luc |first3=Pauline |last4=Miech |first4=Antoine |last5=Barr |first5=Iain |last6=Hasson |first6=Yana |last7=Lenc |first7=Karel |last8=Mensch |first8=Arthur |last9=Millican |first9=Katie |last10=Reynolds |first10=Malcolm |last11=Ring |first11=Roman |last12=Rutherford |first12=Eliza |last13=Cabi |first13=Serkan |last14=Han |first14=Tengda |last15=Gong |first15=Zhitao |last16=Samangooei |first16=Sina |last17=Monteiro |first17=Marianne |last18=Menick |first18=Jacob |last19=Borgeaud |first19=Sebastian |last20=Brock |first20=Andrew |last21=Nematzadeh |first21=Aida |last22=Sharifzadeh |first22=Sahand |last23=Binkowski |first23=Mikolaj |last24=Barreira |first24=Ricardo |last25=Vinyals |first25=Oriol |last26=Zisserman |first26=Andrew |last27=Simonyan |first27=Karen }}</ref> However, critics argue that current LLMs lack true understanding and are merely sophisticated pattern matchers, raising questions about their suitability as a path to ASI.<ref>{{Cite journal |last=Marcus |first=Gary |date=August 11, 2022 |title=Deep Learning Alone Isn't Getting Us To Human-Like AI |url=https://www.noemamag.com/deep-learning-alone-isnt-getting-us-to-human-like-ai/ |website=Noema}}</ref>
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)