The Possibilities of AI [Entire Talk] - Sam Altman (OpenAI)
Summary
TLDR在斯坦福大学举办的企业家思想领袖研讨会上,Sam Altman,OpenAI的联合创始人兼首席执行官,分享了他对于人工智能发展、创业精神以及个人成长的见解。Altman讨论了他对OpenAI的愿景,包括创建一个能够惠及全人类的通用人工智能(AGI)。他还强调了在技术快速发展的时代,作为企业家和研究者,需要拥有的韧性和自我意识。此外,Altman也提到了对社会影响的考虑,以及如何在创新与责任之间找到平衡。
Takeaways
- 😀 Sam Altman是OpenAI的联合创始人兼CEO,OpenAI是一个研究和部署通用人工智能(AI)的非营利性研究实验室,旨在惠及全人类。
- 🌟 OpenAI推出了聊天机器人Chat GPT,该应用在推出后两个月内就达到了1亿活跃用户,成为历史上增长最快的应用程序。
- 🏆 Sam Altman曾被《时代》杂志评为全球100位最具影响力的人物之一,并在2023年被评为年度CEO,同时他也是福布斯全球亿万富翁名单上的一员。
- 🎓 在斯坦福大学学习计算机科学期间,Sam Altman参与了Y Combinator,并创立了社交移动应用公司Loopt,该公司后来被收购。
- 🤖 Sam Altman认为,目前是至少自互联网以来,或者可能是技术历史上最好的创业时机,因为AI的可能性每年都在变得更加非凡。
- 🚀 如果Sam Altman回到19岁,他会选择从事AI研究,并且可能会选择在产业界而非学术界进行研究,因为产业界拥有大量的计算资源。
- 💡 Sam Altman鼓励人们追求非共识的想法,并相信自己的直觉和思考过程,即使这些想法在一开始可能并不显而易见。
- 🔮 他强调,尽管OpenAI的模型如Chat GPT非常强大,但它们仍然有改进的空间,并且OpenAI致力于通过迭代部署来学习并做得更好。
- 🌐 Sam Altman认为,全球公平地使用AI技术非常重要,并且随着AI模型变得更加强大,我们需要更细致地考虑如何负责任地部署它们。
- 🔑 他提到,随着AI的发展,我们可能会需要重新考虑社会契约,并关注社会如何适应这些新变化,以及我们如何管理这种变化的速度。
- 💡 关于AGI(人工通用智能),Sam Altman认为我们每年都将拥有比当前更强大的系统,但他也提醒人们,对于构建比人类更聪明的系统,我们需要更精确的定义和理解。
Q & A
Sam Altman 是谁,他有哪些成就?
-Sam Altman 是 OpenAI 的联合创始人和 CEO,OpenAI 是一个研究和部署通用人工智能(AI)的非营利性研究实验室,旨在让全人类受益。Sam 曾在斯坦福大学学习计算机科学,并在 Y Combinator 担任过总裁。他还被《时代》杂志评为年度最具影响力人物之一,并被列入福布斯世界亿万富翁名单。
OpenAI 背后的核心理念是什么?
-OpenAI 的核心理念是构建通用目的的人工智能(AGI),以造福全人类。OpenAI 旨在推动 AI 研究的边界,并负责任地部署其技术,以确保其对人类社会产生积极的影响。
Sam Altman 在斯坦福大学的经历对他有什么影响?
-Sam Altman 在斯坦福大学学习计算机科学,并参加了斯坦福的创业思想领袖研讨会(ETL)。这段经历为他后来的创业之路打下了基础,尤其是在他加入 Y Combinator 并创立 Loopt 公司之后。
Sam Altman 如何看待当前的创业环境?
-Sam Altman 认为当前是创业的最佳时机之一,尤其是对于 AI 领域的创业。他认为,随着 AI 技术的发展,将会出现更多创新的产品和服务,并且这个时代的创业机会可能比互联网初期还要多。
Sam Altman 对于想要进入 AI 领域的创业者有什么建议?
-Sam Altman 建议创业者应该追求自己独特的想法,而不是追随他人的足迹。他认为,最有价值的创新往往来自于那些非共识的想法,因此创业者应该学会信任自己的直觉,并勇于尝试那些尚未被广泛认可的创意。
OpenAI 如何平衡研发成本和社会责任?
-OpenAI 通过迭代部署和紧密的反馈循环来平衡研发成本和社会责任。Sam Altman 强调,尽管研发成本可能很高,但如果能够为社会创造更大的价值,那么这些投入是值得的。此外,OpenAI 还通过免费提供某些服务来促进全球范围内的公平访问。
Sam Altman 对于 AGI 的未来有哪些预测?
-Sam Altman 认为 AGI 的发展将会是渐进的,每年都会有更加强大的系统出现。他强调,尽管 AGI 可能会带来巨大的变化,但人类社会的日常生活和工作可能会保持相对稳定。
OpenAI 在 AI 安全和伦理方面采取了哪些措施?
-OpenAI 采取了多种措施来确保 AI 的安全和伦理,包括进行红队测试、外部审计以及与社会紧密合作,以确保 AI 技术的负责任部署。
Sam Altman 如何看待 AI 对全球政治格局的影响?
-Sam Altman 认为 AI 可能会极大地改变全球政治格局,但他没有给出具体的影响方式。他强调 AI 将是一个重要的全球性技术,可能会对国家间的力量平衡产生深远的影响。
OpenAI 如何处理技术进步与社会适应之间的关系?
-OpenAI 认为技术进步应该与社会适应同步进行。Sam Altman 强调,随着 AI 模型变得更加强大,需要更加细致地迭代部署,并保持与社会的紧密反馈,以确保技术的发展能够被社会所接受和适应。
Sam Altman 对于年轻创业者有什么建议?
-Sam Altman 建议年轻创业者要勇于追求自己的独特想法,并且学会信任自己的直觉。他认为,创业者应该在社会中寻找自己的位置,并为社会做出贡献。同时,他也鼓励创业者在面对挑战时展现出韧性和适应性。
Outlines
🎓 斯坦福大学创业思想领袖研讨会开场
斯坦福大学的创业思想领袖(ETL)研讨会由斯坦福技术创新工程中心(STVP)和斯坦福商业协会共同举办。Rvie Balani作为管理科学与工程系讲师及Alchemist加速器的负责人,欢迎了OpenAI的联合创始人兼CEO Sam Altman。Sam Altman在技术和创业领域有着卓越成就,包括在Y Combinator的角色以及创立OpenAI。OpenAI是开发了聊天机器人ChatGBT等革命性技术的非盈利研究实验室。Sam Altman的生活和职业经历表明了他对超越可能性和打破界限的承诺。
🤖 对话Sam Altman:OpenAI的愿景与挑战
在与Sam Altman的对话中,讨论了OpenAI的使命——创建造福全人类的通用人工智能(AGI)。Sam分享了他对于AI未来的看法,包括对AGI的期待和OpenAI在AI技术快速发展中所面临的挑战。他强调了在AI研究和产品开发中保持迭代和社会责任的重要性,同时也提到了OpenAI如何通过提供强大的工具来赋予人们创造未来的能力。
🚀 对话Sam Altman:AI技术的未来与机遇
Sam Altman在讨论中提到了AI技术未来的巨大潜力,他认为当前是创业的最佳时期,特别是对于AI领域。他鼓励年轻创业者和学生把握时机,投身于AI研究和创业。Sam还分享了他对于如何识别和追求非共识想法的看法,强调了独立思考和创新的重要性。
💡 OpenAI的创新与社会影响
Sam Altman讨论了OpenAI如何通过创新对社会产生积极影响。他提到了OpenAI的快速增长,特别是ChatGBT在短时间内达到1亿活跃用户。Sam还提到了他个人的一些荣誉,包括被《时代》杂志评为年度最具影响力人物之一。他强调了OpenAI在推动技术发展的同时,也致力于确保技术进步能够惠及全人类。
🌐 全球创新与AI基础设施的挑战
Sam Altman讨论了全球创新在AI领域面临的挑战,特别是关于如何使世界各地的人们都能访问和利用AI技术。他强调了提供公平的AI技术访问权的重要性,并提到了OpenAI如何通过免费提供ChatGBT等工具来支持这一点。此外,Sam还提到了不同国家建立自己的AI基础设施的重要性,以及OpenAI在帮助实现这一目标方面可能扮演的角色。
🛰️ AI在空间探索和殖民化中的潜在作用
Sam Altman探讨了AI在未来空间探索和殖民化中的潜在作用。他认为,鉴于太空对生物体并不友好,使用机器人进行探索和殖民可能更为合适。Sam对AI在这一领域的应用持乐观态度,并认为AI技术将有助于推动人类在太空探索方面取得更大进展。
💡 创业、创新与非共识思维的重要性
Sam Altman分享了他对于创业和创新的看法,特别是关于如何识别和追求非共识的想法。他强调了在共识和非共识之间找到平衡的重要性,并提到了在追求创新时,需要有独立思考和信任自己直觉的能力。Sam还讨论了如何通过与志同道合的人合作,同时保持独立思考,来推动创新。
🔋 能源需求的未来与可再生能源的角色
Sam Altman对能源需求的未来趋势和可再生能源的角色进行了讨论。他认为能源需求将会增长,并表达了对核聚变作为未来主要电力来源的乐观态度。Sam还提到了太阳能加储能作为另一种可能的能源解决方案,并强调了实现全球规模、成本低廉的能源供应的重要性。
🤝 OpenAI的组织结构和使命
Sam Altman解释了OpenAI独特的组织结构,包括非营利组织拥有营利性公司的模式。他讨论了这种结构的起源和演变,以及它是如何适应OpenAI的使命和目标的。Sam强调了OpenAI的使命——推动AI研究的发展,并确保AI技术的发展能够造福全人类。他还提到了OpenAI如何适应不断变化的环境,并在未来继续调整其结构。
🌍 AGI对地缘政治和全球力量平衡的影响
Sam Altman探讨了AGI可能对全球政治和力量平衡产生的影响。尽管他没有给出具体的答案,但他承认AGI可能会在很大程度上改变国际关系和全球力量的平衡。Sam强调了AI技术的重要性,以及它如何可能被用于塑造未来世界的政治面貌。
🤖 AGI的自我认识和负责任的AI部署
Sam Altman讨论了AGI系统能够认识到自身不确定性和与外界沟通的重要性。他认为,构建能够进行自我反思和识别错误的AI系统是非常重要的。此外,他还强调了在部署AGI时采取负责任的方法的重要性,包括确保AI系统的使用是安全和负责任的。
👥 OpenAI的企业文化和团队精神
Sam Altman分享了OpenAI的企业文化和团队精神。他强调了团队成员之间的紧密合作和对共同使命的忠诚。Sam提到了OpenAI员工在面对挑战时的韧性,以及他们如何共同努力实现公司的长期目标。他还讨论了如何通过共享的使命感和目标来建立强大的企业文化。
🚨 AI的滥用和短期内的挑战
Sam Altman讨论了AI在短期内可能被滥用的问题,以及行业、政府和个人如何共同努力应对这些挑战。他强调了AI技术既有积极的一面也有可能被滥用的风险,并提出了建立反馈循环和平衡安全与自由的重要性。Sam还提到了社会如何共同努力,以确保AI技术被用于促进人类福祉。
🎉 Sam Altman的生日和对未来的展望
在Sam Altman的生日之际,他分享了对未来的展望,特别是对创建比人类更智能的AGI的期待和担忧。Sam认为,虽然创造超越人类的智能可能会让人感到害怕,但这是人类进步和社会发展的一部分。他强调了社会如何通过建立更智能的基础设施来使下一代更加强大,并鼓励人们以积极的态度看待AI技术的发展。
Mindmap
Keywords
💡创业思想领袖研讨会(ETL)
💡OpenAI
💡Y Combinator
💡通用人工智能(AGI)
💡迭代部署
💡计算成本
💡半导体晶圆厂
💡非共识思维
💡人工智能的伦理
💡技术基础设施
💡人类适应性
💡全球创新
💡可再生能源
💡人工智能的自我意识
💡OpenAI的组织结构
💡人工智能的地缘政治影响
💡人工智能的滥用
💡人工智能的自我改进
Highlights
山姆·阿尔特曼(Sam Altman)作为OpenAI的联合创始人和CEO,介绍了OpenAI的使命:构建能够惠及全人类的通用人工智能。
OpenAI推出了聊天机器人Chat GPT,该应用在推出后两个月内就达到了1亿活跃用户,成为历史上增长最快的应用程序。
山姆·阿尔特曼被评为《时代》杂志年度最具影响力人物之一,并在2023年被《时代》杂志评为年度CEO。
山姆·阿尔特曼强调,当前可能是自互联网以来,甚至可能是技术历史上最好的创业时期,因为人工智能的可能性每年都在变得更加非凡。
山姆·阿尔特曼认为,如果回到19岁,他会选择投身于人工智能研究,因为这是一个能够产生巨大影响的领域。
山姆·阿尔特曼提到,他对于建立非常大的计算机和人工智能系统的愿景,这将是人类技术历史上的一个重要里程碑。
OpenAI的商业模式和资金问题被讨论,山姆·阿尔特曼表示,他们更关注于创造社会价值而不是短期的盈利。
山姆·阿尔特曼讨论了人工智能的未来发展,他预测未来几年我们将拥有越来越强大的系统。
关于人工智能的潜在危险,山姆·阿尔特曼表示他更担心的是那些我们可能忽视的微妙危险,而不是那些已经被广泛讨论的灾难性事件。
山姆·阿尔特曼分享了他对人工智能在太空探索和殖民中可能扮演的角色的看法,他认为机器人可能更适合在恶劣的太空环境中工作。
山姆·阿尔特曼讨论了如何识别和追求非共识的想法,并强调了信任自己直觉的重要性。
山姆·阿尔特曼分享了他对人工智能在能源需求变化中的作用的看法,以及他对未来能源生成方式的预测。
山姆·阿尔特曼讨论了OpenAI的组织结构,包括其非营利组织拥有营利公司的俄罗斯套娃模式。
山姆·阿尔特曼强调了OpenAI团队的韧性和对使命的忠诚,他认为这是公司成功的关键文化力量。
山姆·阿尔特曼讨论了人工智能在短期内可能被滥用的问题,以及行业、政府和个人如何共同努力以负责任地部署AI。
山姆·阿尔特曼分享了他对人工智能如何改变世界政治格局和权力平衡的看法,尽管他没有给出具体的答案。
山姆·阿尔特曼讨论了人工智能系统是否能够识别自身的不确定性,并将其传达给外部世界的重要性。
山姆·阿尔特曼对于创造比人类更聪明的实体的前景表示了担忧,但同时也认为这是推动社会前进的积极力量。
Transcripts
[Music]
welcome to the entrepreneurial thought
leader seminar at Stanford
University this is the Stanford seminar
for aspiring entrepreneurs ETL is
brought to you by stvp the Stanford
entrepreneurship engineering center and
basis The Business Association of
Stanford entrepreneurial students I'm
rvie balani a lecturer in the management
science and engineering department and
the director of Alchemist and
accelerator for Enterprise startups and
today I have the pleasure of welcoming
Sam Altman to ETL
um Sam is the co-founder and CEO of open
AI open is not a word I would use to
describe the seats in this class and so
I think by virtue of that that everybody
already play knows open AI but for those
who don't openai is the research and
deployment company behind chat gbt Dolly
and Sora um Sam's life is a pattern of
breaking boundaries and transcending
what's possible both for himself and for
the world he grew up in the midwest in
St Louis came to Stanford took ETL as an
undergrad um for any and we we held on
to Stanford or Sam for two years he
studied computer science and then after
his sophomore year he joined the
inaugural class of Y combinator with a
Social Mobile app company called looped
um that then went on to go raise money
from Sequoia and others he then dropped
out of Stanford spent seven years on
looped which got Acquired and then he
rejoined Y combinator in an operational
role he became the president of Y
combinator from 2014 to 2019 and then in
2015 he co-founded open aai as a
nonprofit research lab with the mission
to build general purpose artificial
intelligence that benefits all Humanity
open aai has set the record for the
fastest growing app in history with the
launch of chat gbt which grew to 100
million active users just two months
after launch Sam was named one of
times's 100 most influential people in
the world he was also named times CEO of
the year in 2023 and he was also most
recently added to Forbes list of the
world's billionaires um Sam lives with
his husband in San Francisco and splits
his time between San Francisco and Napa
and he's also a vegetarian and so with
that please join me in welcoming Sam
Altman to the stage
and in full disclosure that was a longer
introduction than Sam probably would
have liked um brevity is the soul of wit
um and so we'll try to make the
questions more concise but this is this
is this is also Sam's birth week it's it
was his birthday on Monday and I
mentioned that just because I think this
is an auspicious moment both in terms of
time you're 39 now and also place you're
at Stanford in ETL that I would be
remiss if this wasn't sort of a moment
of just some reflection and I'm curious
if you reflect back on when you were
half a lifee younger when you were 19 in
ETL um if there were three words to
describe what your felt sense was like
as a Stanford undergrad what would those
three words be it's always hard
questions
um I was like ex uh you want three words
only okay uh you can you can go more Sam
you're you're the king of brevity uh
excited optimistic and curious okay and
what would be your three words
now I guess the same which is terrific
so there's been a constant thread even
though the world has changed and you
know a lot has changed in the last 19
years but that's going to pale in
comparison what's going to happen in the
next 19 yeah and so I need to ask you
for your advice if you were a Stanford
undergrad today so if you had a Freaky
Friday moment tomorrow you wake up and
suddenly you're 19 in inside of Stanford
undergrad knowing everything you know
what would you do would you drop be very
happy um I would feel like I was like
coming of age at the luckiest time
um like in several centuries probably I
think the degree to which the world is
is going to change and the the
opportunity to impact that um starting a
company doing AI research any number of
things is is like quite remarkable I
think this is probably the best time to
start I yeah I think I would say this I
think this is probably the best time to
start a companies since uh the internet
at least and maybe kind of like in the
history of technology I think with what
you can do with AI is like going to just
get more remarkable every year and the
greatest companies get created at times
like this the most impactful new
products get built at times like this so
um I would feel incredibly lucky uh and
I would be determined to make the most
of it and I would go figure out like
where I wanted to contribute and do it
and do you have a bias on where would
you contribute would you want to stay as
a student um would and if so would you
major in a certain major giving the pace
of of change probably I would not stay
as a student but only cuz like I didn't
and I think it's like reasonable to
assume people kind of are going to make
the same decisions they would make again
um I think staying as a student is a
perfectly good thing to do I just I it
would probably not be what I would have
picked no this is you this is you so you
have the Freaky Friday moment it's you
you're reborn and as a 19-year-old and
would you
yeah what I think I would again like I
think this is not a surprise cuz people
kind of are going to do what they're
going to do I think I would go work on
research and and and where might you do
that Sam I think I mean obviously I have
a bias towards open eye but I think
anywhere I could like do meaningful AI
research I would be like very thrilled
about but you'd be agnostic if that's
Academia or Private Industry
um I say this with sadness I think I
would pick
industry realistically um I think it's I
think to you kind of need to be the
place with so much compute M MH okay and
um if you did join um on the research
side would you join so we had kazer here
last week who was a big advocate of not
being a Founder but actually joining an
existing companies sort of learn learn
the chops for the for the students that
are wrestling with should I start a
company now at 19 or 20 or should I go
join another entrepreneurial either
research lab or Venture what advice
would you give them well since he gave
the case to join a company I'll give the
other one um which is I think you learn
a lot just starting a company and if
that's something you want to do at some
point there's this thing Paul Graham
says but I think it's like very deeply
true there's no pre-startup like there
is Premed you kind of just learn how to
run a startup by running a startup and
if if that's what you're pretty sure you
want to do you may as well jump in and
do it and so let's say so if somebody
wants to start a company they want to be
in AI um what do you think are the
biggest near-term challenges that you're
seeing in AI that are the ripest for a
startup and just to scope that what I
mean by that are what are the holes that
you think are the top priority needs for
open AI that open AI will not solve in
the next three years um yeah
so I think this is like a very
reasonable question to ask in some sense
but I think it's I'm not going to answer
it because I think you should
never take this kind of advice about
what startup to start ever from anyone
um I think by the time there's something
that is like the kind of thing that's
obvious enough that me or somebody else
will sit up here and say it it's
probably like not that great of a
startup idea and I totally understand
the impulse and I remember when I was
just like asking people like what
startup should I start
um but I I think like one of the most
important things I believe about having
an impactful career is you have to chart
your own course if if the thing that
you're thinking about is something that
someone else is going to do anyway or
more likely something that a lot of
people are going to do anyway
um you should be like somewhat skeptical
of that and I think a really good muscle
to build is coming up with the ideas
that are not the obvious ones to say so
I don't know what the really important
idea is that I'm not thinking of right
now but I'm very sure someone in this
room does it knows what that answer is
um and I think learning to trust
yourself and come up with your own ideas
and do the very like non-consensus
things like when we started open AI that
was an extremely non-consensus thing to
do and now it's like the very obvious
thing to do um now I only have the
obvious ideas CU I'm just like stuck in
this one frame but I'm sure you all have
the other
ones but are there so can I ask it
another way and I don't know if this is
fair or not but are what questions then
are you wrestling with that no one else
is talking
about how to build really big computers
I mean I think other people are talking
about that but we're probably like
looking at it through a lens that no one
else is quite imagining yet um
I mean we're we're definitely wrestling
with how we when we make not just like
grade school or middle schooler level
intelligence but like PhD level
intelligence and Beyond the best way to
put that into a product the best way to
have a positive impact with that on
society and people's lives we don't know
the answer to that yet so I think that's
like a pretty important thing to figure
out okay and can we continue on that
thread then of how to build really big
computers if that's really what's on
your mind can you share I know there's
been a lot of speculation and probably a
lot of here say too about um the
semiconductor Foundry Endeavor that you
are reportedly embarking on um can you
share what would make what what's the
vision what would make this different
than it's not just foundies although
that that's part of it it's like if if
you believe which we increasingly do at
this point that AI infrastructure is
going to be one of the most important
inputs to the Future this commodity that
everybody's going to want and that is
energy data centers chips chip design
new kinds of networks it's it's how we
look at that entire ecosystem um and how
we make a lot more of that and I don't
think it'll work to just look at one
piece or another but we we got to do the
whole thing okay so there's multiple big
problems yeah um I think like just this
is the Arc of human technological
history as we build bigger and more
complex systems and does it gross so you
know in terms of just like the compute
cost uh correct me if I'm wrong but chat
gbt 3 was I've heard it was $100 million
to do the model um and it was 100 175
billion parameters gbt 4 was cost $400
million with 10x the parameters it was
almost 4X the cost but 10x the
parameters correct me adjust me you know
it I I do know it but I won oh you can
you're invited to this is Stanford Sam
okay um uh but the the even if you don't
want to correct the actual numbers if
that's directionally correct um does the
cost do you think keep growing with each
subsequent yes and does it keep growing
multiplicatively uh probably I mean and
so the question then becomes how do we
how do you capitalize
that well look I I kind of think
that giving people really capable tools
and letting them figure out how they're
going to use this to build the future is
a super good thing to do and is super
valuable and I am super willing to bet
on the Ingenuity of you all and
everybody else in the world to figure
out what to do about this so there is
probably some more business-minded
person than me at open AI somewhere that
is worried about how much we're spending
um but I kind of
don't okay so that doesn't cross it so
you
know open ey is phenomenal chat gbt is
phenomenal um everything else all the
other models are
phenomenal it burned you've earned $520
million of cash last year that doesn't
concern you in terms of thinking about
the economic model of how do you
actually where's going to be the
monetization source well first of all
that's nice of you to say but Chachi PT
is not phenomenal like Chachi PT is like
mildly embarrassing at best um gp4 is
the dumbest model any of you will ever
ever have to use again by a lot um but
you know it's like important to ship
early and often and we believe in
iterative deployment like if we go build
AGI in a basement and then you know the
world is like kind
of blissfully walking blindfolded along
um I don't think that's like I don't
think that makes us like very good
neighbors um so I think it's important
given what we believe is going to happen
to express our view about what we
believe is going to happen um but more
than that the way to do it is to put the
product in people's hands um
and let Society co-evolve with the
technology let Society tell us what it
collectively and people individually
want from the technology how to
productize this in a way that's going to
be useful um where the model works
really well where it doesn't work really
well um give our leaders and
institutions time to react um give
people time to figure out how to
integrate this into their lives to learn
how to use the tool um sure some of you
all like cheat on your homework with it
but some of you all probably do like
very amazing amazing wonderful things
with it too um and as each generation
goes on uh I think that will expand
and and that means that we ship
imperfect products um but we we have a
very tight feedback loop and we learn
and we get better um and it does kind of
suck to ship a product that you're
embarrassed about but it's much better
than the alternative um and in this case
in particular where I think we really
owe it to society to deploy tively
um one thing we've learned is that Ai
and surprise don't go well together
people don't want to be surprised people
want a gradual roll out and the ability
to influence these systems um that's how
we're going to do it and there may
be there could totally be things in the
future that would change where we' think
iterative deployment isn't such a good
strategy um but it does feel like the
current best approach that we have and I
think we've gained a lot um from from
doing this and you know hopefully s the
larger world has gained something too
whether we burn 500 million a year or 5
billion or 50 billion a year I don't
care I genuinely don't as long as we can
I think stay on a trajectory where
eventually we create way more value for
society than that and as long as we can
figure out a way to pay the bills like
we're making AGI it's going to be
expensive it's totally worth it and so
and so do you have a I hear you do you
have a vision in 2030 of what if I say
you crushed it Sam it's 2030 you crushed
it what does the world look like to
you
um you know maybe in some very important
ways not that different uh
like we will be back here there will be
like a new set of students we'll be
talking about how startups are really
important and technology is really cool
we'll have this new great tool in the
world it'll
feel it would feel amazing if we got to
teleport forward six years today and
have this thing that was
like smarter than humans in many
subjects and could do these complicated
tasks for us and um you know like we
could have these like complicated
program written or This research done or
this business
started uh and yet like the Sun keeps
Rising the like people keep having their
human dramas life goes on so sort of
like super different in some sense that
we now have like abundant intelligence
at our fingertips
and then in some other sense like not
different at all okay and you mentioned
artificial general intellig AGI
artificial general intelligence and in
in a previous interview you you define
that as software that could mimic the
median competence of a or the competence
of a median human for tasks yeah um can
you give me is there time if you had to
do a best guess of when you think or
arrange you feel like that's going to
happen I think we need a more precise
definition of AGI for the timing
question um because at at this point
even with like the definition you just
gave which is a reasonable one there's
that's your I'm I'm I'm paring back what
you um said in an interview well that's
good cuz I'm going to criticize myself
okay um it's it's it's it's too loose of
a definition there's too much room for
misinterpretation in there um to I think
be really useful or get at what people
really want like I kind of think what
people want to know when they say like
what's the timeline to AGI is like when
is the world going to be super different
when is the rate of change going to get
super high when is the way the economy
Works going to be really different like
when does my life change
and that for a bunch of reasons may be
very different than we think like I can
totally imagine a world where we build
PhD level intelligence in any area and
you know we can make researchers way
more productive maybe we can even do
some autonomous research and in some
sense
like that sounds like it should change
the world a lot and I can imagine that
we do that and then we can detect no
change in global GDP growth for like
years afterwards something like that um
which is very strange to think about and
it was not my original intuition of how
this was all going to go so I don't know
how to give a precise timeline of when
we get to the Milestone people care
about but when we get to systems that
are way more capable than we have right
now one year and every year after and
that I think is the important point so
I've given up on trying to give the AGI
timeline but I think every year for the
next many we have dramatically more
capable systems every year um I want to
ask about the dangers of of AGI um and
gang I know there's tons of questions
for Sam in a few moments I'll be turning
it up so start start thinking about your
questions um a big focus on Stanford
right now is ethics and um can we talk
about you know how you perceive the
dangers of AGI and specifically do you
think the biggest Danger from AGI is
going to come from a cataclysmic event
which you know makes all the papers or
is it going to be more subtle and
pernicious sort of like you know like
how everybody has ADD right now from you
know using Tik Tok um is it are you more
concerned about the subtle dangers or
the cataclysmic dangers um or neither
I'm more concerned about the subtle
dangers because I think we're more
likely to overlook those the cataclysmic
dangers uh a lot of people talk about
and a lot of people think about and I
don't want to minimize those I think
they're really serious and a real thing
um but I think we at least know to look
out for that and spend a lot of effort
um the example you gave of everybody
getting add from Tik Tok or whatever I
don't think we knew to look out for and
that that's a really hard the the
unknown unknowns are really hard and so
I'd worry more about those although I
worry about both and are they unknown
unknowns are there any that you can name
that you're particularly worried about
well then I would kind of they'd be
unknown unknown um you can
I I am am worried just about so so even
though I think in the short term things
change less than we think as with other
major Technologies in the long term I
think they change more than we think and
I am worried about what rate Society can
adapt to something so new and how long
it'll take us to figure out the new
social contract versus how long we get
to do it um I'm worried about that okay
um I'm going to I'm going to open up so
I want to ask you a question about one
of the key things that we're now trying
to in
into the curriculum as things change so
rapidly is resilience that's really good
and and you
know and the Cornerstone of resilience
uh is is self-awareness and so and I'm
wondering um if you feel that you're
pretty self-aware of your driving
motivations as you are embarking on this
journey so first of all I think um I
believe resilience can be taught uh I
believe it has long been one of the most
important life skills um and in the
future I think in the over the next
couple of decades I think resilience and
adaptability will be more important
theyve been in a very long time so uh I
think that's really great um on the
self-awareness
question I think I'm self aware but I
think like everybody thinks they're
self-aware and whether I am or not is
sort of like hard to say from the inside
and can I ask you sort of the questions
that we ask in our intro classes on self
awareness sure it's like the Peter duer
framework so what do you think your
greatest strengths are
Sam
uh I think I'm not great at many things
but I'm good at a lot of things and I
think breath has become an underrated
thing in the world everyone gets like
hypers specialized so if you're good at
a lot of things you can seek connections
across them um I think you can then kind
of come up with the ideas that are
different than everybody else has or
that sort of experts in one area have
and what are your most dangerous
weaknesses
um most dangerous that's an interesting
framework for it
uh I think I have like a general bias to
be too Pro technology just cuz I'm
curious and I want to see where it goes
and I believe that technology is on the
whole a net good thing but I think that
is a worldview that has overall served
me and others well and thus got like a
lot of positive
reinforcement and is not always true and
when it's not been true has been like
pretty bad for a lot of people and then
Harvard psychologist David mcland has
this framework that all leaders are
driven by one of three Primal needs a
need for affiliation which is a need to
be liked a need for achievement and a
need for power if you had to rank list
those what would be
yours I think at various times in my
career all of those I think there these
like levels that people go through
um at this point I feel driven by like
wanting to do something useful and
interesting okay and I definitely had
like the money and the power and the
status phases okay and then where were
you when you most last felt most like
yourself I I
always and then one last question and
what are you most excited about with
chat gbt five that's coming out that uh
people
don't what are you what are you most
excited about with the of chat gbt that
we're all going to see
uh I don't know yet um I I mean I this
this sounds like a cop out answer but I
think the most important thing about gp5
or whatever we call that is just that
it's going to be smarter and this sounds
like a Dodge but I think that's like
among the most remarkable facts in human
history that we can just do something
and we can say right now with a high
degree of scientific certainty GPT 5 is
going to be smarter than a lot smarter
than GPT 4 GPT 6 going to be a lot
smarter than gbt 5 and we are not near
the top of this curve and we kind of
know what know what to do and this is
not like it's going to get better in one
area this is not like we're going to you
know it's not that it's always going to
get better at this eval or this subject
or this modality it's just going to be
smarter in the general
sense and I think the gravity of that
statement is still like underrated okay
that's great Sam guys Sam is really here
for you he wants to answer your question
so we're going to open it up hello um
thank you so much for joining joining us
uh I'm a junior here at Stanford I sort
of wanted to talk to you about
responsible deployment of AGI so as as
you guys could continually inch closer
to that how do you plan to deploy that
responsibly AI uh at open AI uh you know
to prevent uh you know stifling human
Innovation and continue to Spur that so
I'm actually not worried at all about
stifling of human Innovation I I really
deeply believe that people will just
surprise us on the upside with better
tools I think all of history suggest
that if you give people more leverage
they do more amazing things and that's
kind of like we all get to benefit from
that that's just kind of great I am
though increasingly worried about how
we're going to do this all responsibly I
think as the models get more capable we
have a higher and higher bar we do a lot
of things like uh red teaming and
external Audits and I think those are
all really good but I think as the
models get more capable we'll have to
deploy even more iteratively have an
even tighter feedback loop on looking at
how they're used and where they work and
where they don't work and this this
world that we used to do where we can
release a major model update every
couple of years we probably have to find
ways to like increase the granularity on
that and deploy more iteratively than we
have in the past and it's not super
obvious to us yet how to do that but I
think that'll be key to responsible
deployment and also the way we kind of
have all of the stakeholders negotiate
what the rules of AI need to be uh
that's going to get more comp Lex over
time too thank you next question where
here you mentioned before that there's a
growing need for larger and larger
computers and faster computers however
many parts of the world don't have the
infrastructure to build those data
centers or those large computers how do
you see um Global Innovation being
impacted by that so two parts to that
one
um no matter where the computers are
built I think Global and Equitable
access to use the computers for training
as well inference is super important um
one of the things that's like very C to
our mission is that we make chat GPT
available for free to as many people as
want to use it with the exception of
certain countries where we either can't
or don't for a good reason want to
operate um how we think about making
training compute more available to the
world is is uh going to become
increasingly important I I do think we
get to a world where we sort of think
about it as a human right to get access
to a certain amount of compute and we
got to figure out how to like distribute
that to people all around the world um
there's a second thing though which is I
think countries are going to
increasingly realize the importance of
having their own AI infrastructure and
we want to figure out a way and we're
now spending a lot of time traveling
around the world to build them in uh the
many countries that'll want to build
these and I hope we can play some small
role there in helping that happen trfic
thank
you U my question was what role do you
envision for AI in the future of like
space exploration or like
colonization um I think space is like
not that hospitable for biological life
obviously and so if we can send the
robots that seems
easier hey Sam so my question is for a
lot of the founders in the room and I'm
going to give you the question and then
I'm going to explain why I think it's
complicated um so my question is about
how you know an idea is
non-consensus and the reason I think
it's complicated is cu it's easy to
overthink um I think today even yourself
says AI is the place to start a company
I think that's pretty
consensus maybe rightfully so it's an
inflection point I think it's hard to
know if idea is non-consensus depending
on the group that you're talking about
the general public has a different view
of tech from The Tech Community and even
Tech Elites have a different point of
view from the tech community so I was
wondering how you verify that your idea
is non-consensus enough to
pursue um I mean first of all what you
really want is to be right being
contrarian and wrong still is wrong and
if you predicted like 17 out of the last
two recessions you probably were
contrarian for the two you got right
probably not even necessarily um but you
were wrong 15 other times and and
and so I think it's easy to get too
excited about being contrarian and and
again like the most important thing to
be right and the group is usually right
but where the most value is um is when
you are contrarian and
right
and and that doesn't always happen in
like sort of a zero one kind of way like
everybody in the room can agree that AI
is the right place to start the company
and if one person in the room figures
out the right company to start and then
successfully executes on that and
everybody else thinks ah that wasn't the
best thing you could do that's what
matters so it's okay to kind of like go
with conventional wisdom when it's right
and then find the area where you have
some unique Insight in terms of how to
do that um I do think surrounding
yourself with the right peer group is
really important and finding original
thinkers uh is important but there is
part of this where you kind of have to
do it Solo or at least part of it Solo
or with a few other people who are like
you know going to be your co-founders or
whatever
um and I think by the time you're too
far in the like how can I find the right
peer group you're somehow in the wrong
framework already um so like learning to
trust yourself and your own intuition
and your own thought process which gets
much easier over time no one no matter
what they said they say I think is like
truly great at this this when they're
just starting out you because like you
kind of just haven't built the muscle
and like all of your Social pressure and
all of like the evolutionary pressure
that produced you was against that so
it's it's something that like you get
better at over time and and and don't
hold yourself to too high of a standard
too early on
it Hi Sam um I'm curious to know what
your predictions are for how energy
demand will change in the coming decades
and how we achieve a future where
renewable energy sources are 1 set per
kilowatt
hour
um I mean it will go up for sure well
not for sure you can come up with all
these weird ways in which
like we all depressing future is where
it doesn't go up I would like it to go
up a lot I hope that we hold ourselves
to a high enough standard where it does
go up I I I forget exactly what the kind
of world's electrical gener generating
capacity is right now but let's say it's
like 3,000 4,000 gwatt something like
that even if we add another 100 gwatt
for AI it doesn't materially change it
that much but it changes it some and if
we start at a th gwatt for AI someday it
does that's a material change but there
are a lot of other things that we want
to do and energy does seem to correlate
quite a lot with quality of life we can
deliver for people
um my guess is that Fusion eventually
dominates electrical generation on Earth
um I think it should be the cheapest
most abundant most reliable densest
source
I could could be wrong with that and it
could be solar Plus Storage um and you
know my guess most likely is it's going
to be 820 one way or the other and
there'll be some cases where one of
those is better than the other but uh
those kind of seem like the the two bets
for like really global scale one cent
per kilowatt hour
energy Hi Sam I have a question it's
about op guide drop what happened last
year so what's the less you learn cuz
you talk about resilience so what's the
lesson you learn from left that company
and now coming back and what what made
you com in back because Microsoft also
gave you offer like can you share more
um I mean the best lesson I learned was
that uh we had an incredible team that
totally could have run the company
without me and did did for a couple of
days
um and you never and also that the team
was super resilient like we knew that a
CRA some crazy things and probably more
crazy things will happen to us between
here and AGI um as different parts of
the world have stronger and stronger
emotional reactions and the stakes keep
ratcheting up and you know I thought
that the team would do well under a lot
of pressure but you never really know
until you get to run the experiment and
we got to run the experiment and I
learned that the team was super
resilient and like ready to kind of run
the company um in terms of why I came
back you know I originally when the so
it was like the next morning the board
called me and like what do you think
about coming back and I was like no um
I'm mad um
and and then I thought about it and I
realized just like how much I loved open
AI um how much I loved the people the C
the culture we had built uh the mission
and I kind of like wanted to finish it
Al
together you you you emotionally I just
want to this is obviously a really
sensitive and one of one of oh it's it's
not but was I imagine that was okay well
then can we talk about the structure
about it because this Russian doll
structure of the open AI where you have
the nonprofit owning the for-profit um
you know when we're we're trying to
teach principal ger entrepreneur we got
here we got to the structure gradually
um it's not what I would go back and
pick if we could do it all over again
but we didn't think we were going to
have a product when we started we were
just going to be like a AI research lab
wasn't even clear we had no idea about a
language model or an API or chat GPT so
if if you're going to start a company
you got to have like some theory that
you're going to sell a product someday
and we didn't think we were going to we
didn't realize we're were going to need
so much money for compute we didn't
realize we were going to like have this
nice business um so what was your
intention when you started it we just
wanted to like push AI research forward
we thought that and I know this gets
back to motivations but that's the pure
motivation there's no motivation around
making money or or power I cannot
overstate how foreign of a concept like
I mean for you personally not for open
AI but you you weren't starting well I
had already made a lot of money so it
was not like a big I mean I I like I
don't want to like claim some like moral
Purity here it was just like that was
the of my life a dver driver okay
because there's this so and the reason
why I'm asking is just you know when
we're teaching about principle driven
entrepreneurship here you can you can
understand principles inferred from
organizational structures when the
United States was set up the
architecture of governance is the
Constitution it's got three branches of
government all these checks and balances
and you can infer certain principles
that you know there's a skepticism on
centralizing power that you know things
will move slowly it's hard to get things
to change but it'll be very very
stable if you you know not to parot
Billy eish but if you look at the open
AI structure and you think what was that
made for um it's a you have a like your
near hundred billion dollar valuation
and you've got a very very limited board
that's a nonprofit board which is
supposed to look after it's it's its
fiduciary duties to the again it's not
what we would have done if we knew then
what we know now but you don't get to
like play Life In Reverse and you have
to just like adapt there's a mission we
really cared about we thought we thought
AI was going to be really important we
thought we had an algorithm that learned
we knew it got better with scale we
didn't know how predictably it got
better with scale and we wanted to push
on this we thought this was like going
to be a very important thing in human
history and we didn't get everything
right but we were right on the big stuff
and our mission hasn't changed and we've
adapted the structure as we go and will
adapt it more in the future um but you
know like you
don't like life is not a problem set um
you don't get to like solve everything
really nicely all at once it doesn't
work quite like it works in the
classroom as you're doing it and my
advice is just like trust yourself to
adapt as you go it'll be a little bit
messy but you can do it and I just asked
this because of the significance of open
AI um you have a you have a board which
is all supposed to be independent
financially so that they're making these
decisions as a nonprofit thinking about
the stakeholder their stakeholder that
they are fiduciary of isn't the
shareholders it's Humanity um
everybody's independent there's no
Financial incentive that anybody has
that's on the board including yourself
with hope and AI um well Greg was I okay
first of all I think making money is a
good thing I think capitalism is a good
thing um my co-founders on the board
have had uh financial interest and I've
never once seen them not take the
gravity of the mission seriously um but
you know we've put a structure in place
that we think is a way to get um
incentives aligned and I do believe
incentives are superpowers but I'm sure
we'll evolve it more over time and I
think that's good not bad and with open
AI the new fund you're not you don't get
any carry in that and you're not
following on investments onto those okay
okay okay thank you we can keep talking
about this I I I know you want to go
back to students I do too so we'll go
we'll keep we'll keep going to the
students how do you expect that AGI will
change geopolitics and the balance of
power in the world um like maybe more
than any
other technology um I don't I I think
about that so much and I have such a
hard time saying what it's actually
going to do um I or or maybe more
accurately I have such a hard time
saying what it won't do and we were
talking earlier about how it's like not
going to CH maybe it won't change
day-to-day life that much but the
balance of power in the world it feels
like it does change a lot but I don't
have a deep answer of exactly how
thanks so much um I was wondering sorry
I was wondering in the deployment of
like general intelligence and also
responsible AI how much do you think is
it necessary that AI systems are somehow
capable of recognizing their own
insecurities or like uncertainties and
actually communicating them to the
outside world I I always get nervous
anthropomorphizing AI too much because I
think it like can lead to a bunch of
weird oversights but if we say like how
much can AI recognize its own
flaws uh I think that's very important
to build and right now and the ability
to like recognize an error in reasoning
um and have some sort of like
introspection ability like that that
that seems to me like really important
to
pursue hey s thank you for giving us
some of your time today and coming to
speak from the outside looking in we we
all hear about the culture and together
togetherness of open AI in addition to
the intensity and speed of what you guys
work out clearly seen from CH gbt and
all your breakthroughs and also in when
you were temporarily removed from the
company by the board and how all the all
of your employees tweeted open air is
nothing without its people what would
you say is the reason behind this is it
the binding mission to achieve AGI or
something even deeper what is pushing
the culture every
day I think it is the shared Mission um
I mean I think people like like each
other and we feel like we've you know
we're in the trenches together doing
this really hard thing um
but I think it really is like deep sense
of purpose and loyalty to the mission
and when you can create that I think it
is like the strongest force for Success
at any start at least that I've seen
among startups um and you know we try to
like select for that and people we hire
but even people who come in not really
believing that AGI is going to be such a
big deal and that getting it right is so
important tend to believe it after the
first three months or whatever and so
that's like that's a very powerful
cultural force that we have
thanks um currently there are a lot of
concerns about the misuse of AI in the
immediate term with issues like Global
conflicts and the election coming up
what do you think can be done by the
industry governments and honestly People
Like Us in the immediate term especially
with very strong open- Source
models one thing that I think is
important is not to pretend like this
technology or any other technology is
all good um I believe that AI will be
very net good tremendously net good um
but I think like with any other tool
um it'll be misused like you can do
great things with a hammer and you can
like kill people with a hammer um I
don't think that absolves us or you all
or Society from um trying to mitigate
the bad as much as we can and maximize
the good
but I do think it's important to realize
that with any sufficiently powerful Tool
uh you do put Power in the hands of tool
users or you make some decisions that
constrain what people in society can do
I think we have a voice in that I think
you all have a voice on that I think the
governments and our elected
representatives in Democratic process
processes have the loudest voice in
that but we're not going to get this
perfectly right like we Society are not
going to get this perfectly right
and a tight feedback loop I think is the
best way to get it closest to right um
and the way that that balance gets
negotiated of safety versus freedom and
autonomy um I think it's like worth
studying that with previous Technologies
and we'll do the best we can here we
Society will do the best we can
here um gang actually I've got to cut it
sorry I know um I'm wanty to be very
sensitive to time I know the the
interest far exceeds the time and the
love for Sam um Sam I know it is your
birthday I don't know if you can indulge
us because I know there's a lot of love
for you so I wonder if we can all just
sing Happy Birthday no no no please no
we want to make you very uncomfortable
one more question I'd much rather do one
more
question this is less interesting to you
thank you we can you can do one more
question
quickly day dear
Sam happy birthday to you
20 seconds of awkwardness is there a
burner question somebody who's got a
real burner and we only have 30 seconds
so make it
short um hi I wanted to ask if the
prospect of making something smarter
than any human could possibly be scares
you it of course does and I think it
would be like really weird and uh a bad
sign if it didn't scare me um humans
have gotten dramatically smarter and
more capable over time you are
dramatically more capable than your
great great grandparents and there's
almost no biological drift over that
period like sure you eat a little bit
better and you got better healthare um
maybe you eat worse I don't know um but
that's not the main reason you're more
capable um you are more capable because
the infrastructure of
society is way smarter and way more
capable than any human and and through
that it made you Society people that
came before you um made you uh the
internet the iPhone a huge amount of
knowledge available at your fingertips
and you can do things that your
predecessors would find absolutely
breathtaking
um Society is far smarter than you now
um Society is an AGI as far as you can
tell and the
the way that that happened was not any
individual's brain but the space between
all of us that scaffolding that we build
up um and contribute to Brick by Brick
step by step uh and then we use to go to
far greater Heights for the people that
come after us um things that are smarter
than us will contribute to that same
scaffolding um you will
have your children will have tools
available that you didn't um and that
scaffolding will have gotten built up to
Greater Heights
and that's always a little bit scary um
but I think it's like more way more good
than bad and people will do better
things and solve more problems and the
people of the future will be able to use
these new tools and the new scaffolding
that these new tools contribute to um if
you think about a world that has um AI
making a bunch of scientific discovery
what happens to that scientific progress
is it just gets added to the scaffolding
and then your kids can do new things
with it or you in 10 years can do new
things with it um but the way it's going
to feel to people uh I think is not that
there is this like much smarter entity
uh because we're much smarter in some
sense than the great great great
grandparents are more capable at least
um but that any individual person can
just do
more on that we're going to end it so
let's give Sam a round of applause
[Music]
5.0 / 5 (0 votes)
Sam Altman在斯坦福大学的全面演讲解析:详细探讨GPT4设定的AI新标准和对GPT5的期待,以及人工智能的未来社会影响和商业模式变革
Breaking down Scarlett Johansson's dispute with OpenAI
In conversation | Geoffrey Hinton and Joel Hellermark
6款工具帮你自动赚钱,轻松上手帮你打开全新的收入渠道,赚钱效率高出100倍,用好这几款AI人工智能工具,你会发现赚钱从来没如此简单过
【人工智能】Google大神Jeff Dean最新演讲 | 机器学习令人兴奋的趋势 | 计算的十年飞跃 | 神经网络 | 语言模型十五年发展 | Gemini | ImageNet | AlexNet
Keynote by NVIDIA CEO Jensen Huang at 2024 SIEPR Economic Summit