WHISTLEBLOWER Reveals Complete AGI TIMELINE, 2024 - 2027 (Q*, QSTAR)
Summary
TLDR这份视频脚本披露了一份据称来自OpenAI内部的机密文件,暗示OpenAI计划在2027年之前实现人工通用智能(AGI)。该文件详细阐述了OpenAI正在训练一个125万亿参数的多模态模型,预计将在2023年12月完成训练。尽管文件中有一些猜测成分,但它提供了相当多的证据和细节,令人怀疑OpenAI可能真的在秘密推进AGI的工作。这份视频探讨了该文件所引发的一系列问题和猜测。
Takeaways
- 😃 视频指出 OpenAI 有一份秘密计划,计划在 2027 年之前实现 AGI(通用人工智能)。
- 🤔 这份文件声称 OpenAI 在 2022 年 8 月开始培训了一个 1.25 万亿参数的多模态模型,并在 2023 年 12 月完成训练。
- 🔍 文件显示 OpenAI 一直在努力创建一个人脑大小(1千万亿参数)的 AI 模型,这是他们实现 AGI 的计划。
- 📈 根据 Chinchilla 定律,即使 1 千万亿参数的模型性能略低于人类,如果使用大量数据进行训练,它就能超过人类水平。
- 😲 一些 AI 领袖如 Hinton 和 Hassabis 近期发出警告,表示 AGI 的到来比预期的要快。
- 🕵️ 微软对 OpenAI 的 100 亿美元投资有望为 OpenAI 提供训练 AGI 系统所需的计算资源。
- 💰 Sam Altman 正在筹集 70 亿美元,可能是为了训练大规模 AGI 系统所需的巨额计算成本。
- ⚠️ 有声音呼吁暂停训练超过 GPT-4 水平的 AI 系统,包括正在训练中的 GPT-5。
- 🔐 OpenAI 计划在 2027 年之前解决"超级对齐"问题,以确保安全释放 AGI。
- 🔜 视频暗示,每年都会有新的GPT模型发布,GPT-7 之后可能就是 AGI 系统。
Q & A
OpenAI计划到2027年创造AGI的文件揭示了哪些关键信息?
-该文件透露,OpenAI从2022年8月开始训练一个具有125万亿参数的多模态模型,首个阶段命名为Q星,模型在2023年12月训练完成,但由于高昂的推理成本,发布被取消。文件还暗示,OpenAI的长期计划是通过逐年发布新模型,最终在2027年达到AGI。
文档中提到的“Q星”、“ARUS”和“GOI”是什么?
-Q星、ARUS和GOI被提及为OpenAI开发的模型名称。其中,ARUS模型的开发被取消,因为它运行效率不高。这些名称被视为OpenAI内部计划和模型的一部分,指示了公司在人工智能领域的研究方向和进展。
为什么OpenAI会将原本计划的GPT-5模型取消或重命名?
-原本计划的GPT-5模型被取消或重命名的具体原因在文档中没有详细说明,但暗示这与模型开发过程中的策略调整和技术挑战有关,可能是由于在推理成本、性能预期或技术突破方面遇到的问题。
Elon Musk对OpenAI的计划提出诉讼有何影响?
-Elon Musk提起的诉讼导致了对OpenAI计划的一些延迟,特别是影响了原计划中的GPT-6和GPT-7的开发和发布。Musk的诉讼主要是基于对OpenAI远离其开源目标和创造高级技术应对公众开放的承诺的担忧。
什么是AGI,以及它与当前AI技术的区别是什么?
-AGI(人工通用智能)是指能够执行任何智力任务的人工智能,与人类智力水平相当。它与当前的AI技术(通常专注于特定任务的解决方案)的主要区别在于其通用性和灵活性,AGI能够在没有特定培训的情况下理解和执行各种复杂任务。
文档如何定义人类水平的AGI,以及它的实现对社会有什么潜在影响?
-文档定义人类水平的AGI为能够执行任何一个聪明人类能够完成的智力任务的人工智能。它的实现可能会彻底改变社会,包括经济、就业、教育和技术发展等方面,同时也引发了对安全、伦理和社会影响的广泛关注。
为什么说模型参数数量是预测AI性能的关键指标?
-模型参数数量被视为预测AI性能的关键指标,因为参数越多,模型处理和理解复杂数据的能力通常越强。参数数量与生物大脑中的突触数目相类比,用来估计AI模型的复杂度和潜在的智能水平。
OpenAI如何通过参数计数和数据量来逼近人类大脑的性能?
-OpenAI通过增加模型的参数计数并训练它们使用大量的数据来逼近人类大脑的性能。通过模仿人脑中神经元间突触的连接方式,以及利用海量数据进行训练,OpenAI旨在创造出能够模拟人类认知过程的AI模型。
文档中提到的“Chinchilla法则”是什么,它如何影响AI模型的训练?
-Chinchilla法则基于DeepMind的研究,指出当前模型训练方式在数据使用方面存在低效,通过使用更多的数据进行训练,可以显著提高AI模型的性能。这一发现促使OpenAI和其他AI研究机构重新评估其训练策略,使用更多数据以期望达到更优的训练效果。
为什么说AI领域的研究是动态且迅速发展的?
-AI领域的研究被认为是动态且迅速发展的,因为每天都有大量的新研究和技术突破被发布,不断推动了AI技术的极限,并改变了我们对可能实现的智能水平的理解。这一领域的快速进步要求研究者、开发者和利益相关者持续关注最新动态,以适应不断变化的技术环境。
Outlines
📄 开篇介绍
视频开始讨论一个揭示OpenAI到2027年创建人工通用智能(AGI)计划的文档。强调了该文档包含许多推测和不完全是事实,但提到了一些关键内容,比如OpenAI于2022年8月开始训练一个125万亿参数的多模态模型,以及由于高昂的推理成本而取消了其发布。视频作者提醒观众以怀疑的态度看待这些信息,同时强调了对OpenAI未来计划的兴趣。
🔍 文档深入解析
深入分析了文档中提及的OpenAI的计划,包括关于GPT 4.5被重命名为GPT 5,以及GPT 5原定于2025年发布但被取消的信息。讨论了一些被提及的模型,如arus和GOI,及其与OpenAI相关文章的联系。同时,文档提及了GPT 6(现称GPT 7)的发布被埃隆·马斯克的诉讼所延迟,以及关于OpenAI的公开资源目标的讨论。
🚀 向AGI迈进
该段落讨论了100万亿参数模型的潜力以及通过参数数量预测AI性能的可行性。提到了一个重要的研究,展示了随着参数数量的增加,AI在不同任务上的表现也随之提升,尽管收益递减。这段落也提到了如何通过增加数据量来弥补100万亿参数模型在性能上的不足,以及OpenAI使用的工程技术来桥接这一差距。
🧠 关于模型的进一步洞察
深入探讨了OpenAI对GPT 4模型的开发计划,包括关于100万亿参数模型的误解和澄清。提到了OpenAI利用更少的模型参数实现更高性能的方法,以及在网络上训练模型以处理更复杂问题的挑战。还讨论了关于GPT 4的早期泄露信息,以及OpenAI如何回应这些泄露信息。
🔧 模型测试与预期
讨论了GPT 4在2022年10月和11月被测试的信息,以及这些测试与先前泄露的一致性。文档还揭示了OpenAI官方对于100万亿参数GPT 4模型的立场,以及关于模型视频处理能力的泄露信息。这段落展示了关于AI性能的预期,尤其是与视频生成相关的能力。
🤖 机器人学与未来展望
这一部分讨论了将AI应用于机器人技术的潜力,尤其是关于视觉数据的重要性和通过大规模模型训练实现复杂任务的可能性。提及了特斯拉在其自动驾驶技术中完全依赖视觉数据的决定,以及通过视觉和语言数据训练的大模型如何处理机器人任务。这强调了通过训练更大模型来实现更高级别的AI性能的趋势。
🌐 AGI和超级对齐的道路
分析了OpenAI对于创建AGI的长期计划和目标,特别是关于通过提高模型的参数数量和训练数据量来实现更接近人类水平的AI性能。这部分还涉及了关于AI研究方向和超级对齐(super alignment)问题的深入讨论,以及知名AI研究人员对于AI发展速度和潜在风险的担忧。
💡 综合理解与未来展望
最后一段深入探讨了OpenAI在创建AGI方面的总体战略和计划,包括预计每年发布一个新的GPT模型,直到2027年。讨论了OpenAI对超级对齐技术的期望,以及如何通过逐步发布更新和更强大的模型来为社会和技术环境提供适应AGI的时间。同时,也反思了关于AGI发展的不确定性和可能的未来趋势。
Mindmap
Keywords
💡人工通用智能(AGI)
💡GPT(生成式预训练转换器)
💡参数量
💡计算资源
💡数据
💡多模态
💡超级对齐(Superintelligence Alignment)
💡泄露(Leak)
💡缩放定律(Scaling Laws)
💡风险意识
Highlights
There was a recent document that apparently reveals OpenAI's secret plan to create AGI by 2027.
The document states that OpenAI started training a 125 trillion parameter multimodal model called 'qstar' in August 2022, which finished training in December 2023 but the launch was cancelled due to high inference cost.
Multiple AI researchers and entrepreneurs claim to have had inside information about OpenAI training models with over 100 trillion parameters, intended for AGI.
OpenAI's president Greg Brockman stated in 2019 that their plan was to build a human brain sized model within 5 years to achieve AGI.
AI leaders like Demis Hassabis and Geoffrey Hinton have recently expressed growing concerns about the potential risks of advanced AI capabilities.
After the release of GPT-4, the Future of Life Institute released an open letter calling for a 6-month pause on training systems more powerful than GPT-4, including the planned GPT-5.
Sam Altman confidently stated that there is enough data on the internet to create AGI.
OpenAI realized their initial scaling laws were flawed and have adjusted to take into account DeepMind's 'Chinchilla' laws, which show vastly more data can lead to massive performance boosts.
While a 100 trillion parameter model may be slightly suboptimal, OpenAI plans to use the Chinchilla scaling laws and train on vastly more data to exceed human-level performance.
Microsoft invested $10 billion into OpenAI in early 2023, providing funds to train a compute-optimal 100 trillion parameter model.
An OpenAI researcher's note reveals they were working on preventing an AI system called 'qstar' from potentially destructive outcomes.
The document theorizes that OpenAI plans to release a new model each year until 2027, aligning with their 4-year timeline to solve the 'super alignment' problem for safe AGI release.
GPT-7 is speculated to be the last pre-AGI model before GPT-8 achieves full AGI capability in 2027.
Sam Altman mentioned OpenAI can accurately predict model capabilities by training less compute-intensive systems, potentially forecasting the path to AGI.
Altman is reportedly trying to raise $7 trillion, likely to fund the immense compute required for training a brain-scale AGI model using the Chinchilla scaling laws.
Transcripts
so there was a recent document that
actually apparently reveals open ai's
secret plan to create AGI by 2027 now
I'm going to go through this document
with you Page by Page I've read it over
twice and there are some key things that
actually did stand out to me so without
further Ado let's not waste any time and
of course just before we get into this
this is of course going to contain a lot
of speculation remember that this
document isn't completely 100% factual
so just take this video with a huge of
salt so you can see here the document
essentially says revealing open ai's
plan to create AGI by 2027 and that is a
rather important date which we will come
back to if we look at this first thing
you can see there's an introduction okay
and of course remember like I said there
is a lot of speculation in this document
there are a lot of different facts and
of course like I said anyone can write
any document and submit it to um you
know Twitter or Reddit or anything but I
think this document does contain a
little bit more than that so it starts
out by stating that in this document I
will be revealing information I have
gathered regarding opening eyes delayed
plans to create human level AGI by 2027
not all of it will be easily verifiable
but hopefully there's enough evidence to
convince you summary is basically that
openai has started training a 125
trillion parameter multimodal model in
August of 2022 and the first stage was a
rakus also called qstar and the model
finished training in December of 2023
but the launch was cancelled due to the
high inference cost
and before you guys think it's just
document with like just words I'm going
to show you guys later on like all of
the crazy kind of stuff that is kind of
verifiable that does actually um line up
with some of the stuff that I've seen as
someone that's been paying attention to
this stuff so this is literally just the
introduction um the juicier stuff does
come later but essentially they actually
talk about the and this is just like an
overview so you're going to want to
continue watching they essentially state
that you know this is the original GPT 5
which was planned for release in 2025
bobi GPT 4.5 has been renamed name to
gbt 5 because the original gbt 5 has
been cancelled now I got to be honest
this paragraph here is a little bit
confusing um but I do want to say that
the words arus and the words GOI are
definitely models that were referred to
by several articles that were referring
to leaks from open eye and I think they
were actually on the information so this
is some kind of stuff that I didn't
really hear that much about but the
stuff that I did hear was pretty crazy
so um this arus and this goby thing
although you might not have heard a lot
about it of course it is like like a
kind of like half and half leak but like
I was saying this stuff is kind of true
so you can see here open AI dropped work
on a new arus model in rare AI setback
and this one actually just talks about
um how by the middle of open AI you know
scrapping an araus launch after it
didn't run as efficiently so there's
actually some references to this but a
lot of the stuff is a little bit
confusing but we're going to get on to
the main part of this story now I just
wanted to include that just to show you
that you know these names aren't made up
because if I was watching this video for
the first time and I hadn't seen some of
the prior articles before I'd be
thinking what on Earth is a r what on
Earth is GOI I've only heard about qar
so essentially let's just take a look
and it says the next stage of qstar
originally GPT 6 but since renamed
gpt7 originally for release in 2026 has
been put on hold because of the recent
lawsuit by Elon Musk if you haven't been
paying attention to the space
essentially Elon Musk just f a lawsuit
released a video yesterday um stating
that open ey have strayed far too long
from their goals and if they are
creating some really advanced technology
the public do deserve to have it open
source because that was their goal um
and essentially you can see here it says
qar GBC planned to be released in 2027
achieving full AGI and one thing that I
do want to say about this because
essentially they're stating that you
know they're doing this up to gpt7 and
then after gpt7 they do get to AGI one
thing that I do think okay and I'm going
to come back to this as well is that the
dates kind of do line up and I say kind
of because not like 100% because we
don't know but presuming let's just
presume okay because GPT 4 was released
in 2023 right um let's just say you know
every year release a new model okay um
that would mean that you know in 2024 we
would get gbt 5 in 2025 we get GPT 6 in
2026 we would get GPT 7 and in 2027 we
would get GPT 8 which is of course AGI
now one thing I do think about this that
is kind of interesting and remember I'm
going to come back to this so pay
attention Okay what I'm basically saying
is that if openi are consistent with
their year releases so for example if
they are going to release a new model
every year and if we continue at the
same rate like a new GPC every single
year which is possible him stating that
gp7 being the last release before GPT 8
which is Agi does actually kind of make
sense because once again and I know you
guys are going to hate this but if we
look at the trademarks okay remember
that they trademarked this around the
same time okay around that 2023 time
when all of this crazy stuff was going
on and I think it's important to note as
well is that like there's no gp8 you
might might argue that if they're going
to use all the GPT names why wouldn't
they just trademark GPT a and I think
maybe because like the document States
the model after gpt7 could be AGI and
I'm going to give you guys another
reason on top of that um another reason
is and I'm going to show you guys that
later on in the video but essentially um
open ey's timeline on super alignment
actually does coincide with this Theory
which is a little bit Co coincidental of
course like I said pure speculation
could be completely false open ey like I
said before can go ahead and completely
change their entire plans you know they
can go ahead and drop two models in one
year the point I'm trying to make is
that um certain timelines do align but
just remember this because I'm going to
come back to this because of some
documents stuff that you're going to see
in this document at the end of the video
so anyways um you know it says Elon Musk
caused a delay because of his lawsuit
this why I'm revealing the information
now because no further harm can be done
so I guess Elon musk's lawsuit has kind
of um you know if you wanted bought you
some time so he says I've seen many
definitions of AGI artificial general
intelligence but I will Define AGI
simply as an artificial general
intelligence that can do any
intellectual task a smart human can this
is how most people Define the term now
2020 was the first time that I was
shocked by an AI system so this is just
some um of course you know talk about
his experience with you know AI systems
I'm guessing the person who wrote this
but you know AGI if you don't know AGI
is like an AI system that can do any
task human can but one thing that is
important to discern is that you know
AGI there was a recent paper that
actually talks about the levels of AGI
and I think it's important to remember
that AGI isn't just you know one AI that
can do absolutely everything there are
going to be levels to this AGI system
that we've seen so far and in this paper
levels of AGI they actually talk about
how you know we're already at emerging
AGI which is you know emerging which is
equal or somewhat better than an
unskilled human so we are at level one
AGI and then of course we've got um you
know competent AGI which is going to be
at least 50% of the 50th percentile of
skilled adults and that's competent AGI
that's not yet achieved and then of
course we've got expert AGI which is
90th percentile of skilled adults which
is not yet achieved then we've got
virtuoso AGI which is not yet achieved
which is 99th percentile of all skilled
adults and then we've got artificial
super intelligence which is just 100% so
I think it's important to understand
that there are these levels to AGI
because once someone says AGI I mean is
it 90 9 can it do like half you know
it's like it's just it's just pretty
confusing but I think this is uh a
really good framework for actually
looking at the definition because trust
me it's an industry standard but it is
very very confusing so here's where he
basically says that you know um you know
GPT 3.5 which powered the famous chat
GPT and of course gpt3 which was the not
the successor but the predecessor of 3.5
it says you know these were a massive
step forward towards AGI but the note is
you know gbt2 and all chatbot since
Eliza had no real ability to respond
coherently so while such gpt3 a massive
leap and of course this is where we talk
about parameter count and of course he
says deep learning is a concept that
essentially goes back to the beginning
of AI research in the 1950s first new
network was created in the 50s y y y so
basically this is where he's giving the
description of a parameter and he says
you may already know but to give a brief
digestible summary it's a nalist to a
synapse in a biological brain which is a
connection between neurons and each
neuron in a biological brain has roughly
a thousand connections to other neurons
obviously digital networks OB vular to
biological brains basically saying that
you know of course we're comparing them
but different but um how many synapses
or parameters are in a human brain the
most commonly cited figure for synapse
count in the brain is roughly 100
trillion which would mean each neuron is
100 billion in the human brain has
roughly 1,000 connection and remember um
this number 100 trillion because it's
going to actually be a very very big
number that uh you do need to remember
so of course you can see here the human
brain consists of 100 billion urans and
over 100 trillion synaptic connections
okay and essentially this is trying to
you know um just pair the similarities
between parameters and synapses so
entially stating here that you know if
each neuron in a brain and trust me guys
this is just all going to come into
everything like I know you guys might be
thinking what is the point of talking
about this I just want to hear about qar
but just trust me all of this stuff it
does actually make sense like I've read
this a lot of times so I'm going to skip
some pages but the pages I'm talking
about now just trust me guys you're
going to want to read them it basically
says here if each neuron in a brain has
a th000 connections this means a cat has
roughly 250 billion synapses and a dog
has roughly 530 billion synapses synapse
count generally seems to predict to
intelligence with a few exceptions for
instance elephants techn have a higher
signups count than humans but yet
display lower intelligence of course
basically here's where he's actually
talking about how you know the simplest
explanation for larger signups accounts
with lower intelligence is a smaller
amount of quality data and from an
evolutionary perspective brains are
quote unquote trained on billions of
years of epigenetic data and human
brains evolve from higher quality
socialization communication data than
elephants leading to our Superior
ability to reason but the point he's
trying to make here is that you know
while there are nuances that you know
don't make sense synapse count is
definitely important and I think we've
definitely seen that um with the
similarities in the parameter size with
the explosion of llms and what we've
seen in these multimodal models and
their capabilities and it says again the
explosion in a capabilties since the
early 20110 has been a result of far
more computing power and far more data
gbt2 had 1.5 billion connections Which
is less than a mouse's brain and DBT 3
had 175 billion connections which is get
somewhat closer to a cat's brain and
obviously it's intuitively obvious that
an AI system the size of a cat's brain
would be superior to a system than the
size of a mouse's brain so so here's
where things start to get interesting so
he says in 2020 after the release of the
175 billion parameter gbt 3 many
speculated about the potential
performance of a Model 600 times larger
at 100 trillion parameters just remember
this number because this number is about
to just keep you know repeating in your
head and of course he says the big
question is is it possible to predict AI
performance by parameter count and as it
turns out the answer is yes as you'll
see on the next page and this is where
he actually references this article
which is called extrapolating GPT and
performance by lrien and it was not
score written in 2022 and basically it
talks about how as you scale up in
parameter count you approach Optimal
Performance so essentially this graph
seems to be illustrating the
relationship between neuron networks
measured by the number of parameters
which can be thought of as the strength
of connections between neurons and their
performance on various tasks and these
tasks included language related
challenges like translation read and
comprehension and question and answering
among others and the performance on
these task is measured in the vertical
axis higher values indicating better
performance and the graph shows that as
the number of parameters in increases
the performance on these tasks also
tends to but of course it does have
diminishing returns As you move right
because the curves actually do tend to
Plateau as they reach the higher
parameter counts of course the various
colors on this chart just essentially
represent different tasks and each dot
on those lines represents a neural
network model of a certain size and
certain parameter count being tested on
that you can see right down here this is
where the trained G you can see the gbt
performance so it says flop us TR at gp3
and then of course you can see right
here this is apparently the number of
synap in the brain and just remember the
number 100 trillion or 200 trillion
because it's going to be really
important so s it then says as Lan
Illustrated extrapolations show that air
performance inexplicably seems to reach
human level at the same time as a human
level brain size is matched with the
parameter count his count for the
synapse the brain is roughly 200
trillion parameters as opposed to the
commonly cited 100 trillion figure but
the point still stands 100 trillion
parameters is remarkably close to
Optimal by the way an important thing to
not is that although 100 trillion is
slightly suboptimal in performance there
is an engineering technique that openi
is using to bridge this cap and I'll
explain this towards the very end of
this document because it is crucial to
open ey is building and Lan's post is
one of many similar posts online it's an
extrapolation of Performance Based on
the jump between previous models and
open ey certain has much more detailed
metrics and they've come to the same
conclusion as lanon as I'll show later
in this document so if AI performance is
predictable based on parameter count and
100 trillion parameters is enough for
human level performance when will 100
trillion parameter AI model be released
in the future so here's where we go okay
it says that gbt 5 achieved Proto AGI in
late 2023 with an IQ of 48 now that is a
statement that um you know with the IQ