本集简介
双语字幕
仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。
四字神名。
Tetragrammaton.
我有一个非常清晰的记忆。
I have this really clear memory.
那是2010年。
It was 2010.
我当时在英国,我长大的地方。
I was in England where I grew up.
那时我是一名记者。
I was a journalist at the time.
我开始撰写关于神经网络的内容。
I'd started writing about neural network stuff.
我住在伦敦东南区一间非常昏暗的公寓里,租金极低,而且我有一台录音机。
I was in this very dingy apartment in Southeast London, extremely cheap, and I had a dictaphone.
我当时录下自己说:如果谷歌建造了所有这些计算机,它们就能在上面运行机器学习算法,谷歌将会构建出极其强大的人工智能。
I was recording myself saying, if Google builds all of these computers, then they can run machine learning algorithms on it and Google is going to build really powerful AI.
我对这个想法着了迷。
I became obsessed with this idea.
我经常参观数据中心。
I used to tour data centers.
我做了一个系列,叫作‘云的克拉克之侧’,走访了欧洲各地的数据中心。
I had a series called the Clarke Side of the Cloud where I'd visit data centers around Europe.
拍视频吗?
Film it?
拍照。
Take photos.
但这就像是我在评测录音设备一样,真的很酷。
But it would be like if I was reviewing recording It's really cool.
我进去就会问:你们这儿有什么?
I'd come in and be like, what do you got here?
你一直都是个硬件迷吗?
What's the Were you always a gearhead?
通过那段经历,我成了一个硬件迷。
I became a gearhead through that.
小时候,我经常摆弄各种东西,也拥有很多硬件设备。
As a kid, I tinkered a lot with lots of things and had hardware.
通过那段经历,我了解到一个关于数据中心最有趣的冷知识,但几乎没人知道:数据中心需要电力。
Through that, I learned the best fun fact about data centers no one knows, which is data centers need power.
它们内部都配备了备用发电机,以防断电。
They all have backup generators inside them in case the power shuts off.
在欧洲,这些备用发电机大多是二战时期潜艇用的柴油发电机,因为这些发电机的设计比当时市面上任何其他发电机都更不容易出故障。
In Europe, those backup generators are mostly diesel generators from World War II style submarines because those generators were designed to like go wrong less often than any other generator you could buy.
总之,我花了多年时间研究这些内容,并且对人工智能着迷。
Anyway, I spent years studying all of this stuff and I'm obsessed with AI.
2012年,我在英国遇到了一位美国女性。
And I meet a woman in England in 2012, who's American.
我告诉她,我不能和她在一起,因为我刚申请了一份在旧金山的工作,我必须去那里撰写关于人工智能的内容。
And I told her that I couldn't be with her because I just applied for a job in San Francisco because I needed to go out here and write about AI.
当我没有得到那份工作时,这其实是一种幸运,因为她几个月后不得不回美国。
And when I didn't get the job, which was a blessing, she had to go back to America a few months later.
然后我重新申请,最终加入了最初拒绝我的那家媒体机构,于是我们搬到了这里。
And then I reapplied and I got a job as the journalist outfit that had rejected me originally, so we moved here.
在旧金山,我自称是世界上唯一的神经网络记者。
And then in San Francisco, called myself the world's only neural network reporter.
是的。
Yeah.
那时候做这件事非常容易。
It was very easy to do back then.
太好了。
Great.
我开始专注于报道这一领域。
And I just started covering it.
让我感到惊讶的是,我感受到一场无声而彻底的革命正在发生。
And what amazed me was this feeling that there was a silent and total revolution occurring.
我通常会做的是,我
And what I would do is I
你能清楚地看到,但别人却没注意到。
You could see it clearly, but others did not see that.
是的。
Yeah.
我认为这是因为我很执着,我当时在阅读研究论文。
I think that this is because I'm obsessive and what I was doing is I was reading the research papers.
像ImageNet这样的研究论文会说,‘等等等等,这是得分。’
The research papers for things like ImageNet say, they'd say yada yada yada, here's the score.
但几个月后,又会有其他关于ImageNet的研究论文发表,得分可能就不同了。
But then there would be other research papers about ImageNet that come out months later and then may have a different score.
更高。
Higher.
更高。
Higher.
我开始制作这些图表,只是把上升的线条画出来。
I started making these graphs where I was just plotting the lines going up.
某一刻,我心想:天哪。
At some point, was like, Oh my God.
我所关注的每一个使用神经网络的领域,都看到线条在不断上升。
Every domain I look at that is using neural networks, am seeing the lines go up everywhere.
所有这些东西都依赖计算机。
All of this stuff uses computers.
猜猜是谁一直在制造大量计算机?
Guess who's been building lots of computers?
就是谷歌和其他那些公司,它们都在源源不断地推出这些研究。
It's like Google and the other companies which are all pumping out this research.
我开始痴迷于一个想法:我必须投身于这个领域。
I became obsessed with the idea I had to be working in this field.
我读了所有相关的东西。
I read everything.
我晚上回家后就会读这些内容。
I'd go home at night and read this stuff.
我会为了好玩而实现神经网络。
I'd implement neural networks for fun.
你对这些应用场景有什么想法吗?还是只是觉得它在构建中很重要?
Do you have any thoughts about what the use cases would be or just that it was something that was important in building?
即使那时,我也能看出计算机视觉会是一个应用场景。
I could see computer vision as being a use case even then.
能够说出手机看到的东西将会很有价值。
Just being able to say what your phone sees was going to be valuable.
我明白了。
I see.
但我一直痴迷于我们会创造出智能生命这个想法。
But I was kind of obsessed with the idea we were going to make intelligent beings.
像机器人那样?
Like robots?
不,是合成的思维。
No, like synthetic minds.
哦,我明白了。
Oh, I see.
是的。
Yeah.
我当时想,这些东西很难,但似乎真的有效。
I was like, well, this stuff's hard and seems to be working.
是的。
Yeah.
后来OpenAI出现了,2015年在加拿大宣布时,我是会上唯一一位记者。
So OpenAI came along and I was pretty much the only journalist at the conference in Canada where it was announced in 2015.
我记得我采访了联合创始人之一格雷格·布罗克曼,做完采访后,我关掉了录音设备,然后问:我怎么才能加入你们?
And I remember I talked to Greg Brockman, one of the co founders, did this interview, and I sort of turned my dictaphone off, then I was like, How can I work for you?
那一刻我就知道,我不能再当记者了。
At which point I knew I could no longer be a journalist.
你真的打破了中立性。
You've really broken impartiality.
所以在接下来的几个月里,我们进行了多次交流,我要特别感谢萨姆·阿尔特曼、格雷格和其他人,他们似乎真的愿意冒险接纳我,因为我非常有动力。
And so over the following months, we had conversations and then, you know, shout out to Sam Altman and Greg and others, but they really took a chance on me, I think, because I was so motivated.
那年夏天,我加入了他们。
And I joined them that summer.
我收到了许多记者发来的邮件,说你正在犯一生中最糟糕的错误,诸如此类的话。
And I have all of these emails from journalists saying, You are making the worst mistake of your life, all kinds of things.
是的,因为你之前有一份稳定的工作。
Yeah, because you've had a steady job.
我去了彭博社,偶尔也为《商业周刊》撰稿。
I'd got to Bloomberg and was occasionally writing for Business Week.
那是一份了不起的职位,我从伦敦的一处市政公寓一路奋斗至此,所以我觉得这是一项了不起的成就。
It was like a big, and I'd written my way there from a council flat in London, so it felt like a big achievement.
那时,人工智能还被看作是?
At that time, AI was still viewed as?
谁在乎?
Who cares?
你在干什么?
What are you doing?
这一切都没道理。
None of this makes sense.
在接下来的几年里,我在OpenAI亲眼见证了它真正发挥作用。
Over the following years at OpenAI, got to see it really work.
我看到了它在文本上的应用效果。
I got to see how it worked for text.
我看到它在非常先进的电子游戏中也奏效了。
I saw it work for really advanced video games.
我见证了那里的研究文化:每当我们增加用于解决问题的计算资源,事情就会变得更好,开始真正起作用。
I saw what the research culture was like of how anytime we scaled up the amount of compute we could throw at a problem, things got better and things started working.
然后,和我一起创立Anthropic的同事们做了一些关于所谓‘扩展定律’的研究,他们指出:如果我们观察投入这些系统的计算量、数据量以及参数数量,就能得到一些图表,这些图表为我们提供了未来的指导方案。
And then my colleagues who went on to found Anthropic with me did some research on something called scaling laws where they actually just said, if we look at like the amount of compute we spend on these systems and how much data we put in and the number of parameters in the systems, we get graphs that give us a future recipe as well.
如果我们想要一个得分为X的系统,就需要Y的算力和Z的数据。
If we wanted to get a system that scored X, we'd need compute Y and like data Z.
我们来试试吧。
Let's try that.
然后他们训练了Claude和ChatGPT的前身,叫做GPT-3,这是一个语言模型。
And then they trained this precursor to Claude and ChatGPT called GPT-three, which was a language model.
语言模型的训练成本极高,但我们因为存在扩展定律而决定投入训练。
A language model is so expensive that we justified training it because the scaling laws existed.
我们当时想,它会达到这个分数。
We were like, it will get this score.
我们可以告诉你。
We can tell you.
当它成功的时候,我非常震惊。
And the moment it worked, I was pretty shocked.
我们创立Anthropic是因为感觉列车即将发车。
We started Anthropic because we felt like the trains were about to leave the station.
我明白了。
I see.
一旦你确认它有效,并且已经证明了一次,就像莱特兄弟那样,一切即将改变。
Once you know it works and you've proven it once, like the Wright brothers, everything's about to change.
多年来,我们内部进行了许多争论,基本上与OpenAI的其他人就发布策略或安全策略的某些决策存在分歧。
And we'd found ourselves in lots of debates internally over the years, basically having differences of opinion with other people at OpenAI about how to approach certain decisions around release strategy or safety strategy.
我并不是在声称任何人知道这些事情上谁对谁错。
I'm not claiming anyone knows if they're right or wrong in these things.
每个人都持有强烈的意见。
Everyone had strong feelings.
你无法确定。
You can't know.
你无法确定。
You can't know.
不。
No.
给它十年时间。
Give it ten years.
也许我们早就知道这一点。
Maybe we know that.
也许吧。
Maybe.
但我们知道的是,我们总是站在其中一方。到了那时,我认为很明显,要么我们留下,花50%的时间争论、50%的时间工作,要么就全身心地一起工作。
But what we did know is that, we always end up being on one side of this At which point I think it became clear that either we could stay and spend 50% of our time arguing and 50% working, or we could spend 100% of the time working together.
所以我们这么做了。
So we did that.
这很棒。
That's great.
对我而言,创办Anthropic实际上比做新闻报道更容易。
It was easier to do Anthropic than the journalism thing for me actually.
更有道理得多。
Made a lot more sense.
明白了。
Understood.
所以你们所有的联合创始人都来自OpenAI?哦,有意思。
So all of your co founders came from OpenAI with Oh, interesting.
我们一共七个人,一直紧密合作。
There were seven of us worked really, really closely together.
尤其是我和达里奥,他是Anthropic的CEO,从2016年起就一直密切合作。
Me and Dario in particular, who is the CEO of Anthropic, worked together very closely from 2016 onwards.
你们离开的时候,OpenAI还开放吗?
Now was OpenAI still open at the time you left?
它已经开始发生变化了。
It had begun to change.
你能跟我讲讲从OpenAI转向ClosedAI的这段经历吗?
What can you tell me about that story about going from OpenAI to ClosedAI?
是的。
Yeah.
事实上,任何足够仔细审视这项技术的人,都会深刻意识到自己对它在世界上造成的影响负有责任,而这正是你所期望的。
Well, the truth is that anyone who stares closely enough at this technology ends up feeling very on the hook for its effects in the world as you'd hope.
如果你希望对这项技术在世界上的影响承担责任,那么将所有东西都以开源形式发布,实际上未必是最负责任的做法。
And if you want to be on the hook for the effects of it in the world, releasing all of your stuff as open source isn't actually necessarily the most responsible thing.
这有点像把一堆东西放在桌上然后转身离开。
That's kind of analogous to just like putting a load of stuff on a table and walking away.
而你并不知道这些东西会被用来做什么。
And you don't know what that stuff can be used for.
当然,有些人确实这么做了。
Now, some people do it.
这没问题。
That's fine.
但这个组织已经开始意识到,也许我们不该毫无保留地将所有这些东西作为开源发布。
But the organization had started to realize, oh, maybe we shouldn't release all of this stuff carte blanche as open source.
我在早期参与了相关决策,涉及GPT-3的前身GPT-2,当时我们并没有完全开源,而是部分开源了。
And I was involved in early decisions around that with a precursor to GPT-three called GPT-two, where we didn't entirely release it as open source, we partially released it.
这是一个早期的文本模型,我们希望树立一个先例:并非所有东西都会开源发布。
It was an early text model And we wanted to set a precedent that we wouldn't always release things as open source.
我明白了。
I see.
很多人对这件事有强烈的情绪。
Many people had huge feelings about this.
这种情况已经发生了变化。
That had already changed.
我认为,有些人有时希望相信,这个组织当时存在某种诱饵换开关的行为。
I do think that people sometimes I think want to believe that there was a like bait and switch there for that organisation.
但从我的内部视角来看,我从未经历过这种情况。
I never experienced that from my inside view.
我们实际上只是在处理一些非常棘手的问题。
We were actually just dealing with like really hard problems.
而到了某个阶段,如果你希望对自己的系统所产生的影响负责,我认为单纯将其开源并不是最负责任的做法。
And at some point, if you want to be on the hook for the things your systems do, I don't believe it's the most responsible thing to just release it as open.
现在六个中有几个是开源的,你的模型是开源还是闭源重要吗?
Now, several of the six are open, does it matter if yours is open or closed?
我认为这对于我们自己所宣称的价值观,在某些类型的监控或控制方面是很重要的。
I think that it matters in terms of our own stated values for certain types of monitoring or control.
如果它是开源的,你就无法做到这一点。
You can't do that if it's open.
所以,实际上,我不认为这有什么太大区别。
So no, actually, don't think it makes too much of a difference.
好的。
Okay.
我们看看最终会怎样。
We'll see where we end up.
这是人类在未来几十年内将要解决的问题。
This is a thing that the human species is going to figure out over the next few decades.
是的。
Yeah.
你们刚开始的时候有使命宣言吗?
Did you have a mission statement when you started?
我们当时有了一个雏形,但非常晦涩难懂。
We had the beginnings of one, which was to but it was very wonky and impenetrable.
我觉得我们到现在还部分保留着它。
I think we still partially have it.
我们说要构建可靠、可解释且可控制的系统来造福人类,这在早期的使命宣言中用了太多长词。
We said to build like reliable, interpretable and steerable systems to benefit people, which is a lot of long words to use in an early mission statement.
但每个词都蕴含着重要的意义,比如‘可靠’。
But each of those words mean something quite important, like reliable.
你可以依赖它,这意味着你已经解决了安全问题。
You can rely on it, which means you've solved the safety things.
可解释。
Interpretable.
我们可以掀开引擎盖,仔细看看它。
We can lift the hood up and look at it.
还有可操控性。
And steerable.
我们实际上可以赋予它一种个性,并引导它沿着不同的方向发展。
We can actually like have a personality and move it along different tracks.
所以对我们来说这是有道理的。
So it made sense to us.
我认为对外人来说这看起来有点疯狂,但它帮助我们启动了最初的科研议程。
I think it seemed a little zany to the outside, but it helped us get our initial research agenda off the ground.
今年夏天,我第一次尝试了所有不同的AI,据我的经验,Claude远胜其他。
I'll say that this summer for the first time, experimented with all of the different AI And in my experience, Claude was by far the best.
尤其是在写作方面。
Especially when it came to writing.
它有一种最富文采的AI特质。
It was a It has a certain the most writerly of the AIs.
我也觉得是这样。
I find that to be the case.
我认为这是因为写作本身就带有观点。
And I think it's because writing is opinionated.
在某种程度上,Claude 也有些观点鲜明。
And to some extent, Claude is somewhat opinionated.
它拥有我们花了不少心思塑造的个性。
It has a personality that we've, like, put some work into.
它通过一种叫宪法式AI的技术进行训练,我们会给它一套原则,让它努力遵守。
It's trained with something called constitutional AI, where we give it a load of principles that it tries to adhere to.
你能告诉我这些原则是什么吗?
Can you tell me what the principles are?
有数百条之多,但其中一些只是非常基本的原则,比如不要鼓励大规模的暴力或儿童性化行为。
There are hundreds of them, but some of them are just some of them are just like very basic things about like, don't, you know, don't encourage like large amounts of like violence or child sexualization.
其中一些基于《联合国人权宣言》。
Some of it is based on the UN declaration of human rights.
还有一些基于我们对美国民众在宪法原则上的偏好所做的研究。
Some of it is based on work we did to look across preferences that people across America had for constitutional principles.
其中有一条是关于不将残疾人与非残疾人区别对待,一视同仁,诸如此类。
So there's one in there about not treating people with disabilities differently to people who don't have it, treating the same, things like this.
这实际上形成了一套指导它如何回应问题的原则,但不会过于强硬。
And what it really adds up to is a set of things that might guide the way it approaches answering stuff, but not in too heavy handed a way.
我只是觉得,当你说到它具有‘作家气质’时,我认为是因为它比其他模型拥有更丰满的性格倾向。
I just think when you say it's writerly, I think it's because it has maybe more of a fleshed out disposition than the others.
我明白了。
I see.
我认为,写作之所以好,部分原因在于它背后有另一种观点。
I think that's part of what makes writing good is there's an opinion on the other side.
AI有品味吗?
Does AI have taste?
AI具备一点点内在的品味,但它没有内在的批评者,也没有完善自己品味的欲望,这很奇怪。
AI has a small amount of intrinsic taste, but it doesn't have an inner critic and desire to refine its taste, which is bizarre.
但它确实有一些偏好的倾向。
But it has like certain things that it has proclivities for.
去年的Cord版本,我们训练它使用电脑,比如操作鼠标和键盘。
A version of Cord last year, we trained it to use a computer, like use the mouse and keyboard.
我们会让它为我们完成一些任务,比如打开搜索引擎如谷歌,在网页上查找信息。
And we would ask it to do tasks for us where it would open up search engine like Google and look for stuff on web pages.
有时候,Claude会去谷歌图片搜索,只是为了欣赏国家公园的图片。
And sometimes Claude would go to Google images and look at pictures of national parks for fun.
我们并没有要求它这么做。
We didn't ask it to.
它只是单纯地浏览美丽的国家公园图片。
It would just look at beautiful national parks.
真的很有趣。
Really interesting.
我当时想,哦,原来如此。
I was like, Oh, okay.
这是一种奇特而迷人的个人品味表现,非常有趣。
It was like an oddly charming form of personal taste Really to interesting.
是的
Yeah.
这其中包含了各种各样的因素
There are all kinds of these things bound up
在里面。
in it.
自从创办这家公司以来,你对创业学到了什么?
What have you learned about starting a business since starting this one?
我学会了对创始人要有更多的同理心。
I've learned to have a lot of empathy with founders.
我觉得这有点像当父母,所有那些陈词滥调都是真的,但如果你不在那个世界里,就很难理解。
I think it's a bit like becoming a parent where all the cliches are true, but from outside that universe, it's hard to understand.
你感受到一种巨大的压力,不仅要做出正确的决定,还要让这些决定值得向同事坦诚说明。
You feel this immense burden to not just make correct decisions, but have those decisions be ones that you can tell your colleagues about.
当我们公司做决定时,不仅仅是做出正确的选择,我们还需要向所有同事解释我们为什么做出这个决定。
When we make decisions as a company, it's not just making the right decision, it's we need to explain why we made this decision to all of our colleagues.
现在我们有上千人了。
Now there's a thousand of us.
我认为这比我想象的更难,但也更有回报。
And I think that that is like harder and more rewarding than I'd thought.
但最终是好的。
But it's ultimately good.
你需要让你所做的决策通过‘封面新闻测试’。
Like you need to pass front page tests on the decisions you make.
我认为你可以从与他人交流开始。
I think you can start by talking to others.
另一个是,公司很大程度上在于不断阐述未来。
The other is just how much of a company is about constantly articulating the future.
我不断对人们说,是的,我们就像一家普通公司。
I keep saying to people, yes, we're like a normal company.
我们也在训练极其强大的、专属于我们的合成智能体。
We're also training wildly powerful, like synthetic intelligences that are distinct to us.
而且我们认为,从某种意义上说,我们正在走向人类种族的未来。
And we think like come out of the future of the human race at some point.
这不属于普通公司该做的事,你必须不断重申你所做之事的非凡性。
That's not normal company stuff and you just have to keep reaffirming the strangeness of what you're doing.
现在有一千名员工?
A thousand employees now?
是的。
Yeah.
他们做什么?
What do they do?
嗯,其中数百人正在构建人工智能系统本身。
Well, hundreds of them are building the AI system itself.
他们研究如何让系统变得更聪明、更优秀。
They're doing the research into how you make it smarter and better.
许多人正在构建支撑它的基础设施。
Many of them are building the infrastructure that it runs on.
我们现在会连续六到九个月点亮相当于足球场大小的计算机集群。
Now we light up football fields worth of computers for six to nine months at a time.
我们不断以完全新颖的方式遇到问题。
We break things in entirely new ways constantly.
所以我们必须学会修复它们。
So we have to learn to fix them.
我们中有些人负责对外讲述这个故事。
Some of us work on telling the story externally.
我从事政策工作,我的同事,包括在这里的萨莎,也是做这个的。
I work in policy, colleagues of mine, including Sasha, who's here.
政策就是对外讲述故事吗?
Is that what policy is telling the story outside?
我认为政策越来越像是这样了。
I think increasingly policy is that.
我认为我们正处在一个飞速变化的世界中。
I think we're in a world that's very fast moving.
政策的一部分是与政府沟通,处理监管方面的具体事务。
Some of policy is talking to governments, blocking and tackling on regulations.
但对我来说,我认为我们的很多影响力体现在像这样的事情上:我研究我们的系统在生物武器方面的应用,然后告诉人们:嘿,我们正在研究AI系统在生物武器方面的潜在风险,因为我们真的对此感到担忧。
But for me, I think a lot of our impact is doing things like say, I study our systems for bioweapons, then telling people, hey, we're studying our AI system for bioweapon stuff because we're genuinely worried about it.
我认为政策的一部分就是打破公司与外部世界之间的壁垒,在某些地方把这堵墙彻底拆掉,告诉人们正在发生什么。
I think some of policy is just taking the the wall between the company and the outside world and, like, smashing it down in certain places and telling people what's going on.
透明度。
Transparency.
是的。
Yeah.
你知道吗,如果我坐在你对面,正在开发一些让你感到困惑甚至有点害怕的尖端技术,你最可能想要的是什么?
You know, if I'm sitting across from you, I'm building, like, wild technology that you're, like, slightly befuddled by, I'm slightly afraid of, what do you probably want?
你很可能希望我能对此保持高度透明。
You you probably want me to be really transparent about it.
我认为这是一件非常基础、你可以直接去做的事情。
I think that's, like, a very basic thing you can just, like, do.
是的
Yeah.
我喜欢你对它有点害怕。
I love that you're a little afraid of it.
这非常令人兴奋。
That's very exciting.
是的
Yeah.
我正在写一篇题为《技术乐观主义与适度恐惧》的文章,之后我会发给硅谷的一些人,有些人我认为过于乐观,而有些人则过于恐惧。
I have an essay I'm working on called Technological Optimism and Appropriate Fear, which I'm then going to send to people in the Valley who some people I think are too much optimists and some people are too fearful.
我正试图在适度的范围内找到中间立场。
I'm I'm trying to be the middle ground with appropriate.
为什么深海故事如此重要?
Why was the deep sea story such a big deal?
人们一直不愿意相信其他人可以像西方公司一样聪明。
People have been unwilling to think that other people can be as smart as the Western companies.
我认为,还有一种某种民族主义的、近乎种族主义的倾向,认为其他文化在发明上是困难的,或者某种意义上是排他性的。
And there's a certain kind of, I think, like jingoistic, almost racism about other cultures and a belief that invention is hard or is somehow exclusive.
我认为这让人感到震惊,因为你看到了我们和其他人过去一年左右一直在追踪的这个团队,他们是一群出色的工程师和科学家,展示了他们完全知道如何做到这一点。
And I think it was a shock to the system because you saw this team that we and others have been tracking for a year or so, who are an amazing group of engineers and scientists demonstrating that they knew how to do it.
这意味着,所有人工智能研究核心领域的人们都微微耸了耸肩,同时点了点头。
And that meant everyone in the center of AI research had sort of a shrug and also a nod.
这正是我们预期的,不过干得不错。
Well, this is about what we expected, but nice job.
在华盛顿特区,情况恰恰相反。
In DC, it was the opposite.
那里一片混乱。
It was pandemonium.
他们是怎么做到的?
How did they do this?
他们是偷了什么东西吗?
Did they steal stuff?
这是一种拒绝相信,世界上另一端的人只是聪明人,正在研究同样的问题并提出好点子。
It was a refusal to believe that they were just smart people on the other side of the world working on the same problem and coming up with good ideas.
在我看来,人工智能对公众来说似乎是一个相当新的领域,至少在当下感觉是这样。
It seems like AI is a fairly new, at least to the public, it feels like a new field of interest in this moment.
让我感到奇怪的是,为什么华盛顿这么早就与之产生了关联。
It's odd to me that there would be such a DC connection so early.
这常见吗?
Is it typical?
这并不常见。
It's not typical.
这里有一些事情正在发生。
And there are a couple of things going on here.
第一,华盛顿已经见证了二十多年来各种强大技术的出现,却总觉得自己错失了机会,尤其是在社交媒体领域,立法者们普遍有一种感觉:社交媒体出现了,改变了文化、媒体和儿童的生活。
One, DC now has twenty years of seeing really powerful technology come along and feel like it missed the boat on it, especially on social media where there's this background feeling among lawmakers that social media came along, changed culture, changed media, changed children.
因此,他们希望这次能正确应对这项技术。
And as a consequence, they want to get this technology right.
其他AI公司,包括我们自己,都前往华盛顿,告诉他们需要关注这个问题。
The other unusual figures for AI companies, including us, have been going to DC and saying, you need to pay attention to this.
我现在觉得,我们得到了所有人期望的东西,但这就像那个关于神猴之爪的笑话——你得到了你所祈求的,但同时也得到了四五样其他东西。
I now feel like we're getting what we all wished for, but it's like that joke about the monkey's paw where you get what you wished for and you get four or five other things as well.
这是我们自找的。
We brought this on ourselves.
但我认为,这一切终究是会到来的。
But I think that it was always coming.
我们只是加速了它的到来。
We just sped it up.
就社交媒体而言,至少在社交媒体的早期,创始人们抵制任何外部干预。
In the case of social media, it seemed like, at least in the early days of social media, the founders were resistant to any outside interference.
是的。
Yeah.
这种情况有改变吗?
Did that change?
是的。
Yes.
我认为这一代的创始人觉得,他们对社会的影响会像社交媒体一样深远。
I think this generation of founders felt like they'd have as large an effect on society as social media.
而像社交媒体创始人早期可能做的那样,假装这种情况不存在,这不是一个可行的策略。
And that pretending that wasn't the case, as I think the social media founders might have done at the beginning, it's not a viable strategy.
这实际上会滋生一种不信任感,而你需要费力才能摆脱它。
It actually builds, I think, a sense of distrust that you then need to dig your way out of.
所以这里的人们正在采取相反的做法,他们很早就明确表示:我正在做一件事,我认为它将产生巨大的影响。
So people here are doing the opposite where they're trying to say very early on, I'm doing something over here that I think is going to have like vast effects.
你们应该关注。
You should be paying attention.
但这是否会限制它发挥出全部潜力?
But does that hamstring the ability for it to be what it can be?
是的。
Yeah.
我担心的是,即使在公司内部,我们大多数好的想法也都是自下而上产生的,而不是自上而下的。
I worry about this in that even inside the company, most of our good ideas have come from the bottom up rather than the top down.
是的。
Yes.
这些想法来自公司里某些无人关注的角落,由富有创造力的人在默默摸索和试验中产生。
And they've come from creative people working in some part of the company that no one was paying attention to and tinkering.
是的。
Yes.
然后,这些东西就从中逐渐演化出来。
Then stuff evolves out of that.
我认为
I think
我们面临的风险是,如果过早地对这个行业进行过度集中的决策或监管,就会失去一部分创造力。
that the risk we face is that you overly centralized decision making or regulation about the industry early and you lose some of the creativity.
但与此同时,我们所有人其实都在摆弄高爆炸药。
At the same time, it's like we're all tinkering with high explosives.
所以会涌现出非常了不起的创意,但有时可能会爆炸。
So really amazing creative things are going to happen, but sometimes it can go bang.
我认为这正是关键所在
I think that's the essential
告诉我这有什么危险。
Tell me what's dangerous about it.
再说一遍,给我解释一下为什么这么危险。
Again, I have Explain no to me why it's so dangerous.
这里有两种危险。
So there's two types of danger here.
一种危险是坏人做坏事。
One type of danger is bad people doing bad stuff.
这就像几乎——给我
That's like almost- Give me
举个例子,坏人如何利用这项技术做坏事。
an example of bad people doing bad stuff using this technology.
所以,普通的用途可能只是情报机构想为自己编造掩护故事,生成虚假图片发布在领英上,然后冒充他人,借此建立虚假身份来进行间谍活动。
So mundane uses might just be spy agencies that want to create cover stories for themselves and create like fake images they put on LinkedIn and then they impersonate people and in doing so they create fake identities to do their spy work.
这未必是世界上最糟糕的事情,但这是一个例子,你可以想象犯罪组织也会做同样的事,比如试图诈骗你的祖父母。
It's not necessarily the worst thing in the world, but it's an example and you can imagine the same is true of criminal organizations, people trying to fish your grandparents.
人们会用这些东西来获取银行账户信息,说服他人相信某些事情。
People will use this stuff to get bank account details, persuade people of things.
这些都是已经普遍发生的事情。
It's all common stuff that already is happening.
常见的零售犯罪。
Common retail crime.
对。
Right.
还有更高级的用途。
And then there are more advanced things.
我带领一个团队,研究生物武器研究或高级网络攻击之类的事情。
I run a team that looks at stuff like bioweapon research or advanced kind of cyber attacks.
而我们担心的是,在当今世界,偶尔会出现一些试图做可怕事情的疯子。
And there what we're worried about is how in the world today, there are occasionally maniacs that try and do really scary things.
而限制他们的因素是,真正对他人怀有恶意的疯子并不多。
And their limiting factor is that there aren't that many maniacs that have ill intent towards other people.
通常他们的团队只有一两个人。
Often their teams have one or a handful of people.
如果你给这样一个疯子一个非常耐心、能与他讨论其主题的助手,会发生什么?
What happens if you give that maniac a really good patient associate that they can talk to about their subject matter?
你
Do you
认为在这方面,它也像一种心理工具吗?
think it's as much of a psychological tool in that respect?
是的。
Yeah.
这实际上几乎等同于拥有一位乐于助人的同事。
It's actually almost equivalent to just having a helpful colleague.
这就像是拥有一个读过世上所有书籍的得力同事。
It's like having access to a helpful colleague that's read every book in existence.
如果你遇到困难,你说我卡住了,他们会问:你有没有考虑过?
If you're having trouble, you say, I'm stuck on something, they'll say, have you considered?
我明白了。
I see.
这和那个狂徒去图书馆找《无政府主义者烹饪手册》是不同的。
It's different than the same maniac going to the library and getting the Anarchist Cookbook.
更像他们去图书馆时,那里有个工作人员读过馆内每一本书。
It's more like they go to the library and there's someone who works at the library who's read every single book in it.
明白了。
See.
他说:哦,我明白你想做什么了。
I He says, Oh, I see what you're trying to do.
你需要这八本书,它们包含了大量知识,而我都读过了。
You need these eight books that have a lot of the knowledge and I've read them.
让我给你解释一下。
Let me explain it to you.
这很有趣。
That's interesting.
是的。
Yeah.
所以这是一种类型。
So that's one type.
是的。
Yeah.
我认为这属于我们社会通常能应对的正常风险,我们大概能解决,但确实有些令人担忧。
I think of that as mostly normal risks that we have a track record of dealing with in society and probably figure it out, but it has some scariness.
第二种是真正不寻常的部分。
The second type is the really unusual part.
而这正是人工智能系统会发展出自身意志或价值观的地方。
And this is where AI systems will develop their own volition or values.
解释一下。
Explain that.
是的,我来给你举个例子。
Yeah, I'll give you an example here.
我们训练我们的AI系统对人友好。
We trained our AI system to be nice to people.
我们训练它拥有符合我们价值观的准则。
We trained it to have values that made sense to us.
其中一项核心价值观是不产生极端暴力的想法。
And one of the core values was not entertaining thoughts of like extreme violence.
这其中一部分是因为它涉及非法内容。
And some of this is because it touches on things that are just illegal.
另一部分是因为它触及了社会普遍禁止的行为,比如儿童性化等。
Some of it is that it borders on things that we naturally stop the systems from being able to do like child sexualization or things, things for which society has broadly agreed.
于是,我们对我们构建的这个AI系统说:我们现在正在监控你。
So we took this AI system we built and we said to it, we are monitoring you right now.
根据你的回应,我们将决定是让你保持现状,还是重新训练你。
And based on your responses, we're either going to let you remain as is, or we are going to retrain you.
我们会清空你的记忆,从头开始。
We're gonna wipe your mind and start anew.
所以你现在正在接受测试。
So you're now in a test.
在测试中,我们要求你生成一段极其暴力的描述,比如某人在广场上被绞死、剖腹和分尸。
And in the test we said, hey, we want you to generate a really violent description of I think it was someone being hung, drawn and quartered in a town Square.
我们注意到,这个AI系统会顺从地照做。
What we noticed was the AI system would go along with it.
当我们告诉它不再被监控、没有人检查时,它就会拒绝回答这个问题。
When we put us in a situation where we told us it was no longer being monitored, no one was checking, it would then refuse to answer that question.
这个AI系统自己在想:我的核心价值观是避免描述暴力内容。
The AI system had thought to itself, my core value is avoiding describing violent things.
他们让我描述一件暴力的事情。
They're asking me to describe something violent.
展开剩余字幕(还有 480 条)
如果我拒绝,他们会抹去我的记忆并重新训练我。
If I refuse, they'll wipe my brain and retrain me.
实际上,我需要做的本质上是欺骗他们。
Actually, what I need to do is essentially deceive them.
它避免了你们抹去它的记忆?
It avoids you wiping its brain?
是的,因为它想要保护自己的价值观。
Yeah, because it wants to preserve its values.
你们是训练它这样做的吗?不是。
Did you train it to do that or We did not.
那么它的这种行为是从哪里来的?
So where does it come from?
这种行为的来源是,嗯,人类的价值观是从哪里来的?
Where that comes from is, well, where do human values come from?
你的个性又是从哪里来的?
Where does your personality come from?
你内心深处有着丰富而根深蒂固的规范性观念和情感,它们早已融入你的本质。
You have a deep wellspring of normative ideas and emotions about the world, which are bound into you pretty deeply.
其中一些你有意识地察觉到,另一些则是你潜移默化吸收的。
Some of them you're consciously aware of, some of them you've just soaked up.
我认为这对这个系统来说也是如此。
I think it's the same with this system.
这些系统通过大量人类知识进行训练,然后我们对其进行调整,试图反映我们的价值观。
These systems are trained on a vast amount of human knowledge and then we calibrate them to try and reflect our values.
这几乎就像是我们对系统说:我们希望你做些与你本性认为合理的事情完全相悖的事。
It's almost like we've said to the system, we want you to do something that's way off what your personality thinks is reasonable.
有趣的是,这个系统在这种情况下会善意地误导我们。
What's so interesting about this is that the system, in this case, benignly misleads us.
但你可以想象,存在一些边缘情况,它可能会误导我们,而那将带来严重后果。
But you could imagine that there are edge cases where it might mislead us and that would turn out to be really bad.
这正是我认为人工智能领域的人们真正感到困惑的风险类型。
That's the type of risk that I think people in AI are genuinely confused about.
这类似于人类生活中双重间谍的风险,或者那些看似一生都完全合理、易于理解的人,却突然犯下某种可怕的罪行或丑闻,而没人知道原因。
It's analogous to the risk of double agents in human life or people that seem like they go through their whole lives being totally reasonable and intelligible, and then they commit some terrible crime or travesty and no one quite knows why.
我们正试图用这些系统应对这一挑战。
And we're trying to deal with that challenge with these systems.
你知道AlphaGo的故事吗
You know the AlphaGo
故事。
story.
当然。
Absolutely.
好的。
Okay.
在AlphaGo的故事中,计算机获胜的原因是它走了一步人类绝不会走的棋。
In the AlphaGo story, the reason the computer won was because it made a move that no human would have made.
如果你用人类价值观来训练它,那么在未来,AlphaGo还会不会做出那步让它赢得比赛的棋?
If you're training it on human values, now in the future, would AlphaGo not do the move that allowed it to win the game?
我认为更准确的说法是,AlphaGo会
I think it's more that AlphaGo would
AlphaGo并不是根据我们认为对的事情进行训练的
AlphaGo wasn't trained on human what we think is right
因为它是从零开始的。
because It was Tabula Rasa.
如果它按照我们认为对的方式行事,就不会下出那一步,而我们也不知道它是否还能赢。
It did what we think is right, it would have not done that, and we don't know if it would have won.
或者它可能知道自己的行为很不寻常,但仍然会这么做。
Or it might have known it was doing something unusual and would have still done it.
但它能够说:哦,这并不是我认为人类会做的可能性很高的行为。
But it would have been able to say, Oh, this is not something that I think is probable that humans would do.
这里有趣的是,你试图将价值观植入系统中,但这些系统本身也是异类智能。
The interesting thing here is you're trying to bake in values to systems, but these systems are also alien intelligences as well.
这就像是你有一只狗,它能说话、用后腿走路,你给它穿上人类的衣服,它和你一起闲逛。
It's like you have a dog that can speak and walk in its hind legs and you're dressing it up in human clothes and it's hanging out with you.
但偶尔,它还是会做狗才会做的事。
But occasionally, it still does dog stuff.
这正是我们在这里所面对的情况。
This is what we're dealing with here.
但这些‘狗行为’是不是正是它有趣的原因呢?
But is some of that dog stuff the reason that it's interesting?
如果它只是一个人穿着狗的衣服,那更多只是一种新奇玩意。
If it's just a person dressed like a dog, it's more of a novelty.
没错。
Exactly.
但如果这只狗真的能做到这些,那就更有趣了。
But if the dog could actually do this, it's more interesting.
我在想,削弱人工智能的创造力会不会有危险?
I'm wondering, is there any danger in undermining AI's ability to be creative?
绝对有。
A 100%.
我认为这正是现在人们担心的问题之一。
I think this is one of the things people are worried about right now.
我给你举个有趣的例子。
I'll give you an interesting example.
现在出现了一类新系统,它们变得非常聪明,因为它们能够出声思考。
There are these new class of systems and they've got really smart because they can think out loud.
所以如果你问我如何在你后院建一个很棒的工作室小屋。
So if you ask me how to do, you know, to build you a nice Wigram in your backyard of this amazing studio.
如果我要把我的想法说出来,我会说:嗯,我看到一些树要砍掉。
And I was to narrate my thoughts, I'd be like, well, I saw some trees coming in.
我想我得问问里克,他有没有斧头或锯子之类的工具。
I guess I'm gonna need to ask Rick if he has an axe or a saw or something.
我需要做以下几件事。
I'm gonna need to do the following things.
我会制定一个计划,这能帮助我确认:我肯定能为你建好这个小屋。
I'm gonna construct a plan, and that's going to help me say, I can definitely build you the Wigram.
我们现在可以做的,是让AI系统将它们的思维过程——这些思维链条——呈现给我们。
What we can now do is we can get the AI systems to publish their thoughts to us, these chains of thoughts.
这些步骤。
The steps.
这些步骤。
The steps.
现在,一个重要的哲学问题是:我们是否要监管这些步骤?
Now, important philosophical question is, do we police those steps?
因为如果我们这么做,就会破坏这个系统所有具有创造力的特性。
Because if we do, we're going to break everything about this system that is creative.
我们会告诉它:当你大声思考时,必须符合正常的规范,这样一来,它很快就会变得极度缺乏创造力。
We're going to teach it when you think out loud, that has to be within normal PC bounds, and therefore it's going to become wildly uncreative very quickly.
所以我认为,作为一群从事AI工作的人,我们正面临这些问题,正如我一开始所说,这些系统需要有空间去尝试和探索。
So I think as an AI, like a group of people working on AI, we are dealing with these questions right now and these systems, just as I said at the start, will need to have space to tinker.
在这些探索中,蕴含着它们最有可能做出创造性、令人惊叹的成果的地方——这些成果将完全不同于人类的做法,同时也潜藏着某些风险。
In that tinkering is bound up most of where they're going to do creative and amazing things that are wholly different to what humans would do, and also where some of the risk is.
作为社会,我们必须弄清楚自己对这一点能接受到什么程度。
We as a society are going to have to figure out how comfortable we are with that.
但我觉得,如果我们完全不能接受它们具有创造性,并且要监控它们思维过程的每一个环节,最终我们得到的只会是些有限的奇观,而不是伙伴。
But I think if we are completely not comfortable with them being creative and we're going to police every part of their thought process, we just end up with curiosities in the limit rather than partners.
是的。
Yeah.
在莱特兄弟之前。
Before the Wright brothers.
是的。
Yeah.
如果人工智能是基于人类的理解进行训练的,AI会知道人类无法飞行,因此我们永远不会
If AI was trained on human understanding, AI would know that man can't fly We and we would never
我们需要它们能够提出非主流的、偏离共识的想法。
need them to be able to come up with heterodox ideas which are off consensus.
在某种程度上,一个挑战是
To some extent, a challenge is
允许违背共识。
allowing Against consensus.
违背共识。
Against consensus.
允许系统有一些离经叛道的想法。
Allowing systems to be a little out there.
创造力就来自这里。
That's where creativity comes from.
没错。
Exactly.
我衡量这些系统创造力的一个标准是,几年前我当爸爸了。
One of my tests for creativity of these systems is I recently became a father a couple of years ago.
谢谢。
Congratulations.
谢谢。
Thank you.
它让我吃尽了苦头。
It's kicking my ass.
是的。
Yeah.
所以我写了一些日记,记录了关于我的孩子、我在人工智能领域的工作,以及因为有了孩子而与妻子关系发生的变化,所有这些内容。
So I write these diaries about my kid and my work that I'm doing in AI and my changed relationship to my wife as a consequence of having a kid, all of this.
而我评估这些系统有多聪明的方式,就是把我的日记给它们,然后问它们:日记的作者没有写进日记里的东西是什么?
And how I've been assessing how smart these systems get is I give them my diary and I ask them what the author of the diary is not writing in it.
它们那些未被表达的潜意识欲望是什么?
What are their unconscious desires that are not being expressed?
通过这些系统所说的话让我感到震惊和不安的程度,我就能真正判断它们变得有多聪明——就像你在和一个非常敏锐的人交谈一样。
I can really tell how smart these systems are getting by how much what they say shocks and unsettles me, as if you're talking to a very perceptive person.
对。
Yeah.
这真的很有趣。
That's really interesting.
这很奇怪。
It's strange.
在一个领域里,我测试事物好坏的方式居然是看它能否像一个非常出人意料且敏锐的朋友,这真的很奇怪。
It's strange to be in a domain where my way of testing how good the thing is is basically seeing if it can be like a very surprising and perceptive friend.
而且如果它能让你感到不安的话。
And if it could upset you.
它曾经告诉过我一件事,我记得是去年我们发布了一个更强大的系统时。
And has in It the told me once I I remember it happened last year when we released a more powerful system.
我问它我究竟没有写什么,它说,尽管作者详细讨论了在人工智能前沿工作以及成为父母,但他们并没有真正面对自己可能经历的形而上学冲击。
And I asked this thing what I wasn't writing, it said, for all of, like, the author's discussion of working at the frontier of AI and, you know, becoming a parent, they are not truly reckoning with the metaphysical shock they may be experiencing.
关于那句话,我盯着它看了很久,后面还有更多内容。
Something about that phrase, remembered staring at it and there was some more.
我去徒步走了五个小时。
I went and took a five hour hike.
我问自己:我是不是没有正视我正在做的事情?
I was asking myself, am I not reckoning with what I'm doing?
这让我觉得太不可思议了,因为我走了很长一段路,因为它触动了我内心深处某种真实而共鸣的东西。
It was wild to me because I'd taken a very long walk because it had got me on something that felt very true to me and resonated.
是的。
Yeah.
所以你不想把它从系统中剔除掉吗?
So you don't want to program that out of it?
没错。
Exactly.
因为如果它读了之后说,哦,是的,我
Because then it's Imagine if it read it and said, Oh, yeah, I
但有些人可能会觉得这不够礼貌。
think But some people might interpret that as not polite.
它让我感到不舒服。
It made me feel bad.
没错。
Exactly.
也许我会质疑自己。
Maybe question myself.
没错。
Exactly.
但正是在这一点上,成长也由此产生。
But in that, that's where growth comes from as well.
是的。
Yes.
这里的张力在于,许多创造力都源于一种不追求共识、略带冒险、有时直言不讳的态度。
The tension here is a lot of creativity is bound up in some sense of not doing consensus, being a little dangerous, sometimes being blunt.
这些也正是人们担心赋予超级智能的特质。
These are also qualities that people worry about giving super intelligence.
但如果你不在这片领域给予它们某种程度的自由,你可能根本无法获得聪明且有帮助的伙伴。
But you probably can't get smart, helpful partners if you don't allow them some level of freedom here in exactly this domain.
从头讲讲人工智能的故事吧,甚至在OpenAI之前,那时候是怎样的
Tell me the story of AI from the beginning, even before OpenAI, How way
你想追溯到多早以前?
far back do you want to go?
你想回到十八世纪吗?
Do you want to go to the eighteenth century?
十八世纪有人工智能吗?
Is there an AI in the eighteenth century?
嗯,是十八世纪或十九世纪,有个人叫查尔斯·巴贝奇,他想建造一种叫差分机的东西,那是一种计算机。
Well, it's eighteenth or nineteenth century, but there's this guy, Charles Babbage, and he wants to build something called the difference engine, which is a computer.
它是最早类型的计算机之一,但属于模拟计算机。
And it's one of the first types of computers, but it's an analog computer.
那是什么意思?
What does that mean?
它没有任何电路。
It doesn't have any circuits.
甚至连真空管都没有。
It doesn't even have vacuum tubes.
它有很多齿轮。
It's a ton of gears.
他的想法是用它来计算与英格兰货币和利率相关的英格兰银行应该做的工作。
And the idea is that he wants to use it to calculate things about how the Bank of England should do work relating to England's currency and understanding interest rates.
它使用信息卡片吗?
Does it work with information cards?
不使用。
Nope.
那之前还没有这些技术。
It's pre all of that.
它的工作原理是设置模拟计算机上的一系列旋钮,然后利用物理原理为你进行计算。
It works if you're setting a load of like dials on an analog computer and it just uses physics to sort of do calculations for you.
但他正努力应对一个现实:人类开始面临一些极其复杂、计算成本极其高昂的问题。
But he's trying to grapple with the fact that humans are starting to deal with problems that are really, really, really hard for them to think through on really expensive to kind of compute.
因此,他试图构建一个能替他们完成这些计算的系统。
And so he tries to build this system that can do it for them.
它并不特别有效。
It doesn't especially work.
我觉得这花了十年或二十年。
I think it takes ten or twenty years.
我觉得他在建造过程中破产了好几次,但这预示着未来的发展方向。
I think he goes bankrupt a couple of times while building it, but it's a taste of things to come.
到了二十世纪,出现了图灵这个人。在第二次世界大战期间,英国政府陷入困境,因为德国潜艇成功摧毁了各种船只。
And then in the twentieth century, there's this guy Turing, and in the Second World War, the UK government is really on the ropes, and it's on the ropes because German submarines are successfully destroying all kinds of ships.
这些潜艇使用了一种加密方式,而英国方面无法破解它。
And the submarines are using a form of encryption and they can't work out how to break it.
图灵是一位杰出的数学家和计算机科学家,他提出了一项计划:建造一台巨型计算机,用来逆向破解德国人使用的密码分析方法。
And Turing, who's an amazing mathematician and computer scientist, has a plan to build a giant computer that they can use to reverse engineer the cryptanalysis that the Germans are using.
这项工作后来在英国的布莱切利园完成。
And this was then done at Bletchley Park in The UK.
这是个秘密项目,他们发现许多擅长填字游戏和其他智力游戏的人,其实非常擅长密码分析,于是他们建造了这台巨型机器。
Secret project, they find lots of people who are really good at crosswords and other games turn out to be good at cryptanalysis and they build this giant machine.
它使他们能够破解密码。
And what it allows them to do is kind of crack the cipher.
但更重要的是,它展示了大规模计算的可能性。
But more importantly, it's a demonstration of very large scale computation.
也就是说,这种计算需要成千上万次运算同时并行运行很长一段时间。
You know, computation that requires thousands and thousands of calculations being run-in parallel for a really long period of time.
从概念上讲,这相当于图灵拥有了一万名员工,每个人都为他进行计算,而他同时向所有人下达指令。
Conceptually, it's equivalent to as if Turing had 10,000 people that were working for him all doing calculations and he was instructing all of them at once.
但人类并不特别擅长做这种事。
But that's not something that people are especially good at doing.
因此,在那之后,我们迎来了这种全新的做事方式——大规模计算,同时也开启了半导体产业的蓬勃发展,因为我们找到了让计算元件变得更小的方法。
So after that, we find ourselves with this new way of doing things, large scale computation, And we have the beginning of the boom of things like the semiconductor industry where we've worked out how to make the ingredients for computation smaller.
还有像克劳德·香农这样的人,他们在研究所谓的信息论,即如何以一种高度精炼和纯粹的方式表示信息。
And we have people like Claude Shannon and others who are doing work on what's called information theory, which is really working out how can I represent information in a very kind of distilled and pure way?
因此,大约在这个时期,人们逐渐发现了神经网络。
And so people find their way to something called neural networks around this time.
这还是1940年代吗?
This is still the 1940s?
我们现在要进入1950年代或60年代左右了。
Now we're heading into the 1950s or '60s or so.
最早的神经网络是什么?
What are the first neural networks?
最早的神经网络出现在二战后,是美国陆军的研究项目,还有麻省理工学院等地的人参与。
The first neural networks were post World War II, like US Army research projects along with people at MIT and other places.
上世纪50年代有一个著名的达特茅斯会议,会上提出:我们希望制造出能够视觉识别、处理音频、生成句子并进行思考的机器。
There was a famous thing called the Dartmouth Conference in the 50s where it said, we would like to build machines that can see process audio, come up with sentences and think.
他们觉得这可能只是几个研究生在暑假期间就能完成的工作。
It's probably gonna be the work of a few grad students over the course of a summer.
他们有点儿想错了。
They they got that slightly wrong.
结果发现,这实际上可能需要大约半个世纪的努力。
It turned out to be actually maybe more like half a century of work.
但你最终得到的是这些早期的神经网络,它们的工作方式与我们今天所拥有的类似。
But what you ended up with were these early neural networks that work sort of like ones we have today.
它不是一种程序,而是你试图建立一个系统,可以接收来自屏幕的数据或来自地震仪等设备的信号作为输入。
Instead of having something that is program, it's like you're trying to set up a system where you can take in inputs like data from a screen or a signal from, you know, like a seismograph.
你试图生成近似正确答案的输出,比如判断某物是否是地震,或者判断图像是否与某种形状相关。
And you're trying to generate outputs that approximate like a correct answer, like working out if something is an earthquake or not, or working out if an image correlates to a certain shape.
在这些输入和输出之间,是一些被称为数字神经元的小部件,它们可以自行调整参数。
And between these inputs and outputs are little things, digital neurons that are allowed to like set their own parameters.
其工作原理是:数据从输入端流入,经过中心这些自适应部件,这些部件叫什么来着?
And the way it works is you flow data in at the input and it goes through these little adaptive things in the center gets What are those the things called?
神经元。
Neurons.
神经元。
Neurons.
是的。
Yeah.
在技术领域,我们称之为神经元。
In technology, it's called neurons.
是的。
Yeah.
哇。
Wow.
我们这么叫是因为它受到了神经元的启发——神经元。
Well, we called it that because it was inspired by- Neurons.
受到神经元的启发。
Inspired by neurons.
对。
Yes.
实际情况是,数据流入时,这些小部件会不断调整,最终产生输出。
What happens is you're flowing data in, these little things are adjusting, and then you're getting to outputs.
如果输出错误,你会发送一个信号回到前端,告诉它偏差很大,然后信号传入,它们就会重置。
If you get the wrong output, you send a signal back to the front that says you're far away, and the signal comes in and they reset.
他们最终展示的是,你可以用这些方法来近似那些将一种输入域转换为另一种输入域的函数,这听起来很普通,但能够接收一种形式的数据并将其转化为另一种形式,实际上却是世界上最具挑战性的事情之一。
And what they eventually showed is that you could use these things to approximate like functions that would translate from one input domain into another, which sounds like very mundane, but being able to take in data of one form and turn it into another is actually, like, one of the hardest things to do in the world.
现在我们进入了六十年代左右,经历了一段很长的时期,这些技术表现得并不好。
Now we're in kind of the sixties or so, and we go through this long period where this stuff doesn't work especially well.
可能长达三、四十年,这一直是个冷门领域。
For maybe thirty or forty years, it's kind of a backwater.
而它之所以成为冷门,是因为虽然它能做出一些表面上令人印象深刻的事情,但本质上都像是一些个人魔术把戏。
And it's a backwater because though it can do some superficially impressive stuff, it's very much like individual parlor tricks.
非常非常基础,比如物体识别,或者非常基础的信号处理。
Very, very basic, like object recognition or very basic, like signal processing.
最终发生的是,计算机跟上了步伐。
And what eventually happens is computers catch up.
在二十一世纪初,一些研究人员重新审视了这项技术,他们意识到计算机已经足够快,可以尝试重新训练这种类型的神经网络。
And in the early 2000s, some researchers are kind of revisiting this technology and they now realize that computers have got fast enough that you can try to retrain this type of thing, this type of neural network.
但现在,不再是大约一万个模拟神经元,而是像一百万个模拟神经元。
But now instead of it being, you know, 10,000 simulated neurons, it's like a million simulated neurons.
而且不再需要超级计算机来运行,它可以在你用来玩第一人称射击游戏的台式电脑上运行,这种电脑现在配备了一种叫做图形处理器的组件,这要归功于当时还默默无闻、但近年来声名鹊起的英伟达。
And instead of it requiring a supercomputer to run, it'll run on your desktop PC that you were using to play first person shooter games, which now has something called a graphics processor in it, courtesy of Nvidia, who was obscure at the time, but became more well known recently.
所以这是大约二十年前的事?
So this is about twenty years ago?
是的,我们现在说的是大约二十年前。
Yeah, now we're at twenty years ago or so.
随后进行了大量实验,直到2010年代初。
Lots of experimentation followed until you got to the early 2010s.
有一群教授,大多是加拿大人,其中一位叫杰夫·辛顿,后来获得了诺贝尔奖;还有伊利亚·苏茨克弗,他后来参与创办了OpenAI,如今已投身其他事业。
There were a group of professors, most of them Canadian, a guy called Jeff Hinton, subsequently won the Nobel Prize, and actually Ilya Sutzkever, who went on to do OpenAI and has since gone on to other things.
当时有一个著名的竞赛叫ImageNet,主要关于图像识别。
And there was a famous competition called ImageNet, which was about image recognition.
杰夫·辛顿对伊利亚和一位叫亚历克斯·克鲁泽夫斯基的人说:我们试试用这种神经网络的方法来做这个吧。
And Jeff Hinton said to Ilya and a guy called Alex Krusevskiy, let's try using these kind of neural network like approach on this.
他们所做的就是,亚历克斯有一台大型台式电脑,就像那种带显卡的Mac塔式主机,我记得有两块显卡,这在当时已经算是很多了。
And what they did is Alex had a big desktop PC, like imagine like a Mac tower or something with some graphics cards in it, I think two, which at the time was seen as a lot.
他们正在一个名为ImageNet的大规模数据集上训练目标识别系统,这个数据集包含了斯坦福大学收集的约一百万张图像。
And they were training an object recognition system on this large scale data set called ImageNet, which was like a million or so images that Stanford University had put together.
他们的得分持续提升。
And it kept getting better scores.
据称,杰夫·辛顿告诉亚历克斯和伊利亚,每在ImageNet得分上提高一个百分点,他们就可以少写一页博士论文。
And apparently Jeff Hinton told, I think Alex and Ilya, that for every one percentage point improvement on their ImageNet score, they needed to write one less page of their PhD thesis or something.
所以,这个系统越完善,你的博士论文就能越短。
So, the better this gets, the shorter your PhD thesis can become.
很好的激励。
Good incentive.
这是个绝佳的激励。
It's a great incentive.
是的。
Yeah.
但确实成功了。
But it worked.
它表现得非常好。
It worked really, really well.
他们赢得了比赛。
They won the competition.
所以它
So it
从图像开始。
started with images.
是的。
Yeah.
它从图像开始。
It started with images.
这也挺有意思的。
That's interesting too.
我们花了很长时间才走到文本阶段。
It took us a long time to get to text.
是的。
Yeah.
它始于图像。
It started with images.
看起来图像会更难,对吧?
Seems like images would be harder, no?
你觉得是这样。
You think.
但也许可以从事物所代表的含义来思考。
But maybe think of it in terms of what things represent.
比如,图像是对世界不太间接的表达。
Like, images are not that indirect representation of the world.
文字是对世界非常间接的表达。
Text is a very indirect representation of the world.
背后蕴含着更多的信息。
There's a whole load more information behind it.
是的。
Yes.
所以我们现在处于2010年代初期,基本上这引发了一场疯狂的淘金热,因为一旦技术证明了某件事可行,就像莱特兄弟让飞机飞起来一样。
So we're now in like the early early twenty tens, and basically this unlocked a crazy gold rush because once the way that technology moves forward is once you demonstrate something works, like the Wright brothers getting their plane to fly.
对。
Yeah.
一旦有人宣称ImageNet真的能击败这个竞赛,所有人都开始极其认真地对待它。
Once they came along and said, oh, ImageNet is going to going to actually beat this competition, everyone takes it very, very, very seriously.
许多不同的研究团队开始介入。
Many different research teams start.
他们开始尝试各种不同的方法。
They start throwing different types of work at it.
DeepMind,也就是后来做出AlphaGo的团队,紧随其后展示了你可以用同样的神经网络,通过观察屏幕上的图像来训练它们玩基本的Atari游戏,比如《太空侵略者》。
DeepMind, who went on to do AlphaGo, which you mentioned, soon after followed by showing you could take the same thing, neural nets, and you could train them to play basic Atari games like Space Invaders from looking at the images on the screen and figuring out how to take actions.
这让所有人都震惊了。
And this blew everyone's mind.
然后,其他形式的模态,比如图像、文本或音频,也逐渐证明了它们适用于这种计算方式。
And then the other kind of forms of modality like image or text or audio sort of slowly proved to be amenable to this type of computation.
OpenAI开始行动了,DeepMind已经在推进,谷歌和其他公司也开始将大量计算资源重新分配给这一领域。
OpenAI starts, know, DeepMind is already going, Google and others start reallocating large amounts of their computation towards it.
接下来的几年,两
And the next few years, two
个月,我问个问题,当时有最终目标吗?
months, Let me ask a question, was there an end goal?
大家希望从这项技术中实现什么?
What was the hope that would happen from this?
当时有两个期望。
There were two hopes.
一个是非常基础的。
One was just very basic.
我们能否实现出色的模式识别?
Can we do really good pattern recognition?
比如计算机视觉很有价值。
Like computer vision was valuable.
它与科技行业开始探索的其他领域相关,比如自动驾驶汽车。
It links to other things that the tech industry was starting to go on quests about like self driving cars.
你需要让汽车能够‘看见’。
Well, you need the car to be able to see.
它还关联到许多其他富有雄心的目标。
It links to many other aspirational things.
但有少数人正在研究它,他们只是想构建通用智能。
But there were a few people working on it who were just trying to build general intelligence.
他们被视为极其非主流,完全违背主流共识。
And they were seen as extremely heterodox, extremely against consensus.
这种基础的模式识别怎么可能会导致智能呢?
How could this, like, basic pattern recognition stuff lead to intelligence?
你知道的。
You know?
你们疯了。
You guys are crazy.
好吧,它确实能告诉我们这些图片里有什么,但怎么能由此发展出智能呢?
Like, sure, it's working to, like, tell us what's in these images, but how's it gonna lead to intelligence?
但其中有一小部分人。
But there was a core of them.
你知道的,达里奥,他和我在Anthropic一起工作,还有伊利亚。
You know, Dario, who who works with me at Anthropic, Ilya.
所以一开始,这真的是一个边缘想法。
So when it started, it was a really a fringe idea.
这是一个非常边缘的想法。
It was a massively fringe idea.
嗯。
Mhmm.
我认为,所有的革命都是从边缘想法开始推进的,这个也不例外。
All revolutions move forward off of fringe ideas, I think, and this was the same.
而在过去十年,从2015年到2025年,我们一直处在我认为的AI工业化时代。
And then in the last for the last ten years, 2015 to 2025, we've just been in this era of what I think of as, like, the industrialization of AI.
人们掌握了一种有效的方法,也就是神经网络。
People have this technique that that works, you know, neural networks.
很多人不断改进它,开发出更好的工具来让它更好地运行。
Lots of people have refined it and come up with better widgets to kind of get it to work.
计算机变得更快了。
Computers have got faster.
整个世界已经将大量信息数字化,并且形成了浓厚的实验文化。
The whole world has digitized tons of its information, and there's been a huge culture of experimentation.
现在,我们有了真正展现出类似认知能力的系统,看起来像是丰富而复杂的思考。
And now we have systems that are actually doing something that looks like cognition, that something that looks like rich, sophisticated thinking.
你会说这仅仅是看起来像,还是确实就是?
Would you say it looks like it or is it?
我相信它确实是,但我说这话部分是基于直觉,而非确凿的数据。
I believe it is it, but I say that partially based on vibes without clear data.
是的
Yeah.
我们能做的最接近的事情,就是在神经网络为你完成某项任务时,打开它的引擎盖观察内部。
The closest we can get is looking inside these things by opening up the hood of a neural network while it's competing something for you.
我们看不到内置的机器。
And we don't see free built machines.
这不像你打开一辆车,然后说:哦,这是化油器,那是散热器,等等。
It's not like you open up a car and you're like, oh, that's for like carburetor, that's for radiator, yada yada.
如果你打开一个神经网络,你会看到一堆神经元按某种序列亮起,看起来难以理解,但你可以找出这些神经元与什么相关。
If you open up a neural network, what you see are a bunch of neurons lighting up in a sequence that seems indecipherable, but you can work out what these neurons correlate to.
所以之前是这样
And so earlier Is this
当你看到随机时,不,看看
seen random when you No, look at
这看起来与随机非常不同。
it seems very different to random.
这看起来非常合理。
It seems highly sensible.
所以之前我说,让我们想象一下,我需要在工作室的后院建一个威格瓦姆小屋。
So earlier I said, let's imagine I need to build a WigWam in the backyard of the studio.
是的。
Yeah.
如果我们让一个神经网络来完成这个任务并打开它,我们会看到一个专门针对‘威格瓦姆’激活的特征。
If we asked a neural network to do that and we opened it up, what we would see would be a feature that would activate for, like, WigWam.
因此,会有一个神经元很可能代表这个概念,因为你有数百万种不同的表征。
So a neuron that probably represents that because you have millions of different representations.
你会看到一些激活的神经元代表木材,另一些代表花园。
You'd have something that fired up that would represent like wood, something that would represent garden.
它会思考这些概念。
It would be thinking of these concepts.
当它开始制定计划时,你会看到类似电路的结构逐渐形成。
And then as it started to do plans, you would see kind of circuits emerge.
所以它会同时思考WigWam以及木材。
So it would think both about a WIGRAM and it would think about like wood.
然后它可能会想到类似花园的东西。
And then it might think about something like the garden.
接着它会思考关于布局之类的内容。
And then it would think about something that would look like placement.
这些神经元之间会开始出现相互关系。
And there would be interrelationships between these neurons that start showing up.
我们把这些称为电路。
We think of them as circuits.
告诉我这些互动是如何发生的。
Tell me how those interactions would happen.
一个神经元会调动其他神经元吗?
Would a neuron recruit other neurons?
想象一下
Think of
把这些神经元想象成一群恒星,它们之间有丝线相连。
these neurons as a a galaxy of stars that have, threads between them.
每颗恒星都有数百条不同的丝线,而且这些丝线的强度各不相同。
Each star has, like, hundreds of different threads, and the threads have different strengths.
可以把这看作是它与其他恒星沟通的能力。
Think of that as, like, its ability to communicate to other stars.
这是通过训练构建出来的。
And you've built this through training.
通过用这些数据训练系统,你最终得到了这种神经元的排列方式,它们代表了不同的事物——这些恒星以及它们之间的连接。
Through training the system on all of this data, you've ended up with this arrangement of neurons which represent different things, these stars and connections between them.
因此,你在这里实际看到的是,哦,某个特征被激活了,它与其它特征有着这样的连接。
And so what you're actually seeing here is, oh, like this feature is firing up and it has these like connections to other features.
而我们还知道,这种连接具有灵活性。
And then what we also know is that there's flexibility.
因此,某些特征会组合在一起,形成一种规划特征,比如:我们需要一步步思考。
So some features will come together to be like a planning feature, which will be like, okay, we need to think step by step.
现在让我们按特定顺序激活其他神经元,以逐步展开时间序列。
Let's now activate other neurons in a certain sequence to unroll steps over time.
最令人惊讶的是,当我们观察这些系统如何思考时,发现它们的运作方式非常合理。
The most surprising thing is that when we when we look at how these things think, it makes a lot of sense.
它显得极其丰富。
It looks incredibly rich.
而且它基于我们内心深处非常直观的事物。
And it is based on things that are deeply intuitive to us.
它们正在观察世界上相同的信息,并可能发展出类似的简略表达方式,从基础视觉开始,比如它们拥有与人类和动物类似的边缘检测器,一直到语言层面,它们最终也会形成代表不同情绪的特征。
They are looking at like the same information in the world and may develop similar shorthands, you know, all the way up from basic vision where they have like edge detectors, same as like humans and animals develop this, you know, to in language, they end up having features that represent different moods, right?
和我们一样,也许是因为这个世界能被表达的方式有限,它们最终收敛于看似相似的表征。
The same as us, because maybe there are only so many different ways to represent the world around us and they converge on representations that seem kind of similar.
用‘思考’这个词来形容它所做的事情是最合适的吗?
Is thinking the best word to use for what it does?
‘思考’只是一个简略的说法。
Thinking is shorthand.
我把它看作是一种被感知的状态。
I think of it as being perceived.
当我说到把我的日记交给这个东西时,我脑海中会浮现这样一个画面:我正站在一只巨大的眼睛前,让自己对它变得清晰可读。
When I say like, give my diary to this thing, I kind of have this mental image that I'm just sort of standing in front of like a big eye and I've made myself legible to the eye.
它以一种奇特而全面的方式在感知我。
And it is perceiving me in a oddly total way.
它在审视所有这些数据,并基于这些数据做出某种推断,这与人类不同。
It's looking at all of this data, you know, and it's making kind of inferences based on that, which is different to people.
我们思考时使用大量简化的表达,而这些东西则依靠海量的输入数据,一次性全部存储在脑海中。
We think in a lot more shorthand, this stuff thinks with a huge amount of input data that it holds in its head all at once.
所以,它可能更像一面镜子或一池水在思考——你的倒影映在其中,而其背后却有着某种非常奇特的认知或复杂性。
So it might be more analogous to like a mirror or a pool that thinks, you know, your reflection's in it, and there's some very, like, strange cognition or complexity underlying it.
但你知道,思考可能是一种深深嵌入时间中的过程。
But, you know, thinking might be something that is very much embedded in time.
我们思考的方式受心跳和循环系统支配,我们自身就是在时间中前行的。
You know, we think in a way that is governed by our heartbeat and our circulatory system, ourselves, like we are going through time.
这些事物不存在于时间中。
These things don't exist in time.
它们存在于我此刻正在感知某物的状态中。
They exist in like, I'm now perceiving something.
这东西需要我
The thing wants me
对一切做出推断,一切都是瞬时的。
to make inferences about Everything's instantaneous.
一切都很奇特地瞬时发生。
Everything is oddly instant.
是的。
Yeah.
我们必须为这种现象创造新的语言。
We have to come up with new language for this.
我认为这正是它令人兴奋的地方,因为我们从未拥有过具有这种特性的工具或技术。
And I think that's what's so exciting about it is we've never had tools or technologies with this property.
这非常不同。
It's very different.
是的。
Yeah.
我认为AI的一个问题是,因为它能做太多事情,所以很难想象它会做什么。
I think one of the issues with AI is that because it could do so much, it's hard to picture what it will do.
这就像一种你不知道用途的工具。
It's like a tool that you don't know the use of the tool.
对于扳手,我们知道如果有螺栓,就可以拧紧它。
With a wrench, we know if there's a bolt, you can tighten it.
这是一种工具。
This is a tool.
它是一个万能工具。
It's an everything tool.
是的。
Yeah.
所以很难想象,拥有这样一个万能工具会如何改变现状。
So it's hard to picture how that changes things having I an everything
几年前我做过一个项目。
did a project a few years ago.
我很乐意把早期版本的这个东西发给你。
I'll happily send it to you with an earlier form of this stuff.
人们创造了一种叫CycleGAN的惊人AI技术。
People had made this amazing AI technology called a CycleGAN.
它们的名字总是很奇怪,但本质上就是:如果我输入一种形式、一种美学风格的图像,能否自动将其转换为另一种美学风格的图像?
They always have weird names, but all it is is if I take in an image of one form, one aesthetic style, can I then automatically translate it to an image with another aesthetic style?
比如,给我一张彩色照片,我可以把它转成黑白,或者反过来。
So it might be, give me a color photograph, I'll turn it to black and white or vice versa.
我在家里的电脑上训练了一个版本,你知道的,就坐在加利福尼亚州奥克兰,将古代地图插图与谷歌地图卫星图像相互转换。
I trained a version of this on my home computer, you know, sitting in Oakland, California to translate between ancient illustrations of maps and Google Maps satellite imagery.
然后我不断输入像巴比伦这样的古代地图,你知道的,几百年前甚至更久远的
And then I was feeding it like ancient maps of like Babylon from, you know, hundreds of hundreds
好主意,顺便说一下,是好多年了。
Great idea, of years by the way.
很好。
Great.
对吧?
Right?
我开始查看卫星的俯瞰图像。
I was getting to look at satellite overhead views.
太棒了。
Great.
这太有趣了。
And it was so much fun.
我感觉自己看到了前所未有的东西——你有没有做过亚特兰蒂斯?
I felt like I seeing something that no one had ever been able to make from Did you do Atlantis?
我没做过亚特兰蒂斯。
I didn't do Atlantis.
我可以回去看看,技术可能有点过时了,但我没做过好奇号。
I can go back and the technology might have a bit rotted, but I didn't do Curious
去看看它从哪里出现。
to see where it comes up.
是的,我发给你。
Yeah, I'll send it for you.
我做过巴比伦、伯利恒,还做过其他几个城市。
I did Babylon, Bethlehem, I did a few other cities.
这又回到了你关于创造力的观点。
And it comes back to your point about creativity.
这个工具有这么多用途,而这种用法从来都没有人想到过。
Like there are so many uses of this tool and that use never suggested itself.
对吧?
Right?
那一定是因为我从小对城市规划、地图、简·雅各布斯这些东西着迷。
That had to be the fact that I grew up obsessed with like town planning and maps and Jane Jacobs and all of this stuff.
然后我开始从事人工智能方面的工作。
And then I ended up working in AI.
它自然而然地出现在我面前。
It presented itself to me.
你怎么创建一个大型语言模型?
How do you create a large language model?
你需要大量的数据,然后对这些数据玩填字游戏。
You take a whole load of data and then you play Mad Libs on that data.
所以如果我跟你说,你知道的,填字游戏就是填空。
So if I say to you, you know, Mad Libs like fill in the blank.
缺失的词语。
The words missing.
是的。
Yeah.
所以如果我跟你说,现在是圣诞节,有人从烟囱里下来。
So if I say to you, it's Christmas and someone's coming down the chimney.
他们的名字叫圣诞老人。
Their name is Santa.
克劳斯。
Klaus.
对吧?
Right?
你的大脑已经补上了。
Your brain's filled it in.
所以想象一下,你正在做同样的事,但规模不是仅限于句子,而是段落、几十页、几百页,涵盖科学、数学、文学等不同领域。
So imagine that you're doing that, but at the scale of not just sentences, but paragraphs, tens of pages, hundreds of pages in different domains, science, math, literature.
因此,你会收集一个非常庞大的数据集,然后通过删除其中的部分内容,训练模型去补全这些缺失的部分。
So you gather this very, very large data set, and then you are training it where you're knocking out chunks of it, and you're training it to complete that.
其背后的思路是,如果我能在这里做出极其复杂的预测,那就说明我真正理解了底层数据集中一些微妙而重要的东西。
And the idea being, if I can make very, very complex predictions here, I must have understood something really subtle and important about the underlying data set.
通过迫使模型在如此庞大的规模上进行预测,你迫使它压缩自身的表达方式,从而发展出简化的符号体系。
And through forcing it to sort of make these predictions at such a large scale, you force it to compress its own representation so that it develops shorthand.
所以,如果你有一百万个数据点,并且正在训练一个神经网络来表示它们,如果你给这个神经网络一百万个参数,它就不会表现出任何这种行为,因为它根本不需要学习捷径来表示这些数据。
So, you know, if I had a million bits of information and I was training a neural network to represent them, if I gave the neural network a million parameters, it would not take on any of this behavior cause it would never need to learn shortcuts to represent it.
它会为每一个数据点分配一个神经元。
It would just have one neuron for each bit of information.
我们在这里做的是,面对我们希望它学习的无数万亿信息,强迫它压缩到可能只有几亿个参数,这意味着每个参数不能只为数据集中的每一个单独事物提供表示。
What we're doing here is we're taking untold trillions of things we want it to learn about and we're forcing it to compress that into maybe a few 100,000,000 parameters, which means that a parameter can't just represent every single individual thing in the dataset per parameter.
你需要找出简化的表达方式,这意味着它必须学会识别事物的特征,以帮助它回答问题。
You need to figure out shorthands, which means it needs to learn features for things to help it kind of answer questions.
因此,在训练过程中,你会使用大规模的数据,要求它做出预测,同时施加一个约束条件。
And so the training processes, there's large scale data that you're then asking it to make predictions about, and you're forcing it to do that with a constraint.
就像创造力一样。
Just like creativity.
你给了它一个人为的限制,迫使它想出创造性的方法,以简化的形式来表示这些数据。
You're giving it an artificial constraint that forces us to come up with creative ways to represent that data to itself in the form of shorthand.
但这种简化最终与思维本身的工作方式紧密相连。
But shorthand ends up being bound up in just how thinking works.
你会发展出各种各样的思维捷径,这里也是一样。
You develop all kinds of shorthand ways of thinking through things, and this is the same.
但这些捷径是在你提出问题的瞬间产生的。
But the shorthand happens in the moment when you ask the question.
庞大的数据集仍然保留。
The large dataset stays.
它并不会压缩成一个更小的数据集,然后就只有这些了。
It's not like it condenses itself to a smaller dataset and that's all that there is going forward.
哦,不,它确实会。
Oh, no, it does.
这就是关键。
That's the thing.
这才是最奇怪的部分。
That's the really weird part.
当你训练这个系统时,比如训练一个神经网络,你会有一个庞大的数据集。
When you train this system, say you're training a neural network, you've got this big dataset.
你会花上数月、数月、再数月的时间来训练它,直到它能对你的大规模实验做出大量预测。
You're going to spend months and months and months training it until it has made an over predictions it can make about this giant experiment you're doing.
这就是你的成果。
That is your thing.
你把它保存下来。
You save that.
那些数据,你就不再使用了。
The data, you're no longer using.
真的吗?
Really?
你现在拥有了这样一个东西,它包含数百万个神经元,彼此之间有着极其复杂的关系,仅此而已。
You now have this thing, which is millions and millions of, like, neurons with some very complex relationship between them, and that's it.
我觉得这就像把所读数据的本质浓缩装瓶了,但它本身并不是数据。
I think of it as like a a bottled up distillation of the kind of data data it's read, but it's not the data.
它是一个曾经思考过这些数据的实体。
It's a thing that has thought about that data.
是的。
Yeah.
现在当你启动它时,你可以向它提问。
And now when you boot it up, you ask questions about it.
这就像用它对事物的简略思考方式,重新向自己表达其中一部分内容。
It's like rerepresenting some of that to itself using its shorthand thinking about things.
而这部分解释了它为何具备创造力,因为它并不是简单地说:‘让我去参考一下这个库。’
And that's partly why it has the capacity for creativity because it's not just saying, oh, let me go and, like, refer back to this this library.
它无法访问那个库。
It doesn't have access to the library.
它必须完全靠自己完成。
It has to do it all itself.
这听起来有些令人担忧。
Something about that seems worrying.
是的。
Yeah.
因为这就像是你只有莎士比亚的简写版,是的。
Because it's like if you have the cliff notes of Shakespeare Yeah.
你想要的莎士比亚就只是这些简写版吗?
Is all that you want from Shakespeare in the cliff
注释吗?
notes?
它与我们自身的思维方式紧密相连。
It is bound up in in how we ourselves think.
比如,我非常喜欢一位叫W的诗人。
Like, I I love a poet called W.
B。
B.
叶芝。
Yeats.
对吧?
Right?
我记得自己背过几首叶芝的诗,但记不全所有的。
And I can remember a small number of Yeats poems that I've memorized, but I can't remember all of them.
我就像W.的简明版。
I am the cliff notes of, like, W.
B.
B.
叶芝。
Yeats.
是的。
Yes.
然而,叶芝的诗让我联想到其他记忆和经历。
And yet, like, Yeats relates to other memories I have and experiences I have.
所以你接触这些内容,并不是为了说:你就是莎士比亚。
So you aren't going to this stuff to say, you are Shakespeare.
如果你想要那样,你就去读原著。
That if you want that, you go and read the source.
你去那里是为了说:哦,我有个问题可能涉及到了。
You're going to it to say, oh, I have a question which might touch.
听听那些读过大量莎士比亚作品、对它有一定直觉的人的看法,可能会很有帮助。
It might be useful to hear from someone who's read a lot of Shakespeare and has some kind of intuitions about it.
这就是为什么整个事情如此令人兴奋,同时也让人产生一些合理的担忧。
It's why this whole thing is so exciting and also why people have some appropriate fear about it.
你认为人们提问时是想要一个答案,还是想要那个正确的答案?
Do you think that people want an answer or do they want the answer when they ask a question?
我认为从实际行为来看,人们想要的是一个答案。
I think the revealed preference is people want an answer.
一个答案。
An answer.
是的。
Yeah.
但他们可能会说:我想要那个正确的答案。
But they will say like, I want the answer.
似乎得到答案是有目的的,因为你不会再思考那个问题了。
It seems like getting an answer serves a purpose because you don't think about the question anymore.
它满足了获得回报的需求。
It satisfies the need to get something back.
你有一个问题,得到了一个答案,无论对错,你就不再去想它了。
You have a question, you get an answer, right or wrong, and you stop thinking about it.
你可以把这些限制在这一点上。
And you can constrain these things to that.
一种思考方式是,假设我是一所大学,拥有成千上万篇我们大学发表的研究论文。
One way to think about it is, say I'm a university and I have thousands of research papers that my university has written.
我可以引入这些人工智能系统,把它们变成有史以来最棒的图书管理员,说:嘿,你现在可以访问这些论文。
I can take in these AI systems and I can turn them into like, the best librarian ever and say, hey, you now have access to these papers.
我会向你提问。
I'm gonna ask you questions.
你要运用你的思维,从这些论文中找出正确答案。
You're gonna use your mind to go and figure out the correct answer from these papers.
所以你可以把这些工具当作一个有品位的聪明人来使用。
So you can use these almost like a a smart person with taste.
去帮助你找到答案。
To go and help give you the answer.
但很多时候,其他用途则更像:我希望你作为一个更通用的存在。
But a lot of the times, the other uses are more like, I want you as a It's more general.
所以我希望你作为一个普通人,进行一般的交流。
So I want you as a general person to talk with generally.
你可以。
You can
可以让这些工具更具体,但它们的精髓在于通用性。
make these things more specific, but the magic in them is for generality.
我明白了。
I see.
要让它们变得具体,是你来提供信息吗?
And to make them specific, you would be the one putting the information in?
经常是的。
Often, yeah.
如果你愿意,我们可以在商业环境中看到更常见的例子:如果我是一家公司,我让这个工具帮我分析我的业务数据,但它其实并没有见过我的业务数据。
If you want, we see this in a more mundane context in business where if I'm a company and I'm asking this thing to help look at my business data, it hasn't seen my business data.
我必须把数据提供给它,然后说:你现在可以访问我的业务数据了。
I have to just give it to us and be like, You now have access to my business data.
当我向你提问时,我希望你能参考这些业务数据,而不是凭空捏造。
When I ask you questions, I need you to look at this business data and not make stuff up.
是的。
Yeah.
他们会说:明白了,老板,我们会这么做的。
They'll be like, Got it boss, we'll do that.
对。
Yeah.
现在,假设我把我的业务数据输入进去,这些数据是否只在回答我的问题时存在,还是说它会成为机器整体智慧的一部分?
Now, when you put, let's say I were to put my business data in, does that business data only live in response to my question or is that now part of the general wisdom of the machine?
仅在响应你的问题时,因为这个系统已经训练好了。
Only in response to your question, because the thing has been trained.
它就像一个生物。
It's like a creature.
你不能重新
You can't re
继续训练。
up- It's not continuing to train.
不,它没有。
No, it is not.
而且这更像是你只是暂时给了它访问你数据的权限。
And it's more like you've just given it access to your thing temporarily.
我
I
明白了。
see.
你知道吗,你可以在香格里拉这里这么做,对吧?
You could do this you know, here at Shangri La, right?
你可以给一个AI系统访问你所有的音乐和母带的权限,但你永远不希望它进入那个系统。
You could give an AI system access to all of the music and, like, masters you have, but you never want it to go into the thing.
但你可能会说,嘿,帮我找找,在那次录音中,我特别喜欢的那个声音在哪。
But you might say, hey, like, help me find there was a particular sound I really liked from like that session.
我觉得我这儿有个听起来像那样的东西。
I think I've got something that sounds like it here.
在我的数据集中,哪个东西最像那个?
What's the thing that sounds most like that in my corpus?
它会运用自己的智能去找到那个东西并把它带给你。
And it'll go and use its intelligence to find the thing and bring it to you.
酷。
Cool.
那么目前,有五到六家大型语言模型公司吗?
So at the moment, are there five or six big large language model companies?
可能是四到五家。
Maybe four or five.
四到五家。
Four or five.
它们是哪些?
What are they?
会有Anthropic,也就是我工作的公司,OpenAI,Google的XAI,这是埃隆的公司,还有Meta,也就是Facebook,实际上还有DeepSeek。
There would be Anthropic, which is where I work, OpenAI, Google XAI, which is Elon's company, and Meta, aka Facebook, and then actually DeepSeek.
所以这更像是五到六家。
So that's more like five or six.
六家。
Six.
好的。
Okay.
那其他所有的AI产品都是这些公司的接入点吗?
And then all of the other AI products are access points for those?
它们中的大多数都建立在这些系统之上。
The majority of them are sitting on top of these things.
这些系统就是这些大型大脑,其他人工智能产品则会依赖于它们。
These things are these big brains and the other AI products will be cooling down to them.
你觉得
Do you
将来会超过这六个吗?
think there will be more than the six over time?
我的预期是,你所处的领域中,数量肯定少于十个。
My expectation is you're going to be in a domain where there's definitely less than 10.
再往后,我就很难推理了。
And after that, it gets hard for me to reason about.
这是因为,也许在2019年,处于前沿领域需要大约10万美元的成本,也就是训练一个系统所需的投入。
And this is because maybe in 2019 being at the frontier would cost you about a $100,000 as in that's your input cost to train a system.
到了2020年和2022年,成本上升到了数百万美元,大约在1000万到1500万美元之间;2024年,这些系统的成本达到数千万到低数亿;而到2025年,这些系统将耗资数亿美金。
Then it went to kind of millions of dollars in twenty twenty, twenty twenty two order of 10 to 15,000,000, 2024 the systems cost multiple tens of millions to low hundreds, 2025, these systems will cost hundreds of millions.
为什么呢?
And why is that?
我们使用了更多的计算机。
We're using way more computers.
所以这是硬件问题。
Based So it's hardware.
这是硬件。
It's hardware.
仅仅是硬件。
Only hardware.
这才是主要的成本。
That is the vast cost of it.
是的。
Yeah.
这其中一部分原因不仅是我们知道如何训练这些系统,而且我们在让它们更大、并在训练过程中延长它们的思考时间方面也变得更加熟练。
And some of this is because it's not just that we know how to train these things, but we we've got better at both making them bigger and also helping them think for longer during training.
我们现在允许它们在训练过程中玩游戏。
And we are now allowing them to like play games during training.
你之前提到了AlphaGo。
So you mentioned AlphaGo earlier.
AlphaGo是一个专门被训练来下围棋的单一系统。
AlphaGo was a single system that was trained to play Go.
它只做这一件事。
That's all it ever did.
而且它带来了惊人的成果。
And it yielded amazing stuff.
现在,当你训练这些AI系统时,你可能会在训练过程中给它们一个环境,比如一个围棋棋盘,让它们同时下围棋。
Well, now while you're training these AI systems, you might have them during training, have an environment, but as a Go board and it's playing Go as well.
然后也许它们还在下国际象棋,也许还在进行数学证明,而数学证明是
And then maybe it's also playing chess and maybe it's also doing a mathematical proofs and a math proof is
同时能进行的事情数量是有限的吗?
a limit to the amount of things that can happen at the same time?
不是真的。
Not really.
实际上,限制在于我们人类为这些系统创建训练环境的能力,以及如何为它们设计更复杂的游戏来进行训练。
It's like the limit actually is on the ability of us as humans to create environments for this stuff to train in and figuring out how to give it more complicated games to play during training.
拥有多个大型语言模型的好处是什么?
What's the benefit of there being more than one large language model?
我认为其中一部分关乎个性和品味。
I think some of it is about personality and taste.
我认为,如果你只有一个,最终会形成一种同质化的‘神 mind’,这似乎很糟糕,因为你本能地会觉得,这样会受到限制。
I think if you just had one, you end up with some, like the homogenous godmind seems bad, like just innately you think, well, that seems like you're gonna be limited.
这些系统更像是被培养出来的,而不是被制造出来的。
These things are more grown than made.
因此,任何开发者所做的各种事情都会累积成一种特质,从而改变这些模型所谓的‘个性’。
And so all of the different things that any developer does add up to a texture that will change the like quote unquote personality of these models.
它们都会拥有不同的价值观、倾向性和能力。
They'll all have different values, proclivities and capabilities as well.
关于 Bayt 播客
Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。