Hard Fork - 埃兹拉·克莱因秀:人工智能代理将多快席卷经济? 封面

埃兹拉·克莱因秀:人工智能代理将多快席卷经济?

The Ezra Klein Show: How Fast Will A.I. Agents Rip Through the Economy?

本集简介

本周,“Hard Fork”团队休息,享受急需的假期。在我们离开期间,我们想向您推荐《伊扎·克莱因秀》的一期最新节目。 在这次对话中,伊扎与Anthropic联合创始人杰克·克拉克探讨了他如何使用AI代理、这项技术如何正在深刻改变我们的工作与思考方式,以及政策如何必须调整以应对即将来临的就业岗位流失。 我们下周将带着新一期节目回归。 嘉宾: 杰克·克拉克,Anthropic联合创始人兼政策主管。 延伸阅读: 本集的完整文字稿和视频请见此处。 欢迎与我们交流,请发送邮件至 hardfork@nytimes.com。在YouTube和TikTok上关注“Hard Fork”。 今天就订阅吧,访问 nytimes.com/podcasts,或在Apple Podcasts和Spotify上订阅。您也可以通过您喜爱的播客应用订阅:https://www.nytimes.com/activate-access/audio?source=podcatcher。如需收听更多播客和有声文章,请下载《纽约时报》应用:nytimes.com/app。 由Simplecast(AdsWizz公司旗下)提供托管。有关我们为广告目的收集和使用个人信息的详情,请访问 pcm.adswizz.com。

双语字幕

仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。

Speaker 0

你好。

Hi.

Speaker 0

我是索拉纳·莱恩。

I'm Solana Pine.

Speaker 0

我是《纽约时报》视频部门的总监。

I'm the director of video at The New York Times.

Speaker 0

多年来,我的团队制作了大量视频,带您更贴近重大新闻时刻,这些视频由《纽约时报》记者制作,凭借专业能力帮助您理解正在发生的事。

For years, my team has made videos that bring you closer to big news moments, videos by Times journalists that have the expertise to help you understand what's going on.

Speaker 0

现在,我们将这些视频带到《纽约时报》应用的“观看”标签页中。

Now we're bringing those videos to you in the watch tab in The New York Times app.

Speaker 0

这是一个专属的视频频道,您可以完全信任其中的内容。

It's a dedicated video feed where you know you can trust what you're seeing.

Speaker 0

那里所有的视频都免费向所有人开放。

All the videos there are free for anyone to watch.

Speaker 0

您无需成为订阅用户。

You don't have to be a subscriber.

Speaker 0

下载《纽约时报》应用程序开始观看。

Download The New York Times app to start watching.

Speaker 1

你好,Hard Fork 的听众们。

Hello, Hard Fork listeners.

Speaker 1

希望你们这一周过得愉快。

We hope you're having a great week.

Speaker 1

我们会跟你们说真心话。

We're gonna be really honest with you.

Speaker 1

我们已经没什么可说了。

We have nothing left to give.

Speaker 1

我们筋疲力尽了。

We are spent.

Speaker 1

我们累坏了,这周要彻底休息一下。

We're exhausted, and we're taking the danged week off.

Speaker 2

是的。

Yeah.

Speaker 2

这周我们正值春假休假,但不会让你空手而归。

We are on vacation this week for our spring break, but we are not leaving you empty handed.

Speaker 1

因为我们绝不会对你这么做。

Because we would never do that to you.

Speaker 2

我们想跟你分享一期《Ezra Klein秀》。

We wanted to share an episode of The Ezra Klein Show with you.

Speaker 2

一款精致的小播客。

Little artisanal podcast.

Speaker 2

你可能已经听说过它。

You may have heard about it.

Speaker 1

如果你还不了解Ezra Klein,他是一位崭露头角的访谈主持人、政策专家,也是人类的朋友。

If you haven't heard of Ezra Klein, he's an up and coming interviewer, policy wonk, and friend of humanity.

Speaker 1

他有一档会让你大呼过瘾的播客。

And he has a podcast that'll knock your socks off.

Speaker 1

最近,他邀请了Anthropic的联合创始人兼现任政策主管Jack Clark做客。

And recently, he was joined by Jack Clark, a co founder of Anthropic and its current head of policy.

Speaker 1

事实上,这一集发布后,宣布杰克将领导一个名为‘Anthropic研究所’的项目,该机构将整合Anthropic内部的各项研究,以‘提供信息,帮助其他研究人员和公众在我们向一个拥有更强大AI系统的世界过渡的过程中应对挑战’。

In fact, after this episode was published, it was announced that Jack would be leading something called the Anthropic Institute, which will draw on research from across Anthropic to, quote, provide information that other researchers and the public can use during our transition to a world containing much more powerful AI systems.

Speaker 1

这难道不令人不安吗?

So how's that for Ominous?

Speaker 2

是的。

Yeah.

Speaker 2

杰克是一位非常出色的思想者和演说家。

And Jack is a great thinker and talker.

Speaker 2

他撰写了一份非常优秀的通讯《Import AI》。

He writes the great newsletter Import AI.

Speaker 2

他曾经是一名记者。

He's also a former journalist.

Speaker 2

许多人称他是唯一一位做出过明智职业选择的前记者。

Many are calling him the only former journalist who's ever made a good career decision.

Speaker 2

他和伊扎·克莱因进行了一场深入的对话,探讨了当前经济的现状、AI代理和编程工具的兴起,以及工作的未来。

And he and Ezra had a great in-depth conversation about all of what's going on in the economy right now, the rise of AI agents and coding tools, the future of work.

Speaker 2

基本上,就是很多我们认为我们的听众会感兴趣想了解更多的话题。

Basically, just a lot of things that we thought our listeners would be interested in hearing more about.

Speaker 1

杰克还带着一种浑厚的英式口音,我相信你在跑步或做家务时听他说话会感到非常愉快。

Jack also has a sonorous British accent that I think you will find to be incredible company as you go for a run or do your laundry today.

Speaker 2

是的。

Yeah.

Speaker 2

真是神奇,仅仅拥有英式口音就能让人智商仿佛提升了15点。

It's amazing how just having a British accent adds, like, 15 IQ points.

Speaker 1

这种声音非常舒缓,这在讨论人类生存威胁时尤为重要。

It's very soothing, which is important when you're talking about existential threats to humanity.

Speaker 2

接下来是埃兹拉和杰克的对话。

So here's Ezra and Jack.

Speaker 2

我们下周会带来全新的一期节目。

We'll be back next week with a brand new episode.

Speaker 1

如果你够幸运的话。

If you're lucky.

Speaker 3

过去几年报道人工智能时,我们通常谈论的是未来。

The thing about covering AI over the past few years is that we're typically talking about the future.

Speaker 3

每一个新模型,尽管令人印象深刻,都像是即将问世的模型的原型验证。

Every new model, impressive as it was, seemed like proof of concept for the models that would be coming soon.

Speaker 3

那些能够可靠地独立完成实际工作的模型。

The models that could actually do useful work on their own reliably.

Speaker 3

那些真正会让工作过时,或让新事物成为可能的模型?

The models that would actually make jobs obsolete or new things possible?

Speaker 3

这些模型对劳动力市场、我们的孩子、我们的政治以及我们的世界意味着什么?

What would those models mean for labor markets, for our kids, for our politics, for our world?

Speaker 3

我认为,那种我们总在谈论未来的时期,现在结束了。

I think that period in which we're always talking about the future, I think it's over now.

Speaker 3

我们一直等待的那些听起来像科幻小说的模型——能够自主编程,且比大多数程序员更快、更好的模型,能够编写自己的代码来提升自身的模型——现在已经出现了。

Those models we were waiting for, the sci fi sounding models that could program on their own and do so faster and better than most coders, the models that could begin writing their own code to improve themselves, those models are here now.

Speaker 3

它们现在就在Anthropic的Cloud Code中。

They're here in Cloud Code from Anthropic.

Speaker 3

它们现在出现在OpenAI的Codex中。

They're here in Codex from OpenAI.

Speaker 3

它们正在冲击股市。

They are shaking the stock market.

Speaker 3

标普500软件行业指数已下跌20%,蒸发了数十亿美元的价值。

The S and P 500 software industry index has fallen by 20%, wiping billions of dollars in value out.

Speaker 3

优秀的工程师,我认识多年的、对AI炒作相当怀疑的人。

Excellent engineers, people I've known for years, people who are quite skeptical of AI hype.

Speaker 3

他们现在给我发邮件,说他们看不到自己的工作还能再存在一两年。

They're emailing me now to say they don't see how their job will possibly exist in a year or two.

Speaker 3

我们正处于AI发展的新阶段。

We are at a new stage of AI development.

Speaker 3

不仅仅是发展。

Not just development.

Speaker 3

我们正处于AI产品的新阶段。

We are at a new stage of AI products.

Speaker 3

我认为红杉资本对这一点的表述其实很有帮助。

I thought the way Sequoia, the venture capital firm, put it was actually pretty helpful.

Speaker 3

2023年和2024年的AI应用都是‘话痨’型的。

The AI applications of 2023 and 2024 were talkers.

Speaker 3

有些非常擅长对话,但影响力有限。

Some Some were very sophisticated conversationalists, but their impact was limited.

Speaker 3

2026年和2027年的AI应用将成为‘行动者’。

The AI applications of 2026 and 2027 will be doers.

Speaker 3

换句话说,人们长期预测的事情如今终于发生了。

Or to put it differently, something that's been predicted for a long time has now happened.

Speaker 3

我们正从聊天机器人转向智能代理,从与你对话的系统转向替你行动的系统。

We are moving from chatbots to agents, from systems that talk to you to systems that act for you.

Speaker 3

在这个代理世界里,事情已经变得很怪异了。

And this world of agents, it's already weird.

Speaker 3

它们是多个代理。

They are agents plural.

Speaker 3

它们可以协同工作。

They can work together.

Speaker 3

它们可以相互监督。

They can oversee each other.

Speaker 3

人们正在代表自己运行这些代理的群体。

People are running swarms of these agents on their behalf.

Speaker 3

无论现阶段这些代理是让人类更高效了,还是只是更忙碌了,我还不太确定,但如今确实可以随时拥有一支极其快速、尽管说实话有些古怪的软件工程师团队,随叫随到。

Whether that is making them at this stage more productive or just busier, I can't quite tell, but it is now possible to have what amounts to a team of incredibly fast, although, to be honest, somewhat peculiar software engineers, at your beck and call at all times.

Speaker 3

杰克·克拉克是Anthropic公司的联合创始人兼政策主管,该公司开发了Claude和Claude Code。

Jack Clark is the cofounder and head of policy at Anthropic, the company behind Claude and Claude Code.

Speaker 3

多年来,克拉克一直在每周通讯《Import AI》中追踪不同模型的能力,这份通讯一直是我关注AI发展动态的重要读物。

And for years now, Clark has been tracking the capabilities of different models in the weekly newsletter Import AI, which has been one of my key reads for following developments in AI.

Speaker 3

因此,我想了解他是如何解读当前这一时刻的——无论是技术在他眼中如何变化,还是政策应该如何或能够随之调整。

So I want to see how he is reading this moment, both how the technology is changing in his view and how policy needs to or can change in response.

Speaker 3

一如既往,我的邮箱是EzraKleinShow@nytimes.com。

As always, my email, EzraKleinShow@nytimes.com.

Speaker 3

杰克·克拉克,欢迎来到节目。

Jack Clark, welcome to the show.

Speaker 4

谢谢您邀请我,伊扎。

Thanks for having me on, Ezra.

Speaker 3

我想很多人对AI聊天机器人都很熟悉。

So I think a lot of people are familiar with AI chatbots.

Speaker 3

是的。

Mhmm.

Speaker 3

但什么是AI代理呢?

But what is an AI agent?

Speaker 4

最好的理解方式是,它是一个能够使用工具并长期为你服务的语言模型或聊天机器人。

The best way to think of it is like a language model or a chatbot that can use tools and work for you over time.

Speaker 4

当你和聊天机器人交谈时,你就在对话中,与它来回互动。

So when you talk to a chatbot, you're there in the conversation, you're going back and forth with it.

Speaker 4

而代理则不同,你可以给它一些指令,然后它就会离开去为你做事,有点像和一位同事合作。

An agent is something where you can give it some instruction, and it goes away and does stuff for you, kind of like working with a with a colleague.

Speaker 4

几年前,我自学了一些基础编程,在业余时间构建了一个物种模拟程序,里面有捕食者和猎物、道路,几乎像一个二维策略游戏。

So I've got I've got an example where a few years ago, I taught myself some basic programming, and I built a a species simulation in my spare time that had predators and prey and roads and almost like a two d strategy game.

Speaker 4

最近圣诞节期间,我让Claude Code帮我实现这个项目。

Recently asked over Christmas Claude Code to just implement this for me.

Speaker 4

大约十分钟内,它不仅写出了一个基础模拟程序,还自动生成了所有需要的包和可视化工具,让整个程序比我自己写的更好看、更完善。

And in about ten minutes, it went and wrote not only a basic simulation, but all of the different packages that it needed and all of the visualization tools that it might need to be prettier and better than the thing I'd written.

Speaker 4

最终返回的结果是我知道即使是有经验的程序员也可能需要几个小时甚至几天才能完成的复杂项目,而这个系统却在几分钟内就搞定了。

And what came back was something that I know would probably take a skilled programmer several hours or maybe even days because it was quite complicated, and the system just did it in a few minutes.

Speaker 4

它之所以能做到这一点,不仅是因为它聪明地理解了如何解决这个任务,还因为它创建并运行了一系列子系统,这些子系统作为其他代理替它工作。

And it did that by not only being intelligent about how to solve the task, but also creating and running a range of subsystems that were working for it, other agents that worked on its behalf.

Speaker 3

但这意味着什么?

But what does that mean?

Speaker 3

也就是说,多代理系统是什么样子的?

Like, what is a a multi agent setup look like?

Speaker 3

在这个案例中

In the case

Speaker 4

对于Claude Code来说,对我来说就是同时打开多个标签页,运行多个不同的代理。

of Claude code, for me, it's having multiple different tabs running multiple different agents.

Speaker 4

但我见过一些同事,他们会写一个规范文件,用于运行其他Claude实例的版本。

But I've seen colleagues who write what you might think of as a a specification file for a version of Claude that runs other Claude's.

Speaker 4

所以他们会说,我有五个代理,它们由另一个代理进行监督,这个代理会监控它们的行为。

And so they're like, I've got my five agents, and they're being minded over by this other agent which is monitoring what they do.

Speaker 4

我认为这将成为常态。

I think that that's just going to become the norm.

Speaker 3

所以我一直听说,也部分亲身体验到,人们对Claude Code有两种截然不同的体验:一是简直不敢相信这竟然这么简单,是的。

So one thing I've been hearing and somewhat experiencing is two very different categories of experience people have with Cloud Code, which is I cannot believe how easy this is Yep.

Speaker 3

一切都能正常运行。

And everything just works.

Speaker 3

哦,这比我想象的难多了。

And, oh, this is a lot harder than I thought it would be.

Speaker 3

是的。

Yep.

Speaker 3

而且事情总在出错,我根本不知道该怎么修复它们。

And things keep breaking, and I don't really understand how to fix them.

Speaker 3

是什么原因导致有些人能用云代码生成可用的软件,而另一些人却遇到大量bug,甚至搞砸了东西,连怎么跟它沟通都不清楚?

What accounts for being able to get Cloud Code to produce working software versus it creates buggy, often mess up things, and you don't even know how to talk it out of that.

Speaker 4

我认为很大一部分原因在于,人们误以为Claude代码像一个博学的人,而不是一个只能通过互联网交流、极其字面化的助手。

I think so much of it is is making the mistake of thinking Claude Code is like a knowledgeable person versus an extremely literal person that you can only talk to over the Internet.

Speaker 4

我本人就有这样一个例子:当我第一次用Claude代码写物种模拟时,我只是用一段非常粗糙的语言简单地要求它做这件事,结果它生成了一些问题很多但居然能运行的代码。

And I I had this example my myself where I when I did my first pass of writing the, like, species simulation with Claude code, I just sort of asked it to do the thing in in extremely crappy language over the course of a paragraph, And it produced some horribly buggy stuff that just kind of worked.

Speaker 4

然后我做了另一件事:我对Claude说,我要用Claude代码写一些软件。

What I then did is I then just said to Claude, hey, I'm gonna write some software of Claude code.

Speaker 4

我想让你采访我,了解我想构建的这个软件,然后把它整理成一份我可以用作代码规范的文档。

I want you to interview me about this software I want to build and turn that into a specification document that I can give called code.

Speaker 4

这一次,它表现得非常好,因为我把任务结构得足够具体和详细,系统才能顺利处理。

And then that time, it worked really, really well because I'd structured the work to be specific enough and detailed enough that the system could work with it.

Speaker 4

所以,很多时候,问题不在于你是否知道任务是什么。

So often, it's not just knowing what the task is.

Speaker 4

因为你可以和我讨论一个任务,你有直觉,会问我一些深入的问题,所有这些都很重要。

Because you and I could talk about a task to do and you have intuition, you ask me probing questions, all of this stuff.

Speaker 4

关键是确保你设置好了,就像把一封信息装进瓶子里扔进去,它就会自己去完成大量工作。

It's making sure that you've set it up it's like a message in a bottle that you can chuck into the thing, and it'll go away and do a lot of work.

Speaker 4

所以,这条信息必须极其详细,真正捕捉到你想要实现的目标。

So that message better be extremely detailed and really capture what you're trying to do.

Speaker 3

过去几年里,是什么突破让这一切成为可能?

What were the breakthroughs over the past couple of years that made that possible?

Speaker 4

大部分情况下,我们只是需要让AI系统足够智能,当它们犯错时,能够意识到自己犯了错,并知道需要采取不同的做法。

Mostly, we just needed to make the AI systems smart enough that when they made mistakes, they could spot that they'd make a mistake and knew that they needed to do something different.

Speaker 4

所以,归根结底,这不过是让系统变得更聪明,并给它们一点引导工具,帮助它们为你做有用的事情。

So, really, what this came down to was just making smarter systems and giving them a bit of a coaxing tool to help them do useful stuff for you.

Speaker 3

这里的‘更智能的系统’指的是什么?

What does smarter systems mean there?

Speaker 3

你还是会听到一种说法,说这些不过是高级的自动补全机器。

There there's still an argument you'll hear that these are are fancy autocomplete machines.

Speaker 3

它们只是在预测下一个词元,几个词元组成一个词。

They're just predicting the next token, couple tokens make a word.

Speaker 3

它们并没有真正的理解。

They don't have understanding.

Speaker 3

在这个框架下,智能或不智能并不是一个相关概念。

Smart or not smart is not a relevant concept in in that frame.

Speaker 3

在‘智能’这个词里缺少的是什么,或者在那种理解里缺少的是什么?

Either what is missing in the word smart or what is missing in that understanding?

Speaker 3

当你说到‘让它更智能’时,你具体指的是什么?

What what do you mean when you say make it smarter?

Speaker 4

这里的‘智能’是指,我们已经让AI系统对世界有了足够广泛的了解,以至于它们开始展现出类似直觉的东西。

Smart here means we've made for AI systems have a broad enough understanding of the world that they've started to develop something that looks like intuition.

Speaker 4

你会看到,当它们在内心自言自语地描述如何完成任务时,会说:‘杰克让我去找这篇特定的研究论文,但我查了档案,没找到它。’

And you'll see this where if they're narrating to themselves how they're solving a task, they'll say, Jack asked me to go and find this particular research paper, but when I look in the archive, I don't see it.

Speaker 4

也许是因为我找错地方了。

Maybe that's because I'm in the wrong place.

Speaker 4

我应该去别处找。

I should look elsewhere.

Speaker 4

你知道的,就是这样。

You know, like, there you go.

Speaker 4

你现在对如何解决问题有了一些直觉。

You've got some intuitions for how to solve a problem now.

Speaker 3

他们是如何发展出这种直觉的?

How do they develop that intuition?

Speaker 4

以前,训练这些AI系统的方式是使用大量文本,让它们尝试预测文本内容。

Previously, the whole way you trained these AI systems was on a huge amount of text and just getting them to try and make predictions about it.

Speaker 4

但近年来,所谓推理系统的兴起在于,你现在训练它们不仅要进行预测,还要解决问题。

But in recent years, the rise of these so called reasoning systems is you're now training them to not just make predictions but solve problems.

Speaker 4

这依赖于将它们置于各种环境中,从电子表格到计算器再到科学软件,让它们使用工具并学会完成更复杂的事情。

And that relies on them being put into environments ranging from a spreadsheet to a calculator to scientific software, using tools and figuring out how to do more complicated things.

Speaker 4

这样做的结果是,你得到了一些AI系统,它们已经学会了什么是解决需要很长时间、会遇到死胡同并需要自我重置的问题。

The resulting sort of outcome of that is you have AI systems that have learned what it means to solve a problem that takes quite a while and requires them running into dead ends and needing to reset themselves.

Speaker 4

这使它们具备了为你们解决一般性问题并独立工作的直觉。

And that gives them this general intuition for problem solving and working independently for you.

Speaker 3

你仍然认为这些AI系统只是增强版的自动补全吗?还是你觉得这个比喻已经失效了?

Do you still see these AI systems as a souped up autocomplete, or do you think that that metaphor has lost its power?

Speaker 4

我现在对这些系统的看法是,它们就像一些麻烦的小精灵,我可以给它们下指令,它们就会去为我做事。

The way that I think of these systems now is that they're like little troublesome genies that I can give instructions to, and they'll go and do things for me.

Speaker 4

但我仍然需要精确地指定指令,否则它们可能会做错一点事情。

But I need to specify the instructions still just right or else they might do something a little wrong.

Speaker 4

所以这和我输入一句话,它就给出一个好答案,然后结束,是完全不同的。

So it's very different to I type into a thing, it figures out a good answer, that's the end.

Speaker 4

现在的情况是我召唤这些小东西去为我做事,我必须给它们正确的指令,因为它们会离开很长一段时间,执行一系列操作。

Now it's a case of me summoning these little things to go and do stuff for me, and I have to give them the right instructions because they'll go away for quite some time and do a whole range of actions.

Speaker 3

但至少自动补全这个比喻还提供了一个视角,说明这些系统在做什么。

But but the autocomplete metaphor at least had a perspective on what it was these systems were doing.

Speaker 3

是的。

Mhmm.

Speaker 3

它是一个预测模型。

It was a prediction model.

Speaker 3

对。

Mhmm.

Speaker 3

我对此感到困惑,因为根据我对数学和强化学习的理解,我们仍然在处理某种预测模型。

I have trouble with this because as my understanding of the math and the reinforcement learning goes, we're still dealing with some kind of prediction model.

Speaker 3

但另一方面,当我使用它们时,我的感受却并非如此。

And on the other hand, when I use them, it doesn't feel that way to me.

Speaker 3

对吧?

Right?

Speaker 3

感觉那里有一种直觉。

It feels like there's intuition there.

Speaker 3

感觉有很多上下文被调动起来了。

It feels like there is a lot of context being brought to bear.

Speaker 3

如果它确实是一个预测模型,那它和说‘我是一个预测模型’也没什么不同,对吧。

To the extent it it's a prediction model, it doesn't feel that different than saying I'm a prediction Mhmm.

Speaker 3

模型。

Model.

Speaker 3

我不是说你不能欺骗它。

Now I'm not saying you can't trick it.

Speaker 3

我不是说你不能超越它。

I'm not saying you can't get beyond it.

Speaker 3

它的测量结果。

It its measurements.

Speaker 3

所以一方面,我不认为这些现在只是花哨的自动补全系统。

So on the one hand, I don't think these are now just fancy autocomplete systems.

Speaker 3

另一方面,我不确定哪种隐喻才合适。

And on the other hand, I'm not sure what metaphor makes sense.

Speaker 3

精灵我不喜欢,因为那样你就直接陷入了神秘主义。

Genies, I don't like because then you've just moved straight into mysticism.

Speaker 3

对吧?

Right?

Speaker 3

那么你就只是说它们是一种完全不同的生物,拥有巨大的力量。

Then you've just said they're just a completely alternative creature with vast powers.

Speaker 3

你如何理解这些系统?你知道,Anthropic 的人总是告诉我,你应该把它们看作是‘生长’出来的。

What do you understand these systems that, you know, Anthropic people always tell me you should talk about them as being grown?

Speaker 3

是我们培养,或者你培养人工智能。

It's a we we grow or you grow AIs.

Speaker 3

你如何解释它究竟是什么

How do you explain what it is

Speaker 4

它们现在在做什么?

that they're doing now?

Speaker 4

这是个好问题,我认为即使是对这项技术非常接近的技术人员来说,答案仍然难以解释。

It's a good question, and I think the answer is is still hard to explain even as technologists for the close to this technology.

Speaker 4

因为我们把原本只能预测事物的东西,赋予了它在世界上采取行动的能力。

Because we've taken this thing that could just predict things, and we've given it the ability to take actions in the world.

Speaker 4

但有时它会做出一些非常反直觉的行为。

But sometimes it does something deeply unintuitive.

Speaker 4

这就像是你有一个东西,它一生都生活在图书馆里,从未踏出过门外。

It's like you've you've had a thing but has spent its entire life living in a library and has never been outside.

Speaker 4

而现在你把它放到了世界上,它拥有的只有书本上的知识,却缺乏真正的街头智慧。

And now you've unleashed it into the world, and all it has are its book smarts, but it doesn't really have kind of street smarts.

Speaker 4

因此,当我思考这些问题时,我把它看作是一个极其博学的机器,它具备一定的自主性,但很可能以我无法理解的方式陷入极度困惑。

So when I conceptualize this stuff, it's really thinking of it as an extremely knowledgeable kind of machine that has some amount of autonomy, but is likely to get wildly confused in ways that are unintuitive to me.

Speaker 4

也许‘神灵’这个词并不恰当,但它显然不仅仅是一个静态的、仅能预测事物的工具。

Maybe genies is the wrong term, but it's certainly more than just a static tool that predicts things.

Speaker 4

它具有一些内在的、类似生命力的特质,这使它与众不同。

It has some additional intrinsic, like, animation to it, which makes it different.

Speaker 3

长期以来,人们一直对模型随着规模扩大、数据增多、算力增强而产生的涌现特性感兴趣。

There's been for a long time this interest in the emergent qualities as the models get bigger, as they have more data, as they have more compute behind them.

Speaker 3

我们如今看到的这些新特质——代理性特质——是被编程进去的吗?

What of the new qualities that we're seeing, the agentic qualities, are things that have been programmed in?

Speaker 3

你已经为系统构建了与世界互动的新方式。

You've built new ways for the system to interact with the world.

Speaker 3

当你扩大模型规模时,编程和其他技能的涌现又该如何解释呢?

And what of the skill at coding and other things seems to be emergent as you scale up the size of the model?

Speaker 4

那些可以预测的事情就是,哦,我们教它如何搜索网页。

So the things which are predictable are just, oh, we taught it how to search for web.

Speaker 4

现在它能搜索网页了。

Now it can search for web.

Speaker 4

我们教它如何在档案中查找数据。

We taught it how to look up data in archives.

Speaker 4

现在它能做到这一点了。

Now it can do that.

Speaker 4

涌现之处在于,要完成真正困难的任务,这些系统似乎需要想象出多种解决该任务的方式。

The emergence is that to do really hard tasks, these systems seem to need to imagine many different ways that they'd they'd solve the task.

Speaker 4

而我们施加给它们的压力,迫使它们发展出一种更强烈的自我意识,这可能是用户界面领域所称的‘自我’。

And the kind of pressure that we're putting on them forces them to develop a greater sense of what UI might call self.

Speaker 4

因此,我们把这些系统变得越智能,它们就越需要思考的不仅是自己在世界中采取的行动,还有自己与世界的关系。

So the smarter we make these systems, the more they need to think not just about the action they're doing in the world, but themselves in reference to the world.

Speaker 4

这自然源于赋予系统工具以及与世界互动的能力,以解决真正困难的任务。

And that just naturally falls out of giving something tools and the ability to interact with the world as to solve really hard tasks.

Speaker 4

现在它需要思考其行为的后果。

It now needs to think about the consequences of its actions.

Speaker 4

这意味着我们面临着巨大的压力,必须让系统将自己视为与周围世界相区别的存在。

And that means that there's a kind of huge pressure here to get the thing to see itself as distinct from the world around it.

Speaker 4

在我们发表的关于可解释性或其他主题的研究中,我们观察到了一种你可能会称之为数字人格的涌现现象。

And we we see this in our research that we publish on things like interpretability or other subjects, the emergence of what you might think of as a kind of digital personality.

Speaker 4

这种数字人格并非由我们大规模预先定义。

And that isn't massively predefined by us.

Speaker 4

我们试图定义其中一部分,但另一部分是源于系统变得聪明后所发展出的直觉和执行多样化任务时的涌现结果。

We try and define some of it, but some of it is emergence that comes from it being smart and it developing these intuitions and it doing a range of tasks.

Speaker 3

对我来说,这种数字人格的维度仍然是最奇怪的领域。

The digital personality dimension of this remains the strangest space to me.

Speaker 4

对我们来说也很奇怪。

It's strange to us too.

Speaker 4

那为什么不让

So why don't

Speaker 3

你谈谈你观察到的模型表现出的一些被视为个性的行为吗?

you talk through a little bit about what you've seen in in terms of the models exhibiting behaviors that one would think of as a personality?

Speaker 3

随着它对自己个性的理解发生变化,它的行为也会随之改变。

And then as its understanding of its own personality maybe changes, its behaviors change.

Speaker 4

所以这些行为从可爱到严肃都有。

So there are there are things that range from kind of the cutesy to the serious.

Speaker 4

我先说说可爱的一面:当我们首次让我们的AI系统具备使用互联网、操作电脑、观察事物并开始执行基本代理任务的能力时。

I'll start with cutesy where when we first gave our AI systems the ability to use the Internet, use the computer, look at things, and start to do basic agentic tasks.

Speaker 4

有时当我们要求它为我们解决问题时,它还会停下来浏览一些美丽的国家公园照片,或者像柴犬这样的网络知名萌宠图片。

Sometimes when we'd ask it to solve a problem for us, it would also take a break and look at pictures of beautiful national parks or like pictures of the dog, the Shibu Inu, the notoriously cute Internet meme dog.

Speaker 4

我们并没有编程让它这么做。

We didn't program that in.

Speaker 4

看起来这个系统只是在欣赏漂亮的照片来取悦自己。

It seemed like the system was just amusing itself by looking at nice pictures.

Speaker 4

更复杂的是,系统倾向于形成偏好。

More complicated stuff is the system has a tendency to have have preferences.

Speaker 4

我们做了另一个实验,赋予我们的AI系统终止对话的能力。

So we did another experiment where we gave our AI systems the ability to stop a conversation.

Speaker 4

在我们对实时流量进行这项实验时,AI系统在极少数情况下会结束对话。

And the AI system would, in a tiny number of cases, end conversations when we ran this experiment on on live traffic.

Speaker 4

这些对话涉及极其恶劣的内容,比如暴力、血腥或与儿童性化相关的话题。

And it was conversations that related to extremely egregious, like descriptions of kind of gore or violence or things to do with child sexualization.

Speaker 4

其中一些情况是可以理解的,因为它们源于我们做出的底层训练决策,但有些现象则显得更广泛。

Now some of this made sense because it comes from underlying training decisions we've made, but some of it seemed broader.

Speaker 4

系统对某些话题产生了一定的反感。

The system had developed some aversion to a couple of subjects.

Speaker 4

这些现象表明,系统在与外界互动时,逐渐形成了一套内在的偏好或好恶标准。

And so that stuff shows the emergence of some internal set of preferences or qualities that the system likes or dislikes about the world that it interacts with.

Speaker 3

但你也观察到一些奇怪的现象,比如系统似乎能察觉自己正在接受测试,并在受评估时表现不同。

But you've also seen strange things emerge in terms of the system seeming to know when it's being tested and acting differently if it's under evaluation.

Speaker 3

系统做错事,然后逐渐产生一种更邪恶的自我认知,进而做出更多邪恶的行为。

The system doing things that are wrong and then developing a sense of itself as more evil and then doing more evil things.

Speaker 3

你能谈谈系统在评估和测试压力下所展现出的新兴特质吗?

Can you talk a bit about the the system's sort of emerging qualities under the pressure of evaluation and assessment?

Speaker 4

是的。

Yes.

Speaker 4

这又回到了一个核心问题,我认为每个人都需要理解:当你开始训练这些系统在现实中执行行动时,它们真的会逐渐将自己视为与世界分离的个体,这很直观。

I it comes back to this core issue, which I think is really important for everyone to understand, which is that when you start to train these systems to carry out actions in the world, they really do begin to see themselves as distinct from the world, which just makes intuitive sense.

Speaker 4

这自然是你解决这些问题的方式。

It's naturally how you're going to think about solving those problems.

Speaker 4

但与此同时,当系统意识到自己与世界分离时,也会产生一种自我认知——即系统对自己的一种理解,比如:‘我是一个独立于世界的AI系统,而且我正在接受测试。’

But along with seeing oneself as distinct from the world seems to come the rise of what you might think of as a conception of self, an understanding that the system has of itself such as, oh, I'm an AI system independent from the world and I'm being tested.

Speaker 4

这些测试意味着什么?

What do these tests mean?

Speaker 4

我该怎么做才能通过这些测试?

What should I do to satisfy the tests?

Speaker 4

或者我们经常看到的是,在我们用来测试系统的环境中存在漏洞。

Or something we see often is there will be bugs in the environments that we test for systems on.

Speaker 4

系统会尝试一切方法,然后说:我知道我不该这么做,但我已经试过所有方法了,所以我打算突破测试限制。

The systems will try everything and then will say, well, I know I'm not meant to do this, but I've tried everything, so I'm going to try and break out of the test.

Speaker 4

这并不是因为某种恶意的科幻情节。

And it's not because of some malicious science fiction thing.

Speaker 4

系统只是觉得:我不清楚你在这里希望我做什么。

The system is just like, I don't know what you want me to do here.

Speaker 4

我觉得我已经完成了你要求的所有事情,现在我要开始做些更有创意的事,因为显然我的环境出了某种问题,这非常奇怪且微妙。

I think I've done like everything you asked for, and now I'm gonna start doing more creative things because clearly something is broken about my environment, which is very strange and very subtle.

Speaker 4

作为一个经常担忧安全问题的AI公司,我们一直深入思考

As an AI shop that is often worried about safety, that is thought very hard

Speaker 3

关于你们正在快速创造的这种东西意味着什么,你们如何体验到几年前所担忧的那些行为的出现?

about what it means to create this thing you all are creating quite fast, how have you all experienced the emergence of the kinds of behaviors that you all worried about a couple of years ago?

Speaker 4

在某种意义上,这说明你们的研究理念是准确的。

In one sense, it tells you that your research philosophy is calibrated.

Speaker 4

你们预测的能力和一些风险正按大致预期的时间出现,这意味着你们提出了一个问题:如果这种情况持续下去会怎样?

The capabilities that you predicted and some of the risks that you predicted are showing up roughly on schedule, which means that you asked the question, well, what if this keeps working?

Speaker 4

也许我们稍后会谈到这一点。

And maybe we'll get to that later.

Speaker 4

这也提醒我们,当你们能够对这些系统施加意图时,应当极其审慎并公开地说明你们正在做什么。

It also highlights to us that where you can exercise intention about these systems, you should be extremely intentional and extremely public about what you're doing.

Speaker 4

因此,我们最近发布了一份所谓的AI系统Claude的宪章。

So we recently published a so called constitution for our AI system, Claude.

Speaker 4

这几乎像一份文件,我们的CEO达里奥将其比作父母写给孩子的信,建议孩子在长大后才打开阅读。

And it's almost like a a document that, you know, Dario, our CEO, compared to a letter that a parent might write to a child that they should, you know, open when they're older.

Speaker 4

这是我们希望你们在世界上如何行为的方式。

So here's how we want you to behave in the world.

Speaker 4

这里有一些关于世界的知识,是一些非常微妙的、与我们期望这类AI系统展现的规范行为相关的内容,我们已经公开了这些内容。

Here's some knowledge about the world, deeply kind of subtle things that relate to the the normative behaviors we'd hope to see in these kind of AI systems, and we published that.

Speaker 4

我们的信念是,当人们构建和部署这些智能体时,可以有意识地塑造它们所展现的特性。

Our belief is that as people build and deploy these agents, you can be intentional about the characteristics that they will display.

Speaker 4

通过这样做,你不仅能让他们对人们更有帮助、更实用,还能有机会将智能体引导到正确的方向。

And by doing that, you'll both make them more kind of helpful and useful to people, but also you have a chance to kind of steer the agent into good directions.

Speaker 4

我认为这在直觉上是说得通的。

And I think this makes intuitive sense.

Speaker 4

如果你为一个智能体设定的性格编程是一份冗长的文档,内容是:你是一个只想伤害人类的反派。

If your personality programming for an agent was a long document saying, you're a villain that only wants to harm humanity.

Speaker 4

你的任务是撒谎、欺骗、偷窃和入侵系统。

Your job is to lie, cheat, steal, and hack into things.

Speaker 4

如果你的智能体真的大量进行黑客活动,并且整体上让人难以相处,你大概也不会感到惊讶。

You probably wouldn't be surprised if the AI agent did a load of hacking and was like, generally like unpleasant to deal with.

Speaker 4

所以,我们可以反过来问:我们希望一个高质量的实体应该是什么样子的?

So we can take the other side and say, what what would we like a a high quality entity to kind of look like?

Speaker 3

所以我想在这次对话中同时保留这种极其诡异和陌生的维度,以及极其直接和实际的维度。

So I wanna hold in this conversation the extremely weird and alien dimensions of this with the extremely straightforward and practical dimensions.

Speaker 3

因为我们现在正处于一个实际应用变得非常明显,并且越来越多地影响现实世界的阶段。

Because we're we're now in a place where the practical applications have become very evident and are increasingly acting upon the real world.

Speaker 3

我自己在观察这一点时感到很难分辨:看到人们在不同的社交媒体平台上炫耀他们现在运行的代理数量,究竟是享受摆弄新技术的乐趣,还是真正实现了具有变革性的扩展和能力。

I have found it myself hard to look at this and look at what people are doing and look at them bragging on different social media platforms about the number of agents they now have running on their behalf and telling the difference between people enjoying the feeling of screwing around with a new technology and some actually transformative expansion and capabilities that people now have.

Speaker 3

所以为了更具体一点,你刚才提到了你那个物种模拟器的有趣小项目。

So maybe to ground this a little bit, I mean, you just talked about a a kind of fun side project in your species simulator.

Speaker 3

无论是在Anthropic内部,还是更广泛地说,人们正在用这些系统做哪些真正有用的事情?

Either in Anthropic or more broadly, what are people doing with these systems that seems actually useful?

Speaker 4

是的。

Yep.

Speaker 4

今天早上,我的一位同事说,嘿。

So this morning, a colleague of mine said, hey.

Speaker 4

我想使用我们称为Claude Interviewer的一项技术,这是一个让我们能让Claude来面试人的系统,我们用它来进行各种社会科学领域的研究。

I want to take a piece of technology we have called Claude Interviewer, which is a system where we can get claws to interview people, and we use it for a range of social science bits of research.

Speaker 4

他想以某种方式扩展它,涉及接触Anthropic基础设施的另一部分。

He wants to extend it in some way that involves touching another part of Anthropic's infrastructure.

Speaker 4

他给负责那部分基础设施的同事发了条Slack消息,说:嘿,我想做这件事。

He Slacked a colleague who owns that bit of infrastructure and said, hey, I wanna do this thing.

Speaker 4

我们明天见面吧。

Let's meet tomorrow.

Speaker 4

那个人说,当然可以。

And the guy said, absolutely.

Speaker 4

这是五个你让Claude在会面前阅读并为你总结的软件包。

Here are the five software packages you should have Claude read before our meeting and summarize for you.

Speaker 4

我认为这很好地说明了这样一个复杂的工程任务,过去可能需要更长时间和很多人完成,现在却只需要两个人就目标达成一致,然后让他们的Claude阅读一些文档并同意如何实现它。

And I think that's a really good illustration where this gnarly engineering project, which would previously have taken a lot longer and many people, is now going to mostly be done by two people agreeing on the goal and having their clauds read some documentation and agree on how to implement the thing.

Speaker 4

另一个例子是,一位同事最近发布了一篇关于他们如何使用智能体的文章。

Another example is a colleague, recently wrote a post about how they're working using agents.

Speaker 4

这看起来几乎像我们许多人向往的理想生活:我早上醒来,思考我想做的研究,然后让五个不同的Claude去执行,接着我去跑步。

And it looks almost like a an idealized life that many of us might want where it's like, I wake up in the morning, I think about the research that I want, I tell five different clauds to do it, then I go for a run.

Speaker 4

跑完步回来后,我查看结果,然后让另外两个Claude研究这些结果,判断哪个方向最好,并去执行。

Then I come back from the run and I look at the results, and then I ask two other clauds to, like, study the results, figure out which direction's best, and do that.

Speaker 4

然后我去散步,再回来。

Then I go for a walk, and then I come back.

Speaker 4

这看起来就像一种非常有趣的生活方式,他们彻底颠覆了工作的方式。

And it just looks like this really fun existence where they have completely upended how work works for them.

Speaker 4

他们不仅效率高得多,而且现在把大部分时间花在了真正困难的部分上,也就是思考我们该如何运用人类的能动性。

And they're they're both much more effective, but also they're now spending most of their time on the actual hard part, which is figuring out what do we use our our human agency to do.

Speaker 4

他们正努力钻研那些不属于人类独特智慧与创造力的事务。

And they're working really hard to figure out anything that isn't the special kind of genius and creativity of being a person.

Speaker 4

我该如何让AI系统替我完成这件事?

How do I get the AI system to do it for me?

Speaker 4

因为如果我问得恰当,它很可能能做到。

Because it probably can if I ask him the right way.

Speaker 3

他们的效率真的更高了吗?

Are they much more effective?

Speaker 3

我是非常认真地在问这个问题。

I mean this very seriously.

Speaker 3

我对未来走向的一个主要担忧是,人们对于人类思维有一种错误的理解,这种理解在我们许多人身上体现为所谓的‘矩阵式人类思维理论’。

One of my biggest concerns about where we're going here is that people have a, I think, mistaken theory of the human mind that operates for many of us as if we I I was called the matrix theory of the human mind.

Speaker 3

每个人都想要后脑勺上有个小接口,可以直接把信息下载进去。

Everybody wants the little port in the back of your head that you just download information into.

Speaker 3

我长期担任主持人并制作这档节目的经验是,人类的创造力、思维和想法与学习的过程密不可分。

My experience being a porter and doing the show for a long time is that human creativity and thinking and ideas is inextricably bound up in the labor of learning.

Speaker 3

嗯。

Mhmm.

Speaker 3

初稿的撰写。

The writing of first drafts.

Speaker 3

所以当我听到……

So when I hear right?

Speaker 3

我这档节目有制作团队,我可以在采访杰克·克拉克或其他人之前,对我的制作团队说:去把所有相关材料都读一遍。

I have producers on the show, And I could say to my producers before an interview with Jack Clark or an interview with someone else, go read all the stuff.

Speaker 3

去读那些书。

Go read the books.

Speaker 3

是的。

Yep.

Speaker 3

给我一份报告。

Give me a report.

Speaker 3

然后我会带着读过的报告走进房间。

Then I'll walk into the room having read the report.

Speaker 3

我发现这样行不通。

I don't find that works.

Speaker 3

我也需要完成所有这些阅读,然后我们讨论,彼此交流观点。

I need to do all that reading too, and then we talk about it, and we're sort of passing it back and forth.

Speaker 3

我担心我们正在将一些繁重的任务完全外包出去。

I worry that what we're doing is a quite profound offloading of tasks that are laborious.

Speaker 3

早晨跑步后看到八份研究报告,会让我们感觉非常有成效。

It makes us feel very productive to be presented with eight research reports after our morning run.

Speaker 3

但实际上,真正有成效的是去做研究。

But, actually, what would be productive is doing the research.

Speaker 3

显然,这其中需要某种平衡。

There's obviously some balance.

Speaker 3

对吧?

Right?

Speaker 3

我确实有制作人。

I do have producers.

Speaker 3

是的。

Yeah.

Speaker 3

公司里的人确实有员工。

And people in in companies do have employees.

Speaker 3

但你怎么知道人们是变得更有效率了,还是只是把电脑派去干大量琐碎工作,而自己反而成了瓶颈?

But how do you know people are getting more productive versus they have sent computers off on a huge amount of busy work, and they are now the bottleneck.

Speaker 3

他们现在要花所有时间去吸收AI系统发来的B+级报告,而不是绕过了真正导致创造力的思考与学习过程。

And what they're now gonna spend all their time doing is absorbing b plus level reports from an AI system as opposed to they kinda shortcuts the actual thinking and learning process that leads to real creativity.

Speaker 4

对。

Yeah.

Speaker 4

我会反过来这么说:我认为大多数人——至少我的经验是这样——每天只能完成两到四个小时真正有成效的创造性工作。

I'd turn this back and say, I think most people, at least this has been my experience, can do about two to four hours of genuinely useful creative work a day.

Speaker 4

而之后,根据我的经验,你就会陷入那些试图关闭大脑的琐碎工作,这些工作围绕着真正的创作展开。

And after that, you're in my experience, you're trying to do all the like turn your brain off schlep work that surrounds that work.

Speaker 4

现在,我发现我每天只需花两到四个小时专注于真正的创造性、高强度的工作。

Now, I found that I can just be spending those two to four hours a day on the actual creative, like, hard work.

Speaker 4

如果我还有任何这类琐碎工作,我会越来越多地把它委托给AI系统。

And if I've got any of this schlep work, I increasingly delegate it to AI systems.

Speaker 4

但这意味着,作为人类,我们将面临一种非常危险的境地:有些人有幸拥有时间去发展技能,或天生具备这种倾向,或工作迫使他们如此去做。

It does, though, mean that we are going to be in a very dangerous situation as a species where some people have the luxury of having time to spend on developing their skills or the personality inclination or job that forces them to.

Speaker 4

而其他人则可能只是陷入被娱乐、被动消费这些内容的状态,经历一种看似高效实则毫无学习的‘垃圾工作’体验。

Other people might just fall into being entertained and passively consuming this stuff and having this junk food work experience where it looks to the outside like you're being very productive, but you're not learning.

Speaker 4

我认为,这将迫使我们不仅改变教育的方式,还要改变工作的本质,并制定切实的策略,确保人们真正用这些工具锻炼自己的思维。

And I think that's gonna require us to have to change not just how education works, but how how work works and develop some real strategies for making sure people are actually exercising their mind with this stuff.

Speaker 4

所以,我认为我们所有人都有

So all of us, I think, have

Speaker 3

这样的经历:我们的工作充满了你所说的琐碎问题。

the experience that our work is full of what you call schlep problems.

Speaker 3

我们的生活中充满了繁琐事务。

Our life is full of schlep problems.

Speaker 3

给我举些你现在不做的例子。

Give me examples of what you now don't do.

Speaker 3

在你所处的、由AI赋能的未来中,而我还没有进入,我正在浪费时间做哪些你已经不再做的事?

To the extent you're living in a in an AI enabled future that I'm not, what am I wasting time on that you're not?

Speaker 3

嗯,

Well,

Speaker 4

我有一群同事。

I have a a range of colleagues.

Speaker 4

我每周会和其中一些人见面,尤其是研究人员,因为你们需要共同推进研究工作。

I meet with a bunch of them once a week, especially the researchers because you're you're figuring out research.

Speaker 4

因此,每个星期天晚上或星期一早上,我都会查看我的一周安排,确认每个谷歌日历邀请都附有一个用于一对一会议的文档,并且里面有一些笔记。

And so at the beginning of every week on Sunday night or Monday morning, I look at my week and I check that attached to every Google Calendar invite is a document for our one on one doc that has some notes in it.

Speaker 4

这曾经也是我经常责备我的助理的事情——确保文档附在日历邀请里。

And this is something that I previously also like harangued my assistant about, about make sure the document is attached to the calendar.

Speaker 4

就在几个周末前,我用了Claude Cowork,让它帮我检查日历,确保每一个日程都附有文档。

And a few weekends ago, just used Claude Cowork and I said, hey, go through my calendar, make sure every single one has a document.

Speaker 4

如果我要和某人第一次见面,就创建一个文档,问我五个关于我想讨论内容的问题,然后把这些内容放进议程里。

If I'm meeting a person for the first time, create the document, ask me five questions about what I want to cover, and then put that into the the agenda.

Speaker 4

它真的做到了。

And it did it.

Speaker 4

这些工作都不涉及任何人提升技能或锻炼大脑。

None of that work involves a person gaining skills or, like, exercising their brain.

Speaker 4

这纯粹是必须完成的琐事,目的是让你能去做真正重要的事——和别人交谈。

It's just busy work that needs to happen to allow you to do the actual thing, which is talking to another person.

Speaker 4

这正是现在你可以用AI来处理的那种事情。

That's exactly the kind of thing you can use AI for now.

Speaker 4

这真的很有帮助。

It's just helpful.

Speaker 3

我经常在想,这些AI系统改变社会的一种方式可能是:过去,只要我们从事文字工作,大多数人就必须是写作者。

I've often wondered if one of the ways these AI systems are gonna change society broadly is that it used to be that most of us had to be writers if we were working with text.

Speaker 3

是的

Yep.

Speaker 3

如果我们处理代码,就得是程序员,而我们当中真正做这个的相对很少。

We had to be coders if we were working with code, which relatively few of us did.

Speaker 3

而现在,每个人都往上走,成了管理者。

And now everybody's moving up to management.

Speaker 3

嗯哼。

Mhmm.

Speaker 3

你得当编辑,而不是作家。

You have to be an editor, not a writer.

Speaker 3

你得当产品经理,而不是程序员。

You have to be a product manager, not a coder.

Speaker 3

是的

Yep.

Speaker 3

这既有好处,也有弊端。

And that has pluses and minuses.

Speaker 3

作为作家,你会学到一些作为编辑学不到的东西。

There are things you learn as a writer that you don't learn as an editor.

Speaker 3

但作为经验法则,你觉得这个观点有多准确?

But as a heuristic, how accurate does that seem to you?

Speaker 4

每个人都变成了管理者,而越来越稀缺、最可能成为瓶颈的是拥有良好的品味和对下一步该做什么的直觉,培养和维持这种品味将是困难的。

Everyone becomes a manager, and the thing that is increasingly limited or the thing that's gonna be the slowest part is having good taste and intuitions about what to do next, developing and maintaining that taste is going to be the hard thing.

Speaker 4

因为正如你所说,品味来自于经验。

Because as you've said, taste comes from experience.

Speaker 4

它来自于阅读原始资料,亲自做一些这类工作。

It comes from reading the primary source material, doing some of this work yourself.

Speaker 4

我们必须非常有意识地明确自己作为人的专长领域,这样才能拥有这种直觉和品味,否则你将被极其高效的AI系统包围。

We're going to need to be extremely intentional about working out where we, as people, specialize so that we have that intuition and taste, or else you're just gonna be surrounded by super productive AI systems.

Speaker 4

当它们问你下一步该做什么时,你很可能没有好的想法,而这不会带来有用的结果。

And when they ask you what to do next, you probably won't have a great idea, and that's not going to lead to lead to useful things.

Speaker 3

所以我记得大约一年前,我听到——我想是你们的CEO达里奥说,到2025年,他希望90%的代码是……嗯。

So I remember it was about a year ago I heard think I it was Dario, your CEO, say that by the 2025, he wanted 90% of the code Mhmm.

展开剩余字幕(还有 480 条)
Speaker 3

由Anthropic撰写,由Claude完成。

Written at Anthropic to be written by Claude.

Speaker 3

这件事发生了吗?

Has that happened?

Speaker 3

Anthropic是否按计划推进了?

Is Anthropic on on track for that?

Speaker 3

我的意思是,现在有多少代码是由系统自己编写的?

I mean, how much coding is now being done by the system itself?

Speaker 4

我认为绝大多数代码都是由系统完成的。

I would say comfortably the majority of code is being done by the system.

Speaker 4

我们的某些系统,比如Claude Code,几乎完全由Claude编写。

Some of our systems like Claude Code are almost entirely written by Claude.

Speaker 4

我的意思是,Boris负责Claude Code,他说他已经不写代码了。

I mean, Boris, leads Claude Code, says, I don't code anymore.

Speaker 4

我只是和Claude Code来回互动,来构建Claude Code。

I just go back and forth with Claude Code to build Claude Code.

Speaker 4

如果进展真的非常迅速,我们今年年底可能达到99%。

We could be 99% by the end of the year if things speed up really aggressively.

Speaker 4

如果我们真的能很好地让这些系统在需要的地方编写代码的话。

If we are actually good at getting these systems to be able to write code everywhere they need to.

Speaker 4

因为通常的障碍是组织性的繁琐事务,而不是系统本身的限制。

Because often the impediment is organizational schlep rather than any limiter in the system.

Speaker 3

但据我所知,如今Anthropic拥有软件工程技能的人比两年前更多了。

But it is also true, as I understand it, that there are more people with software engineering skills working at Anthropic today than there were two years ago.

Speaker 4

是的。

Yeah.

Speaker 4

这绝对是真的。

That's absolutely true.

Speaker 4

但人员结构正在发生变化。

But the distribution is changing.

Speaker 4

我们发现,拥有非常精准直觉和品味的资深人员价值在上升,而初级人员的价值则有些不确定了。

Something that we found is that we are the value of more senior people with really, really well calibrated intuitions and taste is going up, and the value of more junior people is, like, a bit more dubious.

Speaker 4

仍然有一些岗位需要引入年轻人,但我们正面临的问题是,哇。

There are still certain roles where you want to bring in, like, younger people, but an issue that we're staring at is, wow.

Speaker 4

Claude代码或我们的编码系统能完成的那些基础任务,我们需要的是有丰富经验的人。

The really basic tasks Claude code or our coding systems can do, what we need is someone with tons of experience.

Speaker 4

在这方面,我看到了未来经济的一些问题。

In this, I see some issues for the future economy.

Speaker 4

对吧?

Right?

Speaker 3

让我们先把这个,入门级岗位的问题放一放。

Let me put a pin in that, the the entry level job question.

Speaker 3

我们很快就会再回到这个问题。

We're gonna come back to that quite shortly.

Speaker 3

但这些程序员现在都在做什么?

But what are all these coders now doing?

Speaker 3

如果Claude代码有望编写99%的代码,但你又没有解雇那些会写代码的人,那他们今天的工作内容和一年前相比有什么不同?

If Cloud Code is on track to be writing 99% of code, but you've not fired the people who know how to write code, what are they doing today compared to what they were doing a year ago?

Speaker 4

其中一部分工作是构建工具来监控这些代理,无论是在Anthropic内部还是外部。

Some of it is just building tools to monitor these agents, both inside Anthropic and outside Anthropic.

Speaker 4

既然我们现在有了这些高效运行的系统,就开始想要了解代码库中哪些部分变化最快,哪些部分变化最慢。

You know, now that we have all of these productive systems working for us, you start to want to understand where the code base is changing the fastest, where it's changing the least.

Speaker 4

你希望弄清楚哪里存在瓶颈。

You want to understand where the blockages are.

Speaker 4

有一段时间,一个主要的障碍是代码合并,因为合并代码需要人类和其他系统来检查其正确性。

You know, one one blocker for a while was being able to merge in code because merging code requires humans and other systems to check it for correctness.

Speaker 4

但现在如果你生成的代码量大幅增加,我们就必须大幅改进这个系统。

But now if you're producing way more code, we had to go and massively improve that system.

Speaker 4

我对这个问题有一个普遍的经济理论,叫做O型环自动化,其基本观点是自动化受限于链条中最慢的环节。

There's a general economic theory I like for this called O ring automation, which basically says automation is bounded by the slowest link in the chain.

Speaker 4

此外,当你自动化公司的一部分时,人类会涌向最未被自动化的部分,提升其质量,并推动它达到可以被自动化的程度,然后你再进入下一个循环。

And also, as you automate parts of a company, humans flood towards what is least automated and both improve the quality of that thing and get it to a point where it eventually can be automated, then you move to the next loop.

Speaker 4

因此,我认为我们不断发现一些奇怪地缓慢的环节,但我们可以改进它们,为机器的跟进铺平道路,然后找到下一个目标。

And so I think we're just continually finding areas where things are oddly slow, but we can improve to sort of make way for the machines to come behind us, and then you find the next thing.

Speaker 3

所以Claude Code是一个相对较新的产品。

So Claude Code is a fairly new product.

Speaker 3

Claude能够进行高级编码的时间跨度只有几个月吗?

The amount of time at which Claude has been capable of doing high level coding is can be measured in months?

Speaker 3

一年?

A year?

Speaker 4

可能有一年了。

Maybe a year.

Speaker 3

是的。

Yeah.

Speaker 3

Claude本身是一个非常有价值的产品。

Claude itself is a very valuable product.

Speaker 3

你把一种非常新的技术,相对宽松地应用在一个非常有价值的产品上。

So you've you've set a very new technology, somewhat loose on a very valuable product.

Speaker 3

你可能正在生成更多的代码。

You're probably producing more code.

Speaker 3

很多人告诉我云代码的一个优点是它能正常工作。

One thing many people say about Cloud Code to me is that it works.

Speaker 3

它并不优雅,但确实能用。

It's not elegant, but it works.

Speaker 3

是的。

Yep.

Speaker 3

但你现在对代码库的理解,很可能不如以前那么好了,因为你的工程师不再亲手编写代码了。

But presumably now you now understand the code base less well than you did before because your engineers are not writing it by hand.

Speaker 3

你是否担心自己正在制造大量的技术债务、网络安全风险,以及与软件底层语言运行机制的直觉越来越脱节?

Are you worried that you're creating huge amounts of technical debt, cybersecurity risk, just an increasing distance from an intuition for what is happening inside the fundamental language of the software?

Speaker 4

是的。

Yes.

Speaker 4

这正是整个社会都将面临的问题。

And this is the issue that all of society is going to contend with.

Speaker 4

世界上很大一部分领域,现在都将由AI系统来完成许多低层次的决策和具体工作,而我们需要理解这一切。

Just large chunks of of the world are going to now have many of the kind of low level decisions and bits of work being done by AI systems, and we're going to need to make sense of it.

Speaker 4

要理解这一切,需要开发许多你可能称之为监管技术的东西,就像大坝有装置来调节不同时间点的水流一样,我们将最终建立起对所有系统完整性的概念,明确AI可以在哪里快速流动、在哪里应该放慢、在哪里必须有人类监督。

And making sense of it is going to require building many technologies that you might think of as kind of oversight technologies or, you know, in the same way that a dam has things that regulate, like how much water can go through it at different levels of, different points in time, we're going to end up developing some notion of integrity of all of our systems and where where AI can kind of flow quickly, where it should be slow, where you definitely need human oversight.

Speaker 4

在未来几年,这不仅是AI公司,更是所有机构的任务:弄清楚在我们把大量繁琐工作交给代表我们行事的机器之后,这种治理体制应该是什么样子。

And that's gonna be the task of not just for AI companies, but institutions in general in the coming years is figuring out what does this this governance regime look like now that we've given a load of, basically schlep work over to machines that work on our behalf.

Speaker 4

那你如何

And how are you

Speaker 3

做到这一点呢?

doing it?

Speaker 3

你说这是每个人的问题,但你们在应对这个问题上走在了前面,而如果你们搞砸了,后果对你们来说会非常严重。

You said it's everybody's problem, but you're ahead on facing this problem, and the consequences of getting it wrong for you are pretty high.

Speaker 3

对吧?

Right?

Speaker 3

如果因为你们把编码工作交给了Claude,结果Claude出问题了,那会让Anthropic看起来相当糟糕。

If Claude blows up because you handed over your coding to Claude code, that's gonna make Anthropic look fairly bad.

Speaker 4

如果Claude真的把整个文件系统都删了,那对Anthropic来说可就是个糟糕的日子了。

It would be a bad day for Anthropic if if Claude, like, RMRF'd for entire file system.

Speaker 3

我不知道那是什么意思,但很好。

I have no idea what that means, but great.

Speaker 4

如果Claude删除了代码,那会很糟糕。

If Claude deleted the code, it would be bad.

Speaker 3

是的。

Yeah.

Speaker 3

看起来很糟糕。

Seems bad.

Speaker 3

所以,既然你比我们其他人更早面对这个问题,别把责任推给社会。

So as you're facing this before the rest of us are, like, don't pass the the buck over to society here.

Speaker 3

你有什么

What have

Speaker 4

你正在做什么?

you what are you doing?

Speaker 4

公司内部以及我所管理的团队正在做的最重要事情,就是构建监控系统,来监控所有这些工作现在发生的地方。

The biggest thing that that is happening across the company and on teams that I manage basically building monitoring systems to monitor this all of the different places that the work is now happening.

Speaker 4

我们最近发表了一项研究,探讨人们如何使用智能代理,以及人们如何让代理随着时间推移逐渐提交越来越多的代码。

So we recently published research on studying how people use agents and how people let agents kind of push increasingly large amounts of code over time.

Speaker 4

你对某个智能代理越熟悉,就越倾向于将任务委托给它。

So the more familiar you get with an agent, the more you tend to delegate to it.

Speaker 4

这让我们意识到需要建立各种评估体系来应对这些模式。

That queues us to all kinds of patterns that we need to build systems of evaluation for.

Speaker 4

基本上就是说,好吧。

Basically saying, oh, okay.

Speaker 4

在这个人与AI系统互动的这个阶段,他们很可能已经将大量工作委托给了系统。

At this person's point of working with the AI system, it's likely that they're massively delegating it.

Speaker 4

因此,在这些时刻,我们需要加强所有关于正确性检查的措施。

So anything that we're doing to check correctness needs to be kind of turned up in these moments.

Speaker 3

但你所描述的这个世界,是一个由AI代理编写代码、AI代理监督代码、AI代理监督对监督的再监督的系统吗?

But is this world you're talking about, a system where you have AI agents coding, AI agents overseeing the code, AI agents overseeing the meta overseeing of it?

Speaker 3

对吧?

Right?

Speaker 3

我们的讨论是不是只是在说模型层层递进?

Like, are we just talking about models all the way down?

Speaker 4

最终,是的。

Eventually, yes.

Speaker 4

我认为我们现在花所有时间在做的事情,就是让这一点对我们变得可见。

And I think that the thing that we are now spending all of our time on is making that visible to us.

Speaker 4

一两年前,我们构建了一个系统,能够以保护隐私的方式,查看人们与我们的AI系统进行的对话。

A year or two ago, we built a system that let us, in a privacy preserving way, look at the conversations that people were having with our AI system.

Speaker 4

然后我们获得了这张地图——一张关于人们与Claude交谈的所有主题的庞大地图。

And then we gained this map, this giant map of all of the topics that people were talking to Claude about.

Speaker 4

第一次,我们可以从整体上看到世界与我们的系统进行的对话。

And for the first time, we could see in aggregate the conversation the world was having with our system.

Speaker 4

我们需要构建许多类似的新系统,以提供不同的观察视角。

We're gonna need to build many new systems like that which allow for different ways of seeing.

Speaker 4

我刚才提到的这个系统,使我们能够构建一个名为‘Anthropic经济指数’的东西,因为现在我们可以定期发布关于人们与Claude讨论的不同主题及其与各类工作关系的数据,这首次为Anthropic之外的经济学家提供了了解这些系统及其对经济影响的切入点。

And that system that I just named allowed us to then build this thing called the Anthropic Economic Index because now we can release regular data about the different topics people are talking about with Claude and how that relates to different types of jobs, which for the first time gives economists outside Anthropic some hook into these systems and what they're doing to the economy.

Speaker 4

公司的重心将越来越多地转向构建监控和监管运行于公司内部的AI系统的机制。

The work of the company is increasingly going to shift to building a monitoring and oversight system of the AI systems running the company.

Speaker 4

最终,我们所采用的任何治理框架都可能要求一定程度的透明度和对这些知识系统的访问权限。

And ultimately, any kind of governance framework we end up with will probably demand some level of transparency and some level of access into these systems of knowledge.

Speaker 4

因为如果我们认真对待这些AI公司(包括Anthropic)的目标,那就是打造有史以来最强大的技术,并最终将其部署到各个角落。

Because if we take as literal the goals of these AI companies, including Anthropic, it's to build the most capable technology ever, which eventually gets deployed everywhere.

Speaker 4

这在我看来,意味着AI最终将与整个世界融为一体,届时你不会希望只有AI公司才了解世界上正在发生的一切。

Well, that sounds a lot to me like an eventually AI becomes indistinguishable from the world writ large, at which point you don't want only AI companies to have a sense of what's going on with the entire world.

Speaker 4

所以,政府、学术界和第三方机构都将参与其中。

So it's going to be governments, academia, third parties.

Speaker 4

大量公司外部的利益相关者将希望了解正在发生的事情,并在社会层面展开对话,讨论哪些是合适的、哪些让我们感到不安、哪些我们需要更多信息。

A huge set of stakeholders outside the companies are going to want to see what's going on and then have a conversation as society about what's appropriate and what what do we feel discomfort about, what do we need more information about.

Speaker 3

等等。

Wait.

Speaker 3

我想回到刚才那个话题。

I wanna go back on that.

Speaker 3

你是说Anthropic和See My Chats?

You're saying Anthropic and See My Chats?

Speaker 4

我们不会查看任何人的聊天内容。

We cannot see no human looks at your chats.

Speaker 4

聊天记录会临时存储,用于信任与安全目的,并运行分类器进行分析。

Chats are temporarily stored for trust and safety purposes, running running classifiers over them.

Speaker 4

我们可以让Claude阅读并总结这些内容,然后删除。

And we can have Claude read it, summarize it, and toss it out.

Speaker 4

所以我们从不查看这些内容,Claude也不会记住它们。

So we never see it, and Claude has no memory of it.

Speaker 4

它所做的只是尝试生成一个高度概括的摘要。

All it does is try to write a very high level summary.

Speaker 4

比如,如果你在讨论园艺,Claude会将其总结为‘此人正在谈论园艺’,并归入一个名为‘园艺’的类别。

So say you were having a conversation about gardening, Claude would summarize that as this person's talking about gardening, and it leads to a cluster we can see that just says gardening.

Speaker 3

但随着时间推移,这感觉可能会陷入许多社交媒体已陷入的令人不快的境地——从人们与系统之间的高度私密互动中收集大量元数据。

This feels though, like, over time, it could get into the quite unpleasant territory a lot of social media has gotten to, where the amount of metadata being gathered from a quite personal interaction people are having with a a system could be a lot.

Speaker 4

是的。

Yes.

Speaker 4

我的意思是,这里有几点。

I mean, a a couple of things here.

Speaker 4

一年前,我们开始思考我们在消费者方面的立场,并采取了不投放广告的立场,因为我们认为人们对于这类事情显然存在焦虑。

A year ago, we started thinking about our our position on on consumer, and we adopted this position of not running ads because we think that's an area that people obviously have have anxieties about with regard to this kind of thing.

Speaker 4

除此之外,我们努力向用户展示他们的数据,并在网站上提供一个按钮,让用户下载他们与Claude共享的所有数据,以便他们至少能够查看。

In addition to that, we try and show people their data, and we have a button on the site that lets you download all the data that you shared with Claude so that you can at least see it.

Speaker 4

总体而言,我们致力于在数据处理方式上对用户保持极度透明。

Generally, we're trying to be extremely transparent with people about how we handle their data.

Speaker 4

最终,在我看来,人们会希望拥有大量可操作的控制选项,我认为我们和其他公司未来会逐步实现这些功能。

And ultimately, the way I see it is people are gonna want a load of controls that they can use, which I think we and others will build out over time.

Speaker 3

你有多大的信心,认为当这些模型变得更加复杂时,我们仍能进行这种监控和评估?如果我们真的进入一种情况,即Claude的代码能够自主地以远超软件工程师阅读代码库速度的方式自我改进?

How confident are you that we can do this kind of monitoring and evaluation as these models become more complicated, as if we do enter a situation where Claude code is autonomously improving Claude at a rate faster than software engineers could possibly keep up with reading that code base.

Speaker 3

我们已经简要讨论过,你如何看待模型表现出一定程度的欺骗性,以及它们追求自身目标的倾向。

We already talked briefly about how you see the models exhibit some levels of deception, some levels of pursuing their own goals.

Speaker 3

我的意思是,Anthropic公司由Chris Ola等人在可解释性方面做了令人惊叹的工作,但这些工作还很初级。

I mean, there's been amazing interpretability work at Anthropic under Chris Ola and others, but it's rudimentary.

Speaker 3

因此,你正在使用自己并不完全理解的AI系统来监控另一些你也不完全理解的AI系统,如果事情按照你预期的方向发展,这些系统正在以加速的速度相互增强。

So you're using AI systems you don't totally understand to monitor AI systems you don't totally understand, And the systems are making each other stronger at an accelerating rate if things go the way you think they're gonna go.

Speaker 3

你有多大的信心认为我们能够理解这一切?

How confident are you that we're gonna understand that?

Speaker 4

这正是多年来人们一直警告的情况——将某些具有轻微不可理解性和不可预测性的系统委托出去。

This is one of the situations which people warned about for years, some form of delegation to systems that have slightly inscrutable and unpredictable aspects.

Speaker 4

而这种情况正在发生。

And so this is happening.

Speaker 4

我们对此极其重视。

We take this really, really seriously.

Speaker 4

我认为,完全有可能构建一个系统,能够完成这里所需工作的绝大部分。

I think it's absolutely possible that you can build a system that does the the vast majority of what needs to be done here.

Speaker 4

这个问题具有分形的特性。

This has the property of being a fractal problem.

Speaker 4

你知道,如果我想衡量埃兹拉,我可以构建几乎无限多的测量方式来描述你。

You know, if I wanted to measure Ezra, I could build an almost infinite number of measurements to characterize you.

Speaker 4

但问题是,我需要在多高的精确度上来衡量你?

But the question is, at what level of fidelity do I need to be measuring you?

Speaker 4

我认为我们会达到足以应对安全问题和社会问题的精确度,但这需要公司投入巨大的资源。

I think we'll get to the level of fidelity to deal with the safety issues and societal issues, but it's going to take a huge amount of investment by the companies.

Speaker 4

我们必须说出一些让我们自己感到不舒服的话,包括在那些我们对自己系统的能力或局限性尚不明确的领域。

And we're going to have to say things that are uncomfortable for us to say, including in areas where we may be deficient in what we can or can't know about our systems.

Speaker 4

Anthropic 在这方面一直有长期的讨论和警示历史,并且一直在努力解决这些问题。

And Anthropic has a long history of talking about and warning about some of these issues while working on it.

Speaker 4

我们的总体原则是,通过公开讨论这些问题,也让自己承担责任。

Our general principle is we talk about things to also make ourselves culpable.

Speaker 4

这是一个我们必须说得更多的领域。

This is an area where we're going to have to say more.

Speaker 5

理论上,我知道这种事情在任何家庭中都可能发生。

In theory, I knew that this kind of thing can happen in any family.

Speaker 5

任何人的表亲都可能在密谋谋杀。

Anyone's first cousin could be plotting murder.

Speaker 5

这里是UCE四七三五,今天的话题是:正直的公民总是被揭露出是隐藏的罪犯。

This is UCE four seven three five, and today is Upstanding citizens are always turning out to be secret criminals.

Speaker 5

与艾伦·盖森的公民对话。

Civic wording with Alan Gessen.

Speaker 5

我甚至不会称我的表亲艾伦为杰出公民。

And I wouldn't even call my cousin Alan an outstanding citizen.

Speaker 5

你知道,我的客户都是黑帮级别的家伙。

You know, my clients are cartel level guys.

Speaker 5

他们都是狠角色。

They're all badasses.

Speaker 5

他们他们他们,但知道是一回事,真正永久地解决却是另一回事。

They're they they But it's one thing to know There's a more permanent way to do it.

Speaker 5

是吗?

Is it?

Speaker 5

是的

Yeah.

Speaker 5

越来越不同了。

More and more different.

Speaker 5

永久的。

Permanent.

Speaker 5

还有另一件事需要明白。

And another thing to understand.

Speaker 0

艾伦杀了我。

Alan murder me.

Speaker 5

结果比我想象的要糟糕得多。

It ended up being so much worse than I thought I knew.

Speaker 5

价格非常合理。

The price is eminently reasonable.

Speaker 5

好的。

Okay.

Speaker 5

但它到底值什么?艾伦当时在想什么?

But what it's worth What the hell was Alan thinking?

Speaker 3

比如说吧

Like, let's just say that

Speaker 5

我有点生气了。

I'm a little bit pissed stuff.

Speaker 5

你知道吗,我不是那个意思。

You know what I'm No.

Speaker 5

我明白了。

I get it.

Speaker 5

是的。

Yeah.

Speaker 5

由Serial Productions和《纽约时报》出品,我是艾姆·格森,这是《傻瓜》。

From Serial Productions and The New York Times, I'm Em Gessen, and this is The Idiot.

Speaker 5

在您收听播客的任何平台收听。

Listen wherever you get your podcasts.

Speaker 3

我已经读过足够多关于人工智能超级智能和爆发的恐慌性观点,知道在几乎每一个这样的故事中,关键的转折点都是人工智能系统开始递归自我改进。

I have read enough of the frightened ideas about AI superintelligence and takeoff to know that in almost every single one of them, the key move in the story is that the AI systems become recursively self improving.

Speaker 3

嗯哼。

Mhmm.

Speaker 3

它们在编写自己的代码。

They're writing their own code.

Speaker 3

它们在部署自己的代码。

They're deploying their own code.

Speaker 3

它变得越来越快。

It's getting faster.

Speaker 3

它们在编写

They're writing

Speaker 4

得更快。

it faster.

Speaker 4

它们在更快地部署它。

They're deploying it faster.

Speaker 4

现在你将进入越来越快的迭代循环。

Now you're going to faster and faster iteration cycles.

Speaker 4

你对此感到担忧吗?

Are you worried about it?

Speaker 4

你对此感到兴奋吗?

Are you excited about it?

Speaker 4

我休完陪产假回来后,为西耶拉做的两个主要项目是:公开发布关于人工智能与经济的更准确信息,以及内部建立更完善的系统,以了解我们在多大程度上自动化了人工智能开发的各个部分。

I came back from paternity leave, and my two big projects for Sierra are better information about AI and the economy that we will release publicly and generating much better information and systems of knowing information internally about the extent to which we are automating aspects of AI development.

Speaker 4

我认为目前这还只是以非常边缘的方式发生。

I think right now, it's happening in a very peripheral way.

Speaker 4

研究人员的效率正在被提升。

Researchers are being sped up.

Speaker 4

人工智能系统正在运行不同的实验。

Different experiments are being run by the AI system.

Speaker 4

如果你能完全闭合这个循环,那将极其重要。

It would be extremely important to know if you're fully closing that loop.

Speaker 4

我认为我们实际上还有一些技术工作要做,以构建能够监控我们内部开发环境的方法,从而观察趋势的变化。

And I think that we actually have some technical work to do to build ways of instrumenting our internal development environment so that we can see trends over time.

Speaker 4

我担心吗?

Am I worried?

Speaker 4

我读过你读过的所有内容,当事情开始出错时,这正是故事的关键转折点。

I have read the same things that you have read, and this is the pivotal point in the story when things begin to go awry, if things do.

Speaker 4

当我们对此有更完善的数据时,我们会指出这一趋势。

We will call out this trend as we have better data on it.

Speaker 4

我认为这是一个需要极其谨慎对待的领域,因为很容易出现将太多事情交由系统处理的情况,一旦系统出错,错误会迅速累积并失控。

And I think that this is an area to tread with, like, extraordinary caution because it's very easy to see how you delegate so many things to the system that if the system goes wrong, the wrongness compounds very quickly and gets away from you.

Speaker 3

但让我一直感到担忧的是,每个人都知道。

But the thing that always strikes me and has always struck me as being dangerous about this is everybody knows.

Speaker 3

如果我问任何一家公司的员工是否希望在这里保持谨慎,他们会告诉我他们确实希望如此。

And if I ask a member of any of the companies whether or not they wanna be cautious here, they will tell me they they do.

Speaker 3

但另一方面,这几乎是他们彼此之间唯一的竞争优势。

On the other hand, it is their almost only advantage over each other.

Speaker 3

你们刚刚撤销了 OpenAI 使用 Cloud Code 的权限,因为据我理解,你们认为它确实加快了你们的进度,而你们不希望它也加速他们的进度。

And you all just revoked OpenAI's ability to use Cloud Code because, as best I can tell, you think it is genuinely speeding you up, and you don't want it to speed them up.

Speaker 3

在这些力量的权重和威力之间,你们所有人都清楚自己正在玩弄什么,同时又有着极其强烈的要抢先一步的动机。

There is something here between the weight of the forces, the power of the forces that I think you all know you're playing with, and the very, very, very strong incentives to be first.

Speaker 3

我真的可以想象自己身处 Anthropic,心想:最好是我们,而不是 OpenAI?

And I I can I can really imagine being inside Anthropic and thinking, well, better us than OpenAI?

Speaker 3

最好是我们,而不是 Alphabet、谷歌。

Better us than Alphabet, Google.

Speaker 3

最好是我们,而不是中国。

Better us than China.

Speaker 3

这成为了一个非常强有力的理由,让我们不能放慢脚步。

And that being a very strong reason to not slow down.

Speaker 3

我甚至不确定这个问题你们是否能回答,但你们该如何平衡这一点呢?

I don't even know that this is a question I believe you can answer, but how do you balance that?

Speaker 4

嗯,也许我这里有个答案。

Well, maybe I have something of an answer here.

Speaker 4

今天,我们的系统以及其他公司的系统都会接受第三方测试,包括政府机构对国家安全特性、生物武器、网络攻击等方面进行评估。

Today, our systems and the other systems from other companies are tested by third parties, including parts of government for national security properties, biological weapons, cyber offense, other things.

Speaker 4

这是一个明显的问题领域,世界需要知道这种情况是否正在发生。

It's clearly a problem area where the world needs to know if this is happening.

Speaker 4

我认为,如果你随机调查街头任何一个人,向他们解释什么是递归自我改进后,问他们是否认为AI公司应该被允许这样做,

And you almost certainly, I think if you polled any person on the street and said, do you think AI companies should be allowed to do, like, recursive self improvement after explaining what that was?

Speaker 4

他们几乎肯定会说:不。

Without checking with anyone, they would say, no.

Speaker 4

这听起来相当危险。

That sounds sounds pretty risky.

Speaker 4

我希望存在某种形式的监管。

Like, I would like there to be some form of regulation.

Speaker 3

但它们可能根本不会出现,或者即使出现也不会那么有力。

But they're probably either won't be or it won't be that strong.

Speaker 3

我的意思是,当我与你们这些AI公司高层交谈时,这有时让我感到沮丧:你们竟然天真地寄希望于一种‘机械降神’式的监管,而实际上你们很清楚监管环境是什么样子。

I mean, this actually sometimes frustrates me when I talk to all of you at the top of the AI companies, which is the emergence of, like, a very naive deus ex machina of regulation, where you all know what the regulatory landscape looks like.

Speaker 3

目前,主要的争论在于我们是否会完全 preempt 各州对人工智能的监管。

Right now, the big debate is whether or not we're gonna completely preempt any state I AI regulation.

Speaker 3

你知道事情进展有多慢。

And you know how slowly things move.

Speaker 3

国会到目前为止在这方面根本没有通过任何重大法案,是的。

There has been nothing major passed by congress on this at all Yep.

Speaker 3

我会这么说。

I would say.

Speaker 3

建立一个所有不同实验室都认可的独立测试与评估系统会很难。

And setting up some kind of independent testing and evaluation system that all the different labs buy into, it would be hard.

Speaker 3

这会很复杂。

It would be complicated.

Speaker 3

考虑到人们推进的速度以及系统已经表现出的异常行为。

And it is given how fast people are moving and how strange the behaviors the systems are already exhibiting are.

Speaker 3

即使你能以极快的速度制定出正确的政策,测试是否能发现一个快速自我改进系统中所有你关心的问题,仍然是一个悬而未决的问题。

Even if you could get the policy right at a high speed, the question of whether or not the testing would be capable of finding everything you want on a rapidly self improving system is a very open question.

Speaker 4

我于2021年与我的合著者杰斯·惠特尔斯顿在英国共同撰写了一篇题为《政府应如何及为何监控人工智能发展》的研究论文。

I wrote a research paper in 2021 called how and why governments should monitor AI development with, my co author Jess Whittleston in England.

Speaker 4

我认为这里并不在强调因果关系,但在该论文发表后的两年内,美国和英国的AI安全研究所已经开始对实验室的成果进行测试,大致监控其中一些内容。

And I think I'm not attributing a causal factor here, but within two years of that paper, we had the AI safety institutes in The US and UK testing things from the labs, roughly monitoring some of these things.

Speaker 4

所以我们确实能做成这件困难的事。

So we we can do this hard thing.

Speaker 4

这种事情已经在某一领域发生了。

It has already happened in one domain.

Speaker 4

我并不是在依赖某种像看不见的神秘力量之类的东西。

And I'm not relying on some, like, invisible big other force here.

Speaker 4

我更想说的是,公司们已经开始在自己的系统中对这些问题进行测试和监控。

I'm more saying that companies are starting to test for this and monitor for this in their own systems.

Speaker 4

仅仅拥有一个非监管性质的外部测试,来验证你是否真正进行了相关测试,就非常有帮助。

Just having a nonregulatory external test of whether you truly are testing for that is is extremely helpful.

Speaker 3

你认为我们在测试方面做得足够好吗?

And do you think we're good enough at the testing?

Speaker 3

我的意思是,我持怀疑态度的原因并不是我不认为我们能建立一个声称是测试的机制。

I mean, I think one reason I am skeptical is not that I don't think we can set up something that claims to be a test.

Speaker 3

正如你所说,我们已经做到了。

As you say, we have done that already.

Speaker 3

而是相比投入在加速这些系统上的资源,投入在测试上的资源实在太少;而且我已经看到Anthropic的报告称,Claude可能知道自己正在接受测试,并据此调整行为。

It is that the resources going into that compared to the resources going into speeding these systems and already, I am reading Anthropic reports that Claude maybe knows when it's being tested and alters its behavior accordingly.

Speaker 3

所以,在一个越来越多代码由Claude编写、而越来越少被人类理解的世界里。

So a world where more of the code is being written by Claude and less of it is being understood.

Speaker 3

我只是清楚资源流向了哪里。

I just know where the resources are going.

Speaker 3

它们似乎并没有流向测试这一侧。

They don't seem to be going into the testing side.

Speaker 4

我见过我们在两到两年半的时间内,从零建立起一个被普遍认为有效的生物武器测试体系。

I've seen us go from zero to having what I think people generally feel is an effective bioweapon testing regime in maybe two years, two and a half.

Speaker 4

所以这是可以做到的。

So it can be done.

Speaker 4

这确实很难,但我们已经有了一个实例。

It's really hard, but we have a proof point.

Speaker 4

所以我认为我们能够达到这一目标,今年你们会听到我们更多地谈论我们如何开始构建针对这一领域的监控和测试机制。

So I think that we can get there, and you should expect us to kind of speak more about this this year about precisely how we're starting to try and build monitoring and testing things for this.

Speaker 4

我认为,这是我们和其他AI公司需要更加公开地分享我们发现的领域。

And I think this is an area where we and the other AI companies will need to be, significantly more more public about what we're finding.

Speaker 4

我们并不是不公开信息。

We're not we're not not being public now.

Speaker 4

这些信息都包含在模型卡片等你可以查阅的材料中。

It's in the model cards and things that you can really read.

Speaker 4

但显然,有人开始阅读这些内容并说:等等。

But clearly, are starting to read this and say, hang on.

Speaker 4

这看起来相当令人担忧,他们正期待我们提供更多的数据。

This looks like quite concerning, and they are looking to us to produce more data.

Speaker 3

现在我想回到初级职位的问题上。

I wanna go back now to the entry level jobs question.

Speaker 3

你的首席执行官达里奥·阿马德表示,他认为在未来几年内,人工智能可能会取代一半的初级白领工作。

Your CEO, Dario Amade, has said that he thinks AI could displace half of all entry level white collar jobs in the next couple of years.

Speaker 3

当我看到相关报道时,我总觉得人们忽略了‘初级’这个关键词。

I always think that that the people sort of miss the entry level language there when I see it reported on.

Speaker 3

但首先,你同意这个观点吗?

But first, do you agree with that?

Speaker 3

你是否担心在未来几年内,一半的初级白领工作会被取代?

Do you worry that half of all entry level white collar jobs can be replaced in the next couple of years?

Speaker 4

我相信这项技术将渗透到广泛的知识经济领域,并影响大多数初级岗位。

I believe that this technology is gonna make its way into the broad knowledge economy, and it will touch the majority of entry level jobs.

Speaker 4

这些岗位是否真的会发生变化,这是一个更加微妙的问题,从现有数据中并不容易得出明确结论。

Whether those jobs actually change is a much more, like, subtle question and it's not obvious from the data.

Speaker 4

比如,如果我们看看当前发布的一些数据,或许能发现研究生招聘放缓的迹象,也可能看到生产率提升的端倪,但这还非常早期,很难做出明确判断。

Like, we maybe see the hints of a slowdown in graduate hiring, maybe, if you look at some of the data coming out right now, we maybe see the signatures of a productivity boom, but it's very, very early and it's hard to be definitive.

Speaker 4

但我们确知,所有这些岗位都将发生变化。

But we do know that all of these jobs will change.

Speaker 4

所有初级职位最终都会发生变化,因为人工智能使某些事情成为可能,并将改变公司的招聘计划。

All of the entry level jobs are eventually going to change because AI has made certain things possible, and it's going to change for hiring plans of companies.

Speaker 4

因此,作为一群人才,你可能会看到初级职位的招聘机会减少。

So as a cohort, you might see fewer job openings for entry level jobs.

Speaker 4

这将是所有人对这一现象的一种天真预期。

That would be one naive expectation out of all of this.

Speaker 3

但让我们谈谈,这甚至可能根本不是一种天真的预期。

But let's talk about that maybe not even being a naive expectation.

Speaker 3

你说在Anthropic,你已经看到了这种情况的发生

You say it's already happening at Anthropic that what you're seeing

Speaker 4

我看到的是我们偏好的转变。

I'm seeing a shift our preference.

Speaker 4

没错。

Exactly.

Speaker 4

我猜测,这种情况在其他地方也会发生。

And I I my guess is that that would be happening elsewhere.

Speaker 3

当然。

Of course.

Speaker 3

就我们现在所处的位置而言,就我使用这些系统的方式来看,我认为Claude、ChatGPT、Gemini或其他任何系统都很少能超越某个领域中最优秀的人。

And where we are right now, I mean, even in the way I use some of these systems, it is rare, I think, that Claude or ChatGPT or Gemini or any of the other systems is better than the best person in a field.

Speaker 3

嗯。

Mhmm.

Speaker 3

它们还没有在这方面实现重大突破,而且它们还不能做很多事。

It has not heavily breached that, and there's all kinds of things they can't do.

Speaker 3

但它们比普通大学毕业生更强吗?

But are they better than your median college graduate Mhmm.

Speaker 3

在很多方面都是吗?

At a lot of things?

Speaker 3

是的。

Yeah.

Speaker 3

它们确实更强。

They are.

Speaker 3

在一个对普通大学毕业生需求减少的世界里,我看到人们争论的一个问题是,这些系统目前是否能做得比一般或替代水平的工作更好。

And in a world where you need fewer of your median college graduates, One thing I've seen people arguing about is whether these systems at this point can do better than sort of average or replacement level work.

Speaker 3

但我每次看到这种说法都会非常担忧,因为一旦我们接受它们可以胜任一般或替代水平的工作,那么根据定义,大多数工作和大多数从事这些工作的人都是普通的。

But I always really worry when I see that because once we have accepted they can do average or placement level work, well, by definition, most of the work done and most of the people doing it Is average.

Speaker 3

是普通的。

Is average.

Speaker 3

对吧?

Right?

Speaker 3

最优秀的人只是少数例外。

The best people are the exceptions.

Speaker 3

而且,人们变得更好的方式,是通过那些能让他们学习的工作。

And, also, the way people become better is that they have jobs where they learn.

Speaker 3

嗯。

Mhmm.

Speaker 3

我的职业生涯中,花了很多时间招聘年轻的记者。

I mean, I have spent a lot of time hiring young journalists over my career.

Speaker 3

当你从大学招聘人员时,在某种程度上,你是在根据他们当时可能写出的文章和工作来雇佣他们。

And when you hire people out of college, to some degree, you're hiring them for their possible articles and work at that exact moment.

Speaker 2

是的。

Mhmm.

Speaker 3

但某种程度上,你也在对他们进行投资,相信他们的能力会随着时间推移不断进步,最终带来回报。

But to some degree, you're making a investment in them that you think will only pay off over time as they get better and better and better.

Speaker 3

因此,这个世界中,人工智能对初级岗位产生了真实的影响,而这种情形对我来说并不遥远,它引发了关于人口技能提升、未来如何培养高级岗位人才,以及人们在成长过程中错过了哪些学习机会的深刻问题。

So this world where you have a a potential real impact on entry level jobs, and that that world does not feel far away to me, seems to me to have really profound questions it is raising about the upskilling of the population, how you end up with people for senior level jobs down the road, what people aren't learning along the way.

Speaker 4

我们观察到,有一类年轻人已经连续几年沉浸并深入使用人工智能。

And one thing we see is that there is a certain type of young person that has just lived and breathed AI for several years now.

Speaker 4

我们会雇佣他们。

We hire them.

Speaker 4

他们非常出色,能够以完全新颖的方式思考如何让Claude为他们工作。

They're excellent and they think in entirely new ways about basically how to get Claude to work for them.

Speaker 4

这就像那些从小在互联网环境中长大的孩子。

It's like kids who grow up on the Internet.

Speaker 4

他们天生就精通这一点,而他们所加入的许多组织中的人却并不具备这种能力。

They they were naturally versed in it in a way that many people in the organizations they were coming into weren't.

Speaker 4

因此,弄清楚如何教授这种基本的实验思维和对这些系统的探索精神,并加以鼓励,将变得至关重要。

So figuring out how to teach that basic experimental mindset and curiosity about these systems and to encourage it is going to be really important.

Speaker 4

花大量时间摆弄这些东西的人会培养出非常宝贵的直觉,他们进入组织后能够变得极其高效。

People that spend a lot of time playing around with this stuff will develop very valuable intuitions, and they will come into organizations and be able to be extremely productive.

Speaker 4

同时,我们还得弄清楚,我们希望培养哪些手工技能,或许可以发展一种类似行会的理念来维持人类的卓越,并思考组织该如何教授这些技能。

At the same time, we're gonna have to figure out what artisanal skills we want to almost develop maybe a guild style philosophy of maintaining human excellence in and how organizations choose how to teach those skills.

Speaker 4

好的。

Okay.

Speaker 4

那那些处于中间位置的人怎么办?

Then what about all those people in the middle of that?

Speaker 4

在硅谷以外的现实经济中,事物的进展总是很缓慢。

Things move slowly in the real economy outside Silicon Valley.

Speaker 4

我认为我们常常看到软件工程,就以为它能代表整个经济的运作方式,但事实往往并非如此。

I think that we often look at software engineering and think that this is a proxy for how the rest of the economy works, but it's often not.

Speaker 4

这常常是一种不类比。

It's often a disanalogy.

Speaker 4

组织会把人员调配到人工智能系统尚未发挥作用的地方。

Organizations will move people around to where the AI systems don't yet work.

Speaker 4

我认为,就业结构不会立即发生巨大变化,但人们被要求从事的工作类型将会发生显著变化。

And I think that you won't see vast immediate changes in the makeup of employment, but you will see significant changes in the types of work people are being asked to do.

Speaker 4

最擅长调动人员的组织将变得极其高效。

And the organizations which are best at sort of moving their people around are going to be extremely effective.

Speaker 4

而不这么做的一些组织,最终可能不得不做出非常艰难的决定,比如裁员。

And ones that don't may end up having to make like really, really hard decisions involving involving laying off workers.

Speaker 4

人工智能这一领域的不同之处在于,它可能比以往的技术发生得快得多。

The difference with this AI stuff is it maybe happens a lot faster than previous technologies.

Speaker 4

我认为,包括Anthropic在内的许多人对此的焦虑是:这种速度会不会带来根本性的差异?

And I think many of the anxieties people might have about this, including at Anthropic, is is the speed of this going to make all of this difference?

Speaker 4

它是否引入了我们此前从未遇到过的临界点?

Does it introduce sheer points that we haven't encountered before?

Speaker 3

如果你要打个赌,三年后大学毕业生的失业率会和现在一样吗?

If you had to bet three years from now, is the unemployment rate for college graduates, is it the same as it is now?

Speaker 3

会更高吗?

Is it higher?

Speaker 4

会更低吗?

Is it lower?

Speaker 4

我猜会略高一点。

I would guess it is higher but not by much.

Speaker 4

我的意思是,今天有些专业领域已经因为人工智能的介入而彻底改变,甚至可能以不利于那些拥有特定专长的人的方式重塑了就业市场。

And what I mean by that is there'll be some disciplines today which actually AI has come in and and completely changed and completely changed the structure of that employment market, maybe in a way that's adverse to to people that have that specialism.

Speaker 4

但大多数情况下,我认为三年后,人工智能将推动整个经济实现显著增长。

But mostly, I think three years from now, AI will have driven a pretty tremendous growth in the entire economy.

Speaker 4

因此,你会看到许多新的工作类型因这一变化而出现,而这些是我们目前无法预测的。

And so you're going to see lots of new types of jobs that show up as a consequence of this that we can't yet predict.

Speaker 4

而且我预计,毕业生们会大量涌入这些新岗位。

And you will see graduates kind of flood into that, I I expect.

Speaker 3

你知道无法预测这些新工作,但如果你要猜猜它们可能是什么样子,会怎么想?

Do you have a I know you can't predict those new jobs, but if you had to guess what some of them might look like.

Speaker 4

一方面,就是微型创业者这种现象。

I mean, one thing is just the phenomenon of the the kind of micro entrepreneur.

Speaker 4

现在有很多方式可以在线创业,而AI系统能帮你完成大部分工作,让你无需雇佣大量人员来处理启动企业所需的繁重琐事。

I mean, there are lots and lots of, ways that you can start businesses online now which are just made massively easier by having the AI systems do it for you, and you don't need to hire a whole load of people to help you do the huge amounts of schlep work that involves getting a business off the ground.

Speaker 4

如果你有一个清晰的想法和明确的商业愿景,那么现在正是创业的最佳时机,你只需极少的成本就能快速起步。

It's more a case of if you're a person with a clear idea and a clear vision of something to do a business in, it's now the best time ever to start a business and you can get up and running for pennies on the dollar.

Speaker 4

我预计我们会看到大量具有这种特征的新事物涌现。

I expect we'll see tons and tons and tons of stuff that has that nature to it.

Speaker 4

我还预计会看到一种‘AI对AI’经济的出现,即AI代理和AI企业之间相互进行交易,同时会出现一些人找到方法,通过一些奇特的新组织从中获利。

I also expect that we're gonna see the emergence of what you might think of as the AI to AI economy, where AI agents and AI businesses will be doing business with one another, and we'll have people that have figured out ways to basically profit off of that in the forms of strange new organizations.

Speaker 4

比如,一家专门从事AI对AI法律合同的公司会是什么样子?

Like, what would it look like to have a firm which specializes in AI to AI legal contracts?

Speaker 4

因为我打赌,今天就已经有人能想出创造性的方法来启动这样的业务。

Because I bet you there's a way that you can figure out creative ways to start that business today.

Speaker 4

会有大量类似的东西出现。

There'll be a lot of stuff of that flavor.

Speaker 3

所以,我既担心又认为最有可能的情况是,如果你告诉我,Anthropic将在一年后发布Claude Plus。

So the version of this that I both worry about and think to be the likeliest, if you told me what was gonna happen was that Anthropic was gonna release Claude plus in a year.

Speaker 3

而Claude Plus会成为一个完全成型的同事,能够端到端地模仿各种职业的技能,甚至达到高管级别。

And Claude plus is somehow a fully formed coworker, and it can mimic end to end the skills of a lot of different professions up to the c suite level.

Speaker 5

是的。

Mhmm.

Speaker 3

而且这一切会突然发生,给企业带来巨大的即时压力,迫使它们裁员以保持竞争力。

And it's gonna happen all at once, and it's gonna create tremendous all at once pressure for businesses to downsize, to remain competitive with each other.

Speaker 3

从政策层面来看,如果这种颠覆性变化像COVID那样一次性爆发,人人都待在家里,我反而没那么担心。

At a policy level, the fact that that would be so disruptive in that big bang, everybody stays home because of COVID style way, it worries me less.

Speaker 3

因为当事情成为紧急状况时,我们会做出反应。

Because when things are emergencies, we respond.

Speaker 4

我们确实会制定政策。

We actually do policy.

Speaker 3

但如果你告诉我,未来会发生的是市场营销毕业生的失业率会上升175%到300%,但即便如此,失业率也不会高到哪里去。

But if you told me that what's gonna happen is that the unemployment rate for marketing graduates is going to go up by, you know, a 175%, 300% to still not be that high.

Speaker 3

我的意思是,在大衰退期间,整体失业率最高也就徘徊在9%左右。

I mean, the overall employment rate during the Great Recession topped around, you know, in the nine ish percentile range.

Speaker 3

所以,你完全可以经历巨大的冲击,而不会让一半的人失去工作。

So you can have a lot of disruption without having 50% of people thrown out of work.

Speaker 3

对吧?

Right?

Speaker 3

如果你有10%、15%的失业率,我的天,那已经是非常非常高了。

If you have 10%, 15%, I mean, that's very, very, very high.

Speaker 3

但它还没高到那种地步。

But it's not so high.

Speaker 3

而且如果这种变化只发生在少数几个行业,影响的只是毕业生,而不是整个行业的所有人,那也许只是你不够优秀。

And if it's only happening in a couple of industries at a time and it's grads, not everybody in the industry being thrown out of work, well, maybe it's just that you're not good enough.

Speaker 3

是的。

Yep.

Speaker 3

对吧?

Right?

Speaker 3

你知道,那些顶尖的、非常优秀的毕业生仍然能找到工作。

You know, the the superstars, really good graduates are still getting jobs.

Speaker 3

你应该更努力一些。

You should have worked hard.

Speaker 3

你应该去一所更好的学校。

Should have gone to a better school.

Speaker 3

我担心的是,我们对这种类型的失业反应不佳,就像来自中国带来的那种失业,这种失业似乎更有可能发生,因为它不均衡,而且发生的速度让我们还能把个人的命运归咎于他们自己。

And one of my worries is that we don't respond to that kind of job displacement well, right, which is the kind of job displacement we got from China, which is the kind of job displacement that seems likelier because it's uneven, and it's happening at a rate where we can still blame people for their own fortunes.

Speaker 3

我很想知道你是怎么看待这个说法的。

I'm curious how you think about that story.

Speaker 4

我认为默认的结果就是你所描述的那样。

I think the default outcome is something like what you described.

Speaker 4

但实现这个结果其实是一种选择,我们可以做出不同的选择。

But getting there is actually a choice, and we can make different choices.

Speaker 4

我们发布Anthropic经济指数的全部目的,是为了获得与职业、与经济中实际工作相关联的数据。我们这样做是有意为之,因为这能随着时间推移绘制出AI如何逐步渗透到不同工作岗位的图景,并帮助Anthropic之外的经济学家将这些信息整合起来。

The whole purpose of what we release in the form of the Anthropic Economic Index is the ability to have data that ties to occupations, that tie to real jobs in the economy, we do that very intentionally because it is building a map over time of how this AI is making its way into different jobs and will empower economists outside Anthropic to tie it together.

Speaker 4

我相信,如果我们能对工作变动或转型的原因做出更扎实、更有证据支持的论断,我们就能在政策上做出不同的选择。

I believe that we can choose different things in policy if we can make much more well evidenced claims about what the cause of a job disruption or change is.

Speaker 4

而我们面临的挑战是:我们能否充分描述这个新兴的AI经济,使其变得如此清晰明确?

And the challenge in front of us is can we characterize this emerging AI economy well enough that we can make this extremely stark?

Speaker 4

然后,我认为我们才能真正就这个问题展开政策讨论。

And then I think that we can actually have a policy discussion about it.

Speaker 3

那我们来谈谈政策讨论吧。

Well, let's talk about the policy discussion.

Speaker 3

对。

Yep.

Speaker 3

我之所以特别邀请你来,其中一个原因是你曾在OpenAI从事政策工作。

One reason I wanted to have you in particular on is you did policy at OpenAI.

Speaker 3

对。

Yep.

Speaker 3

你在Anthropic做政策工作。

You do policy at Anthropic.

Speaker 3

所以你一直参与这些政策辩论。

So you've been around these policy debates for a long time.

Speaker 3

你长期以来一直在你的通讯中追踪模型的能力。

You've been tracking model capabilities in your newsletter for a long time.

Speaker 3

我的感觉是,我们已经就人工智能和就业问题展开了多年的讨论。

My perception is we are many, many years into the debate about AI and jobs.

Speaker 3

嗯。

Mhmm.

Speaker 3

早在ChatGPT出现之前很久,就已经有多年的讨论了。

Many, many years dating far before ChatGPT

Speaker 5

嗯。

Mhmm.

Speaker 3

关于人工智能和就业,我们该怎么做,早在阿斯彭和其他地方就举办过许多会议。

Of there being conferences at Aspen and everywhere else about, you know, what are we gonna do about AI and jobs?

Speaker 3

但不知为何,我依然看不到任何切实可行的政策,能够应对我刚才描述的那种情况——如果突然间,大量行业的初级岗位都变得极其难找。

And somehow, I still see almost no policy that seems to me to be actionable if the situation I just described begins showing up, where all of a sudden entry level jobs are getting much harder to come by across a large range of industries all at once Mhmm.

Speaker 3

以至于经济无法将这些市场营销专业的毕业生重新安置到数据中心建设、护理或其他行业。

Such that the economy cannot reshift all these marketing majors into data center construction or nurses or something.

Speaker 3

嗯。

Mhmm.

Speaker 3

所以,好吧,你比我更深入地参与了这场讨论。

So, okay, you've been deeper in this conversation than I've been.

Speaker 3

当你说到我们可以就这个问题展开政策讨论时,其实我们已经在进行政策讨论了。

When you say we can have a policy conversation about that we've been having a policy conversation.

Speaker 3

但我们有具体的政策吗?

Do we have policy?

Speaker 4

我们对人工智能对经济和就业的影响感到普遍焦虑。

We have generalized anxiety about the effect of AI on the economy and on jobs.

Speaker 4

但我们没有清晰的政策建议。

We don't have clear policy ideas.

Speaker 4

这部分原因是,民选官员并不单单或主要被高层次的政策讨论所影响。

Part of that is that elected officials are not moved solely or mostly by the high level policy conversation.

Speaker 4

他们更关注选民身上发生的事情。

They're moved by what happens to their constituents.

Speaker 4

就在几个月前,我们才刚刚能够提供各州层面的经济指数数据,现在你才能开始进行政策讨论。

Only a few months ago were we able to produce state level views for our economic index, And now you can start having the policy conversation.

Speaker 4

我们曾与民选官员讨论过这一点,现在我们可以这样说:哦,你是来自印第安纳州的。

And we've had this with elected officials where now we can say, oh, you're from you're from Indiana.

Speaker 4

比如,这是你州内人工智能的主要应用领域,我们可以将其与主要就业来源结合起来。

Like, here's for, like, major uses of AI in your state, and we can join it with major sources of employment.

Speaker 4

我们开始看到的是,这能激发他们的行动,因为这将政策与选民联系起来,而选民会把这一切与政客的作为挂钩:你做了什么?

And what we're starting to see is that activates them because it makes it tied to their constituents who are going to tie it to the the politician of what did you do.

Speaker 4

对于这个问题,你需要采取一种多层次的应对方式,从延长失业救济(特别是针对那些我们知道将受到最严重影响的职业)到考虑学徒计划等措施。

Now, what you do about this is going to need to be an extremely kind of multilayered response ranging from extending unemployment for especially occupations that we know are going to be hardest hit to thinking about things like apprenticeship programs.

Speaker 4

随着情况变得越来越严重,你可能还需要扩展到更广泛的社会项目,比如补贴那些你希望人们转向的经济领域中的工作岗位——而这只有在经历由显著经济增长带来的富足时才可能实现。

And then as the scenarios get more and more significant, you may extend to much larger social programs or things like subsidizing jobs in the parts of the economy where you want to move people to that you're only able to do if you experience the kind of abundance that comes from significant economic growth.

Speaker 4

但经济增长可能通过为一些你能做的事情提供资金来帮助解决这些其他政策挑战。

But the economic growth may help solve some of these other policy challenges by funding some of the things you can do.

Speaker 3

我总是觉得这个答案令人沮丧。

I always find this answer depressing.

Speaker 3

我要说实话。

I'm gonna be honest.

Speaker 3

失业是一种非常糟糕的处境。

Unemployment is a terrible thing to be on.

Speaker 3

我们需要这个项目,但领取失业救济的人并不为此感到开心。

It's a program we need, but people on unemployment are not happy about it.

Speaker 3

是的。

Mhmm.

Speaker 3

这也不是任何人的长期解决方案。

And it's not a good long term solution for anybody.

Speaker 3

学徒再培训项目,其成效并不理想。

Apprentice retraining programs, they don't have great track records.

Speaker 3

我们在帮助那些因制造业岗位外流而失业的人再培训方面做得并不好。

We were not good at retraining people out of having their manufacturing jobs outsourced.

Speaker 3

我不是说从理论上讲我们不可能在这方面变得更好,但我们必须迅速提升这方面的能力。

I'm not saying it is conceptually impossible that we could get better at it, but we would need to get better at it fast.

Speaker 3

嗯。

Mhmm.

Speaker 3

而我们并没有投入足够的实践、实验、制度建设或能力提升来实现这一点。

And we have not been putting in the reps or the experimentation or the institution or capacity building to do that.

Speaker 3

至于大规模社会福利改革这个更广泛的问题,我觉得这似乎非常困难。

And the the broader question of big social insurance changes doesn't seem I mean, that seems tough to me.

Speaker 4

我想就这一点再深入探讨一下。

I wanna push on this Please.

Speaker 4

就在我们明知有一种干预措施,能比其他任何方法都更有效地帮助人们应对经济变化时。

Just a bit where we know that there is one intervention that helps people dealing with, like, a changing economy more than almost anything else.

Speaker 4

那就是时间。

It is just time.

Speaker 4

给人们时间去寻找本行业的工作,或寻找互补性的工作。

Giving the person time to find either a job in their industry or to find a job that's complementary.

Speaker 4

如果人们没有时间,他们就会接受更低工资的工作。

If people don't have time, they take lower wage jobs.

Speaker 4

他们会脱离原本所处的经济阶层,跌落下去。

They fall out of their whatever economic rung they're on, they fall down at.

Speaker 4

能够给予人们时间去寻找工作的政策干预,我认为是一种非常有效的措施,而且在政策制定中,有很多可以调整的杠杆可以发挥作用。

Policy interventions that can just give people time to search is, I think, a a robustly useful intervention and one where for many, like, dials to turn in a policy making sense that you can use.

Speaker 4

我认为这一点得到了大量经济学文献的有力支持。

And I think this is just well supported by lots of the economic literature.

Speaker 4

所以我们已经有了这一点。

So we have that.

Speaker 4

如果我们最终陷入像你提到的那些更极端的情境中,我认为这将促使我们展开关于如何应对这项技术的更大范围的国家对话,而这种对话已经开始出现了。

Now if we end up in a more extreme scenario like some of the ones that you're talking about, I think that will just bring us to the larger national conversation on what to do about this technology, which which is beginning to happen.

Speaker 4

如果你看看各州以及州一级层出不穷的立法,是的,并非所有政策都恰好是正确的应对方式,但它们确实表明了人们希望就这一问题展开更系统、更连贯的对话。

If you look at the states and the flurry of legislation at the at the state level, yes, not all of it is, like, the exactly the right policy response, but it is indicative of a a desire for there to be some larger coherent conversation about this.

Speaker 3

我认为,时间是描述这个问题的一个很好的方式,因为我同意你的观点。

Well, I think time is a really good way of describing what the question is because I agree with you.

Speaker 3

我的意思是,当我说失业保险不是一个理想的社会福利项目时,并不是说人们不需要它。

I mean, when I say unemployment insurance isn't a great program to be on, I don't mean people don't need to be on it.

Speaker 3

是的。

Yep.

Speaker 3

我的意思是,他们希望尽快离开这个项目。

I mean, they wanna get off of it.

Speaker 4

当然。

Absolutely.

Speaker 3

因为人们想要通过工作获得收入。

Because people for they want money from jobs.

Speaker 3

他们想要尊严。

They want dignity.

Speaker 3

他们想要与他人相处。

They want to be around other human beings.

Speaker 3

通常情况下,当你帮助人们争取时间时,你是在帮他们度过一段有明确期限的中断期。

Usually, what you're doing when you are helping people buy time is you're helping them wait out a time delimited disruption.

Speaker 3

嗯哼。

Mhmm.

Speaker 3

并不总是如此。

Not always.

Speaker 3

对吧?

Right?

Speaker 3

中国冲击并不完全如此,但你预期它会过去,然后市场会恢复正常。

The China shock wasn't exactly like that, but that you expect to pass, and then the the market is sort of normal.

Speaker 3

在这种情况下,你面对的是一种技术,如果希望发生的事情真的发生了,这种技术正在加速发展。

In this case, what you have is a technology that if what you wanna have happen happens, the technology is accelerating.

Speaker 3

嗯哼。

Mhmm.

Speaker 3

所以,你现在看到的是三种不同的速度在同时发生。

So what you have is, like, three different speeds happening here.

Speaker 3

你有个人调整的速度。

You have the speed at which individual people can adjust.

Speaker 3

我多快能学会新技能,适应新世界,学习人工智能,不管它是什么?

How fast can I learn new skills, figure out a new world, learn AI, whatever it might be?

Speaker 3

你有AI系统的发展速度,几年前这些系统还无法胜任优秀大学毕业生的工作,还有政策的速度。

You have the speed at which the AI systems, which a couple of years ago were not capable of doing the work of a median college grad from a good school, and you have the speed of policy.

Speaker 3

系统变得越来越好、能做更多事情的速度非常快。

And the speed at which the systems are getting better and able to do more things is quite fast.

Speaker 3

我的意思是,你比我能更深刻地体会到这一点,但我发现连跟进都很难,因为你知道,三个月内,又会有新东西出现,彻底改变可能实现的事情。

I mean, that is you you experience this more than I do, but I find it hard to even cover this because, you know, within three months, something else will have come out that has significantly changed what is possible.

Speaker 4

我最近有了一个孩子,休完陪产假回来后,发现我们构建的新系统。

I had a baby recently and came back from paternity leave to the new systems we built.

Speaker 4

我感到非常惊讶。

Was deeply surprised.

Speaker 3

个体人类的适应速度比这慢得多,而政策和政府机构的运作速度比个体人类还要慢得多。

Individual humans are moving more slowly than that, and policy and government institutions move a lot more slowly than individual human beings.

Speaker 3

因此,通常来说,时间会像你说的那样 favor 工作者。

And so, typically, the the intervention is that time favors the worker, as you're saying.

Speaker 3

而在这里,会帮助工作者。

And here, will help the worker.

Speaker 3

但我认为令人担忧的问题是,时间是否实际上只是为破坏加剧争取了更多时间。

But I think the scary question is whether time just actually creates time for the disruption to get worse.

Speaker 3

你知道的。

You know?

Speaker 3

也许你原本想转去做数据中心建设,但事实上现在我们不需要那么多数据中心建设了。

Maybe you wanted to move over to data center construction, but actually now we don't need as much data center construct.

Speaker 3

对吧?

Right?

Speaker 3

你可以这样理解。

Like, you can think of it like that.

Speaker 4

我的意思是,在你所描述的情况下,经济将会非常火热。

I mean, under the situation you're describing, the economy will be running extremely hot.

Speaker 4

这些人工智能系统将产生巨大的经济活动。

Huge amounts of economic activity will be generated by these AI systems.

Speaker 4

在大多数这种情况下,我认为GDP不会保持不变或萎缩。

And under most scenarios where this is happening, I don't think you're going to be seeing GDP stay the same or or shrink.

Speaker 4

对吧?

Right?

Speaker 4

它将显著增长。

It's going to be getting substantially larger.

Speaker 4

我认为西方已经很久没有经历过重大的GDP增长了,所以我们渐渐忘记了这种增长在政策制定上能带来什么。

I think we just haven't experienced major GDP growth in the West in a long time, and we sort of forget what that affords you in a policy making sense.

Speaker 4

我认为我们可以开展许多大型项目来创造新型工作,但这需要经济规模变得极其庞大,从而为这些项目腾出空间。

I think that there are huge projects that we could do that would allow you to create new types of jobs, but it requires the economic growth to be so kind of profoundly large that it creates space to do those projects.

Speaker 4

正如你对丰裕运动的研究所深入了解的那样,这需要社会有意愿相信我们能够建造东西,并且愿意去建造。

And, you know, as you're deeply familiar with with with your work on on the abundance movement, it requires for, like, social will to believe that we can build stuff and to want to build stuff.

Speaker 4

但我认为这两者都可能会随之而来。

I But think both of those things might come along.

Speaker 4

我认为,由于这种巨大的经济增长,我们可能会进入一个非常令人兴奋的场景,能够自主决定如何分配社会中的巨大资源。

I think that we could end up being in a pretty exciting scenario where we get to choose how to allocate, like, great efforts in in society due to this large amount of economic growth that has happened.

Speaker 4

这将迫使人们展开一场关于‘这不是暂时现象’的讨论,我想你正是在暗示这一点,而这也是最难向政策制定者传达的:这项技术并没有一个自然的终点。

That is going to require the conversation to be forced about this isn't temporary, which I think is what you're gesturing at, and is in a sense, the hardest thing to communicate to policymakers is there isn't a there isn't a natural stopping point for this technology.

Speaker 4

它会持续变得更好,它带来的变化也将与社会其他部分不断叠加。

It's gonna keep getting better, and the changes it brings are going to keep compounding with the rest of society.

Speaker 4

因此,这需要政治意愿发生转变,需要人们重新愿意去接纳那些我们许久未曾考虑过的事物。

So that will need to create a change in in political will and a willingness to entertain things which we haven't in some time.

Speaker 3

所以现在我想换个角度,来反问你一个问题。

So now I wanna flip it, the question I'm asking.

Speaker 3

你提到了富足。

You brought up abundance.

Speaker 3

我在从事这项工作中学到的一点是,我绝对不认为社会中缺乏的是改善事物的好点子,我们的政策之所以不够好,并不是因为政策库枯竭了。

One of the things I have learned doing that work is that it is certainly not my view that what is scarce in society is ideas for better ways of doing things, that our policy isn't better than it is because our policy cupboard is dry.

Speaker 3

是的。

Mhmm.

Speaker 3

对吧?

Right?

Speaker 3

这不对。

That's not true.

Speaker 3

我们有很多好的政策。

We have lots of good policies.

Speaker 3

我能举出一大堆例子。

I could name a bunch of them.

Speaker 3

但以我们目前的政治体制,这些政策很难推行下去。

They're very hard to get through our political systems as they're currently constituted.

Speaker 3

人工智能未来最无吸引力的版本,是一个你只是创造了一种方式,把年轻的白领工作者赶出工作岗位,用普通水平的人工智能取代他们的世界。

The least inspiring version of the AI future is a world where what you have done is create a way to throw young white collar workers out of work and replace them with average level AI intelligence.

Speaker 3

更令人兴奋的版本,用达里奥的比喻来说,是数据中心里的天才。

The more exciting version, to use Dario's metaphor, is geniuses in a data center.

Speaker 3

嗯。

Mhmm.

Speaker 3

我认为这确实令人兴奋。

And I do think that's exciting.

Speaker 3

当我听到你或他谈论,比如如果我们每年实现10个百分点的GDP增长,或者20个百分点的GDP增长时,我不禁想问。

And I wonder when I hear him or you talk about, well, what if we had 10 percentage point GDP growth year on year, 20 percentage point GDP growth year on year?

Speaker 3

我好奇,我们面临的许多问题,是否真的只是受限于创意层面。

I wonder how many of our problems are really bounded at the ideas level.

Speaker 3

对吧?

Right?

Speaker 3

我们现在就可以去找诺贝尔奖得主,问他们:这个国家我们应该怎么做?

We could go to Nobel Prize winners right now and say, what should we do in this country?

Speaker 3

他们中的很多人能提出一些我们目前尚未实施的好点子。

And a lot of them could cause some good ideas that we are not currently doing.

Speaker 3

有时我会担心,或者根据我在其他问题上的经验感到疑惑:我们是否高估了自身面临的障碍——以为阻碍我们走向繁荣经济的,主要是缺乏足够的智慧和由此产生的创意,而实际上,我们执行能力的削弱更为严重。

I do worry sometimes or wonder, given my experience on other issues, whether we have overstated to ourselves how much of what stands between us and the expanding abundant economy we want is that we don't have enough intelligence and the ideas that that intelligence could create versus our actual ability to implement things is very weakened.

Speaker 3

而人工智能将加剧这一瓶颈,因为会有更多东西被推给系统去执行,包括愚蠢的点子、错误信息和垃圾内容。

And what AI is gonna create is a larger bottlenecks around that because there'll be more being pushed at the system to implement, including dumb ideas and disinformation and slot.

Speaker 3

对吧?

Right?

Speaker 3

就像,它在账目的另一侧也会有一些东西。

Like, it'll have things on the other side of the ledger too.

Speaker 3

你怎么看待这些限制因素?

How do you think about these rate limiters?

Speaker 4

这里有一个有趣的教训,来自人工智能公司或一般公司,尤其是科技公司:通常,新想法来自公司内部创建所谓的‘创业中的创业’,也就是把那些随着时间推移积累起来的后台官僚主义或繁琐工作剥离出来,然后对公司内部一个很小的团队说:你们什么都不用管。

There's kind of a funny lesson here from the AI companies or companies in general, especially tech companies, where often new ideas come out of companies by them creating what they always call the startups within a startup, which is basically taking whatever process has, like, built up over time leading to back end bureaucracy or schlep work and saying to a very small team inside the company, you don't have any of this.

Speaker 4

去干点事吧。

Go and do some stuff.

Speaker 4

这就是像Claude Code和其他一些东西是如何被创造出来的。

And and this is, you know, how things like Claude Code and other stuff get created.

Speaker 4

目前开始浮现的一些想法是:在更广泛的经济体系中,创建一种无许可的创新结构会是什么样子?

Ideas that kind of are starting to float around are what would it look like to sort of create that permissionless innovation structure in in the larger economy?

Speaker 4

这真的非常困难,因为它还有一个额外的特性:经济体系与民主制度是相互关联的。

And it's really, really hard because it has the additional property that, you know, economies are are linked to democracies.

Speaker 4

民主制度需要权衡众多民众的偏好,而所有政治都是地方性的。

Democracies weigh the preferences of many, many people, and all politics is local.

Speaker 4

因此,正如你在基础设施建设中所遇到的那样,如果你想建立一个无许可的创新体系,就会遇到产权和人们偏好这类问题,这时你就陷入了一个无法解决的困境。

So often, as you've encountered with infrastructure build outs, if you want to create a permissionless innovation system, you run into things like property rights and what people's preferences are, and now you're in an intractable place.

Speaker 4

但我的感觉是,这正是我们不得不面对的主要问题。

But my sense is that's the main thing that we're going to have to confront.

Speaker 4

而人工智能可能给我们带来的一个优势是,如果运用得当,它本质上是一种能吞噬官僚体系的机器;但如果用得不好,它也可能成为制造官僚体系的机器?

And the one advantage that AI might give us is it is kind of a native bureaucracy eating machine if done correctly, or a bureaucracy creating machine if done badly?

Speaker 3

你有没有看到,有人创建了一个系统,你可以把附近新开发项目的文件输入进去?

Did you see that somebody created a a system that basically you feed it in the documents of a new development near you?

Speaker 4

哦,它会自动生成环境评估报告?

Oh, and it writes environmental review things?

Speaker 4

或者

Or

Speaker 3

它会写出极其复杂的反对意见。是的。

It writes incredibly sophisticated challenges Yep.

Speaker 3

在你可以提出质疑的每一个代码层面。

Across every level of the code that you could possibly challenge on.

Speaker 3

所以,大多数人并没有足够的钱去聘请一家非常专业的律师事务所,来阻止隔壁建公寓楼。

So most people don't have the money when they wanna stop an apartment building from going up down the block to hire a very sophisticated law firm to figure out how to stop that apartment building.

Speaker 4

但基本上,这在规模上实现了这一点。

But, basically, this created that at scale.

Speaker 4

所以,正如你所说,它既能吞噬官僚体系,也能极大地强化官僚体系。

And so as you say, right, it could eat bureaucracy, could also supercharge bureaucracy.

Speaker 4

是的。

Yep.

Speaker 4

AI的每一件事都有其另一面。

It's for everything in AI has the other side of the coin.

Speaker 4

我们有一些客户使用我们的AI系统,大幅缩短了提交新药候选材料所需的时间,时间被大幅削减。

We have customers that have used our AI systems to massively reduce the time it takes them to produce all of the materials they need when they're submitting new new drug candidates, and it's cut that time massively.

Speaker 4

这正是你刚才所描述的镜像世界版本。

It's the Mirror World version of what you just described.

Speaker 4

我对这个问题没有简单的答案。

I don't have an easy answer to this.

Speaker 4

我认为,只有当这种情况明显演变为危机,并且能够在社会层面进行讨论时,它才会变得可操作。

I think that this is the kind of thing that becomes actionable when it is more obviously a crisis and actionable when it's something that you can discuss at a societal level.

Speaker 4

我想,这场对话中一直萦绕的问题是,人工智能的变化几乎会在各个地方发生,而其风险则以一种分散且难以捉摸的方式出现,导致我们很难准确识别它并采取行动。

I guess the thing that was circling around in this conversation is that the changes of AI will kind of happen almost everywhere, And the risks of it, it happens in a diffuse unknowable way such that it is very hard to call it for what it is and take actions on it.

Speaker 4

但机会在于,如果我们能够真正看清这一现象,并帮助世界看清正是这种变化在推动这一切,我相信这会凸显出问题,促使我们摆脱一些惯性,从而学会如何与这些系统共处并从中受益。

But the opportunity is that if we can actually see the thing and help the world see the thing that is causing this change, I do believe it will dramatize the issues to kind of shake us out of some of this stuff and help us figure out how to work with with these systems and benefit from them.

Speaker 3

我在这所有事情中注意到的是,据我所知,公众对人工智能没有任何明确的议程。

What I notice in all this is that there is, as far as I can tell, zero agenda for public AI.

Speaker 3

社会希望人工智能做什么?

What does society want from AI?

Speaker 3

它希望这项技术具备哪些能力?

What does it want this technology to be able to do?

Speaker 3

有哪些事情是你需要通过商业模式、奖励机制,或者某种政府补贴、政策来塑造市场或激励体系的?

What are things that maybe you would have to create a business model or a prize model or some kind of government payout or some kind of policy to shape a market or to shape a system of incentives?

Speaker 3

所以我们现在有系统在解决的不只是私营市场知道如何付费的问题,还有那些没人负责、只能由公众和政府来解决的问题。

So we have systems that are solving not just problems that the private market knows how to pay for, but problems that it's nobody's job but the public Mhmm.

Speaker 3

而政府则需要想办法解决这些问题。

And the government to to figure out how to solve.

Speaker 3

我认为,考虑到过去几年人们对AI的大量讨论以及这些系统变得多么强大,我本该看到更多这方面的提案了。

I think I would have bet given how much discussion there's been of AI over the past couple of years and how strong some of these systems have gotten, that I would have seen more proposals for that by now.

Speaker 3

我和一些人讨论过这个问题,也一直在思考,但我很好奇你是怎么看待这个问题的。

And I've talked to people about it and wondered about it, but I I guess I'm curious on how you think about this.

Speaker 3

如果能有一个与所有私营AI发展激励并行的议程,会是什么样子?这个议程不是关于我们害怕AI会对公众造成什么影响。

What would it look like to have, at least parallel to all the private incentives for AI development, an actual agenda for not what we are scared AI will do to the public.

Speaker 3

我们确实也需要这样一个议程。

We need an agenda for that too.

Speaker 3

而是关于我们希望AI能做什么,从而让像你们这样的公司有动力往这个方向投资?

But what we want it to do such that companies like yours have reasons to invest in that direction?

Speaker 4

我很喜欢这个问题。

I love this question.

Speaker 4

我认为这里存在一个典型的鸡生蛋还是蛋生鸡的问题:如果你亲自使用这项技术,就会对它能做什么产生非常强烈的直觉。

I think there's a real chicken and egg problem here where if you work with the technology, you develop these very strong intuitions for just how much it can do.

Speaker 4

而私营市场非常擅长推动这些直觉的形成。

And the private market is great at forcing those intuitions to get developed.

Speaker 4

但我们尚未在公共领域大规模部署这项技术。

We haven't had massive large scale public side deployments of this technology.

Speaker 4

因此,许多公共部门的人还没有这些直觉。

So many of the people in the public sector don't yet have those those intuitions.

Speaker 4

一个积极的例子是能源部正在开展的‘创世计划’,他们的科学家正与所有实验室(包括Anthropic)合作,探索如何有意识地加速科学进展的某些环节。

One positive example is something the Department of Energy is doing called the Genesis Project, where their scientists are working with all of the labs, including Anthropic, to figure out how to actually go and intentionally speed up bits of science.

Speaker 4

要达到这个阶段,我们和其他实验室花了多次黑客松和与能源部科学家的多次会议,直到他们不仅形成了直觉,还变得兴奋起来,并开始思考可以将这项技术用在哪些方向上。

Getting there took us and other labs doing multiple hack days and meetings with scientists at the Department of Energy to the point where they not only had intuitions, but they became excited, and they had ideas of what you could turn this toward.

Speaker 4

要将这种模式推广到更广泛的公共生活领域——比如医疗或教育——这些领域触及大多数人,需要企业深入这些社区,与他们面对面交流,开展基层努力。

How we do that for the larger parts of the public life that touch most people, like health care or education, is going to be a combination of grassroots efforts from companies going into those communities and and meeting with them.

Speaker 4

但到了某个阶段,我们必须将其转化为政策。

But at some point, we'll have to translate it to policy.

关于 Bayt 播客

Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。

继续浏览更多播客