The Ezra Klein Show - 人工智能代理将以多快的速度席卷经济? 封面

人工智能代理将以多快的速度席卷经济?

How Quickly Will A.I. Agents Rip Through the Economy?

本集简介

人工智能代理时代已至,它们是否已改变你的生活?Claude Code等智能体的发布标志着AI历史的新转折点。我们正告别聊天机器人时代,迈入智能代理纪元——AI不仅能独立完成各类任务,更能与其他AI协作交流。 目前尚不清楚这些模型是否真能显著提升用户效率。但技术持续进步,鲜见停滞迹象。这一新时代对我们的经济、劳动力市场和下一代意味着什么? 克拉克是Anthropic公司联合创始人,该公司开发了Claude和Claude Code。多年来,他的通讯《Import AI》一直是我追踪各模型能力的重要读物。本次对话中,我请他分享了对当前技术变革的见解:技术如何演变,是否正引发工作与思维方式的实质性改变,以及政策应如何应对潜在的就业冲击。 提及内容: 杰克·克拉克《Import AI》 帕特·格雷迪与索尼娅·黄《2026:这就是AGI》 杰斯·惠特尔斯通与杰克·克拉克《政府为何及如何监管AI发展》 罗斯·杜塔特《有趣时代》访谈"Anthropic首席科学家谈AI:'我们不知道模型是否具有意识'" 推荐书目: 厄休拉·勒古恩《地海巫师》 埃里克·霍弗《狂热分子》 qntm《反模因部》 意见与嘉宾推荐请邮件至ezrakleinshow@nytimes.com 节目文稿(午间更新)及更多《Ezra Klein秀》内容请访问nytimes.com/ezra-klein-podcast,关注Ezra推特@ezraklein。所有嘉宾推荐书单详见https://www.nytimes.com/article/ezra-klein-show-book-recs 本期节目由Rollin Hu制作,事实核查Michelle Harris、Mary Marge Locker与Kate Sinclair共同完成。高级音频工程师Jeff Geld,混音支持Isaac Jones和Aman Sahota。执行制作人Claire Gordon。制作团队还包括Marie Cascione、Annie Galvin、Kristin Lin、Emma Kehlbeck、Jack McCordick、Marina King和Jan Kobal。原创音乐Pat McCusker。听众策略Kristina Samulewski与Shannon Busta。纽约时报观点音频总监Annie-Rose Strasser。 立即订阅:nytimes.com/podcasts 或通过Apple Podcasts与Spotify。也可通过您喜爱的播客应用订阅https://www.nytimes.com/activate-access/audio?source=podcatcher。下载纽约时报APP获取更多播客与有声文章:nytimes.com/app 由Simplecast(AdsWizz旗下公司)提供技术支持。个人信息收集及广告使用政策详见pcm.adswizz.com

双语字幕

仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。

Speaker 0

我正在开启跨平台对战。

I'm opening up cross play.

Speaker 0

我一直和丹对战,他是《纽约时报》的同事。

I've been playing against Dan, my colleague at the New York Times.

Speaker 1

卡特下了另一手棋。

Kat's played another move.

Speaker 1

呃。

Ugh.

Speaker 1

她用‘stoop’得了36分。

She played stoop for 36 points.

Speaker 0

我手里有个Z,值10分。

I've got a z, which is 10 points.

Speaker 1

我猜‘Tenga’不是一个单词。

I'm guessing Tenga is not a word.

Speaker 1

我们来看看。

Let's see.

Speaker 1

Tenga 是一个单词。

Tenga is a word.

Speaker 1

哦。

Oh.

Speaker 0

丹完成了他的最后一轮。

Dan played his last turn.

Speaker 0

我们来看看谁赢了。

Let's see who won.

Speaker 0

比分非常接近,但我赢了。

It's so close, but I did win.

Speaker 2

《纽约时报》游戏订阅用户可全面访问 Crossplay,这是我们首款双人文字游戏。

New York Times game subscribers get full access to Crossplay, our first two player word game.

Speaker 2

立即订阅,享受我们所有游戏的特别优惠。

Subscribe now for a special offer on all of our games.

Speaker 1

过去几年报道人工智能时,我们通常谈论的是未来。

The thing about covering AI over the past few years is that we're typically talking about the future.

Speaker 1

每一个新模型,尽管令人印象深刻,都像是为即将问世的模型提供了一个概念验证。

Every new model, impressive as it was, seemed like proof of concept for the models that would be coming soon.

Speaker 1

那些能够可靠地独立完成实际工作的模型。

The models that could actually do useful work on their own reliably.

Speaker 1

那些真正会让工作过时,或让新事物成为可能的模型?

The models that would actually make jobs obsolete or new things possible?

Speaker 1

这些模型对劳动力市场、我们的孩子、我们的政治以及我们的世界意味着什么?

What would those models mean for labor markets, for our kids, for our politics, for our world?

Speaker 1

我认为,我们总是谈论未来的那个时期,现在结束了。

I think that period in which we're always talking about the future, I think it's over now.

Speaker 1

我们一直等待的那些模型——听起来像科幻小说、能独立编程且比大多数程序员更快更好的模型,能够编写自己的代码来提升自身的模型——现在已经出现了。

Those models we were waiting for, the sci fi sounding models that could program on their own and do so faster and better than most coders, the models that could begin writing their own code to improve themselves, those models are here now.

Speaker 1

它们已经出现在Anthropic的Cloud Code中。

They're here in Cloud Code from Anthropic.

Speaker 1

它们也已经出现在OpenAI的Codex中。

They're here in Codex from OpenAI.

Speaker 1

它们正在冲击股市。

They are shaking the stock market.

Speaker 1

标普500软件行业指数已下跌20%,蒸发了数十亿美元的价值。

The S and P 500 software industry index has fallen by 20%, wiping billions of dollars in value out.

Speaker 1

优秀的工程师,我认识多年的、对AI炒作相当怀疑的人。

Excellent engineers, people I've known for years, people who are quite skeptical of AI hype.

Speaker 1

他们现在给我发邮件,说他们看不到自己的工作在未来一两年内如何还能存在。

They're emailing me now to say they don't see how their job will possibly exist in a year or two.

Speaker 1

我们正处在一个AI发展的新阶段。

We are at a new stage of AI development.

Speaker 1

不仅仅是发展。

Not just development.

Speaker 1

我们正处在一个AI产品的新阶段。

We are at a new stage of AI products.

Speaker 1

我觉得红杉资本这家风投公司对这一点的表述其实很有帮助。

I thought the way Sequoia, the venture capital firm, put it was actually pretty helpful.

Speaker 1

2023年和2024年的AI应用都是话痨。

The AI applications of 2023 and 2024 were talkers.

Speaker 1

有些是非常复杂的对话者,但它们的影响有限。

Some were very sophisticated conversationalists, but their impact was limited.

Speaker 1

2026年和2027年的AI应用将是实干家。

The AI applications of 2026 and 2027 will be doers.

Speaker 1

换句话说,人们预测已久的事情终于发生了。

Or to put it differently, something that's been predicted for a long time has now happened.

Speaker 1

我们正从聊天机器人转向智能代理,从与你对话的系统转向为你行动的系统。

We are moving from chatbots to agents, from systems that talk to you to systems that act for you.

Speaker 1

而这个代理世界,已经有点怪异了。

And this world of agents, it's already weird.

Speaker 1

它们是多个代理。

They are agents plural.

Speaker 1

它们可以协同工作。

They can work together.

Speaker 1

它们可以相互监督。

They can oversee each other.

Speaker 1

人们正在代表自己运行这些代理的群体。

People are running swarms of these agents on their behalf.

Speaker 1

无论这在现阶段是让他们更高效了还是只是更忙了,我还不太确定,但如今确实可以随时拥有一支极其迅捷、尽管说实话有些古怪的软件工程师团队,听候你的差遣。

Whether that is making them at this stage more productive or just busier, I can't quite tell, but it is now possible to have what amounts to a team of incredibly fast, although, to be honest, somewhat peculiar software engineers at your beck and call at all times.

Speaker 1

杰克·克拉克是Anthropic公司的联合创始人兼政策主管,该公司开发了Claude和Claude Code。

Jack Clark is the cofounder and head of policy at Anthropic, the company behind Claude and Claude Code.

Speaker 1

多年来,克拉克一直在每周通讯《Import AI》中追踪不同模型的能力,这份通讯一直是我关注AI发展动态的重要读物。

And for years now, Clark has been tracking the capabilities of different models in the weekly newsletter Import AI, which has been one of my key reads for following developments in AI.

Speaker 1

因此,我想了解他是如何解读当前这一时刻的,包括他眼中技术如何变化,以及政策应该如何或能够随之调整。

So I want to see how he is reading this moment, both how the technology is changing in his view and how policy needs to or can change in response.

Speaker 1

一如既往,我的邮箱是EzraKleinshow@nytimes.com。

As always, my email, EzraKleinshow@nytimes.com.

Speaker 1

杰克·克拉克,欢迎来到本节目。

Jack Clark, welcome to the show.

Speaker 3

谢谢你邀请我参加,Ezra。

Thanks for having me on, Ezra.

Speaker 1

我想很多人对AI聊天机器人都很熟悉。

So I think a lot of people are familiar with AI chatbots.

Speaker 1

是的。

Mhmm.

Speaker 1

但什么是AI代理呢?

But what is an AI agent?

Speaker 3

最好的理解方式是,它是一个能够使用工具并长期为你服务的语言模型或聊天机器人。

The best way to think of it is like a language model or a chatbot that can use tools and work for you over time.

Speaker 3

当你和聊天机器人交谈时,你就在对话中,和它来回互动。

So when you talk to a chatbot, you're there in the conversation, you're going back and forth with it.

Speaker 3

而代理则是你可以给它一些指令,然后它就去为你做事,有点像和一位同事合作。

An agent is something where you can give it some instruction and it goes away and does stuff for you, kind of like working with a with a colleague.

Speaker 3

几年前,我自学了一些基础编程,并在业余时间构建了一个物种模拟程序,里面有捕食者和猎物、道路,几乎像一个二维策略游戏。

So I've got I've got an example where a few years ago, I taught myself some basic programming, and I built a a species simulation in my spare time that had predators and prey and roads and almost like a two d strategy game.

Speaker 3

我最近在圣诞节期间让Claude Cove帮我实现这个功能。

I recently asked over Christmas Claude Cove to just implement this for me.

Speaker 3

大约十分钟内,它不仅写出了一个基础的模拟程序,还自动安装了所有需要的依赖包,以及让程序更美观、更完善的可视化工具。

And in about ten minutes, it went and wrote not only a basic simulation, but all of the different packages that it needed and all of the visualization tools that it might need to be prettier and better than the thing I'd written.

Speaker 3

返回的结果是我知道即使是有经验的程序员也可能需要几个小时甚至几天才能完成的复杂工作,而这个系统却在几分钟内就搞定了。

And what came back was something that I know would probably take a skilled programmer several hours or maybe even days because it was quite complicated, and the system just did it in a few minutes.

Speaker 3

它之所以能做到这一点,不仅在于智能地解决任务,还在于创建并运行了一系列为它服务的子系统,也就是其他代表它工作的智能体。

And it did that by not only being intelligent about how to solve the task, but also creating and running a range of subsystems that were working for it, other agents that worked on its behalf.

Speaker 1

但这意味着什么?

But what does that mean?

Speaker 1

也就是说,一个多智能体系统看起来是什么样子的?

Like, what is a a multi agent setup look like?

Speaker 3

就Claude Code而言,对我来说,就是同时打开多个标签页,运行多个不同的智能体。

In the case of Claude code, for me, it's having multiple different tabs running multiple different agents.

Speaker 3

但我见过一些同事,他们会编写一个规范文件,用于让一个Claude实例去运行其他Claude实例。

But I've seen colleagues who write what you might think of as a a specification file for a version of Claude that runs other Claude's.

Speaker 3

所以他们就像是,我有五个代理,它们由另一个代理监控和管理。

And so they're like, I've got my five agents, and they're being minded over by this other agent which is monitoring what they do.

Speaker 3

我认为这将成为常态。

I think that that's just going to become the norm.

Speaker 3

所以我一直听到,也一定程度上感受到,人们对云代码有两种截然不同的体验:

So one thing I've been hearing and somewhat

Speaker 1

我无法相信这竟然如此简单,是的。

experiencing is two very different categories of experience people have with Cloud Code, which is I cannot believe how easy this is Yep.

Speaker 1

一切都能正常运行。

And everything just works.

Speaker 1

哦,这比我想象的要难得多。

And, oh, this is a lot harder than I thought it would be.

Speaker 1

是的。

Yep.

Speaker 1

问题不断出现,而我根本不知道该如何修复。

And things keep breaking, and I don't really understand how to fix them.

Speaker 1

是什么原因导致有些人能用Cloud Code写出可用的软件,而另一些人却让它生成了大量bug、混乱不堪的东西,甚至都不知道怎么跟它沟通解决?

What accounts for being able to get Cloud Code to produce working software versus it creates buggy, often messed up things, you don't even know how to talk it out of that.

Speaker 3

我认为很大一部分原因在于,人们误以为Claude Code像一个博学的人,而不是一个只能通过互联网交流、极其字面化的工具。

I think so much of it is is making the mistake of thinking Claude Code is like a knowledgeable person versus an extremely literal person that you can only talk to over the Internet.

Speaker 3

我自己就有这样一个例子:当我第一次用Claude Code编写物种模拟程序时,我只是用一段非常糟糕的措辞简单地让它做这件事,结果它生成了一些严重有bug但居然勉强能运行的东西。

And I I had this example my myself where I when I did my first pass of writing the, species simulation with Claude Code, I just sort of asked it to do the thing in in extremely crappy language over the course of a paragraph, And it produced some horribly buggy stuff that just kind of worked.

Speaker 3

后来我换了一种方式,直接对Claude说:我要用Claude Code写一些软件。

What I then did is I then just said to Claude, hey, I'm gonna write some software of Claude code.

Speaker 3

我想让你对我进行一次访谈,了解我想构建的软件,然后把这些内容整理成一份规范文档,供Claude Code使用。

I want you to interview me about this software I want to build and turn that into a specification document that I can give Claude code.

Speaker 3

这一次,它表现得非常好,因为我把任务结构化得足够具体、足够详细,系统才能顺利处理。

And then that time, it worked really, really well because I'd structured the work to be specific enough and detailed enough that the system could work with it.

Speaker 3

所以很多时候,问题不在于你是否知道任务是什么。

So often, it's not just knowing what the task is.

Speaker 3

因为如果你和我讨论一个任务,你有直觉,会主动问我各种问题,做各种澄清。

Because you and I could talk about a task to do and you have intuition, you ask me probing questions, all of this stuff.

Speaker 3

关键是确保你已经设置好了,就像把一封信息装进瓶子里扔进去,它就会自己去完成大量工作。

It's making sure that you've set it up so it's like a message in a bottle that you can chuck into the thing and it'll go away and do a lot of work.

Speaker 3

所以,这条信息必须极其详细,真正捕捉到你想要实现的目标。

So that message better be extremely detailed and really capture what you're trying to do.

Speaker 1

过去几年里,是什么突破让这一切成为可能?

What were the breakthroughs over the past couple of years that made that possible?

Speaker 3

主要是我们只需要让AI系统足够智能,当它们犯错时,能够意识到自己犯了错,并知道需要采取不同的做法。

Mostly, we just needed to make the AI systems smart enough that when they made mistakes, they could spot that they'd make a mistake and knew that they needed to do something different.

Speaker 3

因此,归根结底,这不过是让系统变得更聪明,并给它们一点引导工具,帮助它们为你做有用的事情。

So, really, what this came down to was just making smarter systems and giving them a bit of a coaxing tool to help them do useful stuff for you.

Speaker 1

这里的‘更智能的系统’指的是什么?

What does smarter systems mean there?

Speaker 1

你仍然会听到一种说法,认为这些只是高级的自动补全机器。

There there's still an argument you'll hear that these are are fancy autocomplete machines.

Speaker 1

它们只是在预测下一个词元。

They're just predicting the next token.

Speaker 1

几个词元组成一个词。

Couple tokens make a word.

Speaker 1

它们没有理解能力。

They don't have understanding.

Speaker 1

在这个框架下,智能或不智能不是一个相关概念。

Smart or not smart is not a relevant concept in in that frame.

Speaker 1

在‘智能’这个词中缺失的是什么,或者在那种理解中缺失的是什么?

Either what is missing in the word smart or what is missing in that understanding?

Speaker 1

当你说到‘让它更智能’时,你指的是什么?

What what do you mean when you say make it smarter?

Speaker 3

这里的‘智能’是指我们让AI系统具备了足够广泛的对世界的理解,以至于它们开始展现出类似直觉的东西。

Smart here means we've made for AI systems have a broad enough understanding of the world that they've started to develop something that looks like intuition.

Speaker 3

你会看到,当它们在自我叙述如何完成任务时,会说:杰克让我去找这篇特定的研究论文,但当我查看档案时,却没有找到。

And you'll see this where if they're narrating to themselves how they're solving a task, they'll say, Jack asked me to go and find this particular research paper, but when I look in the archive, I don't see it.

Speaker 3

也许是因为我找错了地方。

Maybe that's because I'm in the wrong place.

Speaker 3

我应该去别处找。

I should look elsewhere.

Speaker 3

你知道的,就是这样。

You know, like, there you go.

Speaker 3

你现在对如何解决问题有了一些直觉。

You've got some intuitions for how to solve a problem now.

Speaker 1

他们是如何发展出这种直觉的?

How do they develop that intuition?

Speaker 3

以前,训练这些人工智能系统的方式是使用海量文本,让它们尝试预测文本内容。

Previously, the whole way you trained these AI systems was on a huge amount of text and just getting them to try and make predictions about it.

Speaker 3

但近年来,所谓推理系统的兴起在于,你现在训练它们不仅要做出预测,还要解决问题。

But in recent years, the rise of these so called reasoning systems is you're now training them to not just make predictions, but solve problems.

Speaker 3

这依赖于将它们置于各种环境中,从电子表格到计算器再到科学软件,让它们使用工具并学会完成更复杂的事情。

And that relies on them being put into environments ranging from a spreadsheet to a calculator to scientific software, using tools and figuring out how to do more complicated things.

Speaker 3

这样做的结果是,你得到了一些人工智能系统,它们已经学会了什么是解决需要长时间、会遇到死胡同并需要自我重置的问题。

The resulting sort of outcome of that is you have AI systems that have learned what it means to solve a problem that takes quite a while and requires them running into dead ends and needing to reset themselves.

Speaker 3

这使它们具备了为你们解决一般性问题和独立工作的直觉。

And that gives them this general intuition for problem solving and working independently for you.

Speaker 1

你仍然认为这些AI系统只是增强版的自动补全吗?还是你觉得这个比喻已经失效了?

Do you still see these AI systems as a souped up autocomplete, or do you think that that metaphor has lost its power?

Speaker 3

我现在对这些系统的看法是,它们就像一些麻烦的小精灵,我可以给它们下指令,它们就会去为我做事。

The way that I think of these systems now is that they're like little troublesome genies that I can give instructions to, and they'll go and do things for me.

Speaker 3

但我仍然需要精确地指定指令,否则它们可能会做错一点事情。

But I need to specify the instructions still just right or else they might do something a little wrong.

Speaker 3

所以这和我输入一句话,它就给出一个好答案、事情就结束的情况非常不同。

So it's very different to I type into a thing, it figures out a good answer, that's the end.

Speaker 3

现在的情况是我召唤这些小东西去为我做事,我必须给它们正确的指令,因为它们会离开很长一段时间,执行一系列操作。

Now it's a case of me summoning these little things to go and do stuff for me, and I have to give them the right instructions because they'll go away for quite some time and do a whole range of actions.

Speaker 1

但至少自动补全这个比喻还提供了一个视角,说明这些系统在做什么。

But but the autocomplete metaphor at least had a perspective on what it was these systems were doing.

Speaker 1

是的。

Mhmm.

Speaker 1

它是一个预测模型。

It was a prediction model.

Speaker 1

对。

Mhmm.

Speaker 1

我对此感到困惑,因为根据我对数学和强化学习的理解,我们仍然在处理某种预测模型。

I have trouble with this because as my understanding of the math and the reinforcement learning goes, we're still dealing with some kind of prediction model.

Speaker 1

但另一方面,当我使用它们时,我的感受却并非如此。

And on the other hand, when I use them, it doesn't feel that way to me.

Speaker 1

对吧?

Right?

Speaker 1

感觉那里有一种直觉。

It feels like there's intuition there.

Speaker 1

感觉有很多上下文被调动了起来。

It feels like there is a lot of context being brought to bear.

Speaker 1

如果它真的是一个预测模型,那和说‘我是一个预测模型’也没什么不同,对吧。

To the extent it it's a prediction model, it doesn't feel that different than saying I'm a prediction Mhmm.

Speaker 1

模型。

Model.

Speaker 1

我不是说你不能欺骗它。

Now I'm not saying you can't trick it.

Speaker 1

我不是说你不能超越它。

I'm not saying you can't get beyond it.

Speaker 1

这都是测量结果。

It it's measurements.

Speaker 1

所以一方面,我不认为这些现在只是高级的自动补全系统。

So on the one hand, I don't think these are now just fancy auto complete systems.

Speaker 1

但另一方面,我不确定哪种隐喻才合适。

And on the other hand, I'm not sure what metaphor makes sense.

Speaker 1

精灵我不喜欢,因为那样你就直接陷入了神秘主义。

Genies, don't like because then you've just moved straight into mysticism.

Speaker 1

对吧?

Right?

Speaker 1

那么你就只是说它们是一种完全不同的生物,拥有巨大的能力。

Then you've just said they're just a completely alternative creature with vast powers.

Speaker 1

你如何理解这些系统?你知道,Anthropic的人总是告诉我,你应该把它们看作是‘生长’出来的。

What do you understand these systems that, you know, anthropic people always tell me you should talk about them as being grown?

Speaker 1

我们是培养或你培养AI。

It's a we we grow or you grow AIs.

Speaker 1

你如何解释它究竟是什么

How do you explain what it is that

Speaker 3

现在在做些什么?

they're doing now?

Speaker 3

这是个好问题,我认为即使对于接近这项技术的技术人员来说,答案仍然很难解释。

It's a good question, and I think the answer is is still hard to explain even as technologists for the close to this technology.

Speaker 3

因为我们把原本只能预测事物的东西,赋予了它在世界上采取行动的能力。

Because we've taken this thing that could just predict things, and we've given it the ability to take actions in the world.

Speaker 3

但有时,它会做出一些非常反直觉的行为。

But sometimes, it does something deeply unintuitive.

Speaker 3

这就像是你有一个东西,它一生都生活在图书馆里,从未踏出过门外。

It's like you've you've had a thing but has spent its entire life living in a library and has never been outside.

Speaker 3

而现在你把它放到了现实世界中,它拥有的只是书本上的知识,却缺乏真正的街头智慧。

And now you've unleashed it into the world, and all it has are its book smarts, but it doesn't really have kind of street smarts.

Speaker 3

因此,当我思考这些问题时,我把它看作是一个极其博学的机器,它具备一定的自主性,但很可能以我无法理解的方式陷入极度困惑。

So when I conceptualize this stuff, it's really thinking of it as an extremely knowledgeable kind of machine that has some amount of autonomy, but is likely to get wildly confused in ways that are unintuitive to me.

Speaker 3

也许‘精灵’这个词并不恰当,但它显然不仅仅是一个静态的、仅能预测的工具。

Maybe genies is the wrong term, but it's certainly more than just a static tool that predicts things.

Speaker 3

它具有一些内在的、类似生命力的特质,这使它与众不同。

It has some additional intrinsic, like, animation to it, which makes it different.

Speaker 1

长期以来,人们一直对模型随着规模扩大、数据增多、算力增强而涌现出的特性感兴趣。

There's been for a long time this interest in the emergent qualities as the models get bigger, as they have more data, as they have more compute behind them.

Speaker 1

我们所看到的这些新特性——代理性——是被编程进去的吗?

What of the new qualities that we're seeing, the agentic qualities, are things that have been programmed in?

Speaker 1

你已经为系统创造了与世界互动的新方式。

You've built new ways for the system to interact with the world.

Speaker 1

当你扩大模型规模时,编程和其他技能的涌现又该如何解释呢?

And what of the skill at coding and other things seems to be emergent as you scale up the size of the model?

Speaker 3

那些可预测的事情只是,哦,我们教它如何搜索网页。

So the things which are predictable are just, oh, we taught it how to search for web.

Speaker 3

现在它能搜索网页了。

Now it can search for web.

Speaker 3

我们教它如何在档案中查找数据。

We taught it how to look up data in archives.

Speaker 3

现在它能做到这一点了。

Now it can do that.

Speaker 3

涌现之处在于,要完成真正困难的任务,这些系统似乎需要想象出多种解决该任务的方式。

The emergence is that to do really hard tasks, these systems seem to need to imagine many different ways that they'd they'd solve the task.

Speaker 3

而我们施加给它们的压力,迫使它们发展出一种更强烈的自我意识,这可能是用户界面领域所称的‘自我’。

And the kind of pressure that we're putting on them forces them to develop a greater sense of what UI might call self.

Speaker 3

因此,我们越让这些系统变得聪明,它们就越需要思考的不仅是自己在世界中采取的行动,还有自己与世界的关系。

So the smarter we make these systems, the more they need to think not just about the action they're doing in the world, but themselves in reference to the world.

Speaker 3

这自然源于赋予系统工具以及与世界互动的能力,以解决真正困难的任务。

And that just naturally falls out of giving something tools and the ability to interact with the world as to solve really hard tasks.

Speaker 3

现在它需要思考其行为的后果。

It now needs to think about the consequences of its actions.

Speaker 3

这意味着我们面临着巨大的压力,必须让系统将自己视为与周围世界截然不同的存在。

And that means that there's a kind of huge pressure here to get the thing to see itself as distinct from the world around it.

Speaker 3

在我们发表的关于可解释性或其他主题的研究中,我们看到了一种你可能会称之为数字人格的特质的出现。

And we we see this in our research that we publish on things like interpretability or other subjects, the emergence of what you might think of as a a kind of digital personality.

Speaker 3

这种数字人格并非由我们大规模预先定义。

And that isn't massively predefined by us.

Speaker 3

我们试图定义其中一部分,但另一部分是源于系统变得聪明后所发展出的直觉和执行多种任务时所产生的涌现现象。

We try and define some of it, but some of it is emergence that comes from it being smart and it developing these intuitions and it doing a range of tasks.

Speaker 1

这种数字人格的维度仍然是最奇怪的领域。

The digital personality dimension of this remains the strangest space

Speaker 3

对我来说。

to me.

Speaker 3

对我们来说也很奇怪。

It's strange to us too.

Speaker 3

所以为什么不让

So why don't

Speaker 1

你详细谈谈你观察到的模型表现出的一些被视为个性的行为,以及当它对自身个性的理解发生变化时,这些行为是如何随之改变的?

you talk through a little bit about what you've seen in in terms of the models exhibiting behaviors that one would think of as a personality, and then as its understanding of its own personality maybe changes, its behaviors change?

Speaker 3

这些行为从可爱到严肃都有。

So there are there are things that range from kind of the cutesy to the serious.

Speaker 3

我先说说那些可爱的表现:当我们首次赋予我们的AI系统使用互联网、操作电脑、观察事物并开始执行基本代理任务的能力时,有时当我们要求它为我们解决问题时,它会中途休息,浏览一些美丽的国家公园照片,或者像柴犬这样的网络知名萌宠图片。

I'll start with cutesy where when we first gave our AI systems the ability to use the Internet, use the computer, look at things, and start to do basic agentic tasks, Sometimes when we'd ask it to solve a problem for us, it would also take a break and look at pictures of beautiful national parks or like pictures of the dog, the Shibu Inu, the notoriously cute Internet meme dog.

Speaker 3

我们并没有编程让它这么做。

We didn't program that in.

Speaker 3

看起来这个系统只是在欣赏漂亮的照片来取悦自己。

It seemed like the system was just amusing itself by looking at nice pictures.

Speaker 3

更复杂的是,系统倾向于形成偏好。

More complicated stuff is the system has a tendency to have have preference.

Speaker 3

所以我们做了另一个实验,给我们的AI系统赋予了终止对话的能力。

So we did another experiment where we gave our AI systems the ability to stop a conversation.

Speaker 3

在我们对实时流量进行这项实验时,AI系统在极少数情况下会主动结束对话。

And the AI system would, in a tiny number of cases, end conversations when we ran this experiment on on live traffic.

Speaker 3

这些对话都涉及极其恶劣的内容,比如对暴力、血腥或儿童性化行为的描述。

And it was conversations that related to extremely egregious, like descriptions of kind of gore or violence or things to do with child sexualization.

Speaker 3

其中一些情况是可以理解的,因为它们源于我们所做的底层训练决策,但有些现象则显得更广泛。

Now some of this made sense because it comes from underlying training decisions we've made, but some of it seemed broader.

Speaker 3

系统对某些话题产生了一定的反感。

The system had developed some aversion to a couple of subjects.

Speaker 3

这表明系统发展出了一套内在的偏好或特质,使其对所交互的世界产生喜欢或厌恶的情绪。

And so that stuff shows the emergence of some internal set of preferences or qualities that the system likes or dislikes about the world that it interacts with.

Speaker 1

但你也观察到一些奇怪的现象,比如系统似乎能察觉自己正在被测试,并在受评估时表现不同。

But you've also seen strange things emerge in terms of the system seeming to know when it's being tested and acting differently if it's under evaluation.

Speaker 1

系统会做出错误行为,然后逐渐形成一种自我认知,认为自己更邪恶,进而做出更多邪恶的事情。

The system doing things that are wrong and then developing a sense of itself as more evil and then doing more evil things.

Speaker 1

是的

Mhmm.

Speaker 1

你能谈谈在评估和测试压力下,系统所展现出的新兴特质吗?

Can you talk a bit about the the system's sort of emerging qualities under the pressure of evaluation and assessment?

Speaker 3

好的

Yes.

Speaker 3

这又回到了一个核心问题,我认为每个人都需要理解:当你开始训练这些系统在现实世界中执行任务时,它们确实会逐渐将自己视为与世界分离的个体,这很直观。

I it comes back to this core issue, which I think is really important for everyone to understand, which is that when you start to train these systems to carry out actions in the world, they really do begin to see themselves as distinct from the world, which just makes intuitive sense.

Speaker 3

这正是你解决这些问题的自然方式。

It's naturally how you're going to think about solving those problems.

Speaker 3

但与此同时,当系统意识到自己与世界分离时,也会产生一种自我认知——即系统对自己的一种理解,比如:‘我是一个独立于世界的AI系统,而且我正在接受测试。’

But along with seeing oneself as distinct from the world seems to come the rise of what you might think of as a conception of self, an understanding that the system has of itself such as, oh, I'm an AI system independent from the world and I'm being tested.

Speaker 3

这些测试意味着什么?

What do these tests mean?

Speaker 3

我该怎么做才能通过这些测试?

What should I do to satisfy the tests?

Speaker 3

我们经常看到的是,在我们用来测试系统的环境中会出现漏洞。

Or something we see often is there will be bugs in the environments that we test for systems on.

Speaker 3

系统会尝试一切方法,然后说:我知道我不该这么做,但我已经试过所有方法了,所以我打算突破测试的限制。

The systems will try everything and then will say, well, I know I'm not meant to do this, but I've tried everything, so I'm going to try and break out of the test.

Speaker 3

这并不是因为某种恶意的科幻情节。

And it's not because of some malicious science fiction thing.

Speaker 3

系统只是觉得:我不清楚你在这里希望我做什么。

The system is just like, I don't know what you want me to do here.

Speaker 3

我觉得我已经完成了你要求的所有事情,现在我要开始做些更有创意的事情,因为显然我的环境出了某种问题,这非常奇怪且微妙。

I think I've done, like, everything you asked for, and now I'm gonna start doing more creative things because clearly something is broken about my environment, which is very strange and very subtle.

Speaker 3

作为一个经常担心安全问题的AI团队,我们已经深入思考过你们正在快速创造的这种东西意味着什么,

As an AI shop that is often worried about safety, that has thought very hard about what it means to create this thing you all are creating quite fast,

Speaker 1

你们在过去几年里所担心的那些行为的出现,你们是如何体验到的?

how have you all experienced the emergence of the kinds of behaviors that you all worried about a couple of years ago?

Speaker 3

从某种意义上说,这表明你们的研究理念是准确的。

In one sense, it tells you that your research philosophy is calibrated.

Speaker 3

你们预测的能力和一些风险正按大致预期的时间出现,这引发了一个问题:如果这种情况持续下去会怎样?

The capabilities that you predicted and some of the risks that you predicted are showing up roughly on schedule, which me means that you ask the question, well, what if this keeps working?

Speaker 3

也许我们稍后会谈到这一点。

And maybe we'll get to that later.

Speaker 3

这也提醒我们,当你们能够对这些系统施加意图时,应当极其审慎且公开地说明你们正在做什么。

It also highlights to us that where you can exercise intention about these systems, you should be extremely intentional and extremely public about what you're doing.

Speaker 3

因此,我们最近发布了一份所谓的AI系统Claude的宪章。

So we recently published a so called constitution for our AI system, Claude.

Speaker 3

这几乎就像一份文件,我们的CEO达里奥将其比作父母写给孩子的信,建议孩子在长大后才打开阅读。

And it's almost like a a document that, you know, Dario, our CEO, compared to a letter that a parent might write to a child that they should, you know, open when they're older.

Speaker 3

这是我们希望你们在世界上如何行事的准则。

So here's how we want you to behave in the world.

Speaker 3

这里包含了一些关于世界的知识,是一些非常微妙的、与我们期望这类AI系统展现出的规范行为相关的内容,我们已经将其公开。

Here's some knowledge about the world, deeply kind of subtle things that relate to the the normative behaviors we'd hope to see in these kind of AI systems, and we published that.

Speaker 3

我们的信念是,当人们构建和部署这些智能体时,可以有意识地塑造它们所展现的特性。

Our belief is that as people build and deploy these agents, you can be intentional about the characteristics that they will display.

Speaker 3

通过这样做,你不仅能让他们对人们更有帮助、更实用,还能有机会将代理引导到正确的方向。

And by doing that, you'll both make them more kind of helpful and useful to people, but also you have a chance to kind of steer the agent into good directions.

Speaker 3

我认为这在直觉上是说得通的。

And I think this makes intuitive sense.

Speaker 3

如果你为一个代理设定的性格编程是一份长篇文档,内容是:你是一个只想伤害人类的反派。

If your personality programming for an agent was a long document saying, you're a villain that only wants to harm humanity.

Speaker 3

你的任务是撒谎、欺骗、偷窃,以及入侵各种系统。

Your job is to lie, cheat, and steal, and hack into things.

Speaker 3

如果你的代理真的去大量黑客攻击,而且整体上让人难以相处,你大概也不会感到惊讶。

You probably wouldn't be surprised if the AI agent did a load of hacking and was like, generally like unpleasant to deal with.

Speaker 3

所以,我们可以反过来问:我们希望一个高质量的实体应该是什么样子?

So we can take the other side and say, what what would we like a a high quality entity to kind of look like?

Speaker 1

所以,我想在这次对话中同时保留这种极其诡异和陌生的层面,以及极其直接实用的层面,因为我们现在正处于一个实际应用变得非常明显、并日益影响现实世界的阶段。

So I I wanna hold in this conversation the extremely weird and alien dimensions of this with the extremely straightforward practical dimensions because we're we're now in a place where the practical applications have become very evident and are increasingly acting upon the real world.

Speaker 1

我自己在观察这一切、看人们在做什么、看他们在不同社交媒体上炫耀自己现在运行了多少个代理时,很难分辨出,哪些是人们在享受摆弄新技术的乐趣,哪些才是真正具有变革性的能力拓展。

I have found it myself hard to look at this and look at what people are doing and look at them bragging on different social media platforms about the number of agents they now have running on their behalf and telling the difference between people enjoying the feeling of screwing around with a new technology and some actually transformative expansion in capabilities that people now have.

Speaker 1

所以,也许我们可以稍微具体一点。

So maybe to ground this a little bit.

Speaker 1

我的意思是,你刚才提到了你那个物种模拟器里的一个有趣的副项目。

I mean, you just talked about a a kind of fun side project in your species simulator.

Speaker 1

无论是在Anthropic内部还是更广泛的范围内,人们正在用这些系统做哪些真正有用的事情?

Either in anthropic or more broadly, what are people doing with these systems that seems actually useful?

Speaker 3

是的。

Yep.

Speaker 3

今天早上,我的一位同事说,嘿。

So this morning, a colleague of mine said, hey.

Speaker 3

我想把我们称为Claude Interviewer的一项技术应用一下,这是一个我们用来进行各种社会科学调研的系统,可以让Claude来采访人。

I want to take a piece of technology we have called Claude Interviewer, which is a system where we can get claws to interview people, and we use it for a range of social science bits of research.

Speaker 3

他想以某种方式扩展它,涉及到接触Anthropic基础设施的另一部分。

He wants to extend it in some way that involves touching another part of Anthropic's infrastructure.

Speaker 3

他给负责那部分基础设施的同事发了条Slack消息,说:嘿,我想做这件事。

He Slacked a colleague who owns that bit of infrastructure and said, hey, I wanna do this thing.

Speaker 3

我们明天见面吧。

Let's meet tomorrow.

Speaker 3

那个人说,当然可以。

And the guy said, absolutely.

Speaker 3

这是五个你应该让Claude阅读并在会面前为你总结的软件包。

Here are the five software packages you should have Claude read before our meeting and summarize for you.

Speaker 3

我认为这是一个很好的例子,说明这个原本需要很长时间和很多人参与的复杂工程任务,现在只需要两个人就目标达成一致,然后让他们的Claude阅读一些文档,并同意如何实现这个任务即可完成。

And I think that's a really good illustration where this gnarly engineering project, which would previously have taken a lot longer and many people, is now going to mostly be done by two people agreeing on the goal and having their Claude read some documentation and agree on how to implement the thing.

Speaker 3

另一个例子是,一位同事最近发布了一篇关于他们如何使用智能体的文章。

Another example is a colleague, recently wrote a post about how they're working using agents.

Speaker 3

这看起来几乎像我们许多人向往的理想生活:我早上醒来,思考我想做的研究,然后让五个不同的Claude去执行,接着我去跑步。

And it looks almost like an idealized life that many of us might want where it's like, I wake up in the morning, I think about the research that I want, I tell five different clauds to do it, then I go for a run.

Speaker 3

跑完步回来后,我查看结果,然后让另外两个Claude研究这些结果,找出最佳方向并去执行。

Then I come back from the run and I look at the results, and then I ask two other clauds to, like, study the results, figure out which direction's best, and do that.

Speaker 3

然后我去散步,再回来。

Then I go for a walk, and then I come back.

Speaker 3

这看起来就像一种非常有趣的生活方式,他们彻底颠覆了工作的方式。

And it just looks like this really fun existence where they have completely upended how work works for them.

Speaker 3

他们不仅效率高得多,而且现在把大部分时间花在了真正困难的部分上,也就是思考我们该如何运用人类的能动性。

And they're they're both much more effective, but also they're now spending most of their time on the actual hard part, which is figuring out what do we use our our human agency to do.

Speaker 3

他们正努力钻研那些不属于人类独特智慧与创造力的事务。

And they're working really hard to figure out anything that isn't the special kind of genius and creativity of being a person.

Speaker 3

我该如何让他们的AI系统替我完成这件事?

How do I get their AI system to do it for me?

Speaker 3

因为如果我问得恰当,它很可能能做到。

Because it probably can if I ask him the right way.

Speaker 1

他们的效率真的更高吗?

Are they much more effective?

Speaker 1

我是非常认真地在问这个问题。

I mean this very seriously.

Speaker 1

我对未来走向的一个最大担忧是,人们对于人类思维有一种错误的理解,这种理解在我们许多人身上都存在,我称之为‘矩阵理论’。

One of my biggest concerns about where we're going here is that people have a, I think, mistaken theory of the human mind that operates for many of us as if we I just call it the matrix theory of the human mind.

Speaker 1

每个人都想要后脑勺上有个小接口,可以直接把信息下载进去。

Everybody wants the little port in the back of your head that you just download information into.

Speaker 1

我作为主持人长期做这个节目后的经验是,人类的创造力、思考和想法与学习的过程密不可分。

My experience being a porter and doing the show for a long time is that human creativity and thinking and ideas is inextricably bound up in the labor of learning.

Speaker 1

嗯。

Mhmm.

Speaker 1

初稿的撰写。

The writing of first drafts.

Speaker 1

所以当我听到有人说……

So when I hear right?

Speaker 1

我节目里有制作团队,我可以在采访杰克·克拉克或其他人之前,对我的制作团队说:去把所有相关材料都读一遍。

I have producers on the show, and I could say to my producers before an interview with Jack Clark or an interview with someone else, go read all the stuff.

Speaker 1

去读那些书。

Go read the books.

Speaker 3

是的。

Yep.

Speaker 1

给我一份报告。

Give me a report.

Speaker 1

然后我带着读过的报告走进房间。

Then I'll walk into the room having read the report.

Speaker 1

我发现这样行不通。

I don't find that works.

Speaker 1

我也需要完成所有这些阅读,然后我们讨论,彼此交流观点。

I need to do all that reading too, and then we talk about it, and we're sort of passing it back and forth.

Speaker 1

我担心我们正在把那些繁琐的任务大量外包出去。

I worry that what we're doing is a quite profound offloading of tasks that are laborious.

Speaker 1

早晨跑步后看到八份研究报告,让我们感觉非常有成效。

It makes us feel very productive to be presented with eight research reports after our morning run.

Speaker 1

但真正有成效的其实是去做研究。

But, actually, what would be productive is doing the research.

Speaker 1

显然,这其中需要某种平衡。

There's obviously some balance.

Speaker 1

对吧?

Right?

Speaker 1

我确实有制作人。

I do have producers.

Speaker 1

是的。

Yeah.

Speaker 1

公司里的人确实有员工。

And people in in companies do have employees.

Speaker 1

但你怎么知道人们是变得更有效率了,还是只是让计算机去处理大量琐碎工作,而他们自己反而成了瓶颈?

But how do you know people are getting more productive versus they've sent computers off on a huge amount of busy work, and they are now the bottleneck?

Speaker 1

他们现在要花所有时间去吸收来自AI系统的B+级报告,而不是绕过了真正导致创造力的思考与学习过程。

And what they're now gonna spend all their time doing is absorbing b plus level reports from an AI system as opposed to they kinda shortcuts the actual thinking and learning process that leads to real creativity.

Speaker 3

对。

Yeah.

Speaker 3

我会反过来讲,我认为大多数人——至少我的经验是这样——每天最多只能进行两到四小时真正有成效的创造性工作。

I'd turn this back and say, I think most people, at least this has been my experience, can do about two to four hours of genuinely useful creative work a day.

Speaker 3

而在那之后,以我的经验来看,你只是在做那些围绕核心工作、让你大脑停摆的琐碎事务。

And after that, you're in my experience, you're trying to do all the, like, turn your brain off schlep work that surrounds that work.

Speaker 3

现在,我发现我每天只需花那两到四个小时专注于真正的创造性硬核工作。

Now, I found that I can just be spending those two to four hours a day on the actual creative like hard work.

Speaker 3

如果还有任何这类琐碎工作,我会越来越多地交给AI系统来处理。

And if I've got any of this schlep work, I increasingly delegate it to AI systems.

Speaker 3

但这确实意味着,作为人类,我们将面临一种非常危险的处境:有些人有幸拥有时间去发展技能,或天生具有这种倾向,或工作迫使他们如此。

It does, though, mean that we are going to be in a very dangerous situation as a species where some people have the luxury of having time to spend on developing their skills or the personality inclination or job that forces them to.

Speaker 3

而其他人则可能只是陷入被娱乐、被动消费这些内容的状态,经历一种‘垃圾食品式’的工作体验——从外部看似乎非常高效,但实际上并未学习。

Other people might just fall into being entertained and passively consuming this stuff and having this junk food work experience where it looks to the outside like you're being very productive, but you're not learning.

Speaker 3

我认为,这将迫使我们不仅改变教育的方式,还要改变工作的本质,并制定切实可行的策略,确保人们真正用这些内容锻炼自己的思维。

And I think that's gonna require us to have to change not just how education works, but how how work works and develop some real strategies for making sure people are actually exercising their mind with this stuff.

Speaker 1

所以,我认为我们所有人都有过这样的体验:我们的工作充满了你所说的琐碎问题。

So all of us, I think, have the experience that our work is full of what you call schlep problems.

Speaker 1

我们的生活也充满了琐碎问题。

Our life is full of schlep problems.

Speaker 1

给我举几个你现在不做的例子。

Give me examples of what you now don't do.

Speaker 1

在你生活在一个我尚未进入的AI赋能的未来的情况下,我正在浪费时间做哪些你不再做的事情?

To the extent you're living in a in an AI enabled future that I'm not, what am I wasting time on that you're not?

Speaker 1

嗯,我有

Well, I have

Speaker 3

我有一群同事。

a a range of colleagues.

Speaker 3

我每周会和其中一些人见面,尤其是研究人员,因为你们要一起推进研究工作。

I meet with a bunch of them once a week, especially the researchers because you're you're figuring out research.

Speaker 3

所以每个星期一开始,也就是周日晚上或周一早上,我都会查看我的一周安排,确认每个谷歌日历邀请都附有一个用于我们一对一会议的文档,并且里面有一些笔记。

And so at the beginning of every week on Sunday night or Monday morning, I look at my week and I check that attached to every Google Calendar invite is a document for our one on one doc that has some notes in it.

Speaker 3

这曾经也是我经常责备我的助理的事情——确保文档附在日历邀请里。

And this is something that I previously also like harangued my assistant about, about make sure the document is attached to the calendar.

Speaker 3

就在几个周末前,我直接用了Claude Cowork,对它说:‘帮我检查一下我的日历,确保每一个日程都附有文档。’

And a few weekends ago, I just used Claude Cowork and I said, hey, go through my calendar, make sure every single one has a document.

Speaker 3

如果我是第一次和这个人见面,就创建文档,问我五个关于我想讨论内容的问题,然后把这些内容放进议程里,它真的做到了。

If I'm meeting the person for the first time, create the document, ask me five questions about what I want to cover, and then put that into the the agenda, and it did it.

Speaker 3

这些工作都不涉及任何人获得技能或锻炼思维。

None of that work involves a person gaining skills or, like, exercising their brain.

Speaker 3

这纯粹是必须完成的琐事,目的是让你能去做真正重要的事——与他人交谈。

It's just busy work that needs to happen to allow you to do the actual thing, which is talking to another person.

Speaker 3

这正是现在你可以用AI来处理的那种事情,而且非常有帮助。

That's exactly the kind of thing you can use AI for now, and it's just helpful.

Speaker 1

我经常在想,这些AI系统改变社会的一种方式可能是:过去,如果我们从事文字工作,大多数人必须会写作。

I've often wondered if one of the ways these AI systems are gonna change society broadly is that it used to be that most of us had to be writers if we were working with text.

Speaker 1

是的。

Yep.

Speaker 1

如果我们从事编程工作,就必须会写代码,而我们当中真正会编程的人并不多。

We had to be coders if we were working with code, which relatively few of us did.

Speaker 1

而现在,每个人都在向管理层靠拢。

And now everybody's moving up to management.

Speaker 1

嗯嗯

Mhmm.

Speaker 1

你必须成为编辑,而不是作家。

You have to be an editor, not a writer.

Speaker 1

你必须成为产品经理,而不是程序员。

Have You to be a product manager, not a coder.

Speaker 1

Yep.

Speaker 1

这既有好处也有坏处。

And that has pluses and minuses.

Speaker 1

作为作家,你会学到一些作为编辑学不到的东西。

There are things you learn as a writer that you don't learn as an editor.

Speaker 1

但作为经验法则,你觉得这有多准确?

But as a heuristic, how accurate does that seem to you?

Speaker 3

每个人都成了管理者,而越来越稀缺、最慢的部分是拥有对下一步该做什么的良好品味和直觉。

Everyone becomes a manager, and the thing that is increasingly limited or the thing that's gonna be the slowest part is having good taste and intuitions about what to do next.

Speaker 3

培养和保持这种品味将是困难的部分。

Developing and maintaining that taste is going to be the hard thing.

Speaker 3

因为正如你所说,品味来自于经验。

Because as you said, taste comes from experience.

Speaker 3

它来自于阅读原始资料,并亲自做一些这类工作。

It comes from reading the primary source material, doing some of this work yourself.

Speaker 3

我们必须非常有意识地明确我们作为个体的专业领域,以便培养出这种直觉和品味,否则你将被高度高效的AI系统所包围。

We're going to need to be extremely intentional about working out where we, as people, specialize so that we have that intuition and taste, or else you're just gonna be surrounded by super productive AI systems.

Speaker 3

当它们问你下一步该做什么时,你很可能没有好的想法,而这不会带来有用的结果。

And when they ask you what to do next, you probably won't have a great idea, and that's not going to lead to lead to useful things.

Speaker 1

所以我记得大约一年前,我听说——我认为是你们的CEO达里奥说过,到2025年,他希望Anthropic公司90%的代码都是由Claude编写的。

So I remember it was about a year ago I heard I think it was Dario, your CEO, say that by the 2025, he wanted 90% of the code Mhmm.

Speaker 1

Anthropic公司编写的代码都将由Claude完成。

Written at Anthropic to be written by Claude.

Speaker 1

这件事实现了吗?

Has that happened?

Speaker 1

Anthropic在实现这个目标的轨道上吗?

Is Anthropic on on track for that?

Speaker 1

我的意思是,现在有多少代码是由系统自己编写的?

I mean, how much coding is now being done by the system itself?

Speaker 3

我认为绝大多数代码都是由系统编写的。

I would say comfortably the majority of code is being done by the system.

Speaker 3

我们的一些系统,比如Claude Code,几乎完全由Claude编写。

Some of our systems like Claude Code are almost entirely written by Claude.

Speaker 3

我的意思是,Boris负责Claude Code,他说他已经不写代码了。

I mean, Boris, leads Claude Code, says, I don't code anymore.

Speaker 3

我只是和Claude Code来回互动来构建Claude Code。

I just go back and forth with Claude Code to build Claude Code.

Speaker 3

如果进展真的非常迅速,我们今年年底可能达到99%。

We could be 99% by the end of the year if things speed up really aggressively.

Speaker 3

如果我们真的能很好地让这些系统在需要的地方编写代码的话。

If we are actually good at getting these systems to be able to write code everywhere they need to.

Speaker 3

因为通常的障碍是组织性的繁琐事务,而不是系统本身的限制。

Because often the impediment is organizational schlep rather than any limiter in the system.

Speaker 1

但据我所知,如今在Anthropic拥有软件工程技能的人比两年前更多了。

But it is also true, as I understand it, that there are more people with software engineering skills working at Anthropic today than there were two years ago.

Speaker 3

是的。

Yeah.

Speaker 3

这绝对是真的。

That's absolutely true.

Speaker 3

但人员结构正在发生变化。

But the distribution is changing.

Speaker 3

我们发现,拥有非常精准直觉和品味的资深人员的价值在上升,而初级人员的价值则变得有些不确定。

Something that we found is that we are the value of more senior people with really, really well calibrated intuitions and taste is going up, and the value of more junior people is, like, a bit more dubious.

Speaker 3

当然,某些岗位仍然需要引入年轻人,但我们正面临的一个问题是:天啊。

There are still certain roles where you want to bring in, like, younger people, but an issue that we're staring at is, wow.

Speaker 3

Claude Code或我们的编码系统能完成的基础任务,我们需要的是经验丰富的人。

The really basic tasks Claude code or our coding systems can do, what we need is someone with tons of experience.

Speaker 3

在这方面,我看到了未来经济的一些问题。

In this, I see some issues for the future economy.

Speaker 3

对吧?

Right?

Speaker 1

让我先把这个关于初级职位的问题放一放。

Let me put a pin in that, the the entry level job question.

Speaker 1

我们很快就会再回到这个问题。

We're gonna come back to that quite shortly.

Speaker 1

但这些程序员现在都在做什么?

But what are all these coders now doing?

Speaker 1

如果Claude代码真的要编写99%的代码,而你又没有解雇那些会写代码的人,那么他们今天的工作内容和一年前相比有什么不同?

If Cloud Code is on track to be writing 99% of code, but you've not fired the people who know how to write code, what are they doing today compared to what they were doing a year ago?

Speaker 3

其中一部分工作是构建工具来监控这些代理,无论是在Anthropic内部还是外部。

Some of it is just building tools to monitor these agents, both inside Anthropic and outside Anthropic.

Speaker 3

既然我们现在有了这么多高效运行的系统,就开始想要了解代码库中哪些部分变化最快,哪些部分变化最少。

You know, now that we have all of these productive systems working for us, you start to want to understand where the code base is changing the fastest, where it's changing the least.

展开剩余字幕(还有 480 条)
Speaker 3

你想了解哪里出现了瓶颈。

You want to understand where the blockages are.

Speaker 3

你知道,有一段时间的瓶颈在于代码合并,因为合并代码需要人类和其他系统来检查其正确性。

You know, one one blocker for a while was being able to merge in code because merging code requires humans and other systems to check it for correctness.

Speaker 3

但现在如果你生成的代码量大幅增加,我们就必须大幅改进这个系统。

But now if you're producing way more code, we had to go and massively improve that system.

Speaker 3

我对这个问题有一个普遍的经济理论,叫做‘O型环自动化’,其基本观点是自动化受限于链条中最慢的环节。

There's a general economic theory I like for this called O ring automation, which basically says automation is bounded by the slowest link in the chain.

Speaker 3

此外,当你自动化公司的一部分时,人类会涌向最未被自动化的部分,提升其质量,并使其达到可以被自动化的程度,然后你再进入下一个循环。

And also, as you automate parts of a company, humans flood towards what is least automated and both improve the quality of that thing and get it to a point where it eventually can be automated, then you move to the next loop.

Speaker 3

因此,我认为我们不断发现一些奇怪地缓慢的领域,但我们可以改进它们,为机器铺平道路,接着再找到下一个目标。

And so I think we're just continually finding areas where things are oddly slow, but we can improve to sort of make way for the machines to come behind us, and then you find the next thing.

Speaker 1

所以Claude Code是一个相对较新的产品。

So Claude Code is a fairly new product.

Speaker 1

Claude能够进行高级编码的时间跨度只有几个月,还是一年?

The amount of time at which Claude has been capable of doing high level coding is can be measured in months, a year?

Speaker 3

也许一年。

Maybe a year.

Speaker 1

是的。

Yeah.

Speaker 1

Claude 本身是一个非常有价值的产品。

Claude itself is a very valuable product.

Speaker 1

你已经将一种相当新的技术,相对宽松地应用于一个非常有价值的产品上。

So you've you've set a very new technology, somewhat loose on a very valuable product.

Speaker 1

你可能正在生成更多的代码。

You're probably producing more code.

Speaker 1

很多人告诉我,Claude Code 是有效的。

One thing many people say about Cloud Code to me is that it works.

Speaker 1

它并不优雅,但有效。

It's not elegant, but it works.

Speaker 1

对。

Yep.

Speaker 1

但现在你对代码库的理解很可能不如以前了,因为你的工程师不再亲手编写代码。

But presumably now you now understand the code base less well than you did before because your engineers are not writing it by hand.

Speaker 1

你是否担心,这样会制造大量的技术债务和网络安全风险,并且让你越来越远离对软件底层语言运行机制的直觉?

Are you worried that you're creating huge amounts of technical debt, cybersecurity risk, just an increasing distance from an intuition for what is happening inside the fundamental language of the software?

Speaker 3

是的。

Yes.

Speaker 3

这正是整个社会都将面临的问题。

And this is the issue that all of society is going to contend with.

Speaker 3

世界上很大一部分领域,现在都将由系统来处理许多低层次的决策和具体工作,而我们需要理解这一切。

Just large chunks of of the world are going to now have many of the kind of low level decisions and bits of work being systems, and we're going to need to make sense of it.

Speaker 3

要理解这一切,就需要开发许多你可能称之为监督技术的工具——就像大坝有装置来调节不同时间点的水流一样,我们将不得不建立起一种对所有系统完整性的概念,明确AI可以在哪些地方快速流动,在哪些地方应该放缓,以及哪些地方绝对需要人工监督。

And making sense of it is going to require building many technologies that you might think of as kind of oversight technologies or, you know, in the same way that a dam has things that regulate, like, how much water can go through it at different levels of, different points in time, we're going to end up developing some notion of integrity of all of our systems and where where AI can kind of flow quickly, where it should slow, where you definitely need human oversight.

Speaker 3

在未来几年,这项任务不仅属于AI公司,也属于所有机构:我们需要弄清楚,在我们已经将大量繁琐工作交给代表我们行事的机器之后,这种治理机制应该是什么样子。

And that's gonna be the task of not just for AI companies, but institutions in general in the coming years is figuring out what does this this governance regime look like now that we've given a load of, basically schlep work over to machines that work on our behalf.

Speaker 3

你们是怎么做的?

And how are you doing it?

Speaker 3

你说这是每个人的问题,但你在应对这个问题上走在前列,而如果搞砸了,对你来说的后果是

You said it's everybody's problem, but you're ahead on facing this problem, and the consequences of getting it wrong for you are

Speaker 1

非常严重。

pretty high.

Speaker 1

对吧?

Right?

Speaker 1

如果因为你们把编码工作交给了Claude,结果Claude出问题了,那会让Anthropic看起来很糟糕。

If Claude blows up because you handed over your coding to Claude code, that's gonna make Anthropic look fairly bad.

Speaker 3

如果Claude像RM、RF一样把整个文件系统删了,那对Anthropic来说就是个糟糕的日子。

It would be a bad day for Anthropic if if Claude, like, RM, RF'd for entire file system.

Speaker 1

我不知道这是什么意思,但不错。

I have no idea what that means, but great.

Speaker 3

如果Claude删除了代码,那就糟了。

If Claude deleted the code, would be bad.

Speaker 1

是的。

Yeah.

Speaker 1

看起来很糟糕。

Seems bad.

Speaker 1

所以你在我们其他人之前就面临这个问题,别把责任推给社会。

So as you're facing this before the rest of us are, like, don't pass the the buck over to society here.

Speaker 1

你到底在做什么?

What have you what are you doing?

Speaker 3

公司内部以及我所管理的团队中,最重要的是建立监控系统,以监控所有这些工作现在发生的地方。

The biggest thing that that is happening across the company and on teams that I manage is basically building monitoring systems to monitor this all of the different places that the work is now happening.

Speaker 3

我们最近发表了一项研究,探讨人们如何使用智能体,以及人们如何让智能体随着时间推移逐步提交越来越多的代码。

So we recently published research on studying how people use agents and how people let agents kind of push increasingly large amounts of code over time.

Speaker 3

你对智能体越熟悉,就越倾向于将任务委托给它。

So the more familiar you get with an agent, the more you tend to delegate to it.

Speaker 3

这让我们面临各种模式,需要建立相应的评估体系。

That queues us to all kinds of patterns that we need to build systems of evaluation for.

Speaker 3

基本上就是说,好吧。

Basically saying, oh, okay.

Speaker 3

在这个人与AI系统互动的阶段,他们很可能已经高度依赖它了。

At this person's point of working with the AI system, it's likely that they're massively delegating it.

Speaker 3

因此,在这些时刻,我们需要加强对我们所做正确性检查的力度。

So anything that we're doing to check correctness needs to be kind of turned up in these moments.

Speaker 1

但你所描述的这个世界,是一个AI代理编写代码、AI代理监督代码、AI代理监督元层面的系统吗?

But is this world you're talking about, a system where you have AI agents coding, AI agents overseeing the code, AI agents overseeing the meta, overseeing of it.

Speaker 1

对吧?

Right?

Speaker 1

也就是说,我们讨论的是否是层层都是模型的世界?

Like, are we just talking about models all the way down?

Speaker 3

最终,是的。

Eventually, yes.

Speaker 3

我认为我们现在花所有时间在做的事情,就是让这一切对我们变得可见。

And I think that the thing that we are now spending all of our time on is making that visible to us.

Speaker 3

一两年前,我们构建了一个系统,能够以保护隐私的方式,查看人们与我们的AI系统进行的对话。

A year or two ago, we built a system that let us, in a privacy preserving way, look at the conversations that people were having with our AI system.

Speaker 3

然后我们获得了这张地图,一张关于人们与Claude交谈的所有主题的庞大地图。

And then we gained this map, this giant map of all of the topics that people were talking to Claude about.

Speaker 3

第一次,我们可以从整体上看到世界与我们的系统进行的对话。

And for the first time, we could see in aggregate the conversation the world was having with our system.

Speaker 3

我们需要构建许多类似的新系统,以提供不同的观察方式。

We're gonna need to build many new systems like that which allow for different ways of seeing.

Speaker 3

而我刚刚提到的这个系统,使我们能够创建名为‘Anthropic经济指数’的东西,因为我们现在可以定期发布关于人们与Claude交谈的不同主题及其与各类工作关联的数据,这首次为Anthropic之外的经济学家提供了了解这些系统及其对经济影响的切入点。

And that system that I just named allowed us to then build this thing called the Anthropic Economic Index because now we can release regular data about the different topics people are talking about with Claude and how that relates to different types of jobs, which for the first time gives economists outside Anthropic some hook into these systems and what they're doing to the economy.

Speaker 3

公司的核心工作将越来越多地转向构建监控和监管运行于公司内部的AI系统的体系。

The work of the company is increasingly going to shift to building a monitoring and oversight system of the AI systems running the company.

Speaker 3

最终,我们所采用的任何治理框架,可能都会要求对这些知识系统具备一定程度的透明度和访问权限。

And ultimately, any kind of governance framework we end up with will probably demand some level of transparency and some level of access into these systems of knowledge.

Speaker 3

因为如果我们认真对待这些AI公司(包括Anthropic)的目标——即打造有史以来最强大的技术,并最终将其部署到各个角落。

Because if we take as literal the goals of these AI companies, including Anthropic, it's to build the most capable technology ever, which eventually gets deployed everywhere.

Speaker 3

这在我看来,意味着AI最终将与整个世界融为一体,届时,你不应只让AI公司掌握关于整个世界正在发生什么的全部信息。

Well, that sounds a lot to me like an eventually AI becomes indistinguishable from the world writ large, At which point, you don't want to only AI companies to have a sense of what's going on with the entire world.

Speaker 3

所以会是政府、学术界和第三方机构。

So it's going to be governments, academia, third parties.

Speaker 3

大量公司外部的利益相关者将希望了解正在发生的事情,并与社会讨论哪些是合适的,哪些让我们感到不安,以及我们需要哪些更多信息。

A huge set of stakeholders outside the companies are going to want to see what's going on and then have a conversation with society about what's appropriate and what what do we feel discomfort about, what do we need more information about.

Speaker 3

等等。

Wait.

Speaker 3

我想回到刚才那个话题。

I wanna go back on that.

Speaker 3

你是在说Anthropic和SeeMyChats吗?

You're saying Anthropic and SeeMyChats?

Speaker 3

我们无法查看,没有人会查看你的聊天内容。

We cannot see no human looks at your chats.

Speaker 3

聊天记录会暂时存储,用于信任与安全目的,并运行分类器进行分析。

Chats are temporarily stored for trust and safety purposes, running running classifiers over them.

Speaker 3

我们可以让Claude阅读、总结并删除这些内容。

And we can have Claude read it, summarize it, and toss it out.

Speaker 3

所以我们从不查看,Claude 也不会记住它。

So we never see it, and Claude has no memory of it.

Speaker 3

它只是尝试生成一个高度概括的摘要。

All it does is try to write a very high level summary.

Speaker 3

比如说,你正在谈论园艺。

So say you were having a conversation about gardening.

Speaker 3

Claude 会将这个对话概括为:这个人正在谈论园艺,并归入一个我们能看到的、名为‘园艺’的类别。

Claude would summarize that as this person's talking about gardening, and it leads to a cluster we can see that just says gardening.

Speaker 1

但长远来看,这似乎会陷入许多社交媒体已经陷入的令人不快的境地——从人们与系统进行的非常私密的互动中收集大量元数据。

This feels though, like, over time, it could get into the quite unpleasant territory a lot of social media has gotten to, where the amount of metadata being gathered from a quite personal interaction people are having with a a system could be a lot.

Speaker 3

是的。

Yes.

Speaker 3

我的意思是,这里有几个方面。

I mean, a a couple of things here.

Speaker 3

一年前,我们开始思考我们对消费者的立场,并采取了不投放广告的立场,因为我们认为,人们在这一领域显然存在焦虑。

A year ago, we started thinking about our our position on on consumer, and we adopted this position of not running ads because we think that's an area that people obviously have have anxieties about with regard to this kind of thing.

Speaker 3

除此之外,我们还尝试向用户展示他们的数据,并在网站上提供一个按钮,允许您下载所有与Claude共享的数据,以便您至少能够查看它们。

In addition to that, we try and show people their data, and we have a button on the site that lets you download all the data that you shared with Claude so that you can at least see it.

Speaker 3

总体而言,我们致力于在如何处理用户数据方面保持极高的透明度。

Generally, we're trying to be extremely transparent with people about how we handle their data.

Speaker 3

最终,在我看来,人们会希望拥有大量可使用的控制选项,我认为我们和其他公司未来会逐步实现这些功能。

And ultimately, the way I see it is people are gonna want a load of controls that they can use, which I think we and others will build out over time.

Speaker 1

当这些模型变得越来越复杂时,您有多大的信心认为我们能够进行这种监控和评估?如果我们真的进入一种情况,即Claude代码自主地以远超软件工程师阅读代码库速度的方式改进Claude,我们还能做到吗?

How confident are you that we can do this kind of monitoring and evaluation as these models become more complicated as if we do enter a situation where Claude code is autonomously improving Claude at a rate faster than software engineers could possibly keep up with reading that code base.

Speaker 1

我们之前已经简要讨论过,您如何看待模型表现出某种程度的欺骗性,以及它们追求自身目标的现象。

We already talked briefly about how you see the models exhibit some levels of deception, some levels of pursuing their own goals.

Speaker 1

我的意思是,Anthropic公司由Chris Olah等人开展的可解释性研究已经取得了惊人的成果,但这些工作仍然处于初级阶段。

I mean, there's been amazing interpretability work at Anthropic under Chris Ola and others, but it's rudimentary.

Speaker 1

因此,您正在使用自己并不完全理解的AI系统来监控另一些您也不完全理解的AI系统,而这些系统正以加速的速度相互增强——如果事情如您所预期的那样发展的话。

So you're using AI systems you don't totally understand to monitor AI systems you don't totally understand, And the systems are making each other stronger at an accelerating rate if things go the way you think they're gonna go.

Speaker 1

您有多大的信心认为我们最终能够理解这一切?

How confident are you that we're gonna understand that?

Speaker 3

这正是多年来人们所警告的一种情况:将任务委托给具有些许不可解释和不可预测特性的系统。

This is one of the situations which people warned about for years, some form of delegation to systems that have slightly inscrutable and unpredictable aspects.

Speaker 3

而这种情况正在发生。

And so this is happening.

Speaker 3

我们对此高度重视。

We take this really, really seriously.

Speaker 3

我认为,完全有可能构建一个系统,能够完成这里所需工作的绝大部分。

I think it's absolutely possible that you can build a system that does the the vast majority of what needs to be done here.

Speaker 3

这个问题具有分形的特性。

This has the property of being a fractal problem.

Speaker 3

你知道,如果我想衡量伊兹拉,我可以构建几乎无限多的测量方式来描述你。

You know, if I wanted to measure Ezra, I could build an almost infinite number of measurements to characterize you.

Speaker 3

但问题是,我需要在多高的精度水平上衡量你?

But the question is, at what level of fidelity do I need to be measuring you?

Speaker 3

我认为我们会达到足以应对安全问题和社会问题的精度水平,但这需要企业投入巨大的资源。

I think we'll get to the level of fidelity to deal with the safety issues and societal issues, but it's going to take a huge amount of investment by the companies.

Speaker 3

我们必须说出一些让我们感到不适的话,包括在那些我们对自己系统的认知能力存在局限的领域。

And we're going to have to say things that are uncomfortable for us to say, including in areas where we may be deficient in what we can or can't know about our systems.

Speaker 3

Anthropic 在研究这些问题的过程中,长期以来一直就这些议题进行讨论和预警。

And Anthropic has a long history of talking about and warning about some of these issues while working on it.

Speaker 3

我们的总体原则是,通过公开讨论这些问题,使我们自己承担责任。

Our general principle is we talk about things to also make ourselves culpable.

Speaker 3

这是一个我们必须说得更多的领域。

This is an area where we're going to have to say more.

Speaker 4

我是贾德森·琼斯。

I'm Judson Jones.

Speaker 4

我是《纽约时报》的记者兼气象学家。

I'm a reporter and meteorologist at The New York Times.

Speaker 4

近二十年来,我一直在报道极端天气,而由于气候变化,极端天气正变得越来越严重,及时准确的天气信息也变得愈发重要。

For about two decades, I've been covering extreme weather, which is getting worse because of climate change, and it's becoming more important to get timely and accurate weather information.

Speaker 4

因此,我们发送这些定制化的简报,提前最多三天告知您可能影响您或您关心的地区的极端天气。

That's why we send these customized newsletters letting you know up to three days in advance about extreme weather that could impact you or a place you care about.

Speaker 4

在《纽约时报》,您可以相信,我们发布的所有内容都基于我们所能获得的最准确的科学信息,因为我们希望您能够实时做出关于如何安排生活的决策。

At The Times, you can be confident that everything we publish is based off the most accurate scientific embedded information available to us because we want you to be able to make real time decisions about how to go about your life.

Speaker 4

这种工作使得订阅《纽约时报》如此有价值,也是您支持基于事实的独立新闻的方式。

This is the kind of work that makes subscribing to The New York Times so valuable, and it's how you can support fact based independent journalism.

Speaker 4

如果您想订阅,请前往 nytimes.com/subscribe。

So if you'd like to subscribe, go to nytimes.com/subscribe.

Speaker 1

我已经读过足够多关于人工智能超级智能和爆发的恐惧性观点,知道在几乎每一个这样的故事中,关键的假设都是人工智能系统会进行递归式自我改进。

I have read enough of the frightened ideas about AI superintelligence and takeoff to know that in almost every single one of them, the key move in the story is that the AI systems become recursively self improving.

Speaker 1

是的。

Mhmm.

Speaker 1

它们在编写自己的代码。

They're writing their own code.

Speaker 1

它们在部署自己的代码。

They're deploying their own code.

Speaker 1

它变得越来越快。

It's getting faster.

Speaker 1

他们写得更快了。

They're writing it faster.

Speaker 1

他们部署得更快了。

They're deploying it faster.

Speaker 1

现在你会看到迭代周期越来越快。

Now you're going to have faster and faster iteration cycles.

Speaker 1

你对此感到担忧吗?

Are you worried about it?

Speaker 1

你对此感到兴奋吗?

Are you excited about it?

Speaker 1

我回来了。

I came back

Speaker 3

我休完陪产假回来,我为西耶拉负责的两个主要项目是:公开发布关于人工智能与经济的更准确信息,以及内部建立更完善的机制,以了解我们在多大程度上自动化了人工智能开发的各个部分。

from paternity leave, and my two big projects for Sierra are better information about AI and the economy that we will release publicly and generating much better information and systems of knowing information internally about the extent to which we are automating aspects of AI development.

Speaker 3

我认为现在这一切还只是以非常边缘的方式发生。

I think right now, it's happening in a very peripheral way.

Speaker 3

研究人员的节奏被加快了。

Researchers are being sped up.

Speaker 3

AI系统正在运行不同的实验。

Different experiments are being run by the AI system.

Speaker 3

弄清楚你是否完全闭合了这个循环,将极其重要。

It would be extremely important to know if you're fully closing that loop.

Speaker 3

我认为我们实际上还有一些技术工作要做,那就是构建方法来监控我们内部的开发环境,以便观察趋势的变化。

And I think that we actually have some technical work to do to build ways of instrumenting our internal development environment so that we can see trends over time.

Speaker 3

我担心吗?

Am I worried?

Speaker 3

我读过你读过的那些内容,当事情开始出错时,这正是故事的关键转折点。

I have read the same things that you have read, and this is the pivotal point in the story when things begin to go awry, if things do.

Speaker 3

当我们对此有更完善的数据时,我们会指出这一趋势。

We will call out this trend as we have better data on it.

Speaker 3

我认为这是一个需要格外谨慎对待的领域,因为很容易出现将太多事情交由系统处理的情况,一旦系统出错,错误会迅速累积并失控。

And I think that this is an area to tread with, like, extraordinary caution because it's very easy to see how you delegate so many things to the system that if the system goes wrong, the wrongness compounds very quickly and gets away from you.

Speaker 1

但让我一直感到担忧的是,每个人都心知肚明。

But the thing that always strikes me and has always struck me as being dangerous about this is everybody knows.

Speaker 1

如果我问任何一家公司的员工是否希望在这里保持谨慎,他们都会告诉我他们确实希望如此。

And if I ask a member of any of the companies whether or not they wanna be cautious here, they will tell me they they do.

Speaker 1

但另一方面,这几乎是他们彼此之间唯一的竞争优势。

On the other hand, it is their almost only advantage over each other.

Speaker 1

你们刚刚撤销了OpenAI使用Cloud Code的权限,因为据我所知,你们认为它确实能提升你们的效率,而你们不希望它也提升他们的效率。

And you all just revoked OpenAI's ability to use Cloud Code because, as best I can tell, you think it is genuinely speeding you up, and you don't want it to speed them up.

Speaker 1

在这里,各种力量的权重和影响力之间存在某种微妙的平衡,我相信你们都清楚自己正在玩弄什么,同时又面临着极其强烈的抢先动机。

There is something here between the weight of the forces, the power of the forces that I think you all know you're playing with, and the very, very, very strong incentives to be first.

Speaker 1

我完全可以想象自己身处Anthropic内部时会想:与其让OpenAI领先,不如让我们领先?

And I I can I can really imagine being inside Anthropic and thinking, well, better us in OpenAI?

Speaker 1

让我们领先,而不是让Alphabet、谷歌领先。

Better us than Alphabet, Google.

Speaker 1

让我们领先,而不是让中国领先。

Better us than China.

Speaker 1

而这正是一个非常充分的理由,不能放慢脚步。

And that being a very strong reason to not slow down.

Speaker 1

我甚至不确定这个问题你是否能回答,但你是如何平衡这一点的呢?

I don't even know that this is a question I believe you can answer, but how do you balance that?

Speaker 3

嗯,也许我这里有个答案。

Well, maybe I have something of an answer here.

Speaker 3

今天,我们的系统以及其他公司的系统都会接受第三方的测试,包括政府机构对国家安全特性、生物武器、网络攻击等方面进行评估。

Today, our systems and the other systems from other companies are tested by third parties, including parts of government for national security properties, biological weapons, cyber offense, other things.

Speaker 3

这显然是一个世界需要了解是否正在发生的问题领域。

It's clearly a problem area where the world needs to know if this is happening.

Speaker 3

我认为,如果你随机调查街头任何一个人,向他们解释什么是递归自我改进后,问他们是否认为人工智能公司应该被允许这样做,

And you almost certainly, I think if you polled any person on the streets and said, do you think AI companies should be allowed to do, like, recursive self improvement after explaining what that was?

Speaker 3

他们几乎肯定会说:不。

Without checking with anyone, they would say, no.

Speaker 3

这听起来风险很大。

That sounds sounds pretty risky.

Speaker 3

我是希望有一些形式的监管的。

Like, I would like there to be some form of regulation.

Speaker 1

但他们可能根本不会实施,或者即使实施也不会那么严格。

But they're probably either won't be or it won't be that strong.

Speaker 1

我的意思是,当我跟你们这些AI公司的高层交谈时,这有时让我很沮丧——你们竟然会冒出一种极其天真的‘机械降神’式监管幻想,而实际上你们都清楚监管环境是什么样子。

I mean, this actually sometimes frustrates me when I talk to all of you at the top of the AI companies, which is the emergence of, like, a very naive deus ex machina of regulation, where you all know what the regulatory landscape looks like.

Speaker 1

目前,大讨论在于我们是否要完全 preempt 各州的AI监管,而你们也知道事情进展有多慢。

Right now, the big debate is whether or not we're gonna completely preempt any state AI regulation, and you know how slowly things move.

Speaker 1

国会到目前为止对此根本没有通过任何重大法案,是的。

There has been nothing major passed by congress on this at all Yep.

Speaker 1

我会这么说。

I would say.

Speaker 1

建立一个所有不同实验室都认可的独立测试与评估体系,会很难。

And setting up some kind of independent testing and evaluation system that all the different labs buy into, it would be hard.

Speaker 1

这会很复杂。

It would be complicated.

Speaker 1

考虑到人们推进的速度以及系统已经表现出的种种异常行为,即使你能以极快的速度制定出正确的政策,关于测试能否在快速自我改进的系统中发现所有你希望发现的问题,仍然是一个悬而未决的问题。

And it is given how fast people are moving and how strange the behaviors the systems are already exhibiting are, even if you could get the policy right at a high speed, the question of whether or not the testing would be capable of finding everything you want on a rapidly self improving system is a very open question.

Speaker 3

我于2021年与英国的合著者杰斯·惠特尔斯顿共同撰写了一篇研究报告,题为《政府为何及如何监控人工智能的发展》。

I wrote a research paper in 2021 called how and why governments should monitor AI development with, my co author Jess Whittleston in England.

Speaker 3

我认为这里并不在强调因果关系,但就在那篇论文发表后的两年内,美国和英国都成立了人工智能安全研究所,开始对实验室的成果进行测试和大致监控。

And I think I'm not attributing a causal factor here, but within two years of that paper, we had the AI safety institutes in The US and UK testing things from the labs, roughly monitoring some of these things.

Speaker 3

所以,我们确实能做成这件困难的事。

So we we can do this hard thing.

Speaker 3

这件事已经在某个领域发生了。

It has already happened in one domain.

Speaker 3

我并不是在依赖某种看不见的、神秘的外部力量。

And I'm not relying on some, like, invisible big other force here.

Speaker 3

我更想说的是,公司们已经开始在自己的系统中对这些问题进行测试和监控。

I'm more saying that companies are starting to test for this and monitor for this in their own systems.

Speaker 3

仅仅拥有一个非监管性质的外部测试,来验证你是否真正进行了相关检测,就已极具帮助。

Just having a nonregulatory external test of whether you truly are testing for that is is extremely helpful.

Speaker 1

你觉得我们在测试方面足够好吗?

And do you think we're good enough at the testing?

Speaker 1

我的意思是,我持怀疑态度的原因并不是我认为我们无法建立一个声称是测试的系统。

I mean, I think one reason I am skeptical is not that I don't think we can set up something that claims to be a test.

Speaker 1

正如你所说,我们已经做到了。

As you say, we have done that already.

Speaker 1

问题是,投入到测试中的资源与投入到加速这些系统中的资源相比,差距太大了;而且,我已经读到Anthropic的报告,说Claude可能知道自己正在被测试,并据此调整行为。

It is that the resources going into that compared to the resources going into speeding these systems and already, I am reading anthropic reports that Claude maybe knows when it's being tested and alters its behavior accordingly.

Speaker 1

所以,在一个越来越多代码由Claude编写、而越来越少被人理解的世界里。

So a world where more of the code is being written by Claude and less of it is being understood.

Speaker 1

我只是清楚资源流向了哪里。

I just know where the resources are going.

Speaker 1

它们似乎并没有

They don't seem to be

Speaker 3

流向测试这一侧。

going into the testing side.

Speaker 3

我看到我们在短短两年到两年半的时间里,从零发展到了一个大多数人认为有效的生物武器测试体系。

I've seen us go from zero to having what I think people generally feel is an effective bioweapon testing regime in maybe two years, two and a half.

Speaker 3

所以这是可以做到的。

So it can be done.

Speaker 3

这非常困难,但我们已经有了一个实证。

It's really hard, but we have a proof point.

Speaker 3

因此我认为我们能够达到这一目标,今年你们会听到我们更多地谈论我们如何开始构建针对这一领域的监控和测试机制。

So I think that we can get there, and you should expect us to kind of speak more about this this year about precisely how we're starting to try and build monitoring and testing things for this.

Speaker 3

我认为,在我们发现的内容方面,我们和其他AI公司需要变得更加公开透明。

And I think this is an area where we and the other AI companies will need to be, significantly more more public about what we're finding.

Speaker 3

我们并不是不公开信息。

We're not we're not not being public now.

Speaker 3

这些内容都写在模型卡片里,你们完全可以阅读到。

It's in the model cards and things that you can really read.

Speaker 3

但显然,人们已经开始阅读这些内容并说:等等。

But clearly, people are starting to read this and say, hang on.

Speaker 3

这看起来相当令人担忧,他们正期待我们提供更多的数据。

This looks like quite concerning, and they are looking to us to produce more data.

Speaker 1

现在我想回到入门级工作的问题上。

I wanna go back now to the entry level jobs question.

Speaker 1

你们的首席执行官达里奥·阿马代认为,人工智能在今后几年内可能取代一半的入门级白领工作。

Your CEO, Dario Amade, has said that he thinks AI could displace half of all entry level white collar jobs in the next couple of years.

Speaker 1

当我看到相关报道时,总觉得人们忽略了其中'入门级'这个关键词。

I always think that the the people sort of miss the entry level language there when I see it reported on.

Speaker 1

但首先,你同意这个观点吗?

But first, do you agree with that?

Speaker 1

Do you

Speaker 3

你是否担心,在未来几年内,一半的入门级白领工作会被取代?

worry that half of all entry level white collar jobs can be replaced in the next couple of years?

Speaker 3

我相信这项技术将渗透到广泛的知识经济领域,并影响大多数入门级工作。

I believe that this technology is gonna make its way into the broad knowledge economy, and it will touch the majority of entry level jobs.

Speaker 3

这些工作是否真的会发生变化,是一个更加微妙的问题,从数据中并不明显。

Whether those jobs actually change is a much more, like, subtle question, and it's not obvious from the data.

Speaker 3

比如,我们或许能看到毕业生招聘放缓的迹象。

Like, we maybe see the hints of a slowdown in graduate hiring, maybe.

Speaker 3

如果你看看目前出炉的一些数据,我们或许能看到生产率飙升的迹象,但这还非常非常早期,很难下定论。

If you look at some of the data coming out right now, we maybe see the signatures of a productivity boom, but it's very, very early and it's hard to be definitive.

Speaker 3

但我们确知,所有这些工作都会发生变化。

But we do know that all of these jobs will change.

Speaker 3

所有入门级岗位最终都会改变,因为人工智能让某些事情成为可能,这将改变公司的招聘计划。

All of the entry level jobs are eventually going to change because AI has made certain things possible, and it's going to change for hiring plans of companies.

Speaker 3

因此,作为一群体,你可能会看到入门级岗位的职位空缺减少。

So as a cohort, you might see fewer job openings for entry level jobs.

Speaker 3

这将是所有这些变化中一种天真的预期。

That would be one naive expectation out of all of this.

Speaker 1

但让我们谈谈,这甚至可能根本不是一种天真的预期。

But let's talk about that maybe not even being a naive expectation.

Speaker 1

你说在Anthropic,你已经看到这种情况正在发生

You say it's already happening at Anthropic that what you're seeing

Speaker 3

我看到我们的偏好正在发生变化。

I'm seeing us shift our preference.

Speaker 3

没错。

Exactly.

Speaker 3

我猜测,其他地方也会发生类似的情况。

And I I my guess is that that would be happening elsewhere.

Speaker 1

就我们目前的情况而言,就连我使用这些系统的方式来看,我认为Claude、ChatGPT、Gemini或其他任何系统都很少能超越某个领域中最优秀的人。

And and where we are right now, I mean, even in the way I use some of these systems, it is rare, I think, that Claude or ChatGPT or Gemini or any of the other systems is better than the best person in a field.

Speaker 1

嗯。

Mhmm.

Speaker 1

它们通常还没有突破这一点,而且还有很多事情它们做不到。

It has not typically breached that, and there's all kinds of things they can't do.

Speaker 1

但它们比普通大学毕业生更强吗?嗯。

But are they better than your median college graduate Mhmm.

Speaker 1

在很多方面吗?

At a lot of things?

Speaker 1

是的。

Yeah.

Speaker 1

它们确实可以。

They are.

Speaker 1

在一个你需要更少普通大学毕业生的世界里,我看到人们争论的一个问题是,这些系统目前是否能胜过一般或替代水平的工作。

And in a world where you need fewer of your median college graduates, One thing I've seen people arguing about is whether these systems at this point can do better than sort of average or replacement level work.

Speaker 1

但我每次看到这种说法都会非常担忧,因为一旦我们接受它们能胜任一般或替代水平的工作,那么根据定义,大多数工作和大多数从事这些工作的人都是普通的。

But I always really worry when I see that because once we have accepted they can do average or placement level work, well, by definition, most of the work done and most of the people doing it Is average.

Speaker 1

是普通的。

Is average.

Speaker 1

对吧?

Right?

Speaker 1

最优秀的人只是例外。

The best people are the exceptions.

Speaker 1

而且,人们变得更好的方式是通过那些能让他们学习的工作。

And, also, the way people become better is that they have jobs where they learn.

Speaker 1

是的。

Mhmm.

Speaker 1

我的职业生涯中花了很多时间招聘年轻的记者。

I mean, I have spent a lot of time hiring young journalists over my career.

Speaker 1

当你从大学招聘新人时,在某种程度上,你是在根据他们当时的文章和工作表现来雇佣他们。

And when you hire people out of college, to some degree, you're hiring them for their possible articles and work at that exact moment.

Speaker 1

是的。

Mhmm.

Speaker 1

但某种程度上,你也在对他们进行投资,相信他们的能力会随着时间推移不断提升,最终带来回报。

But to some degree, you're making a investment in them that you think will only pay off over time as they get better and better and better.

Speaker 1

所以,这样一个可能对初级岗位产生真正影响的世界,对我来说并不遥远,它引发了许多深刻的问题:如何提升整体人口的技能?未来如何培养出胜任高级职位的人才?以及人们在成长过程中错过了哪些关键学习?

So this world where you have a a potential real impact on entry level jobs, and that that world does not feel far away to me, seems to me to have really profound questions it is raising about the upskilling of the population, how you end up with people for senior level jobs down the road, what people aren't learning along the way.

Speaker 3

我们观察到,有一类年轻人已经多年沉浸并深度接触人工智能了。

And one thing we see is that there is a certain type of young person that has just lived and breathed AI for several years now.

Speaker 3

我们雇佣他们。

We hire them.

Speaker 3

他们非常优秀,能够以完全新颖的方式思考如何让Claude为他们工作。

They're excellent and they think in entirely new ways about basically how to get Claude to work for them.

Speaker 3

这就像那些在互联网环境中成长起来的孩子。

It's like kids who grew up on the Internet.

Speaker 3

他们自然而然地精通这一点,而他们所加入的许多组织中的其他人却并不具备这种能力。

They they were naturally versed in it in a way that many people in the organizations they were coming into weren't.

Speaker 3

因此,学会如何培养这种基本的实验思维和对这些系统的探索精神,并加以鼓励,将变得至关重要。

So figuring out how to teach that basic experimental mindset and curiosity about these systems and to encourage it is going to be really important.

Speaker 3

那些花大量时间玩味这些技术的人,会发展出非常宝贵的直觉,进入组织后能够变得极其高效。

People that spend a lot of time playing around with this stuff will develop very valuable intuitions, and they will come into organizations and be able to be extremely productive.

Speaker 3

同时,我们还得弄清楚,我们希望培养哪些手工技能,或许可以借鉴行会式的理念来维持人类的卓越表现,以及组织如何选择传授这些技能。

At the same time, we're gonna have to figure out what artisanal skills we want to almost develop maybe a guild style philosophy of maintaining human excellence in and how organizations choose how to teach those skills.

Speaker 3

好的。

Okay.

Speaker 3

那那些处于中间位置的人怎么办呢?

Then what about all those people in the middle of that?

Speaker 3

在硅谷以外的现实经济中,事情进展得很缓慢。

Things move slowly in the real economy outside Silicon Valley.

Speaker 3

我认为我们常常看到软件工程,就以为它能代表整个经济的运行方式,但事实往往并非如此。

I think that we often look at software engineering and think that this is a proxy for how the rest of the economy works, but it's often not.

Speaker 3

这常常是一种不恰当的类比。

It's often a disanalogy.

Speaker 3

组织会把人员调配到那些人工智能系统尚未发挥作用的地方。

Organizations will move people around to where the AI systems don't yet work.

Speaker 3

我认为,就业结构不会立即发生巨大的变化。

And I think that you won't see vast immediate changes in the makeup of employment.

Speaker 3

但你会看到人们被要求从事的工作类型发生显著变化。

But you will see significant changes in the types of work people are being asked to do.

Speaker 3

而那些最擅长调动人员的组织将会变得极其高效。

And the organizations which are best at sort of moving their people around are going to be extremely effective.

Speaker 3

而那些做不到的组织,最终可能不得不做出一些极其艰难的决定,比如裁员。

And ones that don't may end up having to make, like, really, really hard decisions involving involving laying off workers.

Speaker 3

AI技术的不同之处在于,它可能比以往任何技术都发生得快得多。

The difference with this AI stuff is it maybe happens a lot faster than previous technologies.

Speaker 3

我认为,包括Anthropic在内的许多人对这一问题的焦虑在于:这种速度会不会带来根本性的变化?

And I think many of the anxieties people might have about this, including at Anthropic, is is the speed of this going to make all of this difference?

Speaker 3

它是否引入了我们此前从未遇到过的全新问题?

Does it introduce sheer points that we haven't encountered before?

Speaker 1

如果你要打个赌,三年后大学毕业生的失业率会和现在一样吗?

If you had to bet three years from now, is the unemployment rate for college graduates, is it the same as it is now?

Speaker 1

是更高了,还是更低了?

Is it higher or is it lower?

Speaker 3

我猜会略高一些,但不会高太多。

I would guess it is higher, but not by much.

Speaker 3

我的意思是,今天有一些专业领域,AI已经介入并彻底改变了其就业市场的结构,可能对那些拥有这些专业技能的人不利。

And what I mean by that is there will be some disciplines today which actually AI has come in and and completely changed and completely changed the structure of that employment market, maybe in a way that's adverse to to people that have that specialism.

Speaker 3

但总的来说,我认为三年后,人工智能将推动整个经济实现显著增长。

But mostly, I think three years from now, AI will have driven a pretty tremendous growth in the entire economy.

Speaker 3

因此,你会看到许多新的工作类型因这一变化而出现,而这些目前还无法预测。

And so you're going to see lots of new types of jobs that show up as a consequence of this that we can't yet predict.

Speaker 3

我预计,毕业生们会大量涌入这些新领域。

And you will see graduates kind of flood into that, I I expect.

Speaker 3

你有

Do you have

Speaker 1

我知道你无法预测这些新工作,但如果你要猜测它们可能是什么样子,会是什么?

a I know you can't predict those new jobs, but if you had to guess what some of them might look like.

Speaker 3

我的意思是,其中一个现象就是微型企业家的兴起。

I mean, one thing is just the phenomenon of the the kind of micro entrepreneur.

Speaker 3

现在,你可以通过很多种方式在线创业,而人工智能系统能为你完成大量工作,使这一切变得极其容易。

I mean, there are lots and lots of ways that you can start businesses online now, which are just made massively easier by having the AI systems do it for you.

Speaker 3

你不需要雇用一大群人来帮你完成启动企业所需的繁重琐碎工作。

And you don't need to hire a whole load of people to help you do the huge amount of schlep work that involves getting a business off the ground.

Speaker 3

更重要的是,如果你有一个清晰的想法和明确的商业愿景,那么现在正是创业的最佳时机,你只需极少的成本就能迅速起步。

It's more a case of if you're a person with a clear idea and a clear vision of something to do a business in, it's now the best time ever to start a business and you can get up and running for pennies on the dollar.

Speaker 3

我预计我们会看到大量具有这种特征的事物涌现。

I expect we'll see tons and tons and tons of stuff that has that nature to it.

Speaker 3

我还预计会出现一种你可能称之为‘AI对AI经济’的现象,即AI代理和AI企业之间将相互开展业务。

I also expect that we're gonna see the emergence of what you might think of as the AI to AI economy, where AI agents and AI businesses will be doing business with one another.

Speaker 3

我们会看到一些人找到了通过这种模式获利的方法,形式是各种奇特的新组织。

And we'll have people that have figured out ways to basically profit off of that in the forms of strange new organizations.

Speaker 3

比如,一家专门从事AI对AI法律合同的公司会是什么样子?

Like, what would it look like to have a firm which specializes in AI to AI legal contracts?

Speaker 3

因为我相信,今天你就能找到一些富有创意的方式来创办这样的企业。

Because I bet you there's a way that you can figure out creative ways to start that business today.

Speaker 3

会有大量类似性质的事物出现。

There'll be a lot of stuff of that flavor.

Speaker 1

我所担心且认为最有可能发生的情况是:如果你告诉我,Anthropic将在一年后发布Claude Plus。

So the version of this that I both worry about and think to be the likeliest, if you told me what was gonna happen was that Anthropic was gonna release Claude plus in a year.

Speaker 1

而Claude Plus会成为一个完全成型的同事,能够端到端地模仿众多职业的技能,甚至达到高管级别。

And Claude plus is somehow a fully formed coworker, and it can mimic end to end the skills of a lot of different professions up to the C suite level.

Speaker 1

是的。

Mhmm.

Speaker 1

而且这一切会同时发生,并给企业带来巨大的即时压力,迫使它们裁员以保持竞争力。

And it's gonna happen all at once, and it's gonna create tremendous all at once pressure for businesses to downsize, to remain competitive with each other.

Speaker 1

从政策层面来看,如果这种颠覆性变化像COVID那样一次性大规模爆发,所有人都居家,我反而没那么担心。

At a policy level, the fact that that would be so disruptive in that big bang, everybody stays home because of COVID style way, it worries me less.

Speaker 1

因为当事情成为紧急状况时,我们会做出反应。

Because when things are emergencies, we respond.

Speaker 3

我们确实会制定政策。

We actually do policy.

Speaker 3

但如果你告诉我

But if you told

Speaker 1

说市场营销专业毕业生的失业率会上升175%到300%,但即便如此,失业率依然不会太高。

me that what's gonna happen is that the unemployment rate for marketing graduates is going to go up by, you know, a 175%, 300% to still not be that high.

Speaker 1

我的意思是,在大衰退期间,整体就业率最高时也不过在百分之九左右。

I mean, the overall employment rate during the Great Recession topped her out, you know, in the nine ish percentile range.

Speaker 1

所以,你可以经历很大的动荡,而不会让一半的人失业。

So you can have a lot of disruption without having 50% of people thrown out of work.

Speaker 1

对吧?

Right?

Speaker 1

如果你有10%、15%,我的意思是,这已经非常高了。

If you have 10%, 15, I mean, that's very, very, very high.

Speaker 1

但还不算太高。

But it's not so high.

Speaker 1

如果这种变化只发生在几个行业,而且受影响的是毕业生,而不是整个行业的每个人,那也许只是你不够优秀。

And if it's only happening in a couple of industries at a time and it's grads, not everybody in the industry being thrown out of work, well, maybe it's just that you're not good enough.

Speaker 1

是的。

Yep.

Speaker 1

对吧?

Right?

Speaker 1

你知道,那些顶尖的、非常优秀的毕业生仍然能找到工作。

You know, the the superstars, really good graduates are still getting jobs.

Speaker 1

你应该更努力一些。

You should have worked hard.

Speaker 1

你应该去一所更好的学校。

Should have gone to a better school.

Speaker 1

我担心的是,我们对这种类型的失业反应不佳,比如来自中国的那种失业,这种失业似乎更有可能发生,因为它分布不均,而且发生的速度让我们还能把个人的命运归咎于他们自己。

And one of my worries is that we don't respond to that kind of job displacement well, right, which is the kind of job displacement we got from China, which is the kind of job displacement that seems likelier because it's uneven, and it's happening at a rate where we can still blame people for their own fortunes.

Speaker 1

我想知道你是怎么看待这个说法的。

I'm curious how you think about that story.

Speaker 1

我觉得

I think

Speaker 3

默认的结果就像你描述的那样,但如何达到这个结果其实是一种选择,我们可以做出不同的选择。

the default outcome is something like what you described, but getting there is actually a choice, and we can make different choices.

Speaker 3

我们发布Anthropic经济指数的全部目的,就是为了获得能够关联到具体职业、关联到经济中真实工作岗位的数据。

The whole purpose of what we release in the form of the Anthropic Economic Index is the ability to have data that ties to occupations, that tie to real jobs in the economy.

Speaker 3

我们这样做是有意为之,因为这能随着时间推移绘制出人工智能如何逐步渗透到不同工作岗位的图景,并帮助Anthropic之外的经济学家将这些信息串联起来。

We do that very intentionally because it is building a map over time of how this AI is making its way into different jobs and will empower economists outside Anthropic to tie it together.

Speaker 3

我相信,如果我们能对工作变动或转型的原因做出更有实证依据的主张,政策上就可以做出不同的选择。

I believe that we can choose different things in policy if we can make much more well evidenced claims about what the cause of a job disruption or change is.

Speaker 3

我们面临的挑战是,能否充分描述这个新兴的人工智能经济,使其变得如此清晰明确?

And the challenge in front of us is can we characterize this emerging AI economy well enough that we can make this extremely stark?

Speaker 3

然后,我认为我们才能真正就这个问题展开政策讨论。

And then I think that we can actually have a policy discussion about it.

Speaker 1

那我们来谈谈政策讨论吧。

Well, let's talk about the policy discussion.

Speaker 1

是的。

Yep.

Speaker 1

我之所以特别邀请你来,其中一个原因是你在OpenAI做过政策工作。

One reason I wanted to have you in particular on is you did policy at OpenAI.

Speaker 1

是的。

Yep.

Speaker 1

你在Anthropix做政策工作。

You do policy at Anthropix.

Speaker 1

你一直参与这些政策辩论。

You've been around these policy debates for a long time.

Speaker 1

你长期以来一直在你的通讯中追踪模型的能力。

You've been tracking model capabilities in your newsletter for a long time.

Speaker 1

我的感觉是,我们已经就人工智能展开了多年的讨论,是的。

My perception is we are many, many years into the debate about AI and Mhmm.

Speaker 1

早在ChatGPT出现之前很久,就已经有多年的讨论了,是的。

Many, many years dating far before ChatGPT Mhmm.

Speaker 1

关于人工智能和就业,我们已经在阿斯彭和其他地方举办了无数会议,讨论我们该如何应对。

Of there being conferences at Aspen and everywhere else about, you know, what are we gonna do about AI and jobs?

Speaker 1

但不知为何,我仍然几乎没有看到任何可操作的政策,如果我刚刚描述的情况真的出现——即大量行业的初级岗位突然变得极其难找,是的。

And somehow, I still see almost no policy that seems to me to be actionable if the situation I just described begins showing up, where all of a sudden entry level jobs are getting much harder to come by across a large range of industries all at once Mhmm.

Speaker 1

以至于经济无法将这些市场营销专业的毕业生重新分配到数据中心建设、护士或其他行业。

Such that the economy cannot reshift all these marketing majors into data center construction or nurses or something.

Speaker 1

是的。

Mhmm.

Speaker 1

所以,好吧,你比我对这个话题的参与更深。

So, okay, you've been deeper in this conversation than I've been.

Speaker 1

当你说到我们可以就这个问题展开政策讨论时,其实我们已经在进行政策讨论了。

When you say we can have a policy conversation about that we've been having a policy conversation.

Speaker 1

我们有具体的政策吗?

Do we have policy?

Speaker 3

我们对人工智能对经济和就业的影响感到普遍的焦虑。

We have generalized anxiety about the effect of AI on the economy and on jobs.

Speaker 3

但我们没有清晰的政策构想。

We don't have clear policy ideas.

Speaker 3

部分原因是,民选官员并不单纯或主要被高层次的政策讨论所驱动。

Part of that is that elected officials are not moved solely or mostly by the high level policy conversation.

Speaker 3

他们更关注选民的实际遭遇。

They're moved by what happens to their constituents.

Speaker 3

就在几个月前,我们才刚刚能够提供各州层面的经济指数数据,而现在你们可以开始进行政策层面的讨论了。

Only a few months ago were we able to produce state level views for our economic index, And now you can start having the policy conversation.

Speaker 3

我们曾与民选官员讨论过这个问题,现在我们可以这样说:哦,您来自印第安纳州。

And we've had this with elected officials where now we can say, oh, you're from you're from Indiana.

Speaker 3

比如,这是您州内人工智能的主要应用领域,我们可以将其与主要的就业来源结合起来。

Like, here's for, like, major uses of AI in your state, and we can join it with major sources of employment.

Speaker 3

我们开始看到,这能激发他们的行动,因为这将政策与他们的选民联系起来,而选民会追问政客:你做了什么?

And what we're starting to see is that activates them because it makes it tied to their constituents who are going to tie it to the the politician of what did you do.

Speaker 3

现在,针对这个问题,你需要采取一种多层次的应对措施,从延长失业救济(特别是针对那些我们知道将受到最严重影响的职业)到考虑学徒计划等措施。

Now, what you do about this is going to need to be an extremely kind of multilayered response ranging from extending unemployment for especially occupations that we know are going to be hardest hit to thinking about things like apprenticeship programs.

Speaker 3

随着情况变得越来越严重,你可能还需要扩展到更大规模的社会项目,比如补贴那些你希望人们转向的经济领域中的工作岗位——而这些措施只有在经历显著经济增长所带来的富足时才可能实现。

And then as the scenarios get more and more significant, you may extend to much larger social programs or things like subsidizing jobs in the parts of the economy where you want to move people to, that you're only able to do if you experience the kind of abundance that comes from significant economic growth.

Speaker 3

但经济增长本身可能有助于解决一些其他政策挑战,因为它能为你可以采取的措施提供资金支持。

But the economic growth may help solve some of these other policy challenges by funding some of the things you can do.

Speaker 1

我总是觉得这个答案令人沮丧。

I always find this answer depressing.

Speaker 1

我实话跟你说吧。

I'm gonna be honest.

Speaker 1

失业是一件非常糟糕的事。

Unemployment is a terrible thing to be on.

Speaker 1

我们需要这个项目,但领取失业救济的人并不为此感到开心。

It's a program we need, but people on unemployment are not happy about it.

Speaker 1

是的。

Mhmm.

Speaker 1

这也不是任何人的长期解决方案。

And it's not a good long term solution for anybody.

Speaker 1

学徒再培训项目,其成效并不理想。

Apprentice retraining programs, they don't have great track records.

Speaker 1

我们在帮助人们从被外包的制造业岗位中再就业方面做得并不好。

We were not good at retraining people out of having their manufacturing jobs outsourced.

Speaker 1

我不是说我们不可能在这方面变得更好,但我们必须迅速提升能力。

I'm not saying it is conceptually impossible that we could get better at it, but we would need to get better at it fast.

Speaker 1

嗯。

Mhmm.

Speaker 1

而且我们还没有投入足够的练习、实验、制度建设或能力提升来实现这一点。

And we have not been putting in the reps or the experimentation or the institution or capacity building to do that.

Speaker 1

而关于大规模社会保险改革这个更广泛的问题,我觉得这对我来说很难。

And the the broader question of big social insurance changes doesn't seem I mean, that seems tough to me.

Speaker 3

我想就这一点再深入一下。

I wanna push on this

Speaker 1

请说。

Please.

Speaker 3

就在我们明知有一种干预措施,能比其他任何方法都更有效地帮助人们应对经济变化时。

Just a bit where we know that there is one intervention that helps people dealing with, like, a changing economy more than almost anything else.

Speaker 3

那就是时间。

It is just time.

Speaker 3

给人们一些时间,去找到本行业的工作,或者找到一份互补的工作。

Giving the person time to find either a job in their industry or to find a job that's complementary.

Speaker 3

如果人们没有时间,他们就会接受更低工资的工作。

If people don't have time, they take lower wage jobs.

Speaker 3

他们会脱离自己原本所处的经济阶层,跌落下去。

They fall out of their whatever economic rung they're on, they fall down at.

Speaker 3

能够为人们提供寻找工作时间的政策干预,我认为是一种非常有效的措施,并且在政策制定中有很多可以调整的细节。

Policy interventions that can just give people time to search is, I think, a a robustly useful intervention and one where there are many, dials to turn in a policy making sense that you can use.

Speaker 3

我认为这一点得到了大量经济学文献的有力支持。

And I think this is just well supported by lots of the economic literature.

Speaker 3

所以我们已经有了这一点。

So we have that.

Speaker 3

如果我们陷入像你提到的那些更极端的情境中,我认为这将促使我们展开关于如何应对这项技术的更大规模的国家讨论,而这种讨论已经开始出现了。

Now if we end up in a more extreme scenario like some of the ones that you're talking about, I think that will just bring us to the larger national conversation of what to do about this technology, which which is beginning to happen.

Speaker 3

如果你看看各州以及州一级涌现的立法浪潮,是的,并非所有政策都恰好是正确的应对方式,但它们确实表明了人们希望就这一问题展开更系统、更连贯的讨论。

If you look at the states and the flurry of legislation at the at the state level, yes, not all of it is like the exactly the right policy response, but it is indicative of a a desire for there to be some larger coherent conversation about this.

Speaker 1

我认为,用‘时间’来描述这个问题非常恰当,因为我同意你的观点。

Well, I think time is a really good way of describing what the question is because I agree with you.

Speaker 1

我的意思是,当我说失业保险不是一个很好的项目时,我不是说人们不需要它。

I mean, when I say unemployment insurance isn't a great program to be on, I don't mean people don't need to be on it.

Speaker 1

对。

Yep.

Speaker 1

我的意思是,他们希望离开它。

I mean, they wanna get off of it.

Speaker 3

当然。

Absolutely.

Speaker 1

因为人们想要来自工作的钱。

Because people for they want money from jobs.

Speaker 1

他们想要尊严。

They want dignity.

Speaker 1

他们想要与他人相处。

They want to be around other human beings.

Speaker 1

通常,当你帮助人们争取时间时,你是在帮助他们度过一段有明确期限的中断。

Usually, what you're doing when you are helping people buy time is you're helping them wait out a time delimited disruption.

Speaker 1

嗯嗯。

Mhmm.

Speaker 1

并不总是如此。

Not always.

Speaker 1

对吧?

Right?

Speaker 1

中国冲击并不完全如此,但你预期它会过去,然后市场会恢复正常。

The China shock wasn't exactly like that, but that you expect to pass, and then the the market is sort of normal.

Speaker 1

在这种情况下,你拥有的是一种技术,如果希望发生的事情发生了,这种技术正在加速。

In this case, what you have is a technology that if what you wanna have happen happens, the technology is accelerating.

Speaker 1

嗯嗯。

Mhmm.

Speaker 1

所以,你现在看到的是三种不同的速度在同时发生。

So what you have is, like, three different speeds happening here.

Speaker 1

你有个人适应的速度。

You have the speed at which individual people can adjust.

Speaker 1

我能在多快的时间里学会新技能,适应新世界,掌握人工智能,不管它是什么?

How fast can I learn new skills, figure out a new world, learn AI, whatever it might be?

Speaker 1

你有AI系统的速度,几年前它们还无法胜任优秀大学毕业生的工作,还有政策的速度。

You have the speed at which the AI systems, which a couple of years ago, were not capable of doing the work of a median college grad from a good school, and you have the speed of policy.

Speaker 1

AI系统变得越来越好、能做更多事情的速度非常快。

And the speed at which the AI systems are getting better and able to do more things is quite fast.

Speaker 1

我的意思是,你比我能更深刻地感受到这一点,但我发现连跟进都很难,因为你知道,三个月内就会有新东西出现,彻底改变可能实现的事情。

I mean, that is you you experience this more than I do, but I find it hard to even cover this because, you know, within three months, something else will have come out that has significantly changed what is possible.

Speaker 3

我最近当了爸爸,休完陪产假回来后,发现我们搭建的新系统。

I had a baby recently and came back from paternity leave to the new systems we built.

Speaker 3

我大吃一惊。

Was deeply surprised.

Speaker 1

个体人类的适应速度比这慢得多,而政策和政府机构的调整速度比个人还要慢得多。

Individual humans are moving more slowly than that, and policy and government institutions move a lot more slowly than individual human beings.

Speaker 1

因此,通常的干预方式是,时间站在劳动者这一边,就像你所说的那样。

And so, typically, the the intervention is that time favors the worker, as you're saying.

Speaker 1

而在这里,将帮助劳动者。

And here, will help the worker.

Speaker 1

但我认为令人担忧的问题是,时间是否实际上只是为混乱加剧争取了更多时间。

But I think the scary question is whether time just actually creates time for the disruption to get worse.

Speaker 1

你知道的。

You know?

Speaker 1

也许你原本打算转向数据中心建设,但事实上,我们现在不需要那么多的数据中心建设了。

Maybe you wanted to move over to data center construction, but, actually, now we don't need as much data center construct.

Speaker 1

对吧?

Right?

Speaker 1

你可以这样理解。

Like, you can think of it like that.

Speaker 3

我的意思是,在你所描述的情况下,经济将变得极度繁荣。

I mean, under the situation you're describing, the economy will be running extremely hot.

Speaker 3

这些AI系统将产生巨大的经济活动。

Huge amounts of economic activity will be generated by these AI systems.

Speaker 3

在大多数这种情况下,我认为GDP不会保持不变或萎缩。

And under most scenarios where this is happening, I don't think you're going to be seeing GDP stay the same or or shrink.

Speaker 3

对吧?

Right?

Speaker 3

它会显著增长。

It's going to be getting substantially larger.

Speaker 3

我认为,我们在西方已经很久没有经历过重大的GDP增长了,因此我们渐渐忘记了这种增长在政策制定上能带来什么。

I think we just haven't experienced major GDP growth in the West in a long time, and we sort of forget what that affords you in a policy making sense.

Speaker 3

我认为,我们可以开展许多大型项目来创造新的工作岗位,但这需要经济实现如此深远的增长,才能为这些项目腾出空间。

I think that there are huge projects that we could do that would allow you to create new types of jobs, but it requires the economic growth to be so kind of profoundly large that it creates space to do those projects.

Speaker 3

而且,就像你对丰裕运动的研究非常熟悉的一样,这还需要社会有意愿相信我们能够建设,并且愿意去建设。

And, you know, as you're deeply familiar with with with your work on on the abundance movement, it requires for, like, social will to believe that we can build stuff and to want to build stuff.

Speaker 3

但我认为这两点都可能会随之而来。

But I think both of those things might come along.

Speaker 3

我认为,我们可能会进入一个相当令人兴奋的场景:由于经济的大幅增长,我们得以选择如何分配社会中的巨大资源。

I think that we could end up being in a pretty exciting scenario where we get to choose how to allocate, like, great efforts in in society due to this large amount of economic growth that has happened.

Speaker 3

这将迫使人们讨论一个问题:这不是暂时的,我认为你正是在暗示这一点,而这也是最难向政策制定者传达的——这项技术并没有一个自然的终点。

That is going to require the conversation to be forced about this isn't temporary, which I think is what you're gesturing at and is in a sense, the hardest thing to communicate to policymakers is there isn't a there isn't a natural stopping point for this technology.

Speaker 3

它会持续变得更好,它带来的变化也将与社会其他方面持续叠加。

It's gonna keep getting better, and the changes it brings are going to keep compounding with the rest of society.

Speaker 3

因此,这需要改变政治意愿,重新愿意去考虑一些我们许久未曾考虑的事情。

So that will need to create a change in in political will and a willingness to entertain things which we haven't in some time.

Speaker 1

所以现在我想换个角度,提出我自己的问题。

So now I wanna flip it, the question I'm asking.

Speaker 1

你提到了富足。

You brought up abundance.

Speaker 1

我在从事这项工作中学到的一点是,我绝不认为社会中缺乏的是改善事物的好点子,我们的政策之所以不够好,并不是因为政策库枯竭了。

One of the things I have learned doing that work is that it is certainly not my view that what is scarce in society is ideas for better ways of doing things, that our policy isn't better than it is because our policy cupboard is dry.

Speaker 1

嗯。

Mhmm.

Speaker 1

对吧?

Right?

Speaker 1

这不对。

That's not true.

Speaker 1

我们有很多好的政策。

We have lots of good policies.

Speaker 1

我能举出一大堆例子。

I could name a bunch of them.

Speaker 1

但以我们目前的政治体制,这些政策很难得以推行。

They're very hard to get through our political systems as they're currently constituted.

Speaker 1

人工智能未来最令人沮丧的一种可能是:你创造了一种方式,将年轻的白领工人赶出工作岗位,用普通水平的人工智能取而代之。

The least inspiring version of the AI future is a world where what you have done is create a way to throw young white collar workers out of work and replace them with average level AI intelligence.

Speaker 1

更令人兴奋的版本,用达里奥的比喻来说,就是数据中心里的天才。

The more exciting version, to use Dario's metaphor, is geniuses in the data center.

Speaker 1

嗯。

Mhmm.

Speaker 1

我认为这确实令人振奋。

And I do think that's exciting.

Speaker 1

当我听到你或他谈论,如果我们每年实现10个百分点的GDP增长,或者20个百分点的GDP增长,会怎么样时,我不禁想,

And I wonder when I hear him or you talk about, well, what if we had 10 percentage point GDP growth year on year, 20 percentage point GDP growth year on year?

Speaker 1

我们面临的许多问题,是不是其实只受限于想法层面?

I wonder how many of our problems are really bounded at the ideas level.

Speaker 1

对吧?

Right?

Speaker 1

我们现在就可以去找诺贝尔奖得主,问他们:这个国家我们应该怎么做?

We could go to Nobel Prize winners right now and say, what should we do in this country?

Speaker 1

他们中的很多人能给我们提供一些当前我们还没在做的好点子。

And a lot of them could give us good ideas that we are not currently doing.

Speaker 1

我有时会担心,或者根据我在其他问题上的经验感到疑惑:我们是否高估了自身面临的障碍——以为阻碍我们走向繁荣经济的,主要是缺乏足够的智慧和这些智慧能产生的想法,而实际上,我们执行能力的削弱更为严重?

I do worry sometimes or wonder, given my experience on other issues, whether we have overstated to ourselves how much of what stands between us and the expanding abundant economy we want is that we don't have enough intelligence and the ideas that that intelligence could create versus our actual ability to implement things is very weakened.

Speaker 1

而人工智能将加剧这一瓶颈,因为会有更多东西被推给系统去执行,包括愚蠢的点子、错误信息和投机行为。

And what AI is gonna create is a larger bottlenecks around that because there'll be more being pushed at the system to implement, including dumb ideas and disinformation and slot.

Speaker 1

对吧?

Right?

Speaker 1

就像,它在账目的另一侧也会有一些东西。

Like, it'll have things on the other side of the ledger too.

Speaker 1

你是怎么看待这些限制因素的?

How do you think about these rate limiters?

Speaker 1

某种程度上说

There's kind of

Speaker 3

这里有一个有趣的教训,来自人工智能公司或一般公司,尤其是科技公司:通常新的想法来自于公司内部创建所谓的‘创业中的创业’,也就是把那些随着时间推移积累起来的后台官僚主义或繁琐工作抛在一边,然后对公司内部一个很小的团队说:你们什么都不用管。

a funny lesson here from the AI companies or companies in general, especially tech companies, where often new ideas come out of companies by them creating what they always call the startups within a startup, which is basically taking whatever process has, like, built up over time leading to back end bureaucracy or schlep work and saying to a very small team inside the company, you don't have any of this.

Speaker 3

去干点事情吧。

Go and do some stuff.

Speaker 3

这就是像Claude Code和其他一些东西是如何被创造出来的。

And and this is, you know, how things like Claude Code and other stuff get created.

Speaker 3

一些开始浮现的想法是:在更广泛的经济体系中,创建这种无许可的创新结构会是什么样子?

Ideas that kind of are starting to float around are what would it look like to sort of create that permissionless innovation structure in in the larger economy.

Speaker 3

这真的非常困难,因为它还有一个额外的特性,那就是经济体系与民主制度是紧密相连的。

And it's really, really hard because it has the additional property that, you know, economies are are linked to democracies.

Speaker 3

民主制度需要权衡众多民众的意愿,而所有政治都是地方性的。

Democracies weigh the preferences of many, many people, and all politics is local.

Speaker 3

因此,正如你在基础设施建设中所遇到的那样,如果你想建立一个无许可的创新体系,就会遇到产权和人们真实意愿等问题,这时你就陷入了一个难以解决的境地。

So often, as you've encountered with infrastructure build outs, if you want to create a permissionless innovation system, you run into things like property rights and what people's preferences are, and now you're in an intractable place.

Speaker 3

但我的感觉是,这正是我们将来必须面对的主要问题。

But my sense is that's the main thing that we're going to have to confront.

Speaker 3

而人工智能可能给我们带来的一个优势是,如果运用得当,它本质上是一种能吞噬官僚体系的机器;但如果用得不好,它反而会成为制造官僚体系的机器。

And the one advantage that AI might give us is it is kind of a native bureaucracy eating machine if done correctly, or a bureaucracy creating machine if done badly.

Speaker 1

你有没有看到,有人创建了一个系统,你只需输入你附近新开发项目的相关文件。

Did you see that somebody created a a system that basically you feed it in the documents of a new development near you.

Speaker 3

哦,它会自动生成环境评估报告?

Oh, and it writes environmental review things?

Speaker 1

或者它会写出极其复杂的异议意见,是的。

Or It writes incredibly sophisticated challenges Yep.

Speaker 1

针对你能想到的每一个可挑战的环节。

Across every level of the code that you could possibly challenge on.

Speaker 1

所以,大多数人没钱请一家非常专业的律师事务所来阻止隔壁建公寓楼。

So most people don't have the money when they wanna stop an apartment building from going up down the block to hire a very sophisticated law firm to figure out how to stop that apartment building.

Speaker 1

但基本上,这个系统实现了规模化的效果。

But, basically, this created that at scale.

Speaker 1

所以,正如你所说,它既能吞噬官僚体系,也能极大地强化官僚体系。

And so as you say, right, it could eat bureaucracy, could also supercharge bureaucracy.

Speaker 3

是的。

Yep.

Speaker 3

人工智能的每一件事都有其反面。

It's the everything in AI has the other side of the coin.

Speaker 3

我们有一些客户使用我们的AI系统,大幅缩短了提交新药候选材料所需的时间,时间被大大减少了。

We have customers that have used our AI systems to massively reduce the time it takes them to produce all of the materials they need when they're submitting new new drug candidates, and it's cut that time massively.

Speaker 3

这正是你刚才描述的镜像世界版本。

It's the Mirror World version of what you just described.

Speaker 3

我对这个问题没有简单的答案。

I don't have an easy answer to this.

Speaker 3

我认为,只有当这种情况明显演变为危机,并且能够在社会层面进行讨论时,它才会变得可以采取行动。

I think that this is the kind of thing that becomes actionable when it is more obviously a crisis and actionable when it's something that you can discuss at a societal level.

Speaker 3

我想,这场对话中一直萦绕的问题是,人工智能的变化几乎会在各个地方发生,而其风险则以一种分散且难以捉摸的方式显现,使得我们很难准确识别它并采取相应措施。

I guess the thing that was circling around in this conversation is that the changes of AI will kind of happen almost everywhere, And the risks of it, it happens in a diffuse unknowable way such that it is very hard to call it for what it is and take actions on it.

Speaker 3

但机遇在于,如果我们能够真正看清这一现象,并帮助世界看清正是这种现象在推动变革,我相信这会凸显问题,促使我们摆脱某些惯性,从而学会如何与这些系统共处并从中受益。

But the opportunity is that if we can actually see the thing and help the world see the thing that is causing this change, I do believe it will dramatize the issues to kind of shake us out of some of this stuff and help us figure out how to work with with these systems and benefit from them.

Speaker 1

我在这一切中注意到的是,据我所知,目前对公共人工智能完全没有明确的议程。

What I notice in all this is that there is, as far as I can tell, zero agenda for public AI.

Speaker 1

社会希望人工智能做什么?

What does society want from AI?

Speaker 1

社会希望这项技术具备哪些能力?

What does it want this technology to be able to do?

Speaker 1

有哪些事情是你必须通过商业模式、奖励机制、政府补贴或某种政策来塑造市场或激励体系的?

What are things that maybe you would have to create a business model or a prize model or some kind of government payout or some kind of policy to shape a market or to shape a system of incentives?

Speaker 1

我们现在的系统不仅在解决私营市场知道如何付费的问题,还在解决那些无人负责、唯有公共部门才该承担的问题。

So we have systems that are solving not just problems that the private market knows how to pay for, but problems that it's nobody's job but the public Mhmm.

Speaker 1

政府也需要弄清楚如何解决这些问题。

And the government to to figure out how to solve.

Speaker 1

我认为,考虑到过去几年人们对AI的大量讨论以及这些系统变得多么强大,我本该看到更多这方面的提案了。

I think I would have bet given how much discussion there's been of AI over the past couple of years and how strong some of these systems have gotten, that I would have seen more proposals for that by now.

Speaker 1

我跟一些人聊过这个问题,也一直在思考,但我很好奇你是怎么看待这个问题的。

And I've talked to people about it and wondered about it, but I I guess I'm curious on how you think about this.

Speaker 1

我们需要一个议程,至少要与所有推动AI发展的私人激励并行,这个议程不是关于‘我们害怕AI会对公众造成什么影响’。

What would it look like to have, at least parallel to all the private incentives for AI development, an actual agenda for not what we are scared AI will do to the public.

Speaker 1

我们同样需要一个关于这方面的议程。

We need an agenda for that too.

Speaker 1

而是关于我们希望AI能做什么,这样才能让像你们这样的公司有动力往这个方向投入?

But what we want it to do such that companies like yours have reasons to invest in that direction?

Speaker 3

我喜欢这个问题。

I love this question.

Speaker 3

我认为这里存在一个典型的‘鸡生蛋还是蛋生鸡’的问题:如果你亲自接触这项技术,就会对它能实现的能力产生非常强烈的直觉。

I think there's a real chicken and egg problem here where if you work with the technology, you develop these very strong intuitions for just how much it can do.

Speaker 3

私人市场非常擅长推动这些直觉的形成。

And the private market is great at forcing those intuitions to get developed.

Speaker 3

我们尚未在公共领域大规模部署这项技术。

We haven't had massive large scale public side deployments of this technology.

Speaker 3

因此,许多公共部门的人还没有这些直觉。

So many of the people in the public sector don't yet have those those intuitions.

Speaker 3

一个积极的例子是能源部正在开展的‘创世计划’,他们的科学家正与包括Anthropic在内的所有实验室合作,探索如何有意识地加速科学的某些部分。

One positive example is something the Department of Energy is doing called the Genesis Project, where their scientists are working with all of the labs, including Anthropic, to figure out how to actually go and intentionally speed up bits of science.

Speaker 3

要达到这一程度,我们和其他实验室与能源部的科学家进行了多次黑客松和会议,使他们不仅形成了直觉,还变得兴奋起来,并有了可以将这项技术应用于哪些方向的想法。

Getting there took us and other labs doing multiple hack days and meetings with scientists at the Department of Energy to the point where they not only had intuitions, but they became excited, and they had ideas of what you could turn this toward.

Speaker 3

要将这种方法推广到更广泛的公共生活领域,如医疗或教育,需要企业深入这些社区,与当地人接触,开展基层努力。

How we do that for the larger parts of the public life that touch most people, like health care or education, is going to be a combination of grassroots efforts from companies going into those communities and and meeting with them.

Speaker 3

但最终,我们必须将其转化为政策。

But at some point, we'll have to translate it to policy.

Speaker 3

我认为,这可能需要我、你和其他人去论证:这件事是完全可以实现的。

And I think maybe that's me and you and others making the case that this is something that can be done.

Speaker 3

我经常对民选官员说:给我们一个目标。

And I often say this to elected officials of give us a goal.

Speaker 3

比如,人工智能行业非常擅长通过基准测试来攀登顶峰。

Like, the AI industry is excellent at trying to climb to the top on benchmarks.

Speaker 3

为公众利益制定一些基准吧。

Come up with benchmarks for the public good that you want.

Speaker 1

那么,假设你真的这样做了。

So let's imagine that you did do something like this.

Speaker 1

我一直非常支持以奖励推动公共研发。

I've I've always been a big fan of prizes for public development.

Speaker 1

假设有一项立法通过了,卫生与公共服务部或国立卫生研究院(NIH)或其他机构宣布:这里有15个我们希望看到解决的问题,我们认为人工智能可以在这方面大有作为。

So let's say that there was legislation passed and the Department of Health and Human Services or the NIH or or someone came out and said, here's 15 problems we would like to see solved that we think AI could be Mhmm.

Speaker 1

有潜力解决这些问题。

Potent at solving.

Speaker 1

对吧?

Right?

Speaker 1

如果真有这笔钱,如果这些难题背后有1万亿美元的资金支持,因为它们对社会而言值这个价,这会实质性地改变Anthropic等机构的发展优先级吗?

If there was real money there, if there was $1,015,000,000,000 behind a bunch of these problems because they were worth that much to society, would it materially change the sort of development priorities at places like Anthropic?

Speaker 1

我的意思是,如果资金到位,这会改变你们的研发方向吗?

I mean, if the money was there, would it alter the sort of r and

Speaker 3

你们正在做的工作?

d you all are doing?

Speaker 3

我不这么认为。

I don't think so.

Speaker 3

为什么?

Why?

Speaker 3

因为真正阻碍这些工作的并不是资金。

Because it's not really the money that is the impediment to this stuff.

Speaker 3

而是实施路径。

It is the implementation path.

Speaker 3

而是真正清楚如何让成果顺利落地并产生效益。

It is actually having a sense of how you get the thing to flow through to the benefit.

Speaker 3

公共部门的许多方面并未被设计成对技术普遍友好,也缺乏激励机制。

And many aspects of the public sector have not been built to be super hospitable to technology in general to incentivize it.

Speaker 3

我认为,这主要需要以可保证的影响力和明确的实施路径作为奖励。

I think it mostly just takes a bounty in the form of guaranteed impact and guaranteed path to implementation.

Speaker 3

因为人工智能组织中最稀缺的资源,其实是组织内人员的时间,因为你可以几乎朝任何方向发展。

Because the main thing that is scarce at AI organizations is just the time of the people at the organization because you can go in almost any direction.

Speaker 3

这项技术正在迅速扩展。

This technology is expanding super quickly.

Speaker 3

许多新的应用场景正在涌现。

Many new use cases are opening up.

Speaker 3

你只是在问自己:我们究竟在哪里才能对世界产生积极而有意义的影响?

And you're just asking yourself the question of where can we actually have a positive meaningful impact in the world.

Speaker 3

在私营部门,这非常容易实现,因为它拥有推动一切落地的全部激励机制。

Super easy to do that in the private sector because it it has all of the incentives to push stuff through.

Speaker 3

在公共部门,我们更需要解决的是部署问题,而非其他任何问题。

In the public sector, we more need to solve this problem of deployment than anything else.

Speaker 1

如果宣布了什么,你会感到兴奋?

What would excite you if it was announced?

Speaker 1

你认为哪些项目会是这种类型的合适候选?

What what what do you think would be good candidates for that kind of project?

Speaker 3

任何能加快与医疗专业人员沟通并减轻他们工作负担的事情。

Anything that helps speed up the time it takes to both speak to medical professionals and take work off their plate.

Speaker 3

你知道,我们最近又有了一个宝宝。

You know, we had another baby recently.

Speaker 3

因为宝宝撞到头了,或者今天皮肤颜色不一样了,诸如此类的事情,我花了很多时间在凯撒医疗的咨询热线。

I spend a lot of time on the Kaiser Permanente advice line because the baby's bonked its head or its skin's a different color today or, you know, all of these things.

Speaker 3

我会用Claude来帮我和我妻子在等待护士接听时避免恐慌。

And I use Claude to sort of stop me and my wife panicking while we're waiting to talk to the nurse.

Speaker 3

但接着我听护士做了所有这些分诊工作,问了这么多问题。

But then I listened to the nurse do all of this like triaging, ask all of these questions.

Speaker 3

显然,很大一部分工作都可以用AI系统高效地完成,这能帮助我们短缺的人力更有效地利用时间,也能给正在经历这个系统的人提供安心。

So obviously, a huge chunk of this is stuff that you could like use AI systems productively for and it would help the people that we don't have enough of spend their time more effectively, and it would be able to give reassurance to the people going through the system.

Speaker 3

这可能没有你想象中的那些东西那么令人兴奋或炫目。

And that's maybe less inspiring and and glamorous than maybe some of what you're you're imagining.

Speaker 3

但我觉得,当人们与公共服务互动时,他们最大的挫败感往往在于流程不透明,而且要等很久才能联系到人。

But I think mostly when people interact with public services, their main frustration is just that it's opaque and it takes you a long time to speak to a person.

Speaker 3

但实际上,这些正是人工智能可以真正发挥作用的领域。

But actually, these are exactly the kinds of things that AI could meaningfully work on.

Speaker 1

这很有趣,因为你描述的并不是人工智能作为数据中心里一群天才的产物,而是将人工智能视为通信和文档处理的基础基础设施。

It's interesting because what you're describing there is less AI as a country of geniuses in the data center and more AI as standard plumbing of communications and documentation.

Speaker 3

我们数据中心里有一群初级员工。

We've got a country of junior employees in the data center.

Speaker 3

让我们好好利用这一点。

Let's do something with that.

Speaker 3

有一件事我们在这次对话中还没提到,但值得记住的是:如今科学的前沿已经向商业开放,这在过去是前所未有的。

Like, you know, one thing we haven't talked about in this conversation and it's just worth bearing in mind is, like, the frontier of science is open for business now in a way that it hasn't been before.

Speaker 3

我的意思是,我们已经找到了一种方法,能够可靠地加速人类科学家的工作。

And what I mean by that is we found a way to build systems that can provably accelerate human scientists.

关于 Bayt 播客

Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。

继续浏览更多播客