Odd Lots - 这是判断文字是否由AI生成的方法 封面

这是判断文字是否由AI生成的方法

This Is How to Tell if Writing Was Made by AI

本集简介

当考虑到许多人并不知道如何以及在哪里使用逗号时,可以说AI在写作方面已经优于大多数人了。它的文字简洁明了,有时出人意料地具有说服力,甚至有时还富有信息量。但它的文字中常常仍有一种说不清道不明的“不对劲”感。很多人在阅读AI生成的文字时,能很快察觉出来。除了风格之外,AI生成文本的存在还带来各种影响:从让学生更容易作弊,到欺骗性聊天机器人的兴起,再到可能削弱Reddit等网站的体验。那么,你究竟如何判断一篇文字是否由AI生成?在本期节目中,我们采访了Pangram Labs的首席执行官Max Spero,这家公司开发了能够检测内容是否由AI生成的软件。我们讨论了他们使用的先进技术、误报和漏报的风险,以及AI写作对互联网未来整体意味着什么。 阅读更多: 在OpenAI宣布Sora死亡后,AI视频应用正迅速崛起 信用衍生品交易因伊朗战争和AI担忧创下新高 只有Bloomberg商业新闻、股票市场、金融、突发及全球新闻订阅用户才能每周收到Odd Lots简报,并无限访问网站和应用。立即订阅:bloomberg.com/subscriptions/oddlots 订阅Odd Lots简报 加入讨论:discord.gg/oddlots 隐私信息请见:omnystudio.com/listener

双语字幕

仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。

Speaker 0

感谢您收听《奇思妙想》。

Thanks for listening to Odd Thoughts.

Speaker 0

请在亚马逊音乐上关注本节目,以收听更多未来剧集,或直接询问Alexa。

Follow the show on Amazon Music for more future episodes, or just ask Alexa.

Speaker 0

在亚马逊音乐上播放播客《奇思妙想》。

Play the podcast Odd Thoughts on Amazon Music.

Speaker 1

今天的节目由先锋集团赞助。

Today's show is brought to you by Vanguard.

Speaker 1

致所有正在收听的理财顾问,我们来聊一聊债券。

To all the financial advisors listening, let's talk bonds for a minute.

Speaker 1

在固定收益市场中捕捉价值并不容易。

Capturing value in fixed income is not easy.

Speaker 1

债券市场庞大、复杂,坦白说,许多公司只是扔给你几只花哨的基金就完事了。

Bond markets are massive, murky, and let's be real, lots of firms throw a couple flashy funds your way and call it a day.

Speaker 1

但先锋集团不是这样。

But not Vanguard.

Speaker 1

在先锋,机构级品质不是一句口号。

At Vanguard, institutional quality isn't a tagline.

Speaker 1

而是对您客户的承诺。

It's a commitment to your clients.

Speaker 1

我们提供涵盖80多种债券基金的优质产品,由一支由200名全球行业专家、分析师和交易员组成的团队主动管理。

We're talking top grade products across the board of over 80 bond funds, actively managed by a 200 person global squad of sector specialists, analysts, and traders.

Speaker 1

这些人全身心投入固定收益领域。

These folks live and breathe fixed income.

Speaker 1

因此,如果您希望为客户提供年复一年的稳定回报,请亲自前往 vanguard.com/audio 查看其业绩记录。

So, if you're looking to give your clients consistent results year in and year out, go see the record for yourself at vanguard.com/audio.

Speaker 1

网址是 vanguard.com/audio。

That's vanguard.com/audio.

Speaker 1

所有投资均存在风险。

All investing is subject to risk.

Speaker 1

先锋营销公司,分销商。

Vanguard Marketing Corporation distributor.

Speaker 2

彭博音频工作室。

Bloomberg Audio Studios.

Speaker 2

播客。

Podcasts.

Speaker 2

广播。

Radio.

Speaker 2

新闻。

News.

Speaker 3

你好,欢迎收听《奇闻杂谈》播客的另一期节目。

Hello, and welcome to another episode of the Odd Lots podcast.

Speaker 3

我是吉尔·韦因森塔尔。

I'm Jill Weisenthal.

Speaker 0

我是特蕾西·阿拉韦。

I'm Tracy Allaway.

Speaker 3

所以,特蕾西,你有没有遇到过一些文字,说不上来具体哪里不对,但就是觉得:我敢肯定这是AI写的。

So, Tracy, you know, you ever come across some writing and you can't articulate exactly why, but you're like, I'm pretty sure AI wrote this.

Speaker 3

这种情况发生得太多了吗?

Does this happen too much?

Speaker 0

坦白说,我还没怎么认真想过这个问题。

So full disclosure, I haven't really thought about it that much.

Speaker 2

真的吗?

Really?

Speaker 2

是的。

Yeah.

Speaker 2

因为问题是,

Because the thing is,

Speaker 0

也许我应该多想想。

probably should think about it more.

Speaker 0

但现在有很多糟糕的写作,我已经有点着迷了。

But there's a lot of bad writing out there, and I've become sort of a nerd to it.

Speaker 0

而且我也觉得,如今花大量时间去判断某篇文章是否由AI生成,会带来巨大的精神负担,尤其是对我们这些从事新闻行业的人来说。

And I also think that, I don't know, trying to figure out whether or not something was generated by AI nowadays, if you actually dedicate a lot of your own time to doing that, that is a huge mental burden Especially, be you and I are in the journalism industry.

Speaker 0

我们现在从公关公司收到的提案中,你觉得有多少是AI生成的?

How many of the pitches do you think that we get from PRs right now are being generated by AI?

Speaker 0

想象一下,如果你每天都要逐个阅读这些提案,并试图判断它们的来源。

Imagine if you're reading each one of those and trying to figure it out on a daily basis.

Speaker 3

你知道吗,我最常想到的是,有人会回复一条推文。

You know what I suppose I think about it the most is someone will respond to a tweet.

Speaker 3

是的。

Yeah.

Speaker 3

我会想,如果这是个真人,那么这个人或许值得我互动一下,或者我想回应一下。

And I'll be like, well, if this is a real person, then maybe this person deserves some engagement and a question, or I want to respond.

Speaker 0

但如果存在问题

But if there's a problem

Speaker 3

是机器人,那显然我就不会,这就是我想说的。

a bot, then obviously I don't And that's where I'm like, You know what?

Speaker 3

我想弄清楚这一点。

I want to figure it out.

Speaker 3

我想知道答案。

I would like to know the answer.

Speaker 3

顺便说一下,我对AI写作有一个有争议的看法,那就是它相当不错。

You know, I have a controversial view about AI writing, by the way, which is that it's pretty good.

Speaker 3

总的来说,我想我可能在最近的一集中提到过,考虑到大多数人根本不知道句子中逗号该放在哪里。

By and large, and I said this I think maybe in a recent episode, when you consider the fact that, I don't know, the majority of the population doesn't know where to put a comma within the sentence.

Speaker 0

这正是我的观点。

This is my point.

Speaker 3

很好。

Good.

Speaker 3

我要说的是,AI从来不会用错逗号的位置。

Mean, one thing I'll say about AI is it never gets the placement of a comma wrong.

Speaker 3

在某种程度上,它是完美的。

On some level, it's perfect.

Speaker 0

你做的吗?

Did you do that?

Speaker 0

我认为这刊登在《纽约时报》上,

I think it was in the New York Times,

Speaker 3

那个测试。

the test.

Speaker 3

是的。

Yeah.

Speaker 3

我有点讨厌那个。

I kinda hated that.

Speaker 0

好的。

Okay.

Speaker 0

为什么?

Why?

Speaker 3

好吧,我来告诉你为什么。

Well, because I'll tell you why.

Speaker 3

首先,只有五个例子。

First of all, there's only five examples.

Speaker 3

只有两个例子,它问读者更喜欢哪一个,但我认为

There's not very Two, it asked the reader which do you prefer, but I think

Speaker 0

而且它们的主题也不同。

And they were different subjects as well.

Speaker 3

是的。

Yeah.

Speaker 3

而且,我认为大多数人可能把它当成猜哪个是人类的问题。

And also, I think most people probably treated that as can you guess which one is a human?

Speaker 3

因为每个人都想说自己更喜欢人类的。

Because everyone wants to say they prefer the human.

Speaker 3

我觉得这并不是一个很好的测试。

I didn't think it was a great test.

Speaker 3

尽管如此,你看,不仅常常难以分辨,而且写得也常常不错,有时AI甚至能写出非常出色的语句。

Nonetheless, look, not only is it often indistinguishable, not often is it often fine writing, sometimes AI could come up with a really remarkable turn of phrase.

Speaker 3

但我总体上还是不喜欢。

But I still, by and large, don't like it.

Speaker 3

你读到一段文字,尤其是很长的AI生成文本时,即使说不出来具体原因,也会觉得:这感觉像是AI写的。

You read a thing, especially a long text that's AI, and it's like, even if you can't articulate it, it's like, this feels AI.

Speaker 3

它有一种病态的、甜腻的腔调,常常让人感到厌烦。

It has a certain sickliness, sweetness to it that is often annoying.

Speaker 3

它太过于如此了。

It's so So

Speaker 0

我注意到的是,它不太擅长模仿风格。

what I notice about it is it doesn't do style very well.

Speaker 0

对吧?

Right?

Speaker 0

所以如果你让它以某位作家的风格写作,只要不是像莎士比亚那样特别明显的风格,它就会表现得很差。

So if you ask it to write something in the style of a writer, if you choose anything other than something really obvious like Shakespeare, it really it suffers.

Speaker 0

但它实际生成的文字倒是挺清晰的。

But the text that it actually outputs is pretty clear.

Speaker 0

是的。

Yeah.

Speaker 0

这非常清楚。

It's totally clear.

Speaker 0

对于基本理解来说,它可能比很多

For basic understanding, it's probably better than a lot of

Speaker 3

真正需要担心这个问题的是老师,显然是大学和学生律师,也许还有很多。

what's The on the real people who are going to have to worry about this are teachers, obviously, universities and student lawyers, and maybe a lot.

Speaker 3

这没问题。

It's fine.

Speaker 3

但有时候你会想,这到底是人写的吗?

But sometimes it's like, okay, did someone write this or not?

Speaker 3

我们最好能知道答案。

And there has to be it'd be nice if we could know the answer.

Speaker 0

另一件正在发生的事是,你有没有看到一些书,上面明确标注了免责声明,说这本书完全由人类撰写。

Well, the other thing that's starting to happen is have you seen any books out there that actually come with a disclosure or disclaimer that say this book has been written only by humans.

Speaker 0

完全没有使用AI。

No AI used at all.

Speaker 0

我第一次看到这种声明是在我们为《Odd Lots》节目阅读的一本书上。

I saw that for the first time on a book that we actually read for an Odd Lots episode.

Speaker 0

我觉得这本书还没出版,但这种做法让我很意外。

I don't think it's come out yet, but that kind of threw me.

Speaker 3

是的。

Yeah.

Speaker 3

不是。

No.

Speaker 3

这种情况越来越多了。

It's more and more.

Speaker 3

无论如何,当我们进入一个绝大多数文字——如果不是全部的话——都是由AI撰写的的世界时,人们会越来越关注这个问题:我们是否能知道真相。

Anyway, as we enter a world in which the vast majority, if not already, of words written are written by AI, there's gonna be interest in this question of whether we know.

Speaker 3

总之,有一家叫Pangram Labs的公司,他们提供一个小工具,你可以付费使用,也可以免费使用:只需粘贴一段文本,它就会告诉你这段文字是人类还是AI撰写的概率。

Anyway, there's this company called Pangram Labs, and they have a little thing, and you can pay for it, but also a free service where you can drop, like, a text in, and it'll say the odds that it's written by human or AI.

Speaker 3

我对它印象非常深刻。

And I'm pretty impressed by it.

Speaker 3

我做了一些我自己的写作和AI生成内容的样本。

I did some samples of my own writing and then AI outputs.

Speaker 3

它全都判断对了。

It got them all right.

Speaker 3

但后来我又做了一些进一步的尝试,比如我想难倒它:我取了一段AI生成的文字,然后把它翻译成中文。

But then I did some further, like I tried to stump it to see if so what I did was I took a piece of AI writing, and then I had it translated into Chinese.

Speaker 3

接着我又把它翻译成正式的中文。

And then I had it translate that into high Chinese.

Speaker 3

你可以想象,这像是用更正式的语体写成的。

It's like, okay, imagine this is being written by a more formal register.

Speaker 3

然后我把它翻译成希伯来语,再翻译回英语。

And then I had that translated into Hebrew, And then I had that translated into English.

Speaker 3

所以原始内容经过这一系列AI翻译的传递,我把最终的输出再次输入Pangram,它还是判断对了。

So the original thing through this series of AI telephone through various translations, and then I put that output back into Pangram, and I got that right.

Speaker 3

它说这是AI写的。

It said it was AI.

Speaker 3

所以即使经过一系列旨在模糊原文风格的转换,看看它最终是否会变成别的东西。

So even after a series of transformations designed to obfuscate the original style of the piece to see if eventually it would emerge as something else.

Speaker 3

所以我很印象深刻。

So I was pretty impressed.

Speaker 3

它似乎有效。

It seems to work.

Speaker 3

我认为这有趣的原因有几点,那就是也许你能直接分辨出来。

And I think that's interesting for a couple reasons, which is maybe there is something that you can just tell.

Speaker 3

但第二点让我有点担心,因为你知道,已经有一些文章声称,这是由AI撰写的。

But two, it sort of worries me because, you know, you there have been articles, and they'll say, like, this is written by AI.

Speaker 3

我认为我最大的担忧之一是我自己写了一些东西。

And I think one of my big fears would be that I write something.

Speaker 3

我喜欢使用破折号。

I like to use an Emdash.

Speaker 3

我一直都是破折号的粉丝。

I've always been an Emdash fan.

Speaker 0

我爱破折号。

I love Emdashes.

Speaker 0

人们就是这样说话的。

That's how people talk.

Speaker 0

对不起。

I'm sorry.

Speaker 3

然后如果有人说这是你用AI写的,我会说:不是我写的。

And then what if it says you wrote this by AI, and I'm like, didn't.

Speaker 3

然后突然间,这个黑箱成了评判、判决我职业生涯的法官和刽子手。

And then here's this black box that is suddenly a judge, jargon, executioner for my career potentially.

Speaker 3

你用AI写的。

You wrote this via AI.

Speaker 3

实验室说的。

The lab says so.

Speaker 3

你完了。

You are now done.

Speaker 3

这让我很担心。

That worries me.

Speaker 3

所以我认为这引发了许多关于这些模型检测技术的有趣问题,我想更多地了解它们是如何运作的。

So I think this raises a lot of very interesting questions about these model detection thing, and I want to learn more about how works.

Speaker 0

此外,关于我们在写作中真正重视什么,也存在许多哲学性的问题。

Well, there's also a lot of philosophical questions about just what we value in writing True.

Speaker 0

同样,因为没有人会因为你使用拼写检查或类似工具而责备你。

As well, because no one's gonna yell at you for using spell check or something like that.

Speaker 0

对吧?

Right?

Speaker 0

很难想象,声誉风险竟然会取决于你是否可能使用过某个聊天平台来做一些基本的校对。

Like, it's kinda crazy to think that reputational risk is gonna hinge on whether or not you might have used a platform, a chat platform to, like, do some basic copy editing.

Speaker 3

完全同意。

Totally.

Speaker 3

很高兴地说,我们确实请到了一位完美的嘉宾。

Well, very happy to say we do, in fact, have the perfect guest.

Speaker 3

我们将与马克斯·斯佩罗对话。

We're gonna be speaking with Max Spero.

Speaker 3

他是Pangram Labs的创始人兼首席执行官,能够回答我们所有的问题。

He is the founder and CEO of Pangram Labs, and he can answer all of our questions.

Speaker 3

所以,马克斯,非常感谢你做客《Odd Lots》。

So, Max, thank you so much for coming on Odd Lots.

Speaker 4

谢谢你们邀请我。

Thanks for having me.

Speaker 3

你怎么知道它是对的?

How do you know it's right?

Speaker 3

所以,当有人输入一段文字时,我们稍后会讨论具体方法。

So you someone puts in a piece of text, and we'll get into the method in a second.

Speaker 3

但有人输入一段文字,系统就会判断是人类还是AI写的。

But someone puts in a piece of text, and it says human, AI.

Speaker 3

是什么让你相信你们在这一问题上有着非常出色的记录?

What makes you believe that you have a very good track record on this question?

Speaker 4

当我们启动Pangram时,我们首先做了一件事,叫做人类基线,也就是作为人类,我们能多准确地判断某段文字是AI生成的还是人类写的?

When we started Pangram, we started by doing this thing we call a human baseline, which is how well can we, as a human, predict whether something's AI or not?

Speaker 4

学习的第一步是弄清楚这个问题是否可解?

That's the first step at learning, is this problem tractable?

Speaker 4

这个问题是难还是容易?

How hard or easy is it?

Speaker 4

我个人发现,我能达到大约90%的准确率。

And I found, me personally, I was able to get about 90% accuracy.

Speaker 4

所以我们认为,AI模型应该能做得比这好得多。

And so we figured an AI model should be able to do much better than that.

Speaker 0

我有一堆方法论方面的问题,我们可以深入探讨。

So I have a bunch of methodology questions which we can get into.

Speaker 0

但在进入这些之前,依你看来,为什么AI生成的内容是坏事?

But just before we get into any of that, why is AI sought bad, in your opinion?

Speaker 0

为什么我们需要追踪和识别AI生成的内容?

Why does it need to be tracked and identified?

Speaker 4

我认为问题在于,生成这些内容实在太容易了。

I think the problem is it's just so easy to generate.

Speaker 4

因此,很难判断其背后的真实意图。

And so it's very difficult to know what is the intent behind it.

Speaker 4

事实上,我们现在还挺幸运的。

Basically, now, I think we're actually pretty lucky.

Speaker 4

我们生活在一个互联网和信息渠道中信噪比相当高的世界里。

We live in a world where the signal to noise ratio on the internet and in our information channels is pretty high.

Speaker 4

我们的信噪比确实很高。

We have pretty high signal to noise.

Speaker 4

但任何不良行为者都可以涌入,用看似合法的AI垃圾信息淹没我们的信息渠道。

But any bad actor can come in and just flood our information channels with AI slop that looks legitimate.

Speaker 4

它看起来像是有人投入了真正的努力和思考。

It looks like somebody put actual effort and thought into it.

Speaker 4

但实际上,这可能只是单次提示的结果,而且还可以被自动化。

But really, it was just like a single prompt, which could have also been automated.

Speaker 3

这是我经常思考的一个问题:曾经有一段时间,也许现在仍然是,如果你读到的内容语法正确

This is something that I think about a lot, which is that there was a point in time, and maybe still is the point in time, where if you read something that was grammatically correct

Speaker 4

是的。

Mhmm.

Speaker 3

标点准确、拼写无误,就有理由认为作者是一个具有某种严肃性和智慧的人。

Where the punctuation was strong, where the spelling was strong, there was reason to think that the person who wrote it was a person of, like, a certain seriousness and a certain intelligence behind it.

Speaker 3

我认为你所指出的问题是,这种关联现在正在被切断,我们不能再使用诸如文风严谨这样的启发式方法来判断发布者是否是认真、聪明的人。

And I think that the issue that you're identifying is that that link is now being severed so that we can't use these heuristics anymore, such as the strict quality of the prose, to know in fact whether this was published by someone who was a serious actor, intelligent or not.

Speaker 0

现在有些人故意在文本中加入拼写错误,以证明自己是

And now you have people inserting typos into their copy to prove that they are

Speaker 3

一个真实的人。

a fact human.

Speaker 3

博伊德,抱歉。

Boyd, sorry.

Speaker 3

让我回到我最初的问题。

Just to go back to my original question.

Speaker 3

你提到过,你们的系统能正确识别90%的情况,但现在它的使用范围更广了,而且有人在为你的软件付费,大概是教师、记者等人群。

So you mentioned, okay, you're able to get it 90% right, but now it's been used a lot more and you have people paying for your software, presumably teachers and journalists, etcetera.

Speaker 3

考虑到这些情况,从90%提升到100%至关重要——如果每十次就出一次错误,对于一款可能将人误判为AI创作者的商业软件来说,这显然是不可接受的。

Given all of that, getting from 90% to 100, I mean, if you could make one out of 10 is clearly unacceptable error rate for a piece of commercial software that could call someone an AI creator.

Speaker 3

所以你们必须做得远好于90%。

So you have to do a lot better than 90%.

Speaker 3

谈谈你们在将软件作为商业产品发布后,从数据中观察到的现象吧,是什么让你相信这款软件能准确区分这两类内容?

Talk to us about what you've seen so far in your data since releasing it as commercial software that makes you believe the software is doing a correct job of allocating between the two categories.

Speaker 4

我们已经构建了非常全面的评估体系。

We've built out really comprehensive evals.

Speaker 4

有两种类型的错误。

There's two kinds of errors.

Speaker 4

一种是假阳性,即某段文字由人类撰写,但系统却判定为AI生成。

There's a false positive, which is when something is written by a human and then we say that it's written by an AI.

Speaker 4

另一种是假阴性,即某段文字由AI生成,但系统未能识别出来。

There's a false negative, which is if it was AI written and we don't catch it.

Speaker 4

因此,我们追踪这两类数据的数值。

And so we track our numbers for both of these.

Speaker 4

对于人类写作,我们实际上相当幸运。

And for human writing, we're actually pretty fortunate.

Speaker 4

我们拥有数以百万计的样本,因此可以得出一个非常可信的误报率,目前我们的数字大约是万分之一。

We have millions and millions of samples so we can get a false positive number that we have a very high degree of confidence in, and our number right now is about one in ten thousand.

Speaker 4

所以,如果我们扫描一万份文档,平均会有一份被错误地识别为AI生成,而实际上它是人类撰写的。

So if we scan 10,000 documents, on average, one will come back as AI when it was actually human.

Speaker 3

那另一方向呢?

And what about in the other direction?

Speaker 4

误报率,我认为大约是99%的准确率,也就是大约1%的漏报率。

False negative, I would say around ninety nine percent accuracy, so around one percent false negative rate.

Speaker 4

我认为这在一定程度上取决于提示的对抗性有多强,以及用户试图规避的程度

I think this depends a little bit more on how adversarial the prompting is, how much they're trying to

Speaker 3

逃避。

evade.

Speaker 3

没错。

Exactly.

Speaker 3

没错。

Exactly.

Speaker 3

我尝试通过多次过滤来混淆原始输出。

I tried to send it through multiple filtrations to obfuscate the original output.

Speaker 3

这将是对抗性提示的一个例子。

That would be an example of adversarial prompting.

Speaker 4

没错。

Exactly.

Speaker 4

但在一般情况下,当我们只是查看AI的直接输出时,准确率超过99%。

But in the general case where we're just looking at straight outputs from AI, it's above 99.

Speaker 0

好的。

Okay.

Speaker 0

那么,当您的模型评估一段文本时,具体是在寻找什么?

So what is your model looking for exactly when it's evaluating a text?

Speaker 0

因为正如我们在引言中提到的,AI生成的内容在语法和句法上通常都很好。

Because as we mentioned in the intro, syntax and grammar tends to be pretty good on AI generated copy.

Speaker 0

风格有时更像是一个标识符,我认为你的观点是对的,乔。

The style is sometimes more of an identifier, I would argue, your point, Joe.

Speaker 0

有时候,它读起来过于甜腻,过于真诚。

Sometimes it reads very saccharine and overly earnest in some ways.

Speaker 0

你在这里具体关注的是什么?

What exactly are you focusing on here?

Speaker 0

有哪些明显的特征?

What are the tells?

Speaker 4

风格和用词确实是其中的一部分,但我认为很多人没有意识到,他们在撰写文本时实际上做出了大量选择。

The style and the word choices are definitely part of it, but I think what a lot of people don't realize is they're actually making a lot of decisions when they write a piece of text.

Speaker 4

每一个短语都有几十种甚至上百种表达方式,而在50、100或200个词的篇幅中,你实际上做出了成千上万的决定。

There's dozens or hundreds of ways to phrase every single phrase, and over the course of 50 or 100 or 200 words, you're making thousands of decisions actually.

Speaker 4

因此,我们正在学习这些前沿模型是如何做出这些决策的模式。

So what we're doing is we're learning the patterns and how these frontier models make these decisions.

Speaker 4

如果这些决策中的绝大多数都与前沿模型的做法一致,那么这段文字由人类撰写的可能性就微乎其微。

And if the vast majority of these decisions line up with how the frontier models are doing it, then it's vanishingly unlikely that this was written by a human.

Speaker 4

你必须恰好做出与大语言模型完全相同的成百上千个决定。

You would have to just happen to make the same exact decisions that the LLM does hundreds of

Speaker 0

次。

times.

Speaker 0

有意思。

Interesting.

Speaker 4

好的。

Okay.

Speaker 3

但这是一个非常重要的观点。

But this is a really important point.

Speaker 3

所以到目前为止,每个人都对放弃破折号这个特征有所体会了。

So everyone at this point has some feel for let go of the em dash tell.

Speaker 3

对吧?

Right?

Speaker 4

是的

Mhmm.

Speaker 3

但我的理解是,你并不是进去后直接硬编码,比如看到一堆破折号就处理。

But my understanding is it's not like you don't go in and, like, hard code if you see a bunch of em dashes.

Speaker 3

这才是关键。

This is the thing.

Speaker 3

在许多情况下,我想象你和模型本身都无法用英语清楚地说明这些决策是什么。

These decisions, in many cases, I imagine neither you nor the model itself can articulate in English what the decisions are.

Speaker 3

你只知道决策模式确实存在。

All you know is that the decision pattern exists.

Speaker 3

这是正确的吗?

Is this correct?

Speaker 3

这是正确的。

This is correct.

Speaker 3

好的。

Okay.

Speaker 3

你能解释一下吗?因此,你的模型学会了这些决策模式,这意味着什么?

Can you explain, so therefore, what does it mean that your model has learned these decision patterns?

Speaker 4

从广义上讲,我们正在训练一个深度学习模型。

So what we're doing on the very broad scale is we're training a deep learning model.

Speaker 4

这是一个相当大的黑箱,但它基于一个语言模型的基础架构。

So it's a pretty big black box, but it has the base model of a language model.

Speaker 4

然后,它不是预测下一个词元,而是预测这段文本是AI生成的还是人类写的。

Then instead of predicting the next token, it's predicting whether the text is AI or not.

Speaker 4

我们训练它的方法是使用数千万个样本进行训练,因此它会看到数百万个人类撰写的例子。

And how we train it is we train on tens of millions of examples, so it sees millions and millions of human examples.

Speaker 4

对于每一个真实的人类文本示例,我们也会向它展示一个AI生成的示例。

And for each human example, we also show it an AI example.

Speaker 4

比如说,其中一个可能是关于Denny's的78个单词长的五星评价。

So, for example, let's say one of these is a five star review for Denny's that's 78 words long.

Speaker 4

然后我们会让AI根据第一个评价的风格,写一篇同样78个单词长的关于Denny's的五星评价。

Then we'll ask an AI to write a five star review about Denny's that's 78 words long in the style of the first one.

Speaker 4

显然,这两个会有所不同。

And obviously, these two will be different.

Speaker 4

因此,我们的模型能够通过对比学习到这两者之间的差异

And so our model is able to learn through contrast what is the difference between

Speaker 3

这两者。

these two.

Speaker 3

重要的是,抱歉,为了明确一下,你和我可能无法清晰地描述出这种差异。

And the important thing, sorry, just to be clear here, is that you and I might not be able to articulate the difference.

Speaker 3

可能在句子长度上会有一些不同。

There will be some difference in maybe the sentence length.

Speaker 3

在用词上也会有一些差异。

There will be some difference in word choice.

Speaker 3

在标点、语法等方面也会有一些不同。

There'll be some difference in punctuation, syntax, whatever.

Speaker 3

但你和我显然不会注意到这些。

But you and I wouldn't obviously spot it.

Speaker 3

然而,在经历了数百万个这样的对比示例后,模型学会了区分其中的差异。

However, after millions of examples of these side by sides, the model learns what the difference is.

Speaker 4

没错。

Exactly.

Speaker 4

我认为人类所能做的最好方式就是寻找一些非常明显的迹象。

I think the best that a human can do is look for some of these really obvious tells.

Speaker 4

ChatGPT 喜欢使用‘不是x,而是y’这样的表达方式。

ChatGPT loves the it's not just x, it's y framing.

Speaker 4

早期的模型特别喜欢某些特定词汇,比如‘tapestry’、‘intricate’和‘delve’。

Earlier models really liked some specific words like tapestry and intricate and delve.

Speaker 3

是的。

Yeah.

Speaker 3

Delve。

Delve.

Speaker 3

Tapestry。

Tapestry.

Speaker 3

是的。

Yeah.

Speaker 4

但是,是的,我认为通过训练Pangram,我们能够比这更深入,超越文档层面的高级迹象,进行更深层次的分析。

But but, yeah, I think by training Pangram, we're able to go much deeper than this and look deeper than the high level signs at the, like, document level signs.

Speaker 1

今天的节目由先锋集团赞助。

Today's show is brought to you by Vanguard.

Speaker 1

致所有聆听的财务顾问,我们来谈一谈债券。

To all the financial advisors listening, let's talk bonds for a minute.

Speaker 1

在固定收益领域捕捉价值并不容易。

Capturing value in fixed income is not easy.

Speaker 1

债券市场庞大、复杂,说实话,许多公司只是给你几个花哨的基金就完事了。

Bond markets are massive, murky, and let's be real, lots of firms throw a couple flashy funds your way and call it a day.

Speaker 1

但先锋集团不是这样。

But not Vanguard.

Speaker 1

在先锋集团,机构级品质不是一句口号。

At Vanguard, institutional quality isn't a tagline.

Speaker 1

这是对客户的承诺。

It's a commitment to your clients.

Speaker 1

我们提供全方位的优质产品,涵盖80多只债券基金,由一支200人的全球专家团队——包括行业专家、分析师和交易员——主动管理。

We're talking top grade products across the board of over 80 bond funds, actively managed by a 200 person global squad of sector specialists, analysts, and traders.

Speaker 1

这些人全身心投入固定收益领域。

These folks live and breathe fixed income.

Speaker 1

所以,如果您希望为客户提供年复一年的稳定回报,请亲自前往 vanguard.com/audio 查看业绩记录。

So, if you're looking to give your clients consistent results year in and year out, go see the record for yourself at vanguard.com/audio.

Speaker 1

网址是 vanguard.com/audio。

That's vanguard.com/audio.

Speaker 1

所有投资均存在风险,Vanguard Marketing Corporation 为分销商。

All investing is subject to risk, Vanguard Marketing Corporation distributor.

Speaker 5

新闻在周末也不会停止。

The news doesn't stop on the weekends.

Speaker 6

环境不断变化,而彭博社正是您掌握一切动态的最佳平台。

Context changes constantly, and now Bloomberg is the place to stay on top of it all.

Speaker 5

你好。

Hi.

Speaker 5

我是大卫·古拉。

I'm David Gura.

Speaker 5

每周六和周日,请收听全新的《彭博周末》。

Join us every Saturday and Sunday for the new Bloomberg this weekend.

Speaker 2

我是克里斯蒂娜·拉菲尼。

I'm Christina Raffini.

Speaker 2

我们将为您带来最新头条、深度分析和重磅访谈。

We'll bring you the latest headlines, in-depth analysis, and big interviews.

Speaker 2

所有在您休息日触动人心的故事。

All the stories that hit home on your days off.

Speaker 6

我是丽莎·马泰奥。

And I'm Lisa Mateo.

Speaker 6

请观看并收听《彭博周末》,了解关于商业、生活方式、人物与文化的深刻而富有启发性的对话。

Watch and listen to Bloomberg this weekend for thoughtful, enlightening conversations about business, lifestyle, people, and culture.

Speaker 5

在周六早上,我们会将过去一周的事件放在背景下分析,探讨市场和世界发生了什么。

On Saturday mornings, we put the past week's events into context, examining what happened in the markets and the world.

Speaker 2

而在周日,我们会采访记者、专栏作家和重要的政治人物,帮助您为即将到来的一周做好准备。

Then on Sundays, we speak with journalists, columnists, and key political figures to prepare you for the week ahead.

Speaker 6

一醒来就加入我们吧,无论您的周末计划带您去往何处,都带着我们同行。

Join us as soon as you wake up and bring us with you wherever your weekend plans take you.

Speaker 5

请在彭博电视上观看我们。

Watch us on Bloomberg Television.

Speaker 5

请在彭博广播上收听我们。

Listen on Bloomberg Radio.

Speaker 5

通过彭博商业应用实时观看节目,或收听我们的播客。

Stream the show live on the Bloomberg Business app or listen to the podcast.

Speaker 2

这就是《彭博周末》,每周六和周日东部时间早上7点开始。

That's Bloomberg this weekend, Saturdays and Sundays starting at 7AM eastern.

Speaker 6

让彭博电视、广播以及您收听播客的任何平台,都成为您周末日常的一部分。

Make us part of your weekend routine on Bloomberg Television, radio, and wherever you get your podcasts.

Speaker 0

所以这让我想起了一件事,我在想该怎么表达,但它让我想起以前人们常做的一种练习:把很多不同的面孔融合在一起,得出一张所谓的具有吸引力的面孔。

So one thing this kind of reminds me of, and I'm thinking how to phrase this, but it reminds me of, you know, those exercises people used to do where you would take a bunch of different faces and meld them all together and come up with, like, one face that was supposedly attractive.

Speaker 0

换句话说,这在多大程度上是一种分布检测器?你是不是在寻找那些你认为AI会采取的特定路径?

Like, to what extent is this basically a distributional detector in the sense that you're looking for certain paths that you think AI would choose?

Speaker 0

我想问的是,会不会因为有人只是以一种平均再平均的方式表达某个句子,就导致了误报?

And I guess, could you get a false positive just from someone who's choosing the average of the average of the average in a way to state a particular sentence?

Speaker 4

也许吧。

Maybe.

Speaker 4

我们把误报率设定为一万分之一而不是零,是有原因的。

There's a reason we have our false positive rate is one in ten thousand and not zero.

Speaker 4

因为有时候我们看到一个误报,发现它读起来完全像一篇AI生成的评论或论文,但实际上它写于2019年。

It's because sometimes we look at the false positive and it's like, oh, it reads exactly like an AI generated review or essay, except that it was written in 2019.

Speaker 4

所以这很可能是一位人类,只是恰好以一种所有路径都收敛到最常见模式的方式写出了内容。

So it was probably a human who just happened to find the exact mode collapsed type of way that all runs.

Speaker 4

我会说,是的,把写作看作一种分布是个不错的思路——人类写作构成一个广阔的空间,而AI写作只是这个空间中的一个微小点。

I would say, yeah, I think it's a good way to think about the distribution of writing or writing as a distribution where there's a space of all human writing, and then AI writing is really just a small point within this space.

Speaker 4

无论你如何提示它,它都不会偏离其训练时的范围太远。

No matter how much you prompt it, it doesn't go that far from where it was trained to be.

Speaker 0

是的。

Yeah.

Speaker 0

好的。

Okay.

Speaker 3

黑箱是什么?

What's the black box?

Speaker 3

我自己构建了一个小模型。

So I built a little model myself.

Speaker 3

我做了这个东西,可以上传文本,然后判断它是更像书面语还是口语。

I built this thing that detects you can upload text, and it says whether it's more resemblant of the written word or the spoken word.

Speaker 4

哦,我看过那个。

Oh, I saw that.

Speaker 4

好的。

Okay.

Speaker 3

是的。

Yeah.

Speaker 3

我用了BERT,这是谷歌开源的一个模型。

And I used BERT, which is like one of these things, open source one from Google.

Speaker 3

你们训练的核心模型是什么?

What is the core model that you trained on?

Speaker 3

还是说你们自己构建了它?

Or is it something or did you build it yourself?

Speaker 3

跟我们聊聊这个吧。

Like, talk to us about that.

Speaker 4

我们的第一个模型是基于BERT构建的,但后续模型我们需要提升容量。

Our very first model was built on BERT, but future models, we needed to up our capacity.

Speaker 3

为什么?

How come?

Speaker 3

解释一下。

Explain that.

Speaker 4

基本上,我们的模型遇到了容量限制。

Basically, we were running into capacity limits with our model.

Speaker 4

它的假阳性率和假阴性率已经达到了上限。

It was capping out at a certain false positive, false negative rate.

Speaker 4

它无法学习到更深层的特征,所以我们不得不将参数量增加十倍,然后一百倍,以便它能深入理解这些前沿模型的写作方式。

It wasn't learning the deeper signals, so we had to 10x and then 100x the parameter count so that it can learn really deeply how these frontier models write.

Speaker 0

你有没有注意到不同模型在写作上的有趣差异?

Have you noticed any interesting differences between how the models write?

Speaker 0

实际上,你们的模型是否不仅能识别是否为AI生成的内容,还能区分是哪个模型生成的?

Actually, is your model trained to identify different models as well as whether or not this is just broadly AI generated?

Speaker 4

我们并没有专门针对不同模型进行训练。

We don't specifically train it on different models.

Speaker 4

我们不会说,这个是CLOUD三,那个是CHET或者GPT5。

We don't say, Hey, this one is CLOUD three and this one is CHET or GPT5.

Speaker 4

我们做了一些可解释性研究,分析了模型的输出嵌入,发现它实际上能学会判断文本来自哪个模型。

What we've done, we've done some interpretability work to look at basically the output embeddings of the model and we find that it actually learns which model the text came from.

Speaker 4

你可以看到一些小的聚类。

You could see little clusters.

Speaker 4

这是Claude的聚类,所有Claude的文本都集中在这里,然后这些是Deepseq和Quen,而这是ChatGPT。

This is the Claude cluster and all of the Claude cluster around here, then these are the Deepseq and Quen, and then this is ChatGPT.

Speaker 4

它们都聚集在不同的空间和嵌入空间中。

They all cluster into different spaces and embedding space.

Speaker 4

因此,很明显,模型能够学会这些前沿模型之间的差异。

So clearly, the model is able to learn what the difference is between these frontier models.

Speaker 0

实际上,既然你提到了Quen,我非常感兴趣。

Well, actually, since you mentioned Quen, I'm very interested.

Speaker 0

Quen生成文本的方式,与美国开发的平台相比,有什么独特之处吗?

Is there anything distinct in terms of how Quen generates text versus platforms that have been developed in The US?

Speaker 4

我认为Quen的独特之处在于它接受了比其他模型更多的中文和多语言词元训练。

I think Quen is unique because it's trained on a lot more Chinese and multilingual tokens than other models.

Speaker 4

我听中国朋友说,它在中文对话流畅性方面要好得多。

I've heard from Chinese friends that it's much better at being conversationally fluent in Chinese.

Speaker 4

除此之外,我不确定我还能说出什么。

Beyond that, I don't know that I can tell.

Speaker 4

让我看一段文字并说这是Quen生成的,这对我来说很难,但我认为更熟悉它的人可能能做到。

It would be hard for me to look at a text and say, I know that's quen, but I think somebody who is more familiar with it might be able to.

Speaker 3

让我们谈谈这项工作的某些哲学或社会影响。

Let's talk about some of the philosophical or societal implications of this work.

Speaker 3

有没有人曾经被Pangram判定为AI生成的文字,但他们却坚持说:‘我发誓,这真的不是AI写的’?

Have you had anyone whose text has been judged to be AI written by Pangram, and they're like, I swear to God, this isn't.

Speaker 3

他们真的非常坚持。

They really insist.

Speaker 3

你对这种情况怎么看?

What do you think about this situation?

Speaker 3

你怎么做?

What do you do?

Speaker 3

跟我们聊聊这个。

Talk to us about that.

Speaker 4

我遇到过几次这种情况。

I've had a couple times this happened.

Speaker 4

确实有过几次,我真心认为这只是一个误报。

There have been times where I genuinely believe that this is just a false positive.

Speaker 4

我们扫描了数亿份文档,所以在一定规模下,这种情况必然会发生。

We scan hundreds of millions of documents, so at a certain scale, this will happen.

Speaker 4

但我也经常遇到一些人,他们总是说:AI检测器根本没用。

But I also get people who all the time, they're just like, AI detectors don't work.

Speaker 4

这完全是骗局。

It's a total fraud.

Speaker 4

而他们在LinkedIn上发布的内容完全是AI生成的,却只是因为被揭穿而生气。

And then whatever they're putting out on LinkedIn is just 100% AI generated, and they're just mad that they're getting called out.

Speaker 4

如果你回溯他们更早的过往和历史,会发现他们发布的一切内容都是AI生成的,直到2023年左右。

And then you look back farther into their past and their history, everything they're putting out is AI generated until about 2023.

Speaker 4

对每个人来说,如果你从历史角度看,会发现有很多账号都在发布一堆垃圾内容。

For everyone, if you look historically, there's a lot of slop accounts that are putting out total slop.

Speaker 4

你可以看出,他们之前发帖可能不多,如果回溯时间线,会发现他们曾经写过人类的文字。

And you can tell either they weren't posting as much before, and if you scan back in time, then you see that they were writing human text at some point.

Speaker 3

所以有很多账号,正好在2023年左右,如果你扫描他们所有的作品,会非常清楚地看到在2023年初左右出现了一个明显的转变吗?

So there's a number of accounts out there that basically right around the 2023, where if you scan the entire corpus of their work, it very clearly shows a switch right around early twenty twenty three?

Speaker 4

是的。

Yeah.

Speaker 4

这真的取决于具体的账号。

It really depends on the account.

Speaker 4

我们发现一个有趣的情况:有一位《卫报》的记者正在报道冬奥会,有人指出:‘这篇文章完全是AI垃圾。’

I think one thing we saw that was interesting was there was a writer for The Guardian that was covering the Winter Olympics, Somebody was like, Hey, this article is total AI slop.

Speaker 4

用Pangram检测后,确认是AI生成的。

Ran it through Pangram, it was AI.

Speaker 4

《卫报》回应说:‘当然,我们的记者不会使用AI。’

The Guardian was like, No, of course our writers don't use AI.

Speaker 4

于是我们扫描了这位记者的历史记录,发现他们确实是从2024年年中到年底开始逐渐使用AI,并在文章中越来越多地依赖它。

So we scanned this single writer's history, and we found that they really did start picking up AI mid to late twenty twenty four, and we're using it more and more in their articles.

展开剩余字幕(还有 477 条)
Speaker 0

我的意思是,暂且扮演一下反对者的角色,意图在识别AI生成的低质内容时重要吗?比如,你确实可以遇到一些恶意行为者,他们可能试图影响人们对某个话题的看法,于是他们在TwitterX上创建了大量机器人,用AI大量发布支持其观点的低质内容。

I mean, just to play devil's advocate for a second, does intent matter when it comes to identifying AI slop in the sense that, okay, I get you can have a bad actor who's maybe trying to influence how people feel about a particular topic, and maybe they've created a bunch of bots on TwitterX, and they're using AI to just flood the zone with a bunch of AI slop supporting their particular viewpoints.

Speaker 0

另一方面,如果你是一名记者,你的工作就是撰写关于新闻话题的清晰易懂的稿件——我要澄清一下,我完全不支持这种做法。

On the other hand, if you're a journalist and your business is to write, you know, like basic understandable copy about a news topic, Just to be clear, I'm not advocating this at all.

Speaker 0

但这种意图和通过纯粹的数量来影响舆论是完全不同的。

But that intent is very different to I'm gonna try to influence something by just sheer volume.

Speaker 4

是的。

Yeah.

Speaker 4

当然,这两种情况的严重程度相差很大。

Definitely, these are one is a lot more severe than the other.

Speaker 4

但与此同时,如果你是一名记者,却用AI来逃避工作、不亲自撰写内容,我认为这也是个问题。

But I think at the same time, if you're a journalist and you're using AI to basically shirk your work and not do your work, I think that's also a problem.

Speaker 4

这对媒体机构来说也是一种声誉风险,因为读者能察觉出来,而且他们会公开指出。

And I think it's a reputational risk to the outlet because people can tell and people are gonna call you out.

Speaker 4

无论AI生成的内容来自哪里,都有很多人不想阅读这些低质内容。

There's a lot of people who don't wanna read AI slop regardless of where it's from.

Speaker 3

是的。

Yeah.

Speaker 3

这确实是真的。

This is definitely true.

Speaker 3

你最终会用完可以用来训练的人类素材吗?

Are you ever gonna run out of human material to change on?

Speaker 3

如果你找到一段发布于2023年之前、尤其是2019年或更早的互联网文本,你可以非常确定这是人类生成的。

You could be pretty confident that if you find some piece of text that was published on the internet prior to 2023, but certainly prior to 2019 or something like that, you could be extremely sure that this was human generated.

Speaker 3

你是否担心未来会更难确定训练数据的来源?

Do you worry that in the future it's going to be harder to even establish the provenance of your training data?

Speaker 4

是的,这确实是我们关注的问题。

Yeah, it's definitely a concern for us.

Speaker 3

跟我们说说你们是怎么考虑这个问题的。

Talk to us about how you're thinking about that.

Speaker 4

我们拥有一个近乎无限的2023年之前的预训练数据储备。

So we have a near infinite data reservoir of pre-twenty twenty three data.

Speaker 4

我们有足够的数据来长期训练。

There's just more than enough for us to train on for a long, long time.

Speaker 4

但问题的一部分是我们也希望用现代文本进行训练。

But part of the problem is we also want to train on modern text.

Speaker 4

现在大家都在讨论,如果有人在写关于大语言模型或人工智能的内容,我们不希望错误地将这些内容标记为AI生成,因为我们的训练数据对这个话题没有认知。

There's all this talk about if somebody's writing about LLMs or about AI, we don't want to incorrectly flag that as AI because our training data has no sense of this topic.

Speaker 4

我认为我们正在探索不同的方法来解决这个问题,但大多数方法都是在确定谁是可信赖的主体,谁在发布人类撰写的内容,我们可以在一定程度上利用我们的模型来识别这些主体。

I think we're looking at different ways to do this, but most of them are just figuring out who is a trusted actor, who do we know is putting out human written content, and we could use our model for that to some degree.

Speaker 4

所以我们有已知的可信主体,我们知道他们发布的是人类撰写的内容,我们也可以使用他们的数据。

So we have known actors, we know they're putting out human written content, and then we could use their data as well.

Speaker 0

一个稍微随机的问题,使用你们的模型,你们能估算出目前互联网上有多少比例是AI生成的垃圾内容吗?

Slightly random question, using your model, are you able to quantify what percentage of the internet at the moment is AI slop?

Speaker 4

大约占40%。

It's about 40%.

Speaker 3

哇。

Wow.

Speaker 3

基于什么?

Based on what?

Speaker 3

你是怎么得出这个数字的?

How'd you get that number?

Speaker 4

互联网上很多都是SEO优化的文章。

A lot of the internet is just SEO written articles.

Speaker 4

这些文章基本上是为了让网站在搜索中更频繁地出现,通过针对特定关键词来实现的。

It's articles written search basically so that your website comes up more often in search because it's targeting certain keywords.

Speaker 4

很多这个行业已经转向使用AI,因为这样就不必支付写手费用,只需花极少的成本就能大量生成文章。

A lot of that industry has switched over to using AI because then instead of having to pay writers, you could churn out articles for pennies on the dollar.

Speaker 4

但我认为这导致了互联网上大量内容都是AI撰写的。

But I think that kind of results in a lot of the Internet being AI written.

Speaker 4

这还稍微有点取决于平台。

It's a little bit it's also kind of platform dependent.

Speaker 4

从网页角度来看,大约占40%。

It's about 40% from an Internet page perspective.

Speaker 4

大约一年半前,我们研究了Medium,发现超过50%的新发表Medium文章都是由AI生成的,这个比例高得惊人。

About a year and a half ago, we looked at Medium and found that over 50% of newly written Medium articles were AI generated, which was a crazy high number.

Speaker 3

那Reddit呢?

What about Reddit?

Speaker 4

Reddit一年前大概是7%,我认为现在略高于10%。

Reddit, it was 7% a year ago, I believe, a little over 10% today.

Speaker 0

等等。

Wait.

Speaker 0

实际上,这让我想起来一件事。

Actually, this reminds me.

Speaker 0

我经常上Reddit,现在非常享受这个平台,但我确实担心其中有多少内容是AI生成的。

So I'm on Reddit a lot, and I really enjoy it nowadays as a platform, but I do worry about how much of it is being generated by AI.

Speaker 0

我不太明白的是,到底有什么经济动机让人去大量发布AI生成的帖子并获得点赞?

And the thing I don't necessarily understand is what are the economic incentives to actually write a bunch of AI generated posts on Reddit and get upvoted?

Speaker 0

这种系统或动机为什么会存在?

Why does that system or motivation even exist?

Speaker 4

有一些初创公司。

There are startups.

Speaker 4

我不会点名具体公司,因为我不想为它们做宣传,但它们会向企业承诺:我们会为你带来Reddit上的自然曝光。

I'm not going to name names because I don't want to promote them but they will sell a promise to companies that we're going to get you organic mentions on Reddit.

Speaker 4

我们会运行看起来像真人一样的AI机器人,它们会自然地推荐你的产品,或在评论或帖子中提及你的产品。

We're going to run our AI bots that seem organic, and they're just going to naturally recommend your product or just mention your product in the comments or in a post.

Speaker 4

所以我见过这方面的证据。

And so I've seen evidence of this.

Speaker 4

我们可以找到这些机器人,它们本质上就像机器人农场,主要进行看似自然的互动,发一些简短回复,有时还会进行品牌提及。

We can find these they're basically like bot farms that are mostly engaging seemingly organically, just doing a short reply, and then sometimes they're doing this brand mention.

Speaker 4

因此,这些帖子非常有价值。

And so that's why these posts are very valuable.

Speaker 0

这真的很有趣。

That's really interesting.

Speaker 3

我还得想象,这些内容之所以有价值,是因为所有模型都用Reddit的数据进行训练。

I have to also imagine it's valuable because all of the models train on Reddit.

Speaker 3

对吧?

Right?

Speaker 3

如果你希望你的产品名称出现在模型输出中,比如‘最好的鼻毛修剪器’之类的?

And if you want your product's name to appear in model outputs, it's like, what is the best nose hair trimmer or whatever?

Speaker 3

Reddit上有很多机器人讨论过这款鼻毛修剪器,因此它更有可能出现在ChatGPT的回复中。

And there's a bunch of bots that on Reddit talked about this nose hair trimmer, and then that's probably more likely to show up in a ChattyPety request.

Speaker 3

对吧?

Right?

Speaker 4

是的。

Yeah.

Speaker 4

对。

Yeah.

Speaker 4

这已经被奇怪地操纵了。

It's been weirdly gamed.

Speaker 4

你知道,以前你只是在谷歌上搜索‘最好的鼻毛修剪器’,但现在却有成千上万篇文章

You know, you used to just Google best nose hair trimmer, and now there's, like, a thousand articles

Speaker 0

现在这些结果之所以排在前面,是因为人们都在那里寻找。

that results, like, show up first nowadays because that's where people are looking.

Speaker 4

是的。

Yeah.

Speaker 4

然后人们开始搜索‘最佳鼻毛修剪器 Reddit’,以获取关于它的Reddit评论。

And then people started searching best nose trimmer Reddit to to get their Reddit comments on it.

Speaker 4

现在人们意识到,这就是人们在搜索的内容,因此你需要在Reddit上投放广告。

And now it's people have realized that that's what people are searching for, so you need to populate Reddit with your advertisements.

Speaker 3

我是在《男性健康》杂志上看到的。

I'm I'm on the men's health.

Speaker 0

你在找鼻毛修剪器吗?

Are you looking for nose hair trimmers?

Speaker 3

松下耳鼻毛修剪器是《男性健康》杂志的首选。

The Panasonic ear nose hair trimmer is the number one choice on men's health.

Speaker 3

优点是握感舒适。

Pro is easy to hold.

Speaker 3

总之,不是这样的。

Anyway, it's not.

Speaker 4

所有这些联盟链接

All these affiliate links

Speaker 3

彻底毁了互联网。

just destroyed the internet.

Speaker 3

我知道。

I know.

Speaker 3

这真的很遗憾,但也没办法。

It's really too bad, but whatever.

Speaker 3

再多跟我们讲讲整个流程吧。

Talk to us more about the whole pipeline.

Speaker 3

我对这个想法非常感兴趣。

So I'm very fascinated by this idea.

Speaker 3

就像是,你看到了一家丹尼餐厅的评价。

It's like, okay, you see this review for Denny's.

Speaker 3

让AI模型尽可能地复制它。

You have the AI model try to replicate it as best as it could.

Speaker 3

会有一些细微的差异。

There'll be these subtle differences.

Speaker 3

跟我们讲讲整个流程。

Talk to us about the whole pipeline.

Speaker 3

你们还用了哪些其他测试来验证真实性?因为我觉得你们想做的是找到最相似的数据集,其差异几乎难以察觉,从而进行极限压力测试。所以,跟我们详细说说这个完整流程吧。

What are the other tests that you're using to get the true because what I imagine you're trying to do is get the most similar data sets with an almost imperceptible difference to really stress test So the whole talk to us really about this whole pipeline.

Speaker 4

是的。

Yeah.

Speaker 4

所以我们真正想做的是,我们

So what we're really trying to do here is we're

Speaker 3

作为一名模型开发者,我正在尝试,不。

As a model maker myself, I'm trying No.

Speaker 3

不。

No.

Speaker 3

抱歉。

Sorry.

Speaker 4

继续说。

Keep going.

Speaker 4

对。

Yeah.

Speaker 4

作为AI专家。

As an AI expert.

Speaker 3

对。

Yeah.

Speaker 3

对。

Yeah.

Speaker 3

作为AI专家。

As an AI expert.

Speaker 3

我需要听

I need to hear

Speaker 4

一些该领域的技巧。

some tips of the field.

Speaker 4

是的。

Yeah.

Speaker 4

所以我们真正寻找的是那些尽可能接近人类与AI边界的事例,这样我们的模型能学得更好。

So what we're really looking for is examples that are as close to the boundary between human and AI as possible so that our model learns better.

Speaker 4

非常明显是AI生成的内容,我们的模型就学得不多。

Something that's very obviously AI is our model's not learning as much.

Speaker 4

对于明显是人类生成的内容也是如此。

Same thing for something that's obviously human.

Speaker 4

第一步是创建一个包含人类示例的合成镜像的数据集。

Step one is creating this dataset with synthetic mirrors of human examples.

Speaker 4

然后我们训练一个模型。

And then we train a model.

Speaker 4

第二步是所谓的主动学习。

And then step two is something called active learning.

Speaker 4

然后我们用这个模型去扫描一个更大规模的数据集,查找错误、假阳性、假阴性,并将这些样本重新纳入训练集,从而训练出一个更好的模型,因为它已经见过这些错误,而这些错误我们认为更接近人类与AI之间的边界。

So we then take this model and use it to scan a much larger corpus of data and look for errors false positives, false negatives and then we pull those back into our training set are able to train a much better model because it's seen these errors, and these errors, we believe, are just much closer to the boundary between human and AI.

Speaker 3

抱歉,为了明确一下,第一轮是已知的人类写作和已知的AI写作。

Sorry, just to be clear, the first pass is like, okay, you have known human writing and known AI writing.

Speaker 3

你训练一个模型。

You train a model.

Speaker 3

然后第二轮是未知的人类写作和已知的AI写作。

And then the next pass is once again unknown human and known AI writing.

Speaker 3

你已经知道每个样本的答案,因此可以列出哪些地方判断错了,然后把这些结果反馈回第一轮模型。

You already know the answer of each of these, and therefore you could come up with a list of which you got wrong, and then that gets fed back into the first version.

Speaker 4

没错。

Exactly.

Speaker 4

一旦我们重新训练,模型就会变得好得多。

Once we retrain, then the model gets much, much better.

Speaker 4

然后我们可以反复进行这个过程,让模型自我提升,每次训练都能变得更好。

And then we could do this as many times as we want to kinda just have a self improving model that gets better with every training run.

Speaker 4

我还可以再深入讲讲我们如何处理人工智能编辑的问题,因为我认为

I can also tell you go a little bit more into how we deal with AI edits because I think

Speaker 3

这是一个

that's an

Speaker 4

日益重要的问题。

increasingly important problem.

Speaker 4

我认为未来大多数写作都会有人工智能提供辅助。

I think most writing will be AI assisted in the future.

Speaker 4

我觉得这个功能已经在谷歌文档里了,而且我用的谷歌输入法里也有。

I think it's already in Google Docs and it's in my Google keyboard.

Speaker 0

语法检查工具Grammarly其实早在做类似的事情了。

Grammarly arguably has been doing this for a while.

Speaker 4

没错。

Exactly.

Speaker 4

Grammarly在后端就用到了大语言模型,我们总不能直接就断定‘现在所有的写作内容都是AI生成的’吧。

Grammarly uses LLMs on the back end, and we don't want to just say, All writing is AI now.

Speaker 4

我们希望能够区分AI辅助和AI生成的内容。

We want to be able to differentiate between AI assisted and AI generated.

Speaker 4

因此,我们使用了不同的提示语。

So what we do is we also have different prompts.

Speaker 4

在人工审核Denny的评论时,我们不会说‘生成一篇这样的评论’,而是说‘帮助改进这篇’。

So for our human review of Denny's, rather than saying, Generate a review like this, we could say, Help improve this.

Speaker 4

让它更正式一些。

Make it more formal.

Speaker 4

修正语法错误。

Clean up the grammar.

Speaker 4

因此,我们有一份很长的AI编辑提示列表,然后可以查看原始人类文本与...

And so we have a long list of AI editing prompts, and then we're able to look at the cosine difference, the distance between the original human text and

Speaker 3

在高维空间中AI生成文本的距离。

The distance AI in hyper multidimensional space.

Speaker 4

没错。

Exactly.

Speaker 4

那么,AI改变了这段文本多少呢?

So how much did AI change this text?

Speaker 3

而且

And

Speaker 4

然后我们就能训练我们的模型,根据这个距离打点,判断这是中等程度的AI辅助、轻度AI辅助还是重度AI辅助。

then we're able to train our model to say, we're just going to put a point on this distance and say, This is moderate AI assistance, this is light AI assistance, this is heavy AI assistance.

Speaker 0

很有趣。

Interesting.

Speaker 0

我打算做一件我觉得自己以前从未做过的事——问一位创始人关于他们公司的使命。

I'm going to do something I don't think I've ever done before, which is ask a founder about their corporate mission.

Speaker 0

但你创办了这家公司,当你思考自己在这里想要实现的目标时,这只是单纯的AI检测吗?比如,可能只有像教师这样的少数群体觉得这很有价值?

But you've set up this company, and when you think about what you're trying to do here, is it just basic AI detection in the sense that there might be, you know, a few groups of people like teachers that find this very valuable?

Speaker 0

还是说,你的使命更宏大,你实际上是在试图改善互联网以及人们在上面看到的内容?

Or is the mission something broader where you're actually trying to improve the Internet and what people see on it?

Speaker 4

我相信,能够检测AI生成内容的技术具有巨大的价值,这种价值不仅对教师有用,对每个行业的每个人都有用,包括律师、出版商,以及任何在互联网上消费内容的普通用户。

I believe the technology of being able to detect AI generated content is immensely valuable, and it's valuable not just for teachers, but for basically everybody in every profession, lawyers, publishers, just an individual who consumes content on the Internet.

Speaker 4

我认为这对所有这些人来说都很有价值。

I think it's valuable for all these people.

Speaker 4

但最终,是的,我们的高层次目标是帮助缓解日益增长的AI内容带来的某些负面影响。

But ultimately, yeah, our high level goal is to help mitigate some of these negative effects of growing AI content.

Speaker 0

但举例来说,仅以产品评论为例,你们的愿景是否是像Yelp这样的平台希望使用这项技术来确保其系统不被操纵?

But for instance, just using the product review example, is the vision that, like, a Yelp, for instance, would wanna use this technology to make sure that its system isn't being gamed?

Speaker 0

还是说,你们的愿景是,如果我是一个特别认真的消费者,有很多空闲时间,想出去吃饭,我可以把各个餐厅的评论都输入Pangram,从而真正判断出哪些是真实的口碑,哪些是炒作?

Or is the vision, like, if I am a particularly diligent consumer who has a lot of time on my hands and I'm looking to go out to a restaurant, I can run all these individual restaurant reviews through Pangram and then, like, actually figure out if it's real hype or not?

Speaker 4

所以我认为目前主要是前者。

So I think right now it's a lot of the former.

Speaker 4

我们与平台合作。

We work with platforms.

Speaker 4

我们最大的客户之一是Quora,他们将大量内容通过Pangram进行处理。

One of our biggest customers is Quora, and they run a bunch of content through Pangram.

Speaker 4

但我们还有许多其他平台使用Pangram来帮助审核并识别AI不良行为者,将他们从平台上移除。

But we have a lot of different platforms that use Pangram to help moderate and find AI bad actors and get them off their platform.

Speaker 4

但我认为,个人消费者这一场景也在快速增长,我们非常希望在此领域推进。

But I also think, yeah, the individual consumer case has been growing a lot, and we're really interested in pushing here.

Speaker 7

这是汤姆·基恩邀请您收听彭博市场观察播客。

This is Tom Keene inviting you to join us for the Bloomberg Surveillance Podcast.

Speaker 7

我们每天致力于让您在商业上更明智。

It's about making you smarter every business day.

Speaker 8

我是保罗·斯威尼。

I'm Paul Sweeney.

Speaker 8

我们为您提供全面的覆盖,

We bring you complete coverage of

Speaker 3

美国市场开盘。

The US market open.

Speaker 3

我们覆盖股票、债券、商品,甚至加密货币,所有您需要的资讯,助您脱颖而出。

We cover stocks, bonds, commodities, even crypto, all the information you need to excel.

Speaker 9

我是亚历克西斯·克里斯托弗。

And I'm Alexis Christophris.

Speaker 9

Bloomberg Surveillance 还为您带来新闻背后的深度分析。

Bloomberg Surveillance also brings you the analysis behind the headlines.

Speaker 9

我们通过与经济学、金融、投资和国际关系领域最杰出的人物对话来实现这一点。

We do that through conversations with the smartest names in economics, finance, investment, and international relations.

Speaker 7

我们每天工作日都会实时进行这些讨论,为您呈现每日播客中最优质的分析。

We do all this live each and every weekday that bring you the best analysis in our daily podcast.

Speaker 7

请在 Apple、Spotify、YouTube 或您收听播客的任何平台搜索 Bloomberg Surveillance。

Search for Bloomberg Surveillance on Apple, Spotify, YouTube, or anywhere else you listen.

Speaker 8

在东海岸,午餐时间收听。

On the East Coast, listen at lunch.

Speaker 8

而在

And on

Speaker 3

西海岸,一醒来就收听。

the West Coast, listen as soon as you wake up.

Speaker 9

这就是由汤姆·基恩、保罗·斯威尼和我,亚历克西斯·克里斯托弗里斯带来的 Bloomberg Surveillance 播客。

That's the Bloomberg Surveillance Podcast with Tom Keene, Paul Sweeney, and me, Alexis Christofferis.

Speaker 9

今天就订阅吧,无论你在哪个平台收听播客。

Subscribe today wherever you get your podcasts.

Speaker 7

Bloomberg Surveillance,每个交易日都值得收听。

Bloomberg Surveillance, essential listening each and every business day.

Speaker 3

pangram.com 的免费版本,比如你每天只能获得几次测试之类的。

The free version of pangram.com, like you get a handful of tests a day or something like that.

Speaker 3

如果有人拥有无限次数的 Pangram 响应,甚至能以无限规模访问 Pangram API,他们理论上能否训练出一个提示词,然后输入到 AI 中,生成类似人类的写作?

If someone had an unlimited number of Pangram responses and maybe had an access to the Pangram API at infinite scale, could they theoretically learn a prompt that they would then be able to put into an AI to generate human style writing?

Speaker 4

我确实有个朋友做过这件事。

I actually had a friend do that.

Speaker 4

他把他的 Claude 代码设成了循环运行。

He put his Claude code on a loop.

Speaker 4

我给了他一些 API 积分,然后他的 Claude 代码就整晚自动运行,不断尝试生成提示词,试图让输出结果看起来像是人类写的,或者能被 Pangram 识别为人类创作的内容。

I I gave him some API credits, and then his Claude code just basically worked overnight writing a prompt, trying to get it to output something that's human written, or that came back from Pangram as human written.

Speaker 4

它确实成功了,但生成的文本相当混乱。

It got there, but the text was pretty incoherent.

Speaker 4

所以,是的,它生成的基本上都是长篇的胡言乱语。

So so, yeah, it was producing more or less long gibberish.

Speaker 4

它在语法上是错误的。

It was grammatically incorrect.

Speaker 4

很多词根本毫无意义。

A lot of the words just didn't really make sense.

Speaker 3

因为这正是我最初的想法。

Because this was my first thought.

Speaker 3

当我看到这个时,我想,这会是个有趣的实验:看看你是否能收集所有输出结果,找出差异,不断迭代你给AI的提示,最终生成一个让Pangram认为是人类撰写的文本。

When I saw it, I was like, that would be a fun experiment to see if you could take all the outputs, find the difference, and just keep iterating on the prompt you would have to tell AI in order to eventually get an output that looked to Pangram like it was human generated.

Speaker 4

是的。

Yeah.

Speaker 4

我认为如果你还用一个LLM来评估连贯性,并同时用Pangram和连贯性评估器来给文本打分,是有可能实现的。

I think there's a way to do it if you also had an LLM judge on coherency and use Pangram and the coherency judge both to score your text.

Speaker 4

我觉得这绝对可行,我也很期待有人去尝试,因为如果能做到这一点,我们的模型会变得更好、更稳健。

I think it's definitely possible, and I'm excited for someone to try to do it because we could make our model a lot better and more robust if this existed.

Speaker 0

乔,我想知道你现在个人的令牌预算大概是多少,以至于你都会考虑这些事情?

Joe, I wanna know what your personal, like, token budget is nowadays that you're even, like, contemplating some of this You

Speaker 3

你知道我感觉怎么样吗?

know what I feel like?

Speaker 3

我用的是Clod Max套餐。

I have the Clod Max plan.

Speaker 3

你知道的?

You know?

Speaker 3

而且我上班的时候不会用。

And I don't work, like, when I'm at work.

Speaker 3

我不会在任何我的氛围编码项目上工作。

I don't work on any of my vibe coding projects.

Speaker 3

你知道吗,小时候,我记得如果我没吃完饭,总会有人说,世界上还有挨饿的孩子。

And you know, like, when we were kids, I don't know if you remember, like, if you didn't eat all your food, like, someone would say, oh, there's, starving kids in the world usually.

Speaker 0

我就想,哦,还有挨饿的氛围编码者需要火鸡呢。

I'm like, oh, like, starving vibe coders that need the turkeys.

Speaker 3

就像是,你没有意识到,我有四个小时的令牌额度,但我几乎从来不会用满。

It's like, oh, you didn't like, I have this four hour token window, and I'm almost never maxing it out.

Speaker 3

我只是在想,世界上另一边有些孩子,多么希望拥有你的令牌,而你却没能充分利用这期间的全部额度。

And I'm just thinking it's like, there are kids on the other side of the world that wish they had your tokens, and you're you're not using all of your tokens for the window.

Speaker 3

你怎么敢这样?

How dare you?

Speaker 3

当我没有用满我的Clod Max令牌额度时,我会感到有点内疚。

I feel a little guilty when I don't max out my clawed max token program.

Speaker 4

我也有Clod Max,而且确实,大多数日子我根本没怎么写代码。

I also have clawed max, and yeah, most days, I'm not doing much coding at all.

Speaker 4

我没有用满额度,但有些日子我会用得远远超过。

I'm not maxing it out, Then some days I'm going way

Speaker 3

那你对此感到内疚吗?

Do over feel guilty about that though?

Speaker 3

哦,好吧。

Oh, okay.

Speaker 3

是的。

Yeah.

Speaker 3

所以我想问你,写作确实挺有意思,但这种技术在图像和视频生成方面前景如何呢?你肯定经常被问到这个问题。

So can I ask you, writing is kinda interesting, but what are the prospects of this being able to work on, say, and you must get this a lot, image and video generation?

Speaker 3

这在理论上是否类似?

Is it at all theoretically similar?

Speaker 3

有没有理由认为它能够被复制,还是说这是一个完全不同的问题?

Is there a reason to think that it will be replicable, or is this just a different beast of a problem?

Speaker 4

我认为这种方法肯定是可行的。

I think the approach is definitely doable.

Speaker 4

我认为一些经济因素会发生变化,尤其是当我们关注当前视频生成的成本时。

I think some of the economics change, especially if we look at video and the cost of generating video today.

Speaker 3

好的。

Okay.

Speaker 4

我们无法像生成文本那样大规模地生成视频。

We can't generate video at the same scale that we can generate text.

Speaker 4

所以我们可能需要一种不同的方法。

And so we might need a kind of different approach.

Speaker 4

但我相信,如果我们能为图像加上音频解决这个问题,那可能就足以一并解决视频问题了。

But I also believe that if we're able to solve this for image plus maybe audio, that could be enough to just solve it for video as well.

Speaker 4

零样本。

Zero shot.

Speaker 0

你能否想象,比如推出某种视频认证计划?

Could you ever envision, I don't know, launching some sort of certification program for video?

Speaker 0

因为这看起来像是我爸爸这一代人,花很多时间在Facebook上。

Because this seems to be my dad's a boomer, spends a lot of time on Facebook.

Speaker 0

这似乎是社会所需要的。

Like, this seems to be what society needs.

Speaker 0

对吧?

Right?

Speaker 0

比如,一段视频附带一个小标识,说明这不是AI生成的,并且有人真正盖章认证过。

Like, a video that comes with a little thing that says this is not AI generated, and someone has actually rubber stamped that.

Speaker 4

有一个名为C2PA的组织,我认为他们在内容溯源方面做得相当不错。

So there's an organization called C2PA, and I think they're doing pretty good work on content provenance.

Speaker 4

他们基本上在与手机制造商和硬件制造商合作,嵌入硬件签名,以证明图像和视频确实是由该硬件拍摄的。

Basically, they are working with phone makers and hardware makers to basically embed hardware signatures to prove that image and video were truly taken from the hardware.

Speaker 0

basically 就是水印。

Like watermarks, basically.

Speaker 4

是的。

Yeah.

Speaker 4

没错。

Exactly.

Speaker 4

所以,与其标记AI生成的内容,不如……

So so rather than marking the AI outputs Yeah.

Speaker 4

我们反而是在真实拍摄的、真实存在的内容中嵌入一种真实性证明。

We're instead embedding, like, a proof of authenticity in the the, like, thing that's real and was captured in real life.

Speaker 0

这很有趣。

That's interesting.

Speaker 3

好的。

Alright.

Speaker 3

所以从宏观角度看,互联网将走向何方?

So big picture, where's the internet going?

Speaker 3

你提到互联网上已经有40%的内容是广告生成的,但也许这也不是世界末日。

You mentioned 40% of the internet is already ad generated, but maybe that's not the end of the world.

Speaker 3

如果只是些我从不看的SEO页面,管他呢。

If it's just a bunch of SEO pages that I never read, I don't know, whatever.

Speaker 3

给我们谈谈宏观层面的想法吧,无论Pangram和其他AI检测模型的普及程度如何,互联网的发展趋势会怎样?

Give us some thoughts high level about what the trajectory of the internet, regardless of the uptake of Pangram and other AI detection models.

Speaker 4

说实话,我对互联网的现状有点担忧。

I'm a little bit worried about the state of the internet, I'm going be honest.

Speaker 4

完全同意。

Totally.

Speaker 4

我认为,如今互联网的运作在很大程度上依赖于信任和共识,而我们目前并没有准备好应对这种前所未有的、规模巨大的机器人潮。

I think right now, so much of it is built around trust and norms in a way that we're not really well equipped to suddenly deal with an onslaught of bots at a completely different scale than we've dealt with before.

Speaker 4

这可能有好的一面和坏的一面。

There's maybe a good case and a bad case.

Speaker 4

我认为坏的情况是互联网走向了‘死亡互联网’理论所描述的境地。

I would say the bad case is the internet goes the way of dead Internet theory.

Speaker 4

所有开放和可访问的空间都被机器人充斥,人们唯一能进行真实交流的地方,只剩下像Discord服务器这样高度封闭的‘围墙花园’——在那里,每个人的身份都是已知的,你知道这里没有机器人。

Just like every space that's open and accessible is just flooded by bots, and then the only place people are able to communicate authentically is in, like, very walled garden, like, closed servers, like like, Discord servers, for example, where, you know, everybody's identity is known and you know you don't have bots in here.

Speaker 4

所以这可能是那种糟糕的 scenario。

So that's maybe the, like, bad scenario.

Speaker 3

我能跟你说一个我有过的大胆想法吗?

Can I tell you an insane thought that I've had?

Speaker 0

说吧。

Go on.

Speaker 0

就只是

Just

Speaker 3

我们要把这给踢出去。

we're gonna kick out of this.

Speaker 0

就只是只是

Just just

Speaker 3

那一个。

the one.

Speaker 3

你听说过吗,那种针对不良行为者的概念,我忘了他们怎么叫了。

Heard of, like, I forget what they call, like, this idea of, like, for the bad actors.

Speaker 3

它叫什么‘天堂模式’或者‘天堂封禁’。

It's called, like, heaven mode or heaven banning.

Speaker 3

你听说过这个吗?

Have you heard of that?

Speaker 3

是的。

Yeah.

Speaker 3

所以有一个想法,对付网络上的不良行为者的一种方式是,嗯。

So there's this thought that one way you could deal with bad actors Mhmm.

Speaker 3

在互联网上,突然间他们被置于一个类似推特的版本中

On the Internet is suddenly they're on a version of, say, Twitter

Speaker 4

好的。

Okay.

Speaker 3

在那里只有机器人,而且每个人都总是同意他们的一切,这会让他们发疯之类的。

In which they're only bots, and everyone always agrees with them on everything, and it drives them crazy and stuff like that.

Speaker 3

而他们永远不会意识到,因为他们会想,哦,这真棒。

And they would never know it because they're like, oh, that's cool.

Speaker 3

每个人都在,然后慢慢地,是的。

Everyone's and then it's sort like, slowly, like, yeah.

Speaker 3

这是一种惩罚人的方式,把他们放到一个永远不会发生争执的互联网环境中。

They just this is like a way you could punish people by putting them on an Internet where they will never get any fighting.

Speaker 4

你被天堂封禁了,被关进了牢笼。

You get heaven banned and put into basically jail.

Speaker 4

你正在跟一帮傻子说话。

You're talking to a bunch of fuck.

Speaker 3

没错。

That's right.

Speaker 3

那将会是被天堂封禁的监狱。

That would be jail where you're heaven banned.

Speaker 3

但我想说的是,我自己搭建了这个小型AI模型,然后我给朋友们看了。

But I thought and again, this is you know, like, built this little AI model myself, and I, like, showed it to my friends.

Speaker 3

他们说:‘哇,真酷啊,乔。’

They're like, oh, it's really cool, Joe.

Speaker 3

我真的很佩服。

I'm really impressed.

Speaker 3

我真的很佩服你居然能做成这个。

Like, I'm really impressed by, like, that you're able to do this.

Speaker 3

我当时就想:人们是在对我诚实吗?

And I was like, are people being honest with me?

Speaker 3

我是不是已经被天堂封禁了?

Have I been heaven banned?

Speaker 3

因为我觉得,如果这东西很烂,你大可以坦率地告诉我。

Because I just, like like, you can be honest with me if it sucks.

Speaker 3

而且我确实有点这种恐惧。

And I'm and I sort of have this fear.

Speaker 4

这是

This is

Speaker 0

史上最谦虚的炫耀了。

the biggest humble brag ever.

Speaker 0

我是认真的。

I'm very serious.

Speaker 0

我做了这件事,大家都觉得它很棒。

I did this thing, and everyone thought it was great.

Speaker 3

我只是想说,我觉得人们可能是出于觉得‘真酷’才对我客客气气的,我担心的就是这个。

I'm just saying, like, people are, like I think people I'm worried that, like, people are being nice to me because, like, oh, cool.

Speaker 3

是的。

Yeah.

Speaker 3

这真的令人印象深刻。

That's really impressive.

Speaker 3

你真的做了这件事。

You, like, did that.

Speaker 3

我有一种深深的焦虑,担心人们并没有对我坦诚相告。

And I have this, like, deep anxiety that, like, people aren't giving it to me straight about it.

Speaker 3

我知道这听起来像是谦虚地炫耀,但真的不是。

I know that sounds like a humble brag, but it's really not.

Speaker 4

这就是为什么你永远不能太成功。

That's why you can never get too successful.

Speaker 4

坎耶·韦斯特身边全是只会说好话的人。

Kanye West is surrounded by a bunch of yes men.

Speaker 3

他从来得不到这样的反馈:哦,这是他第一次尝试用氛围编码做点什么。

He never gets any Oh, this is his first try at doing something with vibe coding.

Speaker 3

我非常焦虑。

I'm deeply anxious.

Speaker 3

不,你直接告诉我它很烂就行了。

No, you can just tell me if it sucks.

Speaker 3

这没问题。

That's fine.

Speaker 3

这正是我的担忧。

That's my worry.

Speaker 0

我不担心这个。

I I don't worry about this.

Speaker 0

如果我发推说我在吃牛排,会有一百个人来批评我。

If I tweet that I'm eating a steak, I will get, like, a 100 people criticizing me

Speaker 3

为了一些事情。

for something.

Speaker 3

肉类争议。

The meat kerabala.

Speaker 3

是的。

Yeah.

Speaker 3

所以还有另一件事,那就是你永远不该发推的两件事:肉类烹饪和享受生活。

So that's the other thing, which is that the two things you are never allowed to tweet about, meat preparation and enjoying life.

Speaker 3

因为如果你一旦享受生活,或者一旦烹饪肉类,人们就会在网上对你大加抨击。

Because if you ever enjoy life and if you ever enjoy and if you ever prepare meat, people will flip out at you on the Internet.

Speaker 3

这两件事是你在网络上绝对不能做的。

Those are the two things that you're not allowed to do online.

Speaker 0

非常对。

Very true.

Speaker 0

还有一个相关的问题,但让我们回到方法论上。

And sort of related question, but just going back to the methodology.

Speaker 0

如果你专注于这种路径依赖的想法,我想象它就像一棵巨大的决策树。

If if you're focused on this sort of, like, path dependent idea, I'm kind of envisioning it as like a giant decision tree.

Speaker 0

对吧?

Right?

Speaker 0

随着模型变得越来越好,而且我们知道它们已经在输出中注入了一定程度的随机性,这有可能吗?

Is there a possibility that as the models get better and better and we know that they're already injecting, like, some degree of randomness into their output.

Speaker 0

虽然我知道肯定会有人较真,给我发消息说:‘你知道计算机根本无法实现真正的随机性。’

Although I know there's gonna be a pedant out there who messages me and says, Well, you know computers can't do true randomness.

Speaker 0

暂且放下这一点,要知道它们正在调整。

Setting that aside, know that they're adjusting.

Speaker 0

它们正以惊人的速度变得更加复杂。

They're becoming more sophisticated at an incredible rate.

Speaker 0

我们知道它们正试图调整并注入一些随机性,以避免这种类型的检测。

We know that they're trying to adjust and inject some randomness in order to avoid exactly this kind of detection.

Speaker 0

你担心它们自身的适应能力吗?

Do you worry about their own adaptation at all?

Speaker 4

我注意到,随着模型能力的提升,我认为它们的输出分布变得越来越复杂。

I have noticed that the models, as they get more capable, I believe their output distribution gets more complex.

Speaker 4

用一个简单的模型来学习更困难,这就是为什么我们一直在扩大模型规模,以捕捉能够表征大语言模型输出的更高复杂度函数。

It's harder to learn with a simple model, which is why we've been increasing our model size to capture a higher complexity function that can capture the LLM output.

Speaker 4

所以我认为我们可能需要继续提升我们的模型。

So I think we may have to continue to make our models better.

Speaker 4

我们必须努力跟上它们的步伐。

We're gonna have to work to keep up with it.

Speaker 4

我们不能躺在功劳簿上睡大觉。

We can't just rest on our laurels.

Speaker 3

什么是突发性和困惑度?

What are burstiness and perplexity?

Speaker 4

是的。

Yeah.

Speaker 4

这是一个被一些AI检测工具使用但Pangram不使用的指标。

So this is a metric that's used by some AI detectors but not Pangram.

Speaker 3

好的。

Okay.

Speaker 4

我可以解释一下它是如何工作的。

And so I can explain a bit about how it works.

Speaker 4

所以困惑度基本上是一种衡量标准

So perplexity is basically a measure

Speaker 3

这并不是perplexity.ai那个网站。

of And this is not perplexity.ai, the website.

Speaker 3

这是一个技术术语。

This is a technical term.

Speaker 3

好的。

Okay.

Speaker 4

很好。

Good.

Speaker 4

这是一个指标。

This is a metric.

Speaker 4

这是衡量一段文本对语言模型而言有多令人困惑的指标。

This is a measure of how confusing a piece of text is to a language model.

Speaker 4

基本上,对于每个词元,我们都可以计算出某种困惑度,也就是这个词元有多预期。

Basically, for example, with every token, we can calculate some perplexity, which is basically how expected this is.

Speaker 4

例如,如果句子是‘我回家去看我的宠物’,下一个词元是‘龙猫’,那么这个词元的困惑度会比‘我的宠物狗’高得多。

For example, if it's I went home to my pet and then the next token is chinchilla, that would be a much higher perplexity token than my pet dog.

Speaker 4

因此,低困惑度的文本,或者说大语言模型的输出,通常具有低困惑度。

So low perplexity texts or really LLM outputs, tend to be low perplexity.

Speaker 4

它们不会生成让自己感到意外的输出。

They're not going to produce outputs that are surprising to themselves.

Speaker 4

这是一种相当不错的方法,可以实现约90%到95%准确率的AI检测器,但它存在一些问题。

This is a decent way to get an AI detector that's around 90% to 95% accurate, but it has some problems.

Speaker 4

最主要的问题是,你无法在此基础上进一步改进。

The main one is that you can't improve upon it.

Speaker 4

基本上,它会产生误报。

Basically, it has false positives.

Speaker 4

非英语母语者写的文章通常 perplexity 很低,因为当你在

Text written by non native English speakers often is low perplexity just because when you're

Speaker 3

学习时,他们不会冒太多风险。

learning They don't take as many risks.

Speaker 3

没错。

Exactly.

Speaker 4

我很感兴趣。

I'm interested.

Speaker 4

是的

Yeah.

Speaker 4

这就是为什么许多早期的AI检测器对非母语英语使用者产生了大量误报。

So that's why a lot of the early AI detectors had a bunch of false positives with ESL speakers.

Speaker 4

因为他们的文本具有低困惑度。

It's because their text was low perplexity.

Speaker 4

所以我认为这是一个非常棒的指标,但它并不是Pangram的路径。

So I think this is a very cool metric, but it is not the path for Pangram.

Speaker 4

相反,我们采用了深度学习的方法,以便能够做得更好

Instead, we went the deep learning approach so we can do better than

Speaker 3

那什么是突发性?

And what's burstiness?

Speaker 3

这仅仅是另一面吗?

Is that just the opposite side of the coin?

Speaker 4

是的

Yeah.

Speaker 4

突发性实际上我也不知道该怎么定义它。

Burstiness is basically actually, I don't know if I can define it.

Speaker 3

好的。

Okay.

Speaker 3

行。

Fine.

Speaker 3

是的。

Yeah.

Speaker 3

没问题。

All good.

Speaker 0

突发性听起来就像是那种男性社群术语,对吧?

Burstiness just sounds like one of those of, I guess, manosphere terms, doesn't it?

Speaker 3

哦,对。

Oh, yeah.

Speaker 0

他一直在用高突发性进行smaxing

He's been look smaxing with high burstiness

Speaker 3

或者类似什么‘突发性 Mug’的东西。

or something like Burstiness mug.

Speaker 3

是的。

Yeah.

Speaker 3

那太好了。

That's great.

Speaker 4

是的。

Yeah.

Speaker 4

我认为这可能只是衡量句子长度以及文本起伏的一种方式。

I think it might just be a measure of sentence length and how the ups and downs of the text.

Speaker 0

但如果我们假设世界普遍关注AI垃圾,并希望对此采取行动,那么在互联网经济、监管或技术(比如你正在开发的东西)方面,最重大的单一改变会是什么,才能真正帮助减少垃圾内容?

But if we assume that the world is collectively concerned about AI slop and wants to do something about it, what would be, like, the single biggest change to the system, either in terms of, like, the economics of the Internet or regulation or technology, like what you're developing, that would actually help reduce slop?

Speaker 4

是的。

Yeah.

Speaker 4

我认为最重要的是规范。

I think the biggest one is norms.

Speaker 4

已经有一些很棒的博客文章讨论了,未经告知就将AI生成的内容发给别人是一种不礼貌的行为。

There have been a couple great blog posts written about how it is rude to send other people undisclosed AI outputs.

Speaker 4

我认为我完全同意这一点。

And I think I completely agree here.

Speaker 4

我觉得如果有人在互联网上提问,另一个人却去用ChatGPT搜索答案再粘贴回去,这有点不礼貌。

I think if somebody asks a question on the internet and then somebody else goes and puts into ChatGPT and then paste the answer, It's kind of rude.

Speaker 4

我当时是想问问我的朋友或关注者们的看法,而不是去问ChatGPT。

I was going here to ask the opinions of my friends or my followers, not ChatGPT.

Speaker 4

我自己本来就可以做这件事。

I could have done that myself.

Speaker 4

因此,我认为在这样一项新技术出现时,建立这种规范非常重要,我们需要快速行动,但我相信这对社会会有很大帮助。

And so I think building this norm is something that it's very new technology, so we need to do it quickly, but I think this would help a lot for society.

Speaker 3

那么,这实际上引出了我一个疑问:我觉得主要的互联网平台实际上正在朝完全相反的方向发展。

Well then, actually, this gets to a question that I have then, which is I feel as though the major Internet platforms are actually moving the exact opposite direction.

Speaker 3

我的意思是,我感到非常震惊。

I mean, I'm stunned.

Speaker 3

也许我曾经不小心点到了什么,但我经常收到邮件,打开Gmail准备回复时,却看到那里有一段幽灵文字,问:‘你只是想让Gemini来回复这封邮件吗?’

Maybe I accidentally clicked on something at some point, but the frequency with which I get an email and then I open it up to respond in Gmail, and there's that ghost text there that, I do you just want Gemini to respond to this?

Speaker 3

我从未这么做过。

I've never done that.

Speaker 3

我也认为,那样做会极其无礼。

I also consider I think that would be extremely rude.

Speaker 3

我从未用AI生成的回复来回复过任何邮件,但这些系统却在暗示你这么做。

I've never responded to any email with AI response, but they're basically telling you to do that.

Speaker 3

它们的做法恰恰相反。

They're doing the exact opposite.

Speaker 3

它们正在破坏这些准则。

They're blowing up these norms.

Speaker 3

所以我想知道,从你的角度来看,你提到你和Quora合作,但根据你的观察,主要的互联网平台是否认为这是一个值得解决的问题?

And so I'm curious from your perspective, you mentioned you work with Quora, But from your impression, do the major Internet platforms think this is a problem worth solving?

Speaker 3

还是说,它们的顾虑是:‘你知道吗?’

Or from their concern is like, you know what?

Speaker 3

内容越多越好。

The more content, the better.

Speaker 0

对他们来说,这里的激励是混合的。

There's mixed incentives there for them.

Speaker 4

这很有趣,因为谷歌似乎在两头下注。

It's funny because Google seems to be playing both sides.

Speaker 4

一方面,他们做过一个广告,引发了很多争议,广告里说孩子们现在可以通过AI发送他们对英雄的敬意信件,而不是自己亲手写信。

On one hand, they had that advertisement which people blew up about where it's like, Oh, children can now send their heroes notes on how much they respect them by using AI instead of writing the note themselves.

Speaker 4

我觉得这不对。

I'm like, This is wrong.

Speaker 4

这在社会层面上是糟糕的。

This is societally bad.

Speaker 4

但与此同时,他们正在努力解决互联网搜索结果中的AI垃圾内容,确保用户获得的是真实内容,而不是AI生成的垃圾内容。

But at the same time, they're working very hard to deal with the AI slop on the internet in search results to make sure people get served real content and not AI slop content.

Speaker 4

所以我认为,显然有很多激励因素在影响产品人员,他们被推动去推广AI,因为这是公司的使命。

So I think I mean, I think obviously there's a lot of incentives that play around product people who are incentivized to push AI because that is the corporate mandate.

Speaker 4

但我觉得,总的来说,即使在我这一圈AI研究人员中,普遍共识也是AI是个强大的工具,但垃圾内容很糟糕。

But, yeah, I think overall, even in my sphere of a bunch of people who are AI researchers, generally consensus is that AI is a powerful tool, but slop is bad.

Speaker 0

这让我想起一件事。

This reminds me.

Speaker 0

我父母以前总让我亲手制作圣诞贺卡,给所有亲戚朋友。

My parents used to make me do these handmade greeting cards for Christmas, for all relatives and stuff.

Speaker 0

这本意是让我展示自己用心沟通的诚意。

And it was supposed to be a demonstration of my commitment to communicating.

Speaker 0

我觉得这挺好的。

Think that's great.

Speaker 0

但它让我留下了心理阴影。

It traumatized me forever.

Speaker 0

因此我特别讨厌贺卡,就是因为以前花几个小时手工制作这些东西。

And I hate greeting cards as a result of them, of doing this, just spending hours manufacturing these things.

Speaker 0

但更搞笑的是,后来我们有了电子贺卡,我父母立刻就改用电子贺卡了。

But then secondly, the funniest thing was once we got ecards, my parents immediately switched to using ecards.

Speaker 0

而且现在这事儿也特别搞笑。

Just and now this is also the funniest thing.

Speaker 0

我爸爸用电子贺卡,他发现电子贺卡系统能告诉他你有没有打开过。

My dad uses ecard he figured out that the ecard system can tell him whether or not you opened it.

Speaker 0

所以他现在就用它来日常沟通了。

So he just uses it as, like, day to day communication now.

Speaker 3

这太搞笑了。

That's so funny.

Speaker 3

我刚给你女儿发了一封邮件,我是通过电子贺卡发的。

I just sent an email to your daughter, and I do it via e card.

Speaker 0

你看,我都注意到你还没打开我发的国际热狗日电子贺卡。

It's like, noticed you haven't opened up my e card for International Hot Dog Day.

Speaker 0

麻烦告诉我你最近怎么样。

Please, let me know what's going on.

Speaker 3

我小时候字写得特别差,我妈就让我亲手写所有感谢信,感谢别人在我成年礼上送的礼物。

I did terrible handwriting as a kid, and my mother made me write all of these handwritten notes to thank people for the gifts I got for my bar mitzvah.

Speaker 3

是的

Yeah.

Speaker 3

我讨厌那样。

I hated it.

Speaker 3

但你知道吗?

But you know what?

Speaker 3

但你知道吗?我与那些人建立了深厚的联系,这些联系一直延续至今。虽然那痛苦的一周里我手写到手腕酸痛,但我觉得这一切都值得。

I have deep connections with all of those people that have lasted And over the that miserable one week where I just rode and I got, you know, hand cramped, I think it paid off.

Speaker 0

好吧。

Alright.

Speaker 0

想象一下,差不多十六年里,一直不停地做这种事,永无止境。

Well, imagine doing that for, like, sixteen years, basically, in a never ending stream.

Speaker 3

马克斯·斯佩罗,非常感谢你做客《Odd Lots》。

Max Spero, thank you so much for coming on Odd Lots.

Speaker 3

这真是太有趣了。

That was a lot of fun.

Speaker 3

我对这场对话非常着迷。

I'm fascinated by this conversation.

Speaker 4

非常感谢你邀请我。

Thanks so much for having me.

Speaker 4

是的。

Yeah.

Speaker 4

讨论这个话题真的很令人兴奋,我认为垃圾信息正在成为一个日益严重的问题。

Really exciting to talk about this, and I think Slop is a growing problem.

Speaker 4

所以希望我们能做得很好。

So hopefully Awesome.

Speaker 4

我们能够应对这个问题。

We're able to deal with it.

Speaker 0

互联网上有40%的内容。

40% of the Internet.

Speaker 0

我不确定我是对这个数字感到惊讶,还是不惊讶。

I can't tell if I'm surprised by that or not.

Speaker 3

那么明年这个时候会变成什么样呢?

And what's it gonna be next year at this time?

Speaker 4

天哪。

Oh, man.

Speaker 4

我不知道。

I don't know.

Speaker 4

可能会这样,很难说。

It'll be, like Hard to say.

Speaker 3

占多数。

Over a majority

Speaker 4

肯定大部分都是。

of them for sure.

Speaker 4

是的。

Yeah.

Speaker 4

几乎可以肯定。

Almost certainly.

Speaker 4

太疯狂了。

Crazy.

Speaker 3

好的。

Alright.

Speaker 3

谢谢你参加《Odd》节目。

Thanks for coming on Odd

Speaker 4

很多。

Lots.

Speaker 4

谢谢。

Thanks.

Speaker 3

特蕾西,我喜欢这场对话。

Tracy, I love that conversation.

Speaker 3

我只是觉得这就像一个非常有趣的谜题。

I just think it's like a really fun puzzle.

Speaker 3

对吧?

Right?

Speaker 0

不。

No.

Speaker 0

完全正确。

Totally.

Speaker 3

这看起来像是一个很有趣的谜题要解决。

It's very like, it seems like a fun question to solve.

Speaker 3

我对这个想法着迷:无论是人类还是人工智能,我们所知道的和我们能够表达出来的之间,必然存在一道鸿沟。

And I'm fascinated by this idea of how, like, with both humans and AI, there is gonna be this gap inevitable between what we know and what we can articulate.

Speaker 0

因为你

Because you

Speaker 3

我和你,撇开人工智能和文本不谈,我们都知道一些事情。

and I, setting aside AI versus text, there are things that we both know.

Speaker 3

例如,这个有新闻价值,那个没有。

For example, this is newsworthy, this isn't.

Speaker 3

这是一个不错的播客节目。

This is a good episode of a podcast.

Speaker 3

这个不是。

This isn't.

Speaker 3

这个听起来像是个可信的嘉宾,而这个不是。

This is a credible sounding guest, and this isn't.

Speaker 3

这种认知和能够解释清楚其中原因之间的差距,就像是你只是隐约知道而已。

The gap between that and then being able to explain why, it's like, well, you just sort of know it.

Speaker 3

对吧?

Right?

Speaker 3

你就是有种这样的感觉。

You just sort of have this feeling.

Speaker 0

这是一种直觉。

There's an intuition.

Speaker 3

是的。

Yeah.

Speaker 3

这种直觉是通过大量例子积累起来的,某种程度上,这和AI的训练方式是一样的。

And that intuition is built up from numerous examples, which is the same way in a sense that, like, the AI is trained.

Speaker 3

就像这些你只能从模式中感知到的东西,你能看到它们,却无法完全清晰地表达出究竟发生了什么。

It's like these things that you only know from patterns, and you can see them without fully being able to, like, articulate exactly what's going on.

Speaker 0

那么,我对这一点的另一个疑问是,从长远来看,这真的重要吗?毕竟,互联网的很大一部分已经建立在机器人和虚假注意力经济之上。

Well, the other question I would have on that is, is it even gonna matter in the long run if you think about, like, much of the Internet is already built on bots and the sort of, like, false attention economy?

Speaker 0

如果我们的整个世界观都被人工智能驱动的垃圾内容所塑造,

Like, if our entire, like, worldview becomes shaped by AI driven drivel

Speaker 3

是的。

Yeah.

Speaker 0

如果互联网的经济模式仍然依赖于单个机器人账号之类的东西,那也无所谓了。

Doesn't matter if, like, the economics of the Internet are still attached to individual bot accounts and things like that.

Speaker 0

我不确定我有没有把这解释清楚。

I don't know if I'm if I'm explaining this.

Speaker 4

但是

But

Speaker 3

不。

No.

Speaker 3

不。

No.

Speaker 3

我觉得这很有道理。

I think it makes a lot of sense.

Speaker 3

而且我认为,确实很重要,我们必须彻底改变我们的思维方式,就像马克斯开头说的那样,我也一直在思考这个问题:过去,如果你看到一篇文字,标点符号非常规范,拼写也很完美,读起来逻辑清晰,你就会想,嗯,这肯定是聪明人写的。

And I do think, like, it is important like, we're gonna have to change the entire way we think, and Max said at the beginning, which is and I've thought about this, which is that it used to be that if you came across a piece of writing and the punctuation was excellent Mhmm.

Speaker 3

标点和拼写都很出色,读起来也很有条理,你就会觉得,好吧,

And the spelling was excellent and it was, like, cogent sounding, you're like, okay.

Speaker 3

这篇文章肯定是聪明人写的。

I this has been written by a smart person.

Speaker 3

我会认真对待它的内容。

I will take the content seriously.

Speaker 3

对吧?

Right?

Speaker 3

但现在,技巧和输出之间完全脱节了,因为你确实可以做到这一点。

Now there is this complete severance of sort of, like, craft and output because you could and you do this.

Speaker 3

让克劳德写一篇论证,支持最荒谬的主张——比如,为我论证里根在20世纪80年代初推行减税的原因,与70年代的UFO目击报告有关。

Ask Claude to write an argument in favor of the most absurd proposition Ask Claude to write an argument for me that the reason why Reagan wanted to do tax cuts in the early 1980s related to these reports of UFO sightings in the 1970s.

Speaker 3

而它写出的内容不仅语法正确,还会竭力构建出这个论点最完美的版本。

And it will write something that not only is it grammatically correct, it'll actually strain to come up with the best version of this argument before.

Speaker 3

而且,如果你在读完之前还觉得,也许这个人真的认真对待了这个论点,但现在这个论点完全是凭空捏造的,我们就必须彻底改变对这类事情的判断标准。

And again, if prior to that having read it, like, oh, maybe the person like, this person took this argument seriously, but now this argument is just created ex nihilo, we're gonna have to really, like, change our heuristics about this stuff.

Speaker 0

我们创造出了无限多语法极好却纯粹是怪人的声音。

We've created an unlimited stream of basically cranks with really good grammar.

Speaker 3

是的。

Yeah.

Speaker 3

没错。

That's right.

Speaker 3

没错。

That's right.

Speaker 3

因为过去我们能认出怪人,是因为他们语法很差,或者会给我们发邮件,一半单词是黄色的,另一半还带下划线。

Because it used to be we knew the crank because they had bad grammar, or they would email us, and, like, half the words would be in yellow and the other half would be underlined.

Speaker 0

绿色墨水是经典的例子。

Green ink was the classic example.

Speaker 3

我们用来判断一个人是怪人的那些工具,

The tools that we use to just, like, oh, this person's crank.

Speaker 3

比如,他们一半的单词都是大写,诸如此类。

They, like, you know, half their words are in all caps and stuff like that.

Speaker 3

这些方法现在不管用了。

Those don't work anymore.

Speaker 0

好的。

Alright.

Speaker 0

说到这个,我们就这样结束吧?

On that note, shall we leave it there?

Speaker 3

我们就到这里吧。

Let's leave it there.

Speaker 0

以上就是本期《奇怪想法播客》的全部内容。

This has been another episode of the Odd Thoughts Podcast.

Speaker 0

我是特蕾西·阿拉韦。

I'm Tracy Allaway.

Speaker 0

你可以关注我,账号是Tracy Allaway。

You can follow me at Tracy Allaway.

Speaker 3

我是乔·维森塔尔。

And I'm Joe Wiesenthal.

Speaker 3

你可以关注我,账号是The stalwart。

You can follow me at The stalwart.

Speaker 3

关注我们的嘉宾马克斯·斯佩罗。

Follow our guest, Max Spero.

Speaker 3

他的账号是max_spero_。

He's at max underscore Spero underscore.

Speaker 3

关注我们的制作人:卡门·罗德里格斯,账号是CarmenArmen;乔·贝内特,账号是Dashbot;以及凯尔·布鲁克斯,账号是Kale Brooks。

Follow our producers, Carmen Rodriguez at Carmen Armen, dash Joe Bennett at Dashbot, and Kale Brooks at Kale Brooks.

Speaker 3

如需获取更多Odd Lots内容,请访问bloomberg.com/oddlots,那里有我们所有节目的每日简报,你还可以在我们的Discord频道discord.gg/oddlots中24小时不间断地讨论这些话题。

And for more Odd Lots content, go to bloomberg.com/oddlots where have a daily newsletter on all of our episodes, and you could chat about all of these topics twenty four seven in our Discord, discord.gg/oddlots.

Speaker 0

如果你喜欢《Odd Lots》这个节目,喜欢我们讨论互联网有40%是乱七八糟的内容,那就请在你最喜欢的播客平台上给我们一个好评。

And if you enjoy Odd Lots, if you like it when we talk about how the Internet is 40% slopped, then please leave us a positive review on your favorite podcast platform.

Speaker 0

记住,如果你是彭博的订阅用户,你可以免费收听我们所有的节目,没有任何广告。

Remember, And if you are a Bloomberg subscriber, you can listen to all of our episodes absolutely ad free.

Speaker 0

你只需要在Apple Podcasts上找到彭博频道,然后按照那里的说明操作即可。

All you need to do is find the Bloomberg channel on Apple Podcasts and follow the instructions there.

Speaker 0

感谢收听。

Thanks for listening.

Speaker 10

我是弗朗辛·拉克鲁瓦,一位获奖记者,我推出了一档新播客《弗朗辛·拉克鲁瓦的领袖》,由彭博播客出品。

I'm Francine Lacroix, an award winning journalist, and I've got a new podcast, Leaders with Francine Lacroix from Bloomberg Podcasts.

Speaker 10

我采访过从国家元首到时尚偶像等各种人物,探讨当下新闻,但我一直很好奇,这些人在作为领导者时究竟是怎样的。

I've interviewed everyone from heads of state to fashion icons about the news of the moment, but I've always been curious who are these people as leaders.

Speaker 0

我认为并没有一种

I don't think there's one

Speaker 9

正确的领导方式。

right way to be a leader.

Speaker 4

做出决定。

Make decisions.

Speaker 4

一个糟糕的决定总比不做决定要好。

A poor decision is always better than no decision.

Speaker 10

每隔一周的星期一收听新一期节目。

Listen to new episodes every other Monday.

Speaker 10

在您收听播客的平台关注《Francine Lacroix 的领导者》。

Follow leaders with Francine Lacroix wherever you get your

Speaker 6

播客。

podcasts.

关于 Bayt 播客

Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。

继续浏览更多播客