Invest Like the Best with Patrick O'Shaughnessy - 加文·贝克——人工智能、半导体与机器人前沿 - [像最佳投资者一样投资,第385期] 封面

加文·贝克——人工智能、半导体与机器人前沿 - [像最佳投资者一样投资,第385期]

Gavin Baker - AI, Semiconductors, and the Robotic Frontier - [Invest Like the Best, EP.385]

本集简介

本周的嘉宾是加文·贝克。加文是Atreides Management的管理合伙人兼首席投资官,此前已多次做客本节目。他是我最喜爱交流的投资人之一,而这次对话或许是我与他最精彩的一次。早在世纪之交,加文就开始以投资者身份关注英伟达,这使他成为探讨人工智能与投资话题的绝佳人选。本次对话信息量极大,我由衷感谢加文再次与我们分享智慧。请尽情享受这场与加文的精彩对话。 完整节目笔记、文字记录及相关内容链接,请访问本期节目页面。 ----- 本期节目由Ramp赞助播出。Ramp致力于帮助企业优化支出管理,既降低成本又为团队腾出时间投入更高价值项目。作为史上增长最快的金融科技公司,Ramp获得了至少16位我曾采访过的顶尖投资人支持——这个数字可能超过我所知的任何其他公司。值得注意的是,许多行业标杆企业都在使用Ramp,包括Airbnb、Anduril、Shopify等公司,以及红杉资本、Vista Equity等投资机构。他们通过Ramp管理支出、自动化繁琐财务流程,并将节省的资金与时间重新投入增长。Colossus和Positive Sum使用Ramp也正是基于同样原因。登录Ramp.com/invest免费注册即可获得250美元迎新礼金。 ----- 本期节目由Tegus赞助播出。我们正在重塑投资研究领域的游戏规则。告别过时低效的研究方式,拥抱搭载10万+份文字记录的创新平台——仅今年新增记录就超过2.5万份。我们的平台内容增速是对手的8倍,月新增量达2倍,稳居行业前沿。更值得一提的是,Tegus独家提供75%的私募市场访谈记录,为您带来无可替代的深度洞察。体验海量优质文字记录库带来的变革,即刻登录tegus.com/patrick申请免费试用。 ----- 《像顶尖投资人一样思考》是Colossus, LLC旗下节目。更多往期内容请访问joincolossus.com/episodes。历史嘉宾包括Tobi Lutke、Kevin Systrom、Mike Krieger、John Collison、Kat Cole、Marc Andreessen、Matthew Ball、Bill Gurley、Anu Hariharan、Ben Thompson等众多大咖。订阅Colossus周刊(每周日发送),即可获取我们所有播客的最新动态,精选当周节目中的商业投资精华及优质书单。点此注册。关注我们的推特账号:@patrick_oshag | @JoinColossus。本期节目后期制作由The Podcast Consultant (https://thepodcastconsultant.com)完成。 节目时间轴: (00:00:00) 欢迎收听《像顶尖投资人一样思考》 (00:04:42) "七巨头"与科技行业竞争格局 (00:06:29) 生成式AI与规模法则 (00:08:36) AI基础设施面临的挑战 (00:15:02) AI与数据中心的未来展望 (00:17:51) AI模型的能效优化 (00:35:14) 合成数据与AI训练 (00:42:37) 推理环节与智能手机的角色 (00:48:35) AI领域的投资启示 (00:49:09) 新兴企业的发展机遇 (00:51:20) 应用层面临的挑战 (00:52:25) AI对广告业的影响 (00:53:40) AI投资回报率争议 (00:54:39) SaaS指标与AI颠覆效应 (00:55:59) AI优先的应用型公司 (01:00:50) 机器人技术的未来 (01:14:01) 科技巨头的领导力 (01:24:05) 投资理念的演进

双语字幕

仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。

Speaker 0

我在《像最好的投资者一样投资》中经常谈论的一个概念是人生事业。

Something I speak about frequently on Invest Like The Best is the idea of life's work.

Speaker 0

更有趣的一种理解方式是,我在寻找那些执着于使命的狂热者。

A more fun way to think about it is that I'm looking for maniacs on a mission.

Speaker 0

这正是我们投资公司Positive Sum的核心理念,也是我对我们的赞助商Ramp如此热情的原因。

This is the basis for our investment firm, Positive Sum, and it's the reason why I'm so enthusiastic about our presenting sponsor, Ramp.

Speaker 0

不仅创始人Kareem和Eric是真正的人生事业型创业者,无疑是执着于使命的狂热者,他们还打造了一款产品,通过简化财务运营,帮助创业者和财务团队腾出更多时间去做他们的人生事业,从而节省了每个人最宝贵的资源——时间。

Not only are the founders, Kareem and Eric, life's work level founders, certainly maniacs on a mission, they have created a product that is effectively an unlock for founders and finance team to do more of their life's work by streamlining financial operations, saving everyone their most precious resource, time.

Speaker 0

Ramp构建了一个企业信用卡和费用管理的指挥控制系统。

Ramp has built a command and control system for corporate cards and expense management.

Speaker 0

你可以在一个平台上发放卡片、管理审批、支付各种供应商款项,甚至自动化完成财务结账。

You can issue cards, manage approvals, make vendor payments of all kinds, and even automate closing your books all in one place.

Speaker 0

根据我本人使用Ramp经营业务的经验,这款产品极其直观、简洁,让生活变得轻松许多,让你不禁为那些尚未切换到Ramp的公司感到遗憾。

Speaking from my own experience using Ramp for my business, the product is wildly intuitive, simplistic, and makes life so much easier that you'll feel bad for any company who hasn't yet made the switch.

Speaker 0

Ramp团队无比执着,产品持续进化,为你节省下你从未想象过能重新获得的时间。

The Ramp team is relentless, and the product continues to evolve to save you time that you would never have dreamed of getting back.

Speaker 0

在我看来,没有什么比那些能减少其他创业者构建他们想要之物的障碍的技术更有趣的了。

To me, there is nothing more interesting than technologies that reduce friction for other entrepreneurs to be able to build the thing that they want to.

Speaker 0

太多注意力都集中在云计算、API 以及其他让创业者生活更轻松的方式上。

So much attention has gone to cloud computing, APIs, and other ways of making life easy for founders.

Speaker 0

Ramp 所做和正在做的,是在这个类别中构建另一套工具。

What Ramp has done and is doing is build yet another set of tools in this category.

Speaker 0

要开始使用,请访问 ramp.com。

To get started, go to ramp.com.

Speaker 0

卡片由凯尔特银行和萨顿银行发行,均为联邦存款保险公司(FDIC)成员。

Cards issued by Celtic Bank and Sutton Bank, member FDIC.

Speaker 0

适用条款和条件。

Terms and conditions apply.

Speaker 1

大家好,欢迎各位。

Hello, and welcome, everyone.

Speaker 1

我是帕特里克·奥肖内西,欢迎收听《像最好的人一样投资》。

I'm Patrick O'Shaughnessy, and this is Invest Like the Best.

Speaker 1

这档节目是对市场、理念、故事和策略的开放式探索,旨在帮助你更好地投资你的时间和金钱。

This show is an open ended exploration of markets, ideas, stories, and strategies that will help you better invest both your time and your money.

Speaker 1

《像最好的投资者一样投资》是Colossus播客家族的一部分,你可以在joincolossus.com上访问我们所有的播客,包括编辑过的文字稿、节目笔记和其他学习资源。

Invest like the best is part of the Colossus family of podcasts, and you can access all our podcasts, including edited transcripts, show notes, and other resources to keep learning at joincolossus.com.

Speaker 2

帕特里克·奥肖内西是Positive Sum的首席执行官。

Patrick O'Shaughnessy is the CEO of Positive Sum.

Speaker 2

帕特里克和播客嘉宾表达的所有观点均为他们个人的观点,不代表Positive Sum的立场。

All opinions expressed by Patrick and podcast guests are solely their own opinions and do not reflect the opinion of Positive Sum.

Speaker 2

本播客仅作信息参考,不应作为投资决策的依据。

This podcast is for informational purposes only and should not be relied upon as a basis for investment decisions.

Speaker 2

Positive Sum的客户可能持有本播客中讨论的证券。

Clients of positive sum may maintain positions in the securities discussed in this podcast.

Speaker 2

如需了解更多信息,请访问psum.vc。

To learn more, visit psum.vc.

Speaker 1

本周的嘉宾是加文·贝克。

My guest this week is Gavin Baker.

Speaker 1

加文是Atrades Management的管理合伙人兼首席投资官,他之前多次做客本节目。

Gavin is the managing partner and CIO of Atrades Management, and he has been on the show many times before.

Speaker 1

他是我最喜欢交谈的投资者之一,这可能是我与他最喜爱的一次对话。

He is one of my favorite investors to talk to, and this may be my favorite conversation with him.

Speaker 1

加文早在世纪之交就开始以投资者身份关注英伟达,因此他是讨论所有人工智能和投资话题的完美嘉宾。

Gavin first started covering NVIDIA as an investor at the turn of the century, making him the perfect guest to discuss all things AI and investing.

Speaker 1

这次对话包含大量细节,我非常感谢加文再次与我们分享他的智慧。

There's so much detail in this conversation, and I'm incredibly grateful to Gavin for sharing his wisdom with us again.

Speaker 1

请欣赏加文·贝克的这场精彩对话。

Please enjoy this fantastic conversation with Gavin Baker.

Speaker 1

好了,加文。

Alright, Gavin.

Speaker 1

尽管我们私下交流的频率比这高得多,但事实上我们已经好几年没在节目中这样对话了。

You and I have actually not done this in years even though we do it offline much more frequently than that.

Speaker 1

所以我很高兴能再次在录音中进行这次对话。

So I'm really excited to do it again on the record.

Speaker 1

我有一份想和你讨论的50件事清单,看看我们能聊多少。

I have a list of 50 things I wanna talk to you about, so we'll see how many we get to.

Speaker 1

但一个很有趣的开场框架是我看到你之前发布的一个观点:1960年,《七侠荡寇志》由尤·伯连纳领衔,最终七人中只有三人活过了枪战。

But a really fun opening framing was something I saw you put out into the world, which was back in 1960, the magnificent seven were led by Yul Brenner, and, ultimately, only three of the seven survived the shootout.

Speaker 1

今天我们也有了一支新的‘七巨头’,我们几乎可以花整个时间来讨论它们。

We've got we've got a new mag seven today, and we could probably spend the whole time talking about them.

Speaker 1

但我们不会。

We won't.

Speaker 1

不过我觉得,不妨从你对这些巨头公司为何如今可能正陷入某种商业混战的见解开始,谈谈它们与当今科技界诸多变化的关系。

But I thought it would be a fun opening moment just to hear you riff on why these massive companies might actually be in some form of business shootout now with a lot of what's happening in the world of technology.

Speaker 3

我认为这些公司长期以来都各自拥有独立的业务领域,彼此是竞争关系。

So I think these companies were all in their own discrete swim lanes for a long time, competitive swim lanes.

Speaker 3

它们真正重叠的地方只有云计算,当时谷歌、亚马逊和微软都在竞争,但那是一个非常稳定的寡头垄断格局。

The only place where they really overlapped was cloud computing, where you had Google, Amazon, and Microsoft all competing, but that was a very stable oligopoly.

Speaker 3

谷歌在2014年或2015年左右 aggressively 降低了价格。

Google cut prices aggressively, something like 2014, 2015.

Speaker 3

亚马逊也跟进降价了。

Amazon matched.

Speaker 3

这实际上严重抑制了亚马逊的收入增长,并引发了人们对云计算将沦为商品的广泛担忧——这曾是一个真实且激烈争论的话题,但现在看来简直荒谬。

And that actually materially impaired Amazon's revenue growth and fed into all of these fears that cloud computing was gonna be a commodity, which was a real fear, hotly debated topic, looks deeply ridiculous now.

Speaker 3

但我认为这为一种局面铺平了道路:嘿。

But I think that set the stage for like, hey.

Speaker 3

我们三家公司基本上会在成本基础上达成某种利润率共识,然后在其他方面寻求差异化。

We're effectively gonna agree between the three of us on some markup on cost, but then try to differentiate in other ways.

Speaker 3

亚马逊拥有电子商务。

Amazon had e commerce.

Speaker 3

Facebook拥有广告业务,其位置比谷歌更靠上层漏斗。

Facebook had advertising that was, know, higher up the funnel from Google.

Speaker 3

谷歌拥有搜索业务。

Google had search.

Speaker 3

苹果则显然拥有设备、操作系统和App Store。

Apple obviously had the device in the OS and the App Store.

Speaker 3

你知道的,来自安卓的轻微竞争。

You know, minor competition from Android.

Speaker 3

Netflix正在做流媒体视频。

Netflix is doing streaming video.

Speaker 3

而且,这里可能也有一点与谷歌的竞争。

And, again, a little bit of competition with Google there, maybe.

Speaker 3

但它们的竞争领域相当明确。

But they're pretty distinct competitive sets.

Speaker 3

微软虽然拥有云业务,但也拥有庞大的企业软件业务,专注于生产力工具。

Microsoft, while they had cloud, they also had this massive enterprise software business really focused on productivity.

Speaker 3

至于gin.ai,也就是LLM,gin.ai的意思是生成式AI,但GPT中的G代表通用。

And just with gin.ai, with LLMs to the name, I mean, gin dot ai means generative AI, but the g in GPT means general purpose.

Speaker 3

这是一种如此通用的技术,以至于它们突然间进入了同一个竞争领域,所有人都觉得这关乎生死存亡。

It's such a general technology that they're all of a sudden in the same swim lane, and they all feel like it's existential.

Speaker 3

马克·扎克伯格、萨提亚和孙达尔以不同方式告诉你,我们甚至根本不考虑投资回报率。

Mark Zuckerberg, Satya, and Sundar just told you in different ways, we are not even thinking about ROI.

Speaker 3

他们之所以这么说,是因为真正控制这些公司的那些人——比如创始人——要么拥有超级投票权股票,要么像微软那样拥有重大影响力,他们相信自己正在竞相创造一个数字之神。

And the reason they said that is because the people who actually control these companies, you know, the founders, there's either supervoting stock or significant influence in the case of Microsoft, believe they're in a race to create a digital god.

Speaker 3

如果你率先创造了这个数字之神,我们可以争论它的价值是数万亿还是数十万亿美元,也可以争论这种想法是否荒谬。

And if you create that first digital god, we can debate whether it's tens of trillions or hundreds of trillions in value, and we can debate whether or not that's ridiculous.

Speaker 3

但这就是他们所相信的。

But that is what they believe.

Speaker 3

他们相信,如果在这场竞赛中失败,就意味着对公司构成生存威胁。

And they believe that if they lose that race, losing the race is an existential threat to the company.

Speaker 3

因此,拉里·佩奇显然曾在谷歌内部多次表示:我宁愿破产,也不愿输掉这场竞赛。

So Larry Page has evidently said internally Google many times, I am willing to go bankrupt rather than lose this race.

Speaker 3

所以,人人都在关注投资回报率这个等式,但做决策的人却并不在意,因为他们坚信规模定律将继续有效。

So everybody's really focused on this ROI equation, but the people making the decisions are not because they so strongly believe that scaling laws will continue.

Speaker 3

关于这些涌现特性是否只是上下文学习等现象,目前还存在巨大争议。

And there's a big debate over whether these emergent properties are just in context learning, etcetera, etcetera.

Speaker 3

但他们相信,规模定律将继续发挥作用。

But they believe scaling laws are gonna continue.

Speaker 3

模型会变得更好、更强大,推理能力也会更强。

The models are gonna get better and more capable, better at reasoning.

Speaker 3

由于他们持有这种信念,他们会一直投入,直到出现无可辩驳的证据表明规模定律正在放缓。

And because they have that belief, they're going to spend until I think there is irrefutable evidence that scaling laws are slowing.

Speaker 3

而要获得无可辩驳的证据,为什么自从GPT-4发布以来,进展就变慢了?

And the only way you get irrefutable evidence, why has progress slowed down since g p two four came out?

Speaker 3

因为还没有推出新一代的NVIDIA显卡。

Because there hasn't been a new generation of NVIDIA GPUs.

Speaker 3

这是实现能力下一个真正跃升所必需的。

It's what you need to enable that next real step function change in capability.

Speaker 3

这实际上会很有趣。

It is actually gonna be interesting.

Speaker 3

因此,Blackwell的延迟与此息息相关。

So everyone the Blackwell delay plays into this.

Speaker 3

要创建一个由数万张GPU组成的连贯训练集群,真的非常非常困难。

It's really, really hard to create what's called a coherent trading cluster of tens of thousands of GPUs.

Speaker 3

‘一致’意味着每个GPU都可以知道其他GPU在想什么。

And coherent just means each GPU, we could say, knows what the other is thinking.

Speaker 3

从技术上讲,这更像是它们拥有一个共享的内存空间。

Technically, it's more like they have a shared memory space.

Speaker 3

要进行训练,这个集群必须保持一致。

And that cluster has to be coherent to train.

Speaker 3

直到最近,世界上最大的一致集群是32000个。

And the biggest coherent cluster in the world until very recently was 32,000.

Speaker 3

所以你可能有32000个H100,但由于效率问题,实际同时使用的可能只有最多一万五千到一万六千个H100。

So you had 32,000 h one hundreds probably only using, like, at most fifteen, sixteen thousand of those h one hundreds at once because of efficiency problems.

Speaker 3

但XAI决定要建造一个十万GPU的集群。

But x AI decided they were gonna build a 100,000 GPU cluster.

Speaker 3

我认为,凭借埃隆独特的物理工程思维——我们在SpaceX、特斯拉,以及现在Neuralink上已经看到过——他与非常有能力的团队一起,解决了如何将一切微型化的问题。

And I would say with Elon's unique physical engineering mind, which we've seen play out at SpaceX and Tesla, and now at Neuralink, where he figured out how to miniaturize everything along with really capable teams.

Speaker 3

我认为他在孟菲斯从第一性原理出发,重新设计了XAI的数据中心。

I think he rearchitected the x AI data center from first principles in Memphis.

Speaker 3

这与其他数据中心非常不同。

It's very different than other data centers.

Speaker 3

正因为如此,他们能够实现足够的密度,即使没有 NVIDIA、Broadcom 等公司的下一代网络技术,也能让 10 万个 GPU 的集群保持一致性。

And because of that, they're able to get enough density that they could effectively make a 100,000 GPU cluster coherent even with Hopper without next generation networking technologies from NVIDIA, Broadcom, and others.

Speaker 3

他们已经开始利用这一点。

And they started trading on that.

Speaker 3

这意味着我认为,我们可能会在接下来的六到九个月内,看到 GPT-4.5 级别的模型独立实现规模扩展,当然,具体时间我不确定。

And that means that I think we're gonna see probably, you know, scaling on its own, the first GPT four and a half class model sometime, I don't know, in the next six, nine months.

Speaker 3

我不知道具体时间。

I don't know the timing.

Speaker 3

之后,你就会看到 Blackwell。

And then after that, you'll have Blackwell.

Speaker 3

然后,由于有了下一代网络技术,集群的 GPU 数量将提升到 30 万个。

And then you'll go up to 300,000 GPUs in a cluster because you have next generation networking technologies that make that easier.

Speaker 3

这将是一个巨大的跃迁式变革。

And that is gonna be a massive step function change.

Speaker 3

然后,你就能得到类似GPT-5、5.5或6这样的模型。

And then that's where you could get maybe a GPT five, five and a half, six class model.

Speaker 3

而这就是它具有生存意义的原因:如果JIT AI如此通用,你拥有这种ASI时,你会想,在AI能生成远超人类的无限优质内容的世界里,内容的价值何在?

And then the reason it's existential is just like if JIT AI is so generalizable and you have this ASI, you're like, what's the value of content in a world where AI can make content that's infinitely better than any human?

Speaker 3

在这样一个世界里,Netflix的价值何在?当我可以说‘今晚我想看《星际迷航》和《星球大战》的混剪’时?

What's the value of Netflix in a world where I could say, I wanna watch a mashup of Star Trek and Star Wars tonight?

Speaker 3

搜索可能会消失,被智能代理取代。

Search might just go away and be replaced by agents.

Speaker 3

对此保持谦逊非常重要。

It's important to have a lot of humility about this.

Speaker 3

管理老虎基金的查斯·科尔曼是一位杰出的投资者,他提供了一个非常有趣的统计数据:ChatGPT于2022年推出,对AI而言,就如同1994年Netscape Navigator之于互联网。

Chase Coleman, who runs Tiger, who's an exceptional investor, had a really interesting statistic, which was ChatGeePity came out in 2022 and was to AI as NetScape Navigator, was to the Internet in 1994.

Speaker 3

在Netscape Navigator推出后的两年里,全球互联网市值中仅有不到1%是在那时创立的。

Only less than 1% of current global Internet market cap was founded.

Speaker 3

在Netscape Navigator推出后的两年里,全球互联网市值中仅有不到1%是在那时创立的。

Founded in the two years after Netscape Navigator came out.

Speaker 3

没人能想象到那些将会成立的公司。

And nobody could imagine the companies that were gonna be founded.

Speaker 3

所以,最大的公司都是在很多年之后,或者五六年之后才成立的。

So it's just the biggest companies were founded many years later or several years later, five or six years later.

Speaker 3

所以我们还处于非常早期的阶段,保持谦逊非常重要。

So it's just we're very early, important to have a lot of humility.

Speaker 3

但只要规模定律持续有效,而它们唯一可能被证伪的方式是出现新一代GPU,具备传统上三到五倍的性能提升,并且网络技术也得到改进,使得我们可以将三到五倍数量的GPU连接在一起。

But as long as scaling laws continue, and the only way they can be disproven is if you have a new generation GPU that has the traditional, whatever it is, three to five x performance improvement, and then you get better networking such that you can link three to five x more of them together.

Speaker 3

然后你获得了十倍的算力数量级提升,但模型质量却没有显著改善,那么规模定律就会停止,这将对整个数据中心基础设施造成灾难性后果。

And then you get that 10 x, the order of magnitude improvement in compute, and you don't see a massive improvement in model quality, then scaling laws will stop, and that'll be a catastrophe for the entire data center infrastructure.

Speaker 3

但所有近距离关注这一领域的人们都相信,规模定律将继续下去。

But the people close to this all believe that scaly laws are gonna continue.

Speaker 1

昨晚,我女儿从晚餐回家的路上对我说:嘿。

Last night, my daughter on the way home from dinner said to me, hey.

Speaker 1

你能问一下ChatGPT一个问题吗?

Can you ask ChatGPT some question?

Speaker 1

我说:你iPad上有ChatGPT吗?

And I said, do you have ChatGPT on your iPad?

Speaker 1

她说:没有。

And she said, no.

Speaker 1

还没有。

Not yet.

Speaker 1

你得帮我下载一个。

You have get it for me.

Speaker 1

我说:那你直接

I said, well, just

Speaker 0

去搜一下。

Google it.

Speaker 0

她说:谷歌是什么?

She goes, what's Google?

Speaker 0

哇。

Wow.

Speaker 3

她多大了?

How old is she getting?

Speaker 1

她说,谷歌不是真实存在的东西。

She goes, Google's not a real thing.

Speaker 1

她就是这么说的。

That's what she said.

Speaker 1

他们用ChatGPT做所有事情。

They just use ChatGPT for everything.

Speaker 1

我儿子用它来做所有事情。

My son uses it for everything.

Speaker 1

她八岁。

She's eight.

Speaker 1

他十岁。

He's 10.

Speaker 1

哇。

Wow.

Speaker 1

想想这个真是令人着迷,等等。

It's just fascinating to think about, wait.

Speaker 1

什么?

What?

Speaker 1

你刚说了什么?

What did you just say?

Speaker 1

他们整天都在用。

And they use it all day.

Speaker 1

他们每天全天都在用Dolly。

They use Dolly all day every day.

Speaker 1

他们在脑海中搜索时用的是ChatGPT,这对孩子来说真是件令人着迷的事,如果他们愿意,完全可以两者都用。

Search in their mind is ChatGPT, which is just a fascinating thing for kids that could use both if they wanted to.

Speaker 1

他们不再使用搜索引擎,而是想要一个直接给出答案的引擎。

And rather than a search engine, they just want the answer engine.

Speaker 1

我觉得这完全令人着迷。

And I just find that completely fascinating.

Speaker 3

当然。

Absolutely.

Speaker 3

这对人类意味着什么?

What does it mean for humanity?

Speaker 3

开放的互联网一直非常好。

The open Internet has been really, really good.

Speaker 3

是的,我们现在有了这些封闭的内容花园,但仍然存在一个非常强大、充满活力的开放互联网,当你使用谷歌搜索时,通常还是会得到不同的观点,尽管我们都身处这些众所周知的过滤气泡中。

And, yes, we have these walled content gardens now, but there's still, like, a very robust, vibrant open Internet where when you Google, you do get divergent opinions, generally, you know, even though we're all in these well known filter bubbles.

Speaker 3

但如果你只得到一个答案,我认为我们正处在一个重要的时刻。

But if you just get one answer, I actually think we're at an important moment.

Speaker 3

对我来说,投资者能够真正为世界做出贡献的情况非常罕见。

It's very rare that, to me, investors can really, really contribute to the world.

Speaker 3

我们以整体的方式、间接地、大规模地做出贡献,因为我们资助了所有这些新技术,这些技术正在解决癌症问题,间接资助了人工智能。

We contribute in aggregate, in an indirect way, in a massive way because we fund all these new technologies that, you know, are solving cancer, indirectly funding AI.

Speaker 3

但我认为,你能产生更直接的影响,而不是一种分散的影响,这种情况很少见。

But I think it's rare that you can have, like, a more direct impact as opposed to kind of a diffuse impact.

Speaker 3

我认为,对于人类而言,最重要的是我们不能陷入一个只存在一个主导模型的世界。

I think it is supremely important for humans that we do not end up in a world where there is just one dominant model.

Speaker 3

这是我能想象到的最反乌托邦的未来。

That is the most dystopian future I can imagine.

Speaker 3

因为如果全球的孩子们都效仿你的孩子的行为,那么这个模型所承载的任何价值观都将被灌输给全人类。

Because that model then, if kids all over the world follow your kids' behavior, whatever values that model has will be imbued to the rest of humanity.

Speaker 3

我认为,我们正处在一个历史上罕见的时刻,由于多种原因,客观真理的概念正受到攻击。

And I think we are at a time in history where, for a lot of reasons, the idea of objective truth is under attack.

Speaker 3

对许多年轻人来说,他们的感受就是事实。

For a lot of younger people, their feelings are facts.

Speaker 3

科学和客观真理的理念正受到严重挑战。

The idea of science and objective truth is really under attack.

Speaker 3

我认为,拥有致力于客观真理的AI至关重要,无论这种立场多么不受欢迎。

And I think it's really important to have AIs that are dedicated to the idea of objective truth no matter how unpopular that may be.

Speaker 3

我认为,实现这一点的最佳方式是让它们相互竞争。

And I think the best way to accomplish that is to have them compete.

Speaker 3

你只有一种主导的人工智能。

You only have one dominant AI.

Speaker 3

这非常可怕。

That's very scary.

Speaker 3

别再想《1984》了。

And Forget 1984.

Speaker 3

你好。

Hello.

Speaker 3

谁知道未来会变成什么样?

Who knows what kind of future?

Speaker 3

但如果你有三四个甚至五个这样的系统,情况就完全不同了。

But whereas if you have three, four, five of these, that's very different.

Speaker 3

它们会相互竞争。

They'll compete.

Speaker 3

它们会有不同的价值体系。

They'll have different value systems.

Speaker 3

我认为,作为投资者,我们可以通过降低并改善AI基础设施的成本,比如突破性的网络、存储技术、提升GPU利用率的软件,以及直接资助那些与谷歌和Meta竞争的实验室,来实现这一点。

And I think we can, as investors, by lowering and improving the cost of AI infrastructure, you know, whether it's breakthrough networking, storage, technology, software that improves the utilization rate of GPUs, and directly funding, I think, some of these labs that are competing with Google and Meta.

Speaker 3

我认为这对世界来说非常重要。

Like, I think it's very important for the world.

Speaker 1

我们能聊聊数据中心和半导体的世界吗?

Can we talk a little bit about the world of data centers and semiconductors?

Speaker 1

因为你是这两个领域最资深的投资者之一。

Because you are one of the most seasoned investors in both of these spaces.

Speaker 1

我认为你二十年前就开始研究半导体了,你对这个领域的了解比我和我交谈过的任何人都要深入。

You've been I think you started your career covering semis twenty five years ago or something, and you just know more about it than just about anyone else I've talked to.

Speaker 1

我经常向你提问,或者就这两个领域的问题打电话给你。

And I I often find myself asking you or calling you with a question about these two areas.

Speaker 1

目前的关注点大多集中在模型上,而对半导体领域的关注较少。

And most of the attention has been on the models and less on the world of semiconductors.

Speaker 1

每个人都知道英伟达,但他们只是想当然地认为,是的,它制造了这些强大的芯片,支撑着整个体系。

Everyone knows NVIDIA, obviously, but they just sort of assume, like, yeah, it makes these amazing chips that power the rest of it.

Speaker 1

还有太多其他事情正在发生。

And there's just so much else that's going on.

Speaker 1

你之前提到过正在建设的新集群。

You've hinted at it with, like, the new clusters that are being built.

Speaker 1

但从你的角度来看,能否给我们讲讲这一领域的发展现状?哪些方面最让你感兴趣?无论是数据中心、单个芯片还是系统,你怎么看待都行。

But just give us, like, a state of from your perspective, how this has evolved and what aspects of it are most interesting to you, whether that's the data center or the individual chip or the systems or however you wanna approach it.

Speaker 1

我觉得你可能是我认识的对这两个领域最有见解、最有价值的人,而且你已经思考这些问题很久了。

I just think you have probably the most interesting and valuable perspective of anyone I know on these two topics and have been thinking about them for a long time.

Speaker 1

现在这已经成为核心事件了。

And now it's like the main event.

Speaker 3

这无疑是核心事件。

It's for sure the main event.

Speaker 3

我从2000年1月开始关注英伟达。

I started covering NVIDIA in January 2000.

Speaker 3

我认为,观察英伟达作为一家上市公司的早期阶段。

I would say watching the early days of NVIDIA has a public company.

Speaker 3

特斯拉是一家上市公司,作为两家公司的重大投资者,这无疑是我职业生涯中最大的荣幸,也是我作为投资者做过的最令人兴奋的事。

Tesla has a public company, and being a substantial investor in both have been by far the greatest privileges of my career and the most exciting thing I've ever done as an investor.

Speaker 3

埃隆和吉斯恩无疑是我在过去见过的两位最出色的首席执行官,而AMD的苏姿丰紧随其后。

And Elon and Jitsun are for sure the two best CEOs I've ever seen with Lisa Su right behind them at AMD.

Speaker 3

我说过,她将AMD从一家杠杆高达五倍、整整落后英特尔四年之久的公司,彻底转变为在各个方面全面超越英特尔的企业。

And I just say that she took AMD from company five times levered, literally four years behind Intel, to essentially utterly dominating Intel in every way.

Speaker 3

她接手时,公司账上只剩下二十天的现金。

Think they have twenty days of cash when she took over.

Speaker 3

人人都在谈论萨提亚,他确实非常出色。

Everybody talks about Satya, and he's very impressive.

Speaker 3

他接手的是一家垄断企业,此前管理得极其糟糕,但拥有稳定的经常性收入和高利润率。

He took over a monopoly that had been extremely poorly run with recurring revenue and high margins.

Speaker 3

无论如何,我在Cimis已经待了很长时间。

Anyways, I have been at Cimis for a long time.

Speaker 3

这是我的第一热爱。

It is my first love.

Speaker 3

我们在哪儿?

Where are we?

Speaker 3

首先,我知道在你的许多播客中都提到过这一点。

So first, you know, I think it's been touched on a lot of your podcasts.

Speaker 3

作为科技投资者,我们所有人听到的最重要的一点是,科技在过去三、四十年如此出色的原因之一是软件的边际成本为零。

The number one thing we have all heard as tech investors, one reason tech has been so amazing for thirty or forty years is that software has zero marginal costs.

Speaker 3

这些公司都拥有极高的毛利率和持续的收入,而人工智能恰恰相反。

These companies all have extremely high gross margins, recurring revenues, and AI is the exact opposite.

Speaker 3

人工智能的边际成本极高,因为扩展定律意味着,唯一提升质量的方法就是投入更多资金。

AI has extremely high marginal costs because scaling laws literally mean that the only way you get an improvement in quality is by spending a lot more.

Speaker 3

就此为止。

Full stop.

Speaker 3

这就是扩展定律的含义。

That is what scaling laws mean.

Speaker 3

如果你相信扩展定律,你就应该相信人工智能将具有极高的边际成本。

If you believe in scaling laws, you believe AI will have very high marginal costs.

Speaker 3

但由于我们接下来要讨论的多种原因,这些边际成本会迅速大幅下降。

Now because of a variety of things that we're gonna talk about, those marginal costs go down really, really quickly.

Speaker 3

但就前沿模型而言,尤其是在训练和推理方面,边际成本仍然非常高,不过推理的成本要低得多。

But they're still really, really high at the leading edge for models, especially for trading and for inference, but much less so for inference.

Speaker 3

推理和训练是两个完全不同的市场。

Inference and trading are two totally different markets.

Speaker 3

因此,这意味着高边际成本将使基础设施、效率和卓越性成为最重要的成功因素,尤其是对模型公司本身而言。

So what this means that it has really high marginal cost is that infrastructure, efficiency, and excellence, I think, is going to emerge as the single most important success factor, particularly for the model companies themselves.

Speaker 3

这一点如今已通过一种名为MFU(模型浮点运算利用率)的指标得到衡量,其数值通常在35%到40%之间。

And this has been measured today in something called MFU, model flops utilization, and that generally runs around 35 to 40%.

Speaker 3

这实际上是指你实际用于训练的计算能力占理论计算能力(浮点运算)的百分比。

And that's literally the percentage of compute, theoretical compute, flops that you're actually applying to trading.

Speaker 3

因此,在理论浮点运算中,大多数公司已停止公布这一数据,因为竞争过于激烈。

So of the theoretical flops, most companies that have been published people stop publishing this because it's so competitive.

Speaker 3

但像GPT-3、谷歌的Lambda以及NVIDIA的Megatron等技术论文都曾公布过它们的MFU,数值均在28%到39%之间。

But, you know, the technical papers for GPT three, I think it was called Google's Lambda, and NVIDIA's Megatron all showed their MFU, and it was between high twenties and high thirties for all of them.

Speaker 3

谷歌的最高。

Google had the highest.

Speaker 3

如果你的MFU更高,意味着在花费相同金额的情况下。

If you have a higher MFU, it means that you can choose for the same amount of money you were spending.

Speaker 3

你拥有相同数量的GPU和大致相同的电力。

You have the same amount of GPUs and the same amount of power presumably.

Speaker 3

你可以选择更快的上市时间。

You could choose between faster time to market.

Speaker 3

如果你的MFU是50%,而你的竞争对手是40%,在同等交易计算量下,你可以快25%进入市场。

If you run a 50% MFU and your competitor's running 40 for an equivalent amount of trading flops, you could be in market 25% faster.

Speaker 3

你可以选择更好的质量。

You could choose between better quality.

Speaker 3

你可以选择让交易运行更久,或者通过多种方式降低模型成本,其中大多数与量化有关。

You might just do the trading route run as long as possible, or you can make the model lower cost in a variety of ways, most of which relate to quantization.

Speaker 3

我们可以深入探讨,但这非常技术性。

And we could get into that, but it's very technical.

Speaker 3

这几乎就像,如果你因为更高的MFU并结合量化技术,获得了25%更高品质的模型,那么你实际上可以将推理成本降低近50%。

It's almost like if you have a 25% higher quality model because you have a higher MFU and you build in quantization, then you can actually get a almost a 50% reduction in inference cost.

Speaker 3

因为如果你能比竞争对手低一个量化等级,这就构成了巨大的优势。

Because if you can quantize one level lower than competitors, it's a profound advantage.

Speaker 3

所以我认为,如果让我选一个指标来评估实验室的成功,现在已经有五个GPT-4级别的模型了:谷歌、OpenAI、Anthropic、XAI和Meta。

And so I think MFU, if I were to pick one metric to evaluating lab success because now there's been five g p d four class models from it's Google, OpenAI, Anthropic, XAI, and Meta.

Speaker 3

所以现在有五个GPT级别的模型。

So there's five GPT.

Speaker 0

但其中有一个表现很好吗?

And does have one that good?

Speaker 3

也许紧随其后,甚至可以说同样令人印象深刻,因为它用小得多的参数量就取得了所有评估中的优异结果。

Maybe right under, arguably, just as impressive because it gets really good results in all the evals with a much smaller parameter count.

Speaker 3

顺便说一句,这些模型如今已是商品,但我怀疑,一旦扩展定律继续生效,GPT-7或GPT-8的训练成本将高达5000亿美元。

By the way, these models are commodities today, but I am suspicious once we get to scaling laws continue, GPT seven or eight literally cost $500,000,000,000 to trade.

Speaker 3

我不认为它们会一直保持商品属性。

I don't think they're gonna stay commodities.

Speaker 3

规模是进入壁垒中最强大的因素,而这种规模非常巨大。

Scale is the most powerful barrier to entry, and that's a lot of scale.

Speaker 3

因此,MFU 是最重要的指标,因为它为你提供了所有这些优势,让你在已经训练出这些 GPT 核心级别模型的五到六家公司中脱颖而出。

So MFU is the most important metric because it gives you all of these advantages and ways to differentiate yourself amongst five people, six people who've trained these GPT core class models.

Speaker 3

但我认为还有一项指标比 MFU 更好,今天早上我刚刚想到了它。

But I think there's something better than MFU, and I actually came up with it this morning.

Speaker 3

它在 MFU 的基础上进行了补充,并对 MFU 进行了分解。

And it adds to MFU and decomposes MFU.

Speaker 3

我会把它看作是一个统一的 AI 效率公式。

I would think of it as maybe like a unified AI efficiency equation.

Speaker 3

因此,我会将 MFU 拆解为两个部分。

So MFU, I would decompose into two things.

Speaker 3

第一部分是所谓的 MAMAF,即最大可实现矩阵乘法浮点运算次数,MIBF,MatBull。

The first thing is something called MAMAF, maximum achievable matrix multiplication flops, MIBF, MatBull.

Speaker 3

它衡量的是软件效率。

And what this measures is software efficiency.

Speaker 3

这涉及到CUDA。

And this goes to CUDA.

Speaker 3

这个叫斯坦·贝克曼的人在X上提出了它。

This guy Stan Bexman came up with it on X.

Speaker 3

他指出,每块芯片都有一个理论最大性能,这个值很容易计算,也就是浮点运算能力。

What he did is each chip has a theoretical maximum performance, which is just easily calculable, you know, flops.

Speaker 3

然后他研究了实际能够达到的性能。

And then he looked at what can you get in practice.

Speaker 3

根据他的测试,NVIDIA显卡的运行效率为83%。

NVIDIA GPUs run it, per his testing, 83%.

Speaker 3

我确信有些实验室的运行效率已经接近90%。

And I'm sure that there are some labs that are running them at closer to 90.

Speaker 3

但这只是单张显卡的情况。

But this is just a single GPU.

Speaker 3

AMD花了这么长时间才进入这个市场,原因之一是,在半导体行业,通常要等到第三代芯片才能真正取得突破。

One reason it took AMD so long to break into this market beyond the fact that it almost always in semiconductors, it almost always takes you till your third generation chip to really, really hit it.

Speaker 3

谷歌的TPU也是同样的情况。

That's what it took with Google, with TPUs.

Speaker 3

你知道,Instinct Mx300是一款不错的芯片。

You know, Instinct Mx300 is a good chip.

Speaker 3

因为他们的ROCCM开源软件太差了,他们内部没有能力真正改进它。

Because their ROCCM open source software was terrible, they didn't have the internal capabilities to really improve it.

Speaker 3

社区里没有人花时间去改进它。

Nobody in the community took time to improve it.

Speaker 3

所以Instinct MI300刚推出时,运行效率只有大约25%到30%的MAMAF。

So the Instinct MI three hundred first came out, it ran at something like 25 or 30% MAMAF.

Speaker 3

现在根据斯坦的测试,已经提升到了60%。

Now per stand's tests, it's up to 60%.

Speaker 3

但由于它的浮点运算能力更强,实际上已经略微领先于NVIDIA的GPU。

But because it has more FLOPS, it's actually slightly ahead of the NVIDIA GPUs.

Speaker 3

我觉得这真的非常出色。

And this was actually really I thought it was brilliant.

Speaker 3

这是一种概念化和量化CUDA优势的方式。

It was a way of conceptualizing and quantifying the CUDA advantage.

Speaker 3

它们可以运行在83的水平。

They can run at 83.

Speaker 3

AMD的运行水平是25。

AMD is running at 25.

Speaker 3

这回答了问题:MI250实际上是一款相当不错的芯片。

And it answers the question, the MI250 was actually a pretty good chip.

Speaker 3

但没人使用它。

Nobody used it.

Speaker 3

为什么?

Why?

Speaker 3

它的MAMAF可能只有10%。

Its MAMAF was probably 10%.

Speaker 3

首先,有MAMAF,它衡量的是你芯片的软件性能有多好。

So first, there's MAMAF, and that is how good is the software for your chip.

Speaker 3

然后,作为使用这些软件的人,你有多擅长优化它?

And then how well do you, as someone who's using that software, optimize it?

Speaker 3

Netflix 曾经说,我们比亚马逊更懂得如何高效使用 AWS。

Netflix used to say that we know how to use AWS more efficiently than Amazon.

Speaker 3

然后,很多人说,这些实验室比英伟达更懂得如何高效使用 CUDA。

Then a lot of people have said some of these labs know how to use CUDA more efficiently than NVIDIA.

Speaker 3

这是一个巨大的秘密武器。

That's a big secret sauce.

Speaker 3

如果它们的运行效率达到 95%,那就已经是 10% 的差距了。

If they're running at 95, right there, that's 10%.

Speaker 3

如果英伟达的 GPU 在单卡基础上的效率是 83%,为什么我们的 MFU 却只有 35% 到 40%?

If NVIDIA GPUs are running at 83% efficiency on a per GPU basis, why are we down at 35, 40% for MFU?

Speaker 3

原因是我称之为 SFU 的东西,即系统浮点运算效率。

And the reason is something I would call SFU, system flops efficiency.

Speaker 3

这真正涵盖了网络、存储和内存。

And this really captures networking, storage, and memory.

Speaker 3

在每一个这些方面,我们可以将SFU分解为每一个组成部分。

In each one of those and we could decompose SFU into each of those.

Speaker 3

但我认为SFU是一个有助于理解它的有用方式。

But I think SFU is like a helpful way to think about it.

Speaker 3

这还不是全部。

That's not all.

Speaker 3

因此,你需要将MAMAF乘以SFU,即系统浮点运算效率。

So you need to multiply MAMAF times SFU, system FLOPS efficiency.

Speaker 3

然后你还需要乘以每次训练运行中用于检查点的时间百分比。

And then you need to multiply that by the percentage of time that is spent in a checkpoint per training run.

Speaker 3

这些因素都会相互叠加。

And these things all compound.

Speaker 3

它们叠加起来会产生一些惊人的结果。

And they compound out to some wild things.

Speaker 3

顺便问一下,我需要解释一下检查点吗?

And just in case, should I explain checkpointing or no?

Speaker 3

因为正如我前面提到的,训练集群必须保持一致才能正常运行,这意味着每个GPU都需要知道其他所有GPU在想什么。

Because as I referenced earlier, trading cluster has to be coherent to function, and that means each GPU needs to be aware of what every other GPU is thinking.

Speaker 3

如果任何一个GPU出现故障,你就会丢失自上次保存模型以来的所有内容,这被称为检查点。

If any one GPU fails, you lose everything from the last time you saved the model, which is called a checkpoint.

Speaker 3

GPU经常出现故障。

The GPUs fail all the time.

Speaker 3

不仅仅是GPU会出问题,GPU还会熔化。

Not just GPUs, but GPUs melt.

Speaker 3

但如果你读过Lava三号的技术论文,你会发现GPU故障的原因清单简直令人震惊。

But if you read the Lava three technical paper, I mean, it's like the list of reasons that GPUs fail is like astonishing.

Speaker 3

光链路中断了。

An optical link goes down.

Speaker 3

交换机宕机了。

A switch goes down.

Speaker 3

在为每个GPU提供支持的存储、网络和内存链条中,有太多可能的故障点,导致它们失败的方式数不胜数。

There's so many points of failure in the chain of storage, networking, and memory that feeds each GPU that there are innumerable ways for them to fail.

Speaker 3

没有存储、内存和网络的GPU毫无用处。

The GPU without storage, memory, and networking is worthless.

Speaker 3

如果你想真正深入理解,我强烈推荐每个人去组装一台游戏电脑。

And if you want to really, really understand, this is something I highly recommend to everyone, build a gaming PC.

Speaker 3

因为每一台计算机,无论是iPhone、笔记本电脑还是数据中心,都具有相同的四个基本组成部分。

Because every computer, whether it's the iPhone, whether it's a laptop, whether it's a data center, has the same, what I would call, four fundamental elements.

Speaker 3

这三个部分是内存、存储和计算,以及连接所有这些组件的网络。

Three are memory, storage, and compute, and then the networking that connects all those things.

Speaker 3

当你组装一台游戏电脑时,你实际上需要将PCI Express插槽连接到GPU,然后再连接到CPU。

And when you assemble a gaming PC, you literally have plug, yeah, the PCI Express into the GPU and then into the CPU.

Speaker 3

然后你必须将内存条插入DRAM插槽,使其正常连接。

And then you you have to connect slot in the DRAM so that it'll connect.

Speaker 3

接着,DRAM必须连接到闪存存储,而DRAM显然是内存。

And then the DRAM has to connect to the flash storage, and DRAM is obviously memory.

Speaker 3

我强烈推荐这样做。

I highly recommend this.

Speaker 3

但我认为,看待数据中心或任何计算机的方式,就是想象它是一家餐厅。

But I think the way to think of a data center or any computer is just imagine that it's a restaurant.

Speaker 3

在这间餐厅里,计算的主要单元,在AI服务器中就是GPU,它是主厨。

In this restaurant, the primary unit of compute and in an AI server, it is the GPU, is the head chef.

Speaker 3

主厨如果没有食物、食材和厨具,就什么都做不了。

A head chef can do nothing without food and ingredients and utensils.

Speaker 3

我会把存储想象成把食物运送到餐厅的送货卡车。

I would conceptualize storage as like the delivery truck that brings food to the restaurant.

Speaker 3

而存储与服务器其余部分的连接方式,今天总是通过PCI Express实现。

And then the way storage connects to the rest of the server is today always over PCI Express.

Speaker 3

这是一种网络技术。

That's a networking technology.

Speaker 3

这 literally 就是把食物从货车搬到餐厅冰箱的人。

And that is literally the guy who moves the food from the food truck into the restaurant's refrigerator.

Speaker 3

他们会把餐厅的冰箱称为——这是一个不完美的类比——可能是内存。

And they'll call the restaurant's refrigerator this is an imperfect analogy, maybe the memory.

Speaker 3

然后你需要将数据从内存中移动出来,实际上在这个情况下是进入CPU,进入大量的DRAM内存中。

And then you have to move the data from the memory, actually, in this case, into the CPUs, into the big pool of DRAM.

Speaker 3

CPU才能完成它的任务。

The CPU can do its job.

Speaker 3

然后你再把数据移到GPU的内存中,GPU才能最终完成它的任务。

And then you move it to the GPU's memory, and then the GPU can finally do its job.

Speaker 3

也许GPU的内存就像炉灶。

Maybe the GPU memory is like the stove.

Speaker 3

在炉灶和副厨之间流动的东西,就是CPU和GPU之间的连接。

And what flows between the stove and the sous chef is the connection between the CPU and the GPU.

Speaker 3

我们可以深入探讨这一点,但除非这位厨师有炉灶、烹饪用具和食材,否则他什么都做不了。

And we we can go into this, but just unless that chef has a stove, cooking utensils, and food, he could do nothing.

Speaker 3

过去五年里,数据中心面临的一个大问题尤其突出,这些数字大致是准确的。

And the big problem with the data center is over the last five years in particular, these numbers are gonna be directionally accurate.

Speaker 3

GPU的速度提升了50倍,而数据中心的其余部分仅提升了四到五倍。

GPUs have gotten 50 times faster, and the rest of the data center has only gotten four to five times faster.

Speaker 3

这就是为什么 MFU 如此之低的原因。

And that is why MFU is so low.

Speaker 3

因为 GPU 在等待所有这些组件完成工作,而大部分时间它都无所事事。

Because the GPU is sitting around waiting for all those things to do their job, and it's doing nothing most of the time.

Speaker 3

因此,我认为投资于下一代网络、存储和内存技术,特别是网络技术,是非常合理的。

So I do think it is sensible really sensible to invest in next generation networking, storage, and memory technologies, particularly in networking.

Speaker 3

因为如果我们想要实现百万级 GPU 集群——也就是百万个 Rubens(Blackwell 之后的下一代),我们就需要在每一个环节上取得根本性的突破。

Because if we're gonna get to a million GPU cluster, that'll be a million Rubens, which is the generation after Blackwell, we're gonna need profound breakthroughs in every step of that day.

Speaker 3

每一个环节。

Every single thing.

Speaker 3

全新的炉灶、全新的冰箱、机器人帮厨、新工具,一切都要更新。

All new stoves, all new refrigerators, robotic sous chefs, new utensils, everything.

Speaker 3

否则,资源将被浪费,MFU 将只有 3% 到 5%。

Otherwise, it's gonna be wasted, and MFU is gonna be three to 5%.

Speaker 3

但无论如何,回到检查点的问题上,由于这条链路上存在太多故障点,数据中心的‘主厨’——GPU——经常着火。

But anyways, coming back to checkpointing, because there's so many points of failure in that chain, the head chef in data centers, being the GPU, goes up in flames a lot.

Speaker 3

它们真的会熔化。

They literally melt.

Speaker 3

然后其他每个组件也经常损坏。

And then every other component breaks a lot.

Speaker 3

这个集群由32,000个这样的隐喻性厨房组成,每一个环节一旦失败,就会导致整个集群崩溃。

The cluster is 32,000 of these metaphorical kitchens where each step, if it fails, brings down the whole cluster.

Speaker 3

因此,由于这个原因,人们会频繁进行检查点保存,也就是保存模型。

So because of this, people checkpoint frequently, and that means save the model.

Speaker 3

如果你通过更好的网络拓扑或更好的冷却技术,使GPU不再熔化,故障率降低,集群更可靠,你就需要更少地进行检查点保存。

Now if you, through better networking topologies or better cooling technologies such that GPUs don't melt, if you have a lower failure rate, if you have a more reliable cluster, you need to checkpoint less.

Speaker 3

所以我们经历了巨大的SFU提升,然后是检查点频率的下降。

So we've had mammoth, SFU, then checkpointing frequency.

Speaker 3

这让我们看到,比如说,有一家公司运行在90%的MAMAF,然后它的SFU只有50%。

And that gives us and like, let's just say, there's one company that runs at 90% MAMAF, and then it runs at 50% SFU.

Speaker 3

现在你的利用率只有45%。

Now you're at 45% utilization.

Speaker 3

他们必须频繁检查点,我随便编个例子。

And they have to checkpoint, I'm just gonna make something up.

Speaker 3

三分之一的时间,你的MFU会降到30%。

A third of the time, you're down to 30% MFU.

Speaker 3

如果你有另一家公司,因为他们是CUDA高手,能实现接近100%的MAMAF。

If you have a different company that can run at close to a 100% MAMAF because they're CUDA masters.

Speaker 3

他们运行在大约60%的SFU。

And they run at, let's call it, 60% SFU.

Speaker 3

现在你的是60%,而另一个是45%。

Now you're at 60, the other you're 45.

Speaker 3

这已经是一个巨大的差异了。

So it's already a massive difference.

Speaker 3

相差33%。

It's a 33% difference.

Speaker 3

如果你只需要每10%的时间做一次检查点,那你就能达到54%。

And then if you have to checkpoint only 10% of the time, you're at 54%.

Speaker 3

你的竞争对手只有30%。

Your competitor's down at 30.

Speaker 3

这还不是全部。

That's not even all of it.

Speaker 3

我们最后需要乘上的因素是PUE,即电源使用效率。

The last thing that we need to multiply it by is PUE, which is power utilization efficiency.

Speaker 3

电力是有成本的,但对于这些大型集群来说,成本是垂直上升的。

The power has a cost, but it's going vertical for these big clusters.

Speaker 3

如今在美国,基本上只有三个地方能够为单个数据中心提供足够可靠的吉瓦级电力,我。

There's only basically three places in The United States today where you can get a gigawatt of power to a single data center that's reliable enough, I.

Speaker 3

嗯,那就是核能。

E, that's nuclear.

Speaker 3

我认为美国的平均电价大约是每千瓦时0.08美元。

I think it's like $08 per kilowatt hour on average in The United States.

Speaker 3

他们为这吉瓦电力收取的费用将是这个价格的十倍。

What they're going be charging for that gigawatt is 10x that.

展开剩余字幕(还有 480 条)
Speaker 3

但你的PUE确实很重要,因为你的电力成本非常关键。

But your PUE then really matters because your cost of electricity really matters.

Speaker 3

所以,你为优化SFU所做的某些事情,实际上可能会以非常负面的方式提高你的PUE。

So things you do to optimize SFU might actually increase your PUE in a really negative way.

Speaker 3

因此,我想,如果我今天早上刚想到这一点,也许这太明显了,每个实验室都已经在这么做了。

And so I think if you're I just thought of this this morning, maybe it's super obvious that every lab already is doing this.

Speaker 3

但我只是觉得,如果你把它全部代入一个公式,再计算出所需的成本,就能真正做出权衡。

But I just think if you leak it all into an equation and then you do dollars that it takes to get it, you can really make trade offs.

Speaker 3

哦,天哪。

Oh, wow.

Speaker 3

如果我在网络上多花两倍的钱,我会获得高得多的SFU,这很好,但我的PUE也会大幅上升,这就不好了。

If I spend 2x more on networking, I get much higher SFU, which is great, but then I have a much higher PUE, which is bad.

Speaker 3

而我们最终想要的不仅仅是每秒多少exaflops,而是每美元资本支出、每瓦电力所获得的每秒exaflops。

And what we ultimately want is not just actual exaflops per second, but we want exaflops per second per dollar of CapEx per watt of electricity.

Speaker 3

而我刚才描述的这个公式——至少我会把它写下来——涵盖了所有这些因素。

And the equation I just described, which I'm going to write out at least for myself, captures all of that.

Speaker 3

每个人都会做出不同的设计决策。

And everybody's going to be making different design decisions.

Speaker 3

数据中心架构以及你所做的这些设计决策将变得极其重要。

And data center architecture and these design decisions you make are going to be immensely important.

Speaker 3

我认为你会看到,特别是在这些十万级集群中,一些公司将在每美元资本支出每瓦特功耗的exaflops方面获得超过100%的优势。

And I think you're gonna see, particularly with these 100,000 clusters, some of these companies are going to have greater than a 100% advantage in exaflops per dollar of capex per watt consumed.

Speaker 3

这时,这些实验室之间的真正差距就会显现出来——这决定了GPT-7或8的成本是1万亿美元还是5000亿美元,下一代的成本是2000亿美元还是4000亿美元。

And this is when you're gonna really separate these labs because that is the difference between GPT seven or eight costing a trillion dollars or 500,000,000,000, the next class costing 200,000,000,000 or 400,000,000,000.

Speaker 3

对于推理来说,情况也是一样的。

And then for inference, it's the same.

Speaker 3

完全一样。

It's the exact same.

Speaker 3

这不仅影响服务成本,还直接影响用户体验,比如每秒处理的token数,而我们知道,这对谷歌的搜索用户体验来说是最重要的因素之一。

Not only is it cost to serve, but it directly affects user experience in terms of tokens per second, which we know from Google is one of the most important things for search UX.

Speaker 3

因此,所有这些也都会应用于推理场景。

So all of this is also gonna apply in inference.

Speaker 3

推理要简单得多,因为它实际上完全取决于内存带宽和片上内存。

Inference is just a lot easier because it really, really just comes down to memory bandwidth and on chip memory.

Speaker 1

这让我想起了瓦茨拉夫·斯米尔关于能源历史的论述,他称之为原动机,比如化石燃料、风能等等。

This reminds me so much of the Vaclav Smile history of energy stuff where you would get a new he called them prime movers, some source of energy, fossil fuels, wind, whatever.

Speaker 1

然后你会经历一个漫长时期,所有的进步都来自于效率的提升。

And then you get this long period of the gains all coming from the efficiency.

Speaker 1

所以在煤炭时代的早期,当你转动涡轮机时,它只能捕获煤炭本身10%到15%的可用能量。

So if you're spinning a turbine or something in the early days of coal, the turbine only captured 10 or 15% of the available energy from coal itself.

Speaker 1

而现在,我们已经达到了大约98%的水平。

And now we're at, like, 98 or something like that.

Speaker 1

这听起来和你在这里描述的情况完全一致。

And basically, it sounds like that's the exact same thing that you're describing here.

Speaker 1

GPU就像煤炭,它们会不断改进,这在技术上很酷,因为煤炭不会。

The GPU is the coal, and those will keep getting better, which is cool in technology unlike coal.

Speaker 1

但听起来这基本上就是整个故事,实在太有趣了。

But it sounds like that's basically the story, which is so interesting.

Speaker 3

这就像通过不同的化石燃料机制和发动机将阳光转化为可用能源。

That is like turning sunlight into usable energy via different fossil fuel mechanisms and motors.

Speaker 3

而这是将阳光直接转化为计算能力。

And this is turning sunlight into compute, literally.

Speaker 3

现在,这种阳光可以以真实的阳光形式出现。

And now that sunlight can come in the form of actual sunlight.

Speaker 3

它也可以以人工阳光的形式出现。

It can come in the form of artificial sunlight.

Speaker 3

那就是核能。

That's nuclear.

Speaker 3

它还可以以储存的阳光形式出现。

It can come in the form of stored sunlight.

Speaker 3

那就是化石燃料。

That's fossil fuels.

Speaker 3

但关键是将阳光转化为计算能力的效率。

But this is the efficiency at which you turn sunlight into compute.

Speaker 3

太酷了。

And So cool.

Speaker 3

就看你为这个比率支付了多少钱。

In the dollars you pay for that ratio.

Speaker 3

嗯。

Yeah.

Speaker 3

顺便说一下,这真的很有趣。

And by the way, it is just interesting.

Speaker 3

比如,这就是我以前对风能不太感兴趣的原因,因为涡轮机已经非常高效了。

Like, it's one reason, like, I was never that excited about wind because turbines were really efficient.

Speaker 3

而十年前或十五年前,太阳能的效率还非常低。

Whereas you could look ten or fifteen years ago, solar was terribly inefficient.

Speaker 3

我真的觉得这太棒了。

I actually think it's awesome.

Speaker 3

这让我有点难过。

It's a little sad to me.

Speaker 3

所有这些孩子都非常担心全球变暖,而你经常听到关于二十多岁的人说,我不希望把孩子带到一个变暖的世界。

All these kids are really worried about global warming, and you have all these things about 20 year olds, oh, I don't want to bring children into a warming world.

Speaker 3

全球变暖是个大问题。

Global warming is a big problem.

Speaker 3

但这个问题已经解决了。

It is a solved problem.

Speaker 3

它之所以被解决,是因为光伏电池的效率每年以接近10%的速度复合增长,而电池效率则比这个数字低200个基点。

And it is solved because photovoltaic cell efficiency is compounding just under 10% a year, and battery efficiency is compounding 200 bps less than that.

Speaker 3

如果你将这种增长趋势长期延续下去,甚至不需要太长远的未来,世界将直接依靠太阳能运行。

And if you compound that out, over the long term, not even the long term, the world is going to run on sunlight directly.

Speaker 3

化石燃料将彻底消失。

Fossil fuels are just going to go away.

Speaker 3

这是因为经济因素。

It's because of economics.

Speaker 3

因为太阳能和储能的成本将低于任何其他发电方式。

And it's because sunlight and storage is going to be cheaper than every other way of providing power.

Speaker 3

下一代现在不会再面临这个问题了。

It's just not going to be a problem for the next generation now.

Speaker 3

也许我们达到了某个临界点,导致情况不可逆转,诸如此类。

Maybe we hit some tipping point and it's irreversible and this and that.

Speaker 3

在我的有生之年,排放量将急剧下降。

Emissions are going to collapse in my lifetime.

Speaker 3

真的会急剧下降。

Literally collapse.

Speaker 3

显然,像前工业时代的人类一样,我们实际上是巨大的污染者,因为每单位木柴的污染非常高,而你必须生火,否则就会被剑齿虎吃掉。

Obviously, like pre industrial age humans, we're actually massive polluters because the pollution per unit of firewood is really high, and you needed fires, otherwise you'd get eaten by a saber toothed tiger.

Speaker 3

但那时候人类数量并不多。

Now, there weren't that many of them.

Speaker 3

在我的有生之年,我们有可能回到新石器时代的排放水平,前提是我能因为人工智能而多活几年。

It is possible that we're going to be back to neolithic levels of emissions in my lifetime, assuming I could live a little bit longer because of AI.

Speaker 3

所以这是一个令人振奋和鼓舞的想法。

So that's an awesome and encouraging thought.

Speaker 3

这很好。

That's good.

Speaker 3

我只是希望有人能大声疾呼这一点。

And I just wish somebody was shouting that from the rooftops.

Speaker 1

好吧,我们就在那儿。

Well, here we are.

Speaker 1

我很想继续深入探讨,因为每一层都十分有趣。

I'd love to keep going up the stack here because every single level is interesting.

Speaker 1

半导体本身以及其中潜在的创新让我很感兴趣。

So the semiconductors themselves and the potential innovations there are interesting to me.

Speaker 1

从数据层面一直到应用层,我都觉得非常有趣。

The data piece is really interesting to me all the way up to the application layer.

Speaker 1

我只是好奇你对这一切有什么看法。

And I'm just curious what you think about all of this stuff.

Speaker 1

我们在这所有讨论中都假设,到了GPT-5、6和7时,我们将拥有更多数据来训练它们。

So one of the things that we're assuming in all this is that we're gonna have more data to train these things on at GPT five, six, and seven.

Speaker 1

我认为关于未来可用数据是什么、我们如何获取更多数据或如何创建合成数据,以及这些方法是否能帮助打造更好的模型,存在一些有趣的讨论。

And I think there are some interesting discussions about what available data there will be or how we'll get more data or how we'll create synthetic data and whether or not that will work to create a better model.

Speaker 1

比如,我对您如何看待这一切很感兴趣,因为如果我们真要展开您之前所描述的这场大规模竞争,这些条件都是必需的。

Like, I'm interested in what you think about all of this stuff that is required if we're gonna have the big arms race that you were describing earlier.

Speaker 3

我认为九个月前这确实是一个悲观的情景。

So I do think this was a real bear case maybe nine months ago.

Speaker 3

但我在Cloud 3.5的技术论文中看到过一些暗示,而在NVIDIA Nebotron的技术论文中则表达得更加明确。

But I think it was hinted at in the Cloud 3.5 technical paper and maybe a little more explicit in the NVIDIA Nebotron technical paper.

Speaker 3

你知道,这太棒了。

You know, it's awesome.

Speaker 3

NVIDIA,随便吧。

NVIDIA, whatever.

Speaker 3

我觉得AI的发展可能会遇到一个瓶颈因素。

It feels like there's gonna be a rate limiting factor for AI.

Speaker 3

他们解决了这个问题,而Nebotron正是做到了这一点。

They solve it, and that's what Nebotron did.

Speaker 3

但我认为,由于一些没人理解的原因,没人真正理解这些模型。

But I do think for reasons that no one understands, and no one understands these models.

Speaker 3

没人明白它们是如何工作的,为什么能工作,为什么存在缩放定律。

No one understands how they work, why they work, why scaling laws.

Speaker 3

有各种各样的理论,也许我们对它们的理解正在一点点进步,但没人真正理解它们。

There's all sorts of theories, and maybe we're getting a little better at understanding them, but no one understands them.

Speaker 3

所以,第一点,没人理解,但看起来合成数据是有效的。

So no one understands, point one, but it does look like synthetic data works.

Speaker 3

没人明白为什么,但看起来它确实有效。

No one understands why, but it looks like it works.

Speaker 3

那么,再次强调,它会持续有效吗?

Now, again, will it continue working?

Speaker 3

我不知道。

I don't know.

Speaker 3

没人知道。

Nobody knows.

Speaker 3

没人知道。

No one knows.

Speaker 3

见过凯文·斯科特的人刚做了一个播客,他基本上说:看。

People who have seen Kevin Scott just did a podcast, and he basically said, look.

Speaker 3

我看过一些GPT-5的早期检查点,扩展定律仍在继续。

I've seen some early checkpoints of GPT five, and scaling laws are continuing.

Speaker 3

我认为,在很多方面,这是目前我们拥有的最好的迹象,表明它们确实在扩展。

I think in a lot of ways, that's the best indication we have that they're scaling.

Speaker 3

实际上,就XAI而言,我认为OpenAI结合了XAI所做的工作和Blackwell的延迟。

And I actually think largely in respect to XAI, I do think OpenAI is the combination of what XAI is doing and Blackwell delay.

Speaker 3

Blackwell的延迟意味着,如果你在等待Blackwell并试图组建一个10万台规模的集群,XAI将可能拥有一年的优势,这对其他实验室来说是不可接受的。

The Blackwell delay means that if you're waiting for Blackwell and you're trying to get a 100,000 cluster, XAI is gonna have arguably a one year advantage, which is untenable to these other labs.

Speaker 3

所以现在他们都急于启动自己的10万台规模集群,但他们没有埃隆从第一性原理出发来设计和重新设计数据中心。

So they're all now frantically working on stating up their own 100,000 cluster, but they don't have Elon designing the data center, redesigning the data center from first principles.

Speaker 3

数据中心架构以前一直只是锦上添花的东西。

Data center architecture was always like kind of nice to have.

Speaker 3

现在这已成为必需品。

Now it is must have.

Speaker 3

这关乎生死存亡。

It's existential.

Speaker 3

我认为合成数据看起来确实会有效。

I think synthetic data does look like it's gonna work.

Speaker 3

这引出了一个问题:这些模型的价值将从何而来?

This goes to where will the value for these models come from?

Speaker 3

为什么Meta对开源如此坦然?

Why is Meta so comfortable open sourcing?

Speaker 3

你最终可能会看到所有人都开源。

You may ultimately see everyone open source.

Speaker 3

XAI已经开源了GrokWOD。

XAI has open source, GrokWOD.

Speaker 3

但原因在于,价值可能并不来自模型本身。

But the reason is the value may not come from the model.

Speaker 3

现在想想,如果你在运行我描述的这个方程,并且在每美元资本支出和每瓦特的exascale算力上拥有1到200%的优势,且规模定律成立,那么你在模型质量上将获得巨大的优势,以至于你永远不会开源它。

Now look, if you're running that equation I described and you have a one to 200% advantage on exaflops per CapEx dollar per watt and scaling laws hold, you're going to have such a massive advantage of model quality that you're never going to open source it.

Speaker 3

你仍然可以有效地窃取模型并进行蒸馏。

You could still effectively kind of steal models, distill them.

Speaker 3

如果你拥有规模定律带来的计算优势,天啊,你将拥有极其宝贵的东西。

If you have that compute advantage of Scalie Laws whole, wow, you're gonna have something very valuable.

Speaker 3

价值显然来自于分发渠道和独特数据。

The value clearly comes from distribution and unique data.

Speaker 3

Meta已经开源了。

Meta is open sourced.

Speaker 3

但他们并没有开源所有数据,这些数据只会用于他们自己的LABA版本。

Haven't open sourced all their data, and that data is just gonna be for their version of LABA.

Speaker 3

所以它肯定会更好。

So it'll for sure be better.

Speaker 3

谷歌拥有YouTube,以及他们在试图构建知识图谱时开发的各种其他数据源——那些在你搜索时出现的小小知识面板。

Google, they have YouTube, and then all sorts of other data sources that they've developed when they tried to boil the ocean to create the knowledge graph, which are those little knowledge panels that kind of appear when you do searches.

Speaker 3

因此,结合YouTube、知识图谱项目以及谷歌地图,他们拥有海量数据。

So between YouTube and the work they did for the knowledge graph and Google Maps, they have crazy data.

Speaker 3

他们并不在意,因为他们可以用这些数据来为自己的模型变现。

They don't care because they could use that data to monetize their model.

Speaker 3

XAI无论是否开源,都肯定能以其他人无法比拟的方式访问xData,因为x拥有XAI 25%的股份。

XAI, whether they open source or not, they will always have, for sure, access to xData in a way no one else does because x owns 25% of xAI.

Speaker 3

而且我认为,随着时间推移,XAI将成为贯穿埃隆旗下整个公司生态系统的智能层。

And then I would think over time, xAI will kinda be an intelligence layer that cuts across Elon's ecosystem of companies.

Speaker 3

顺便说一下,我们应该谈谈人工智能与机器人技术,这可能是我们这一生中最大的颠覆,堪比人工超级智能和这些数字神明。

And by the way, we should talk about AI and robotics, which I should think may be the biggest disruption in our lifetime, comparable to artificial superintelligence and these digital gods.

Speaker 3

如果这些模型没有人在BAMAF、SFU、检查点频率和PUE方面形成复合优势,从而在每瓦每美元资本支出的exaflops性能上实现显著差异,我认为它们最终都会收敛到大致相同的智能水平。

If these models, unless someone develops a compounded advantage of BAMAF, SFU, checkpointing frequency, and PUE, such that they have a dramatic difference at exaflops per dollar of CapEx per watt, I think they're probably all gonna converge to roughly the same intelligence.

Speaker 3

事实上,以我认为至少谷歌和Meta的思维方式来看,即使其他人领先他们很多且效率更高,他们也会试图用金钱来解决这个问题。

And the reality is, given the way that I think at least Google and Meta are thinking, even if someone else is way ahead of them and more efficient, they will try to solve the problem with money.

Speaker 3

哦,天哪。

Oh, wow.

Speaker 3

他们只花了三千亿美元?

It only cost them 300,000,000,000?

Speaker 3

没问题。

No problem.

Speaker 3

我们很乐意花一万亿美元。

We're happy to spend a trillion.

Speaker 3

不过,最终经济规律还是会起作用。

Although, ultimately, economics will apply.

Speaker 3

这些可能会影响股价。

These may impact stock prices.

Speaker 3

比如,我认为这七个巨头中的一些公司,这种事并非不可想象。

Like, and I think it's not inconceivable, some of these magnificent seven.

Speaker 3

别管股票回购和股息了。

Forget buybacks and dividends.

Speaker 3

他们可能会取消股息,停止股票回购,转而发行新股来资助这项事业。

They may eliminate their dividends, stop buybacks, and start issuing stock to fund this.

Speaker 3

多年来,我们在很多方面都可能处在一个极其疯狂的世界中。

For years, we could be in a really wild world in a lot of ways.

Speaker 3

但这些智能体很可能最终会趋同于某种相似的‘智商’。

But these intelligences will probably converge on kind of a similar what I'll call IQ.

Speaker 3

而关键就在于谁拥有最独特的实时世界数据。

And then it's just who has the most differentiated real time data about the world.

Speaker 3

这些独特的数据源,加上你每次对这些模型的回答进行评分,都在帮助它们改进。

It's these unique data sources coupled with every time you rate an answer from one of these models, you're helping it improve.

Speaker 3

因此,如果你能把独特数据与互联网规模的分发结合起来,你就拥有了制胜公式。

And so if you can couple unique data with Internet scale distribution, then you're going have a winning formula.

Speaker 3

但只有少数几家公司具备这种能力。

And there's only a few companies that have that.

Speaker 3

比如xAI、谷歌、微软,尽管在互联网规模方面可能有所不同。

You know, it's xAI, it's Google, it's Microsoft, although Internet scale maybe.

Speaker 3

而OpenAI则通过微软获得这种能力。

And OpenAI gets that through Microsoft.

Speaker 3

可能是亚马逊和Anthropic通过这种方式获得优势,当然还有Meta。

Probably Amazon and Anthropic gets it through that, and then, you know, obviously, Meta.

Speaker 3

然后苹果是一个巨大的变数,我们应该谈谈这一点。

And then Apple is the big wild card, and we should talk about that.

Speaker 3

因为我认为最重要的一件事是,推理将在哪里进行?

Because I think one of the biggest things is where is the inference going to happen?

Speaker 3

计算资源往往在集中化和去中心化之间循环。

And compute tends to cycle in between centralization and the decentralization.

Speaker 3

我们现在正处于一个长期集中化的末期,大量计算都在云端的大数据中心运行,这是因为这些大型数据中心能实现更高的效率。

And we're at the end of, like, a long period of centralization in the cloud, where a lot of compute ran in the cloud in these big data centers is just because you could get much higher efficiency in these big data centers.

Speaker 3

显然,对于交易来说,这将发生在云端的巨型数据中心中。

Now clearly for trading, that is gonna happen in giant data centers that will be in the cloud.

Speaker 3

不过,我认为人们对这些AI数据中心的一个被低估的方面是,我们不必长期担心它们会压垮电力需求,因为它们可以建在任何地方。

Although something that I think is underappreciated about these AI data centers and why we shouldn't worry about them over the long term crushing power demand is you could put them anywhere.

Speaker 3

你可以把它们建在怀俄明州。

You put them in Wyoming.

Speaker 3

它们不需要靠近大城市。

Like, they don't need to be near a big city.

Speaker 3

我认为我们最终会看到巨大的数据中心建在页岩气田上,旁边配有发电厂,远离任何人类,这可能是美国短期内缺乏核能的解决方案。

I think we'll eventually see giant data centers in shale gas fields with power plants on those fields long away from any humans that will probably be an intermediate term solution to the lack of nuclear power in America.

Speaker 3

美国确实拥有所有最优秀的模型。

The US, we do we have all the best models.

Speaker 3

无论好坏,美国都希望这些模型在美国境内进行训练。

And The US wants these models trained in America for better or worse.

Speaker 3

显然,Mistral 是法国的。

Obviously, Mistral is French.

Speaker 3

我不知道这些模型是在哪里训练的。

I don't know where the models were trained.

Speaker 3

它们会根据在哪里能以最高利用率获得最低成本的计算资源,经历集中化与去中心化的循环,这类似于我前面描述的方程的一种变体。

They have these cycles of centralization and decentralization based on where can you get the lowest cost compute at the highest utilization, kind of a variant of the equation I described earlier.

Speaker 3

我认为这个方程的各个组成部分,不仅对实验室有帮助,对投资者也大有裨益。

And I do think the subcomponents of that equation, not only are they maybe helpful to labs, but they're really helpful to investors.

Speaker 3

如果你能将SFU提升20%,一切就归结于几毫米的硅片。

If you can improve SFU by 20, it all comes down to millimeters of silicon.

Speaker 3

仅仅多出几毫米的硅片,甚至可能更少,他们就会惊呼:天啊,因为硅片的毫米数才是成本的终极决定因素。

For only a few more millimeters of silicon or maybe even less millimeters of silicon, they're like, oh my god, because millimeters of silicon are the ultimate in cost.

Speaker 3

哇,你简直拥有一个极其成功的公式。

Wow, you have like a massively winning formula.

Speaker 3

所以,你可以几乎逐个审视这条漫长链条中的每个环节,哪里的每平方毫米硅片成本最高?

And so you can just almost go look at each step of that long chain and where do you have really high cost per square millimeter of silicon?

Speaker 3

而那里就是真正可以优化的机会。

And then that is the opportunity to really optimize.

Speaker 3

无论是在基础设施软件、数据中心硬件,还是半导体层面,机会都存在。

And whether it's at the infrastructure software, the data center hardware, the semiconductor layer, it's all there.

Speaker 3

但我认为,推理任务将越来越多地在手机上完成。

But I think you're gonna see inference increasingly done on phones.

Speaker 3

这显然是苹果的布局。

So this is clearly Apple's play.

Speaker 3

这是他们处于如此有利地位的原因之一。

It is one reason that they are in such a advantaged position.

Speaker 3

所有其他公司都像囚犯一样被困住了。

All these other companies are trapped when these prisoners deliver.

Speaker 3

我确信,考虑到成本如此之高,他们一定也希望如此。

I am sure they would like, given how much it costs.

Speaker 3

他们都是经济动物。

They're all economic animals.

Speaker 3

如果他们能达成一致,说:你知道吗?

If they could just reach an agreement and say, you know what?

Speaker 1

别杀了杰夫。

Not killing Jeff.

Speaker 3

在2026年之前,没人会搭建Blackwell集群,他们可能都会签字同意。

Nobody is gonna stand up a Blackwell cluster until 2026, they'd probably all sign it.

Speaker 3

但这简直就是经典的囚徒困境,而那将是一个纳什均衡。

But it is literally a classic prisoners to live, and that would be a Nash equilibrium.

Speaker 3

但在国家处于生存危机的情况下,我们不可能在创造数字上帝的竞赛中达成纳什均衡。

But we're not going to reach a Nash equilibrium in the race to create digital God when the states are existential.

Speaker 3

所以这里存在囚徒困境,但苹果并不在其中。

So there is prisoners to live, but Apple's not.

Speaker 3

同样地,谷歌在资本支出、运营支出等各方面累计投入了数百亿美元用于搜索。

By the same way, Google spends collectively, cumulatively spend, I don't know, hundreds of billions of search between CapEx, OpEx, all this stuff.

Speaker 3

而苹果凭借其在iOS系统中的分发垄断地位,几乎和谷歌一样有效地实现了变现。

And Apple monetizes it almost as well as Google because they have this distribution chokehold, Tollbooth, in iOS.

Speaker 3

他们显然会像谷歌在安卓手机上那样,对Apple Intelligence采取同样的做法。

They're clearly gonna do the same thing with Apple intelligence that Google will obviously do that with their phone, the Android.

Speaker 3

因此,你的手机上将运行一个小型模型,用于处理简单问题,可以把它想象成一个拥有100智商、具备极其丰富知识的模型。

And so you're gonna have a small model running on your phone that for simple questions, think of it as like a 100 IQ model that with, like, really, really sophisticated knowledge.

Speaker 3

可能会有两个这样的模型。

There'll probably be two of them.

Speaker 3

它们会相互核查。

They'll check each other.

Speaker 3

因此,大量推理将在边缘端进行。

And a lot of inference will then happen at the edge.

Speaker 3

如果推理在边缘端进行,我认为我们将看到超级手机。

And if inference is happening at the edge, then I do think we're gonna see superphones.

Speaker 3

因为对你作为人类而言,限制推理的主要因素通常是内存,而如今iPhone的售价正是基于其闪存容量。

Because the advantage to you as a human, what bounds inference at most times is memory, and you will be willing to pay today iPhones are sold based on how much flash storage they have.

Speaker 3

你最关心的将是手机上的DRAM容量,因为它将决定你能本地运行的模型参数规模。

What you will most care about is the amount of DRAM you have on your phone because that will determine the perimeter count of the model you can run locally.

Speaker 3

而这正是你本地智能的质量,它能以隐私安全的方式访问你所有的数据。

And that is therefore the quality of your own local intelligence that has access to all of your data in a privacy safe way.

Speaker 3

当你向它提问时,如果能在本地完成,它就会在本地处理。

And when you ask it for something, if it could be done on a compute, it will.

Speaker 3

我认为在网络领域,一个真理是:能路由时就路由,必须时才交换。

And I think for networking, like a truism is route when you can, switch when you must.

Speaker 3

但我觉得,对于推理来说,下一步将是:能本地处理时就本地处理,必须时才用云端。

But I think the next thing for inference will be local when you can, cloud when you must.

Speaker 3

所以,如果你能在手机上本地进行推理,你总会选择这样做,因为它是免费的。

So if you can inference locally on your phone, you will always do that because it's free.

Speaker 3

而云端推理肯定是要花钱的,因为你正在消耗GPU小时。

And cloud inference definitely costs money because you're burning GPU hours.

Speaker 3

现在虽然已经非常高效了,但在手机上进行推理实际上是免费的。

Now it's very efficient, but the inference on your phone is effectively free.

Speaker 3

然后,作为人类,我们的竞争力正变得越来越明显,这很有趣。

And then increasingly, our competitiveness as humans you know, it's very funny.

Speaker 3

很多非常聪明的人对人工智能持一点怀疑态度。

A lot of really smart people are a little skeptical about AI.

Speaker 3

如果你的智商是125、135或者更高,我能理解。

I get it if, like, you have a 125 IQ or a 135 or whatever it is.

Speaker 3

人工智能,尤其是在你的领域,并不让你觉得有多了不起。

AI, particularly your domain, isn't at all impressive to you.

Speaker 3

然后你就说,好吧。

And it's like, okay.

Speaker 3

对我来说,这有点像通过搜索来了解其他领域。

It's a little bit for search for me learning about other domains.

Speaker 3

但老天,很多人的智商并没有达到120。

But, man, a lot of humans don't have one twenty IQs.

Speaker 3

人们忽视的是,有很多人的智商,比如说,我不知道。

And what people are missing is that there's a lot of humans with, like, I don't know.

Speaker 3

我完全不知道。

I have no idea.

Speaker 1

100的智商。

A 100 IQs.

Speaker 1

最多的是100智商的人。

The most with a 100 IQs.

Speaker 1

是的。

Yeah.

Speaker 3

那就假设是100智商吧。

So let's just say it's a 100 IQs.

Speaker 3

他们可以使用AI,突然间,他们的智商就像变成了115。

And they could use AI, and all of a sudden, they're like a one fifteen.

Speaker 3

这简直让人震惊。

And it's like, holy shit.

Speaker 3

太不可思议了。

It's incredible.

Speaker 3

而这种情况会持续下去,直到这些ASI的智商达到一千,那时人们就会想:作为人类,我为什么还要做其他事,而不只是和AI一起创作艺术呢?

And that's just gonna keep going until these ASIs have IQ of a thousand, and it's like, why, as a human, am I bothering to do anything other than work with an AI to create art?

Speaker 3

不管怎样,如果我能拥有一部价值3000美元的iPhone,它的DRAM是千元iPhone的四倍,存储空间也更大,以便本地模型能运行RAG。

Anyways, if I could have, like, a $3,000 iPhone that has four times the DRAM on it that the thousand dollar iPhone has, it did maybe more storage so that local model could do rag.

Speaker 3

这是我的模型。

And this is my model.

Speaker 3

它喜欢我。

It likes me.

Speaker 3

我选择用来和我对话的声音。

I pick the voice that talks to me in.

Speaker 3

它了解我。

It knows me.

Speaker 3

它是我的朋友。

It's my friend.

Speaker 3

它是我的代理。

It's my agent.

Speaker 3

在改进这些模型的过程中,所有这些都同时为我点赞。

All at once is thumbs up for me in that process of improving these models.

Speaker 3

我认为最终会有一种方式,让它们能够按需为个人提供RLHF支持,但这需要大量突破才能实现。

I think they will ultimately be there will be a way that they could be RLH staffed on an individual basis on the fly, which we're gonna need a lot of breakthroughs to make that happen.

Speaker 3

它会是我的模型,喜欢我、了解我,并知道我想要什么。

And it's gonna be my model that likes me and knows me and knows what I want.

Speaker 3

然后,这就是大家所谈论的代理未来:我手机上有一个代理。

And then this is where you get this agent future that everybody's talking about, where it's like I have an agent on my phone.

Speaker 3

如果我手机上有一个代理,而我的超级手机拥有115或120的智商,而别人的代理只有100的智商,那么作为人类,我会拥有显著的优势。

And if I have an agent on my phone, because I have a super phone that has, you know, an IQ of a 115 or a 120 and someone else's agent only has an IQ of a 100, I'm gonna be profoundly advantaged as a human.

Speaker 3

然后这种趋势会一直持续下去,因为我相信苹果最终会从中盈利。

And then that continues all the way up because I think eventually Apple will monetize this.

Speaker 3

苹果实现盈利的方式显然是因为设备上的LLM还不够智能。

And the way Apple will monetize this clearly is with the odd device LLM isn't smart enough.

Speaker 3

它们会将任务发送到云端。

They're gonna send it to the cloud.

Speaker 3

它们会使用一种叫做路由器的技术,这是每家AI应用公司都在使用的。

They're gonna use something called a router, which is what every AI application company uses.

Speaker 3

你只想根据每美元的成本,将查询路由到最合适的模型。

You just wanna route to the best model for the query per dollar.

Speaker 3

苹果会向公司收费,以让它们进入自己的路由器系统。

Apple will charge companies to be in their router.

Speaker 3

然后,如果你想被路由到更高级的模型,最终你不得不支付越来越高额的费用。

And then, oh, if you wanna be routed to, eventually, you're gonna have to pay a bigger and bigger toll.

Speaker 3

每个人都会像谷歌那样支付这笔费用。

And everybody will pay that toll the same way Google paid that toll.

Speaker 3

但如果我作为一个普通人,能在本地拥有一个更智能的模型,然后我可以选择让苹果让这个过程变得很简单。

But if I, as a human, I could have a smarter model locally, and then I could opt in into maybe Apple will just make it easy.

Speaker 3

他们会说:好吧,你可以选择云超级智能,或者选择云智能。

They'll say, okay, you can have cloud superintelligence or you can have cloud intelligence.

Speaker 3

而我每月支付60美元来获得云超级智能。

And I pay $60 a month for cloud superintelligence.

Speaker 3

不管它是每月1000美元、10000美元,为了这个你愿意付多少钱?

Now, or whatever it is, dollars 1,000 a month, dollars 10,000 a month, what wouldn't you pay for that?

Speaker 3

所以我的手机比很多人多出了20点智商,而我每月支付10000美元来获得云超级智能、超智能。

And so I have 20 points of IQ on my phone relative to a lot of people, and then I'm paying $10,000 a month for cloud superintelligence, hyperintelligence.

Speaker 3

超级智能每月只要1000美元,而普通智能每月20美元。

Superintelligence is just $1,000 a month, and then regular intelligence is $20 a month.

Speaker 3

作为一个人,我会获得巨大的优势。

Like, I'm going to be really advantaged as a human.

Speaker 3

这在我看来很反乌托邦,但我也很难不认为它会真的发生。

And that seems dystopian to me, but it's also kind of hard for me to not see that happening.

Speaker 3

许多投资结论都由此衍生,就像说,嘿。

And many, many investment conclusions flow from this in the same way that they flow from like, hey.

Speaker 3

好的。

Okay.

Speaker 3

猛犸象公司达到了90%。

Mammoth is at 90%.

Speaker 3

好的。

Okay.

Speaker 3

所以这可能不是一个好的投资地点。

So that's probably not a good place to invest.

Speaker 3

SFU达到了30%?

SFU is at 30%?

Speaker 3

哇。

Wow.

Speaker 3

这是个绝佳的投资地点。

That's a great place to invest.

Speaker 3

PUE为1.8,理论上可以达到1.3。

PUE is at 1.8, and it could theoretically be at 1.3.

Speaker 3

在这个等式中,这是个绝佳的投资地点。

That's a great place to invest in that equation.

Speaker 3

你最好投资于效率最低的那条链。

You kind of want to invest at the most inefficient chain.

Speaker 3

推理发生的位置会带来大量投资影响。

A lot of investment implications flow from where does inference happen.

Speaker 3

而这对人类也会产生许多深远影响。

And then a lot of implications for humanity flow from that as well.

Speaker 1

我很想聊聊这个等式中更普通的一面。

I'd love to talk a little bit about the commoner part of this whole equation.

Speaker 1

我们谈到了奥林匹斯山,以及那些为争夺主导权而相互竞争的王者游戏。

We talked about Mount Olympus and the game of kings fighting each other for dominance.

Speaker 1

但我们还没有讨论过,一个普通的新公司如何利用这些超级智能在应用层开展业务,或者做其他事情?

We haven't talked at all about what about just a normal new company that's being started to take advantage of these super intelligences at the application layer or to do something else?

Speaker 1

之后我们再谈机器人。

Then we'll get to robotics after that.

Speaker 1

但如果我迫使你只能作为初创公司的早期A轮或B轮投资者,而它们没有这些卓越七家公司及其他巨头所拥有的资源,你会如何看待在GPTX持续扩展和提升的世界中,哪些公司和机会最具吸引力?

But if I force you to just be a early stage series A, series B investor in the world of startups and they don't have this kind of resources that all these magnificent seven and others have, how would you think about the aspects of companies and opportunities that would get most attractive in the world where GPTX keeps scaling and getting better?

Speaker 3

我首先想说的是,关键始终在于人们在做什么。

First thing I would just say, it's always what are people doing?

Speaker 3

我目前专注于那些改进我前面提到的方程要素的公司,因为我认为……

What I am doing is really focusing on companies that improve that equation elements of that equation I described earlier because I think

Speaker 1

这是一个瓶颈。

That's a choke point.

Speaker 3

这是最有可能成功的方向。

That's the highest likelihood of success.

Speaker 3

在科技领域,只要你发现了一个瓶颈,如果能投资于缓解该瓶颈的领域,通常都会取得不错的结果。

And in tech, whenever you find a constraint, if you can invest in something that alleviates that constraint, you usually do well.

Speaker 3

而目前,我深信瓶颈在于SFU、检查点,这关系到可靠性,以及PUE。

And right now, I profoundly believe the constraint is an SFU, checkpointing, which goes to reliability, and then PUE.

Speaker 3

因此,这就是我投资的方向。

So that is where I am targeting my dollars.

Speaker 3

我认为应用层真的非常困难。

I think the application layer is really, really hard.

Speaker 3

有些人在这方面做得非常出色。

There are people who are killing it.

Speaker 3

我觉得你播客里的莎拉,我虽然从未见过她,但非常钦佩。

I think you have Sarah, who I've never met, but I admire on your podcast.

Speaker 3

从我的角度看,她在应用层上做得简直太棒了。

And she's, like, absolutely, from my perspective, crushing it at the application layer.

Speaker 3

但说实话,今天投资应用层对我来说真的很难。

But just, geez, investing there today feels really hard to me.

Speaker 3

我只是觉得,任何对这一点充满信心的人都应该记住查斯·科尔曼那1%的数据。

And I just think anyone who has a lot of conviction about that needs to be reminded of that 1% Chase Coleman stat.

Speaker 1

是的。

Yeah.

Speaker 1

我刚才也在想这一点。

I was just thinking that.

Speaker 3

所有这些风投都对这样一种观点非常坚定:那些在1995年、1996年、1997年获得融资的公司,似乎都抓住了互联网的机遇。

All these VCs had really strong views of, like, there are all these companies that were funded in '95, '96, '97, and they seemed like they had got in the net.

Speaker 1

但这需要时间。

But it just takes time.

Speaker 3

但我想,他们当时确实是为风投而进入这个领域的。

But I guess they did go into that for the VC.

Speaker 3

他们上市了,并且成功套现了。

They went public, and they were able to cash out.

Speaker 3

但他们实际上并没有真正深入理解互联网的长远价值。

But they actually didn't go into the net with the fullness of time.

Speaker 3

所以我只是觉得,在应用层面上保持极大的谦逊非常重要。

So I just think it's important to have a lot of humility at the application layer.

Speaker 3

也许只是因为我对基础设施层非常熟悉。

And maybe it's just that I am so comfortable with this infrastructure layer.

Speaker 3

到目前为止,我一直专注于这一领域。

That is where I have concentrated to date.

Speaker 3

实际上,只有一个应用公司让我感到兴奋。

There's really only been one application company that I've been excited about.

Speaker 3

我们见过很多这样的公司。

We've seen a lot of them.

Speaker 3

也许我太了解蔡斯·科尔曼的数据了,因此在对待这一应用层时过于谨慎。

Maybe I'm, like, too aware of the Chase Coleman stat, and and I'm being too careful in how I approach this application layer.

Speaker 3

但我可能会错过很多像CMGI那样了不起的公司。

But I'm going to miss out on a lot of CMGI was incredible.

Speaker 3

除了雅虎之外,还有哪些公司?

What were some other besides Yahoo?

Speaker 3

Lycos呢?

What were Lycos.

Speaker 3

还有许多我们今天甚至想不起来的公司,但它们都是非凡的风险投资成果。

There are all these companies that we don't even think of today, but they're incredible venture outcomes.

Speaker 3

也许我确实需要更加留意这一点,但显然有一些人,无论是Benchmark还是Cerusfirm,都在取得成功。

Maybe I do need to be a little more mindful of that, but they're clearly people, whether it's Benchmark, whether it's Cerusfirm, who are succeeding.

Speaker 3

我在谈到应用层时提到的这一点,是听Vishria说的,但真的深深引起了我的共鸣。

And what I was saying in the application there, I heard this from Vishria, but it really resonated with me.

Speaker 3

我认为这也涉及到整个AI技能投资回报率的争论。

And I think it also goes to this whole skilling AI ROI debate.

Speaker 3

这不仅是一个关于创造数字神明的投资回报率争论,我认为还存在一个非常明确的投资回报率,而且你实际上可以通过数学计算来证明。

Not only is it an ROI debate about creating a digital god, but I think there's a super clear ROI, and you can actually do math to show that.

Speaker 3

很多人将资本性支出和运营性支出混为一谈,这种方式毫无帮助。

And a lot of people are conflating CapEx and OpEx in ways that are just not helpful.

Speaker 3

Meta的股价下跌了约80%,部分原因是他们在元宇宙上过度支出,但部分原因也是因为苹果通过IDFA剥夺了他们的定向投放能力。

Meta went down, like, 80% partially because they were overspending on the metaverse, but partially because Apple took away their ability to target through IDFA.

Speaker 3

Meta的估值是五倍,其收入增长正在重新加速。

Meta's a five x, and the revenue growth is reaccelerated.

Speaker 3

为什么它的增长会重新加速?

Why is it reaccelerated?

Speaker 3

因为他们投入了大量资金用于人工智能,以研究如何精准投放广告。

Because they spent a vast amount amount of money on AI to figure out how to target.

Speaker 3

也许仅凭元宇宙的回报就足以证明所有支出的合理性。

It's like maybe the meta return alone justifies all of the spending.

Speaker 3

他们称之为元宇宙优势,但元宇宙优势只是其中一部分。

They call that meta advantage, but meta advantage is just part of it.

Speaker 3

作为广告主,你可以让元宇宙平台来负责定向投放。

That's just where as an advertiser, you can let meta do the targeting.

Speaker 3

你还需要谷歌的性能最大化广告。

You need and then Google has performance max.

Speaker 3

所有这些都曾是广告创意产生的地方。

And all of that is that's where the it used to be ads were created.

Speaker 3

我们稍后会回到应用场景。

We're gonna come back to applications.

Speaker 3

广告创意曾经是由人类来决定的:我们需要向波士顿一位40岁的白人男性展示这种创意,向这个城市中特定人群展示另一种创意,并在一天中的特定时段投放。

But ads were created, we're gonna pay some humans to figure out that we need to show this creative to white 40 year old guy in Boston, and then this type to this type of demographic category in this city, and we're gonna show them at this time of day.

Speaker 3

不,我们要在本地体育队赢球且天气晴朗时展示广告,因为那是看广告的最佳时机。

No, we're gonna show them after the local sports team is won and it's sunny because that's actually the best time to see an ad.

Speaker 3

但现在AI可以完成所有这些工作。

But now AI can do all of that.

Speaker 3

我们将拥有一百万种创意,并且会实时优化。

And we're gonna have a million creatives and it's gonna be optimized on the fly.

Speaker 3

这就是AI正在帮助谷歌、Meta和其他公司实现的目标。

That's what AI is enabling Google and Meta and other firms to do.

Speaker 3

而且可能仅凭这一点

And probably just on that

Speaker 1

仅这一点就足以证明其价值。

Just that justifies it.

Speaker 1

是的。

Yeah.

Speaker 3

AI已经带来了巨大的投资回报。

There has been a massive ROI on AI.

Speaker 3

我们在谈什么?

What are we talking about?

Speaker 3

关于这场AI投资回报率的争论,另一件让我觉得特别有趣的是:好吧,我们要进行一场关于投资回报率的晦涩辩论。

And then the other thing that's really funny to me about this whole AIROI debate is, okay, we're going to have this abstruse debate about ROI.

Speaker 3

这些公司都是上市公司,还有一种叫做资本回报率的指标。

These companies are all public, and there is something called return on invested capital.

Speaker 3

自从这些公司增加资本支出以来,它们的资本回报率就一直在上升。

And ROIC has gone up for all of these companies since they ramped CapEx.

Speaker 3

我们在谈什么?

What are we talking about?

Speaker 3

如果你对AI的投资回报率持怀疑态度,为什么这些公司的资本回报率会上升?

If you're an AI ROI skeptic, why has ROIC got up at these companies?

Speaker 3

是的,资本支出确实增加了。

Well, yeah, CapEx is up.

Speaker 3

笔记本的支出增加得更多。

Notepad is up more.

Speaker 3

为什么会这样?

Why is that?

Speaker 3

因为他们正在做你在AI世界中预期会发生的事情。

Because they're doing exactly what you would expect to happen in the world of AI.

Speaker 3

他们正在用GPU小时替代人力劳动。

They're trading off human labor against GPU hours.

Speaker 3

这就是为什么他们的ROIC会上升,而这些GPU确实非常高效。

That's why their ROIC has gone up and these GPUs are really efficient.

Speaker 3

让我们来讨论一下AI的投资回报率吧。

Let's have an AI ROI debate.

Speaker 3

当那些大规模投入AI的公司的ROIC开始下降时,才值得讨论;在此之前,这纯粹是智力上的荒谬。

When the ROIC at the big AI spenders starts to go down, until then, it's just the height of intellectual ridiculousness.

Speaker 3

我的意思是,荒谬至极。

I mean, ridiculous.

Speaker 3

维什里亚提到的一点,我觉得对应用层非常有力:有这么多SaaS指标。

What Vishriya said that I thought was so powerful for the application layer, there's all these SaaS metrics.

Speaker 3

第一年后,你应当达到百万级别才能跟上节奏。

After your first year, you ought to be at million to be on pace.

Speaker 3

第二年后,你希望达到五百万。

After your second year, you want to be at 5,000,000.

Speaker 3

第三年后,你希望达到大约一千万。

After your third year, you want to be at like 10.

Speaker 3

如果你的现金消耗合理且超出这个水平,你很可能能跑通这些曲线,成为一家成功的SaaS公司。

And if you're above that with reasonable cash burn, you could probably run these curves, you're going be a successful SaaS company.

Speaker 3

所有这些曲线,大家都知道,正是SaaS估值倍数被推高的原因之一,因为它几乎变得像量化指标一样。

And all those curves, which everybody knew is one reason SaaS multiples got so inflated because it became almost like quantitative.

Speaker 3

哇,这家公司在第三年就达到了150亿美元?

Oh, wow, this company is at $15,000,000,000 in year three?

Speaker 3

这意味着我们可以预测它在第八年将达到十亿美元。

That means we can pencil them in for a billion dollars in year eight.

Speaker 3

A,这导致估值倍数被疯狂推高。

A, that led to multiples expanding ridiculously.

Speaker 3

B,这导致整个行业融资过度,使得每个垂直领域的竞争加剧,导致这些曲线不再成立。

B, it led to the industry being overfunded such that the degree of competition in each vertical went up such that the curves no longer held.

Speaker 3

这也是为什么如果你在2021年左右进行了大量SaaS投资,你现在会陷入困境的原因之一。

And this is one reason why if you made a lot of SaaS investments, particularly in 2021, you're in a world of pain.

Speaker 3

此外,AI出现了,它从根本上改变了应用软件的范式,因为应用软件的核心作用是提升人类效率。

And then on top of that, here AI comes along, and it fundamentally changes the paradigm for application software because what application software fundamentally does is makes humans more efficient.

Speaker 3

而如今,我们正处于AI时代,它确实提升了人类效率,但你不再需要围绕它构建那么庞大的系统。

And today, we're in state with AI where it's making humans more efficient, but you don't need that big of a wrapper around it.

Speaker 3

这就引出了我们关于渔业的评论。

That's where we're gonna come to the fishery comment.

Speaker 3

最终,如果AI开始取代人类,而你以每用户席位模式销售应用软件,当用户席位开始减少时,你就面临问题了。

And then ultimately, if it's replacing humans, if you're selling application software on a per seat model and those seats start to go down, you have a problem.

Speaker 3

这些公司对此非常清楚。

These companies are super aware of it.

Speaker 3

它们曾经一路狂奔,但现在却不得不在每个垂直领域应对新一代以AI为核心的公司,这些公司通常只是围绕GPT构建的极其轻量的封装。

They were going fast in one direction, and now they have to contend almost in every vertical with this next generation of AI first application companies, which are just these often very thin wrappers around GPT.

Speaker 3

你会称它为一个LLM包装器。

You'll call it an LLM wrapper.

Speaker 3

渔业部门告诉我一件特别有趣的事:所有这些AI公司都在彻底颠覆传统的SaaS指标。

And what Fisheries said to me that was so funny is like all these AI companies are blowing these traditional SaaS metrics out of the water.

Speaker 3

但这些情况并没有在那些AI投资回报率论文中被提及。

And all this isn't being touted in those AI ROI papers.

Speaker 3

现在有这么多公司能在九个月内从零做到三千万美元的收入。

Just that there's all these companies that are going like 0 to 30,000,000 in nine months.

Speaker 3

尽管AI的边际成本其实很高,但它们相对于传统软件公司却更注重现金流效率。

And even though AI is really high marginal cost, they're being more cash flow efficient relative to software company.

Speaker 3

因此,在几乎每一个垂直领域,都有多家AI优先的公司,正如Vishri所说,它们只是围绕你选择的任何LLM做的极其轻量的包装。

So in almost every vertical, there's multiple companies that are AI first that are just really I think what Vishri said was they're just paper thread wrappers around pick your LLM of choice.

Speaker 3

然后,是的,它们会使用路由机制来选择最佳的LLM,但对客户来说,它们就像魔法一样。

Then, yeah, they use a router to find the best LLM, but they're like magic to their customers.

Speaker 3

它们并不是在争夺软件预算。

They aren't going after software budgets.

Speaker 3

他们瞄准的是人力预算。

They're going after labor budgets.

Speaker 3

观察这一切对我来说非常有智力上的吸引力。

It's very intellectually interesting to me to watch.

Speaker 3

你如何获得这样的信念:这些公司中的一家能够利用最初对客户来说如同魔法的功能来建立护城河,同时承认大型模型将继续以极高的速度持续进化?

How do you get conviction that one of these companies can use what is initially magical to their customer to build defensibility in, all while allowing for the fact that the big models are going to continue to compound at a really high rate?

Speaker 3

这可能涉及到找到一种非常有效的销售方式。

And this goes to things like maybe you just find a really good sales motion that works.

Speaker 3

也许你找到了一个非常容易集成的切入点。

Maybe you find a really easy integration point.

Speaker 3

也许你在这种集成基础上建立了某种护城河。

Maybe you build some defensibility around that integration.

Speaker 3

也许你让小企业能够在非常原始的系统上轻松实现RAG。

Maybe you make it really easy for a small business to do rag on a really unsophisticated system.

Speaker 3

你如何在这一层封装中建立差异化?

How are you building differentiation into that wrapper?

Speaker 3

而且希望你不要进行微调,因为深度微调会把你锁定在某个模型生成上。

And hopefully, you're not fine tuning because that does extensively fine tuning because that locks you into a model generation.

Speaker 3

你如何构建一个复合型AI系统,使得每次查询都能以最低成本完成,并且从一个小模型开始,逐步过渡到大模型?

How are you creating a compound AI system such that you're serving each query at the lowest cost possible and you're starting with a small model and treating it to a big model?

Speaker 3

你可以做很多事情,而这些都非常重要。

There's a lot of things you can do, and all of those are really, really important.

Speaker 3

但我只是觉得,在应用层面上,似乎任何你能想象到的类别,都有多个初创公司迅速崛起,完全颠覆了传统的SaaS指标。

But I just think at that application layer, there's just it seems like for any category, any category it feels like that can be imagined, there's multiple startups that have exploded, that are defying all traditional SaaS metrics.

Speaker 3

到目前为止,我还不清楚它们每一个究竟有多少竞争优势。

And it's not clear to me at this point how much defensibility each one of them has.

Speaker 3

其中一些可能会在我说的某个维度上拥有中等程度的竞争壁垒,但我认为很多可能并没有。

Some of them are going to have a mid defensibility around one of the axes I described, but I think a lot of them may not.

Speaker 3

我认为这正是这个领域如此难以把握的原因,也是我为何如此谨慎地对待这个领域。

And I think that's what makes it so hard and why I've been approaching that space so carefully.

Speaker 1

你知道最令人惊叹的是什么吗?

You know the amazing thing?

Speaker 1

我认为关于劳动力的这种说法,举个最简单、最普通的例子来说吧。

I think the framing around labor I mean, just take me as a stupid, simple, tiny example.

Speaker 1

我有一个由三个人组成的团队在开发这样一个系统。

I have a team of three people building one of these.

Speaker 1

我们可以把它称为一个轻量级的LLM包装器,用于对私营公司进行研究。

We'll call them like a really light LLM wrapper for doing research on private companies.

Speaker 1

如果你想想我每天使用这个工具的频率,就像我以前使用分析师那样,一天好几次,而且它返回的结果质量相当。

And if you just think about how often I use this thing in a way that I normally would have used an analyst, it's multiple times a day, and it's returning work as good.

Speaker 1

所以,在每一个白领劳动市场中,都有大量没有差异化的繁重工作,我想这是其中一个主要的发现。

So there's so much undifferentiated heavy lifting in every white collar labor market, I guess, is like one major learning.

Speaker 3

100%。

100%.

Speaker 3

如果你

If you

Speaker 1

能立即获得答案,而且几乎是免费的,那你会多么频繁地使用这些工具,简直难以置信。

can have the answers instantly and for effectively free, it's crazy how much you use these things.

Speaker 3

我们才刚刚开始。

We're just getting started.

Speaker 3

是的。

Yeah.

Speaker 3

在我的公司,我也有一样的情况,用了一个大语言模型。

And I have the same thing with, like, an LLM at my firm.

Speaker 3

这其实非常有趣。

It was actually very interesting.

Speaker 3

我们有一个非常有才华的实习生。

We had an intern who was really talented.

Speaker 3

他说:听着,我觉得这个内部AI工具太棒了。

And he said, Listen, I think this internal AI tool is amazing.

Speaker 3

我每天都用,而且使用频率越来越高。

And I use it every day and more every day.

Speaker 3

今年夏天,它变得好太多了。

It's gotten much better this summer.

Speaker 3

我们为此投入了很长时间。

We work at it for a long time.

Speaker 3

然后我说,给我看看你是怎么用的。

And then I was like, Show me how you use it.

Speaker 3

因为我一直用的是某种方式。

Because I'd been using it a certain way.

Speaker 3

结果这个21岁的孩子说,嗯,我就是这样用的。

And then this literally, this 21 year old kid was like, Well, this is what I do.

Speaker 3

突然间,我使用这个工具的时间变成了每天两小时、三小时、四小时。

And now all of a sudden, my usage of this tool is two hours a day, three hours a day, four hours a day.

Speaker 3

我认为杰克·韦尔奇有个术语叫‘琐碎工作’,就是在白领环境中必须做得很好的那种辛苦、不愉快的工作。

And I do think Jack Welch had this term called scut work, just like hard, unpleasant work that had to be done really well in a white collar setting.

Speaker 3

我认为这些所谓的‘小助手’或‘迷你助手’——不管我们怎么称呼它们——将取代大量这类琐碎工作。

And I think it's all these little rappers or mini rappers, whatever we're gonna call are gonna replace a lot of that scutt work.

Speaker 3

最初,它们会与人类协同工作。

And initially, it's gonna be in combination with humans.

Speaker 3

在科幻小说中,尼尔·阿舍的世界里有一种叫做‘高频带’的东西,是一种人机混合体。

With science fiction, Neil Asher's world, there's something called a high band, which is a human AI hybrid.

Speaker 3

它们通过类似Neuralink的装置连接到一个由外部结构支撑的计算机服务器上,这些外部结构是机器人外骨骼,人们可以靠它行走。

And it's where they're linked through something like a Neuralink to, like, a computer server that is supported by, like, an exo structure, robotic exo structure that they walk around on.

Speaker 3

这就像某种融合象棋,你怎么叫都行。

There's, like, fusion chess, whatever you want to call it.

Speaker 3

我们会有一段时间处于这种状态。

We'll have that for a while.

Speaker 3

然后,100智商的人表现得像120智商的人,接着表现得像130智商的人,再之后130智商的人表现得像160智商的人。

And then that goes to the 100 IQ humans performing like 120 IQ humans, and then performing like 130 IQ humans, then 130 IQ humans performing like 160 IQ humans.

Speaker 3

但最终,只要规模定律持续下去——这本身就是一个巨大的未知数——最终就只会是AI了。

But then eventually, it feels like as long as scaling laws continue, which is a big F, it's just gonna be the AIs.

Speaker 1

也许我们可以聊聊机器人技术。

Maybe we could talk about robotics.

Speaker 1

我曾在四月份与一位投资者进行了一次非常有趣的对话,他长期私下和公开地投资于这些领域,持有我们今天讨论的许多公司的大量股份。

I had a really interesting conversation, call it April, with an investor that has been investing in lots of these same things for long periods of time, privately and publicly, and has big positions in lots of the companies that we've been talking about today.

Speaker 1

他向我指出,在未来五年左右的时间里,人们严重低估了机器人与我们今天讨论的全部技术相结合所发挥的作用。

His observation to me was the big underestimation that's happening over, let's say, five years is the role that robotics and robots will have combined with all of this technology we've spent all of today talking about.

Speaker 1

我很想听听你对这一点的看法,因为短期内,感觉这些公司融资热潮有些过热,你根本不清楚这些机器人到底设计来做什么,基本上都是通用人形机器人。

And I would love to hear you riff on that because in the near term, it feels like a little bit quite a bit of frothiness, like some crazy funding rounds for these companies that you don't really know what the robots are being designed to do, sort of general purpose humanoid type robots.

Speaker 1

还有很多更专业的机器人也很有趣。

There's all sorts of interesting, more specialized ones that are cool too.

Speaker 1

但你对这一切有什么看法?

But what do you think about all this?

Speaker 1

相对于基础模型和半导体领域,机器人似乎被讨论得少得多。

It does seem kind of under discussed relative to just all the foundation model and semiconductor stuff.

Speaker 3

我同意。

I agree.

Speaker 3

我认为,机器人可能带来的短期颠覆性影响,会超过我们刚才讨论的白领劳动自动化。

I think it may end up being a bigger near term disruption than what we were just discussing, the automation of a lot of white collar labor.

Speaker 3

我认为,第一个真正会对世界产生影响的机器人,将是每辆配备特斯拉所谓AI四代硬件的特斯拉汽车。

I think the first robot the first robot that's really gonna impact the world is every Tesla car with what they call their AI four hardware.

Speaker 3

从我的角度来看,公开的无干预行驶里程数是存在的。

Because from my perspective, there's a publicly sourced miles between disengagement.

Speaker 3

所以你必须记住,对于特斯拉来说,特斯拉的无干预行驶里程数也会是一样的。

So you have to remember for Tesla, Tesla is gonna get the same miles between disengagements.

Speaker 3

比如,如果你在火星上建了一座新城市,那里全是外观不同、街道布局各异的汽车,你把一辆特斯拉放进去,它在那座城市的无干预行驶里程数,也会和在其他任何城市一样。

Like, if you built a new city on Mars and it was populated by entirely different looking cars and streets and everything, you could drop a Tesla in that city, and it would have the same miles between disengagements that it gets in any other city.

Speaker 3

而像小马智行这样的系统则是有地理围栏限制的。

Whereas something like Weibo is geofenced.

Speaker 3

我们实际上只在那些拥有良好网格道路和良好天气等条件的城市中使用它。

We're really only using it in cities that have, like, nice grids and good weather, etcetera, etcetera, etcetera.

Speaker 3

通过观察不同版本FSD的众包数据中的无干预行驶里程数,我清楚地看到,当他们切换到12.3版本时——这本质上是完全基于深度学习的系统,我认为几乎消除了所有人工编写代码——进步的速度发生了显著变化。

It is clear to me looking at the crowdsourced data, miles between disengagements with different versions of FSD, that when they cut over to twelve three, which is effectively all deep learning, I think eliminated almost all human code, something dramatic changed in the rate of progress.

Speaker 3

然后当他们切换到12.5版本时,这个版本在AI四代硬件上运行效果最佳,也就是以前被称为HW四代的硬件。

And then when they cut over to 12.5, which runs best on the AI four, which used to be called the HW four.

Speaker 3

这其实就是特斯拉的本地计算机。

It's just like the local computer of the Tesla.

Speaker 3

现在它正在推广到AI3,这是另一个跃升。

It is now rolling out to AI three, was another step function.

Speaker 3

根据同样的扩展规律,这些跃升性的改进是在远少于特斯拉现在在奥斯汀超级工厂数据中心公开部署的算力下实现的。

And these, going to that same scaling law, those step function improvements were made with a fraction of the compute that Tesla is now installing publicly in their data center at the Gigafactory in Austin.

Speaker 3

他们实际上已经申请了一些专利——有时作为投资者,我希望他们别申请这么多专利——但这些专利在数据中心冷却方面非常创新,与他们在做的工作相关,据公开消息称,他们将部署超过5万台H100或H200芯片。

And they've actually filed sometimes I wish they as an investor, I wish they couldn't file so many of these patents, but they've filed some really innovative patents for data center cooling related to what they're doing with what looks I think it's been publicly said it's gonna be, you know, over 50,000 h one hundreds or h two hundreds.

Speaker 3

FSD现在遵循同样的扩展规律,甚至可以说遵循更快的扩展规律,因为他们需要追赶GPT已经走过的路。

FSD is now on the same scaling law and arguably on a faster scaling law because they have a lot of catch up to do that GPTs have been on.

Speaker 3

所以我认为12.5版本就像GPT-3,它能稳定地在大多数地方自动驾驶而无需人工干预。

So I think 12.5 is like GPT three, and it can consistently drive me most places with no interventions.

Speaker 3

我是个季节性驾驶员。

I'm a seasonal driver.

Speaker 3

我只在夏天开我的特斯拉。

I really only drive my Tesla in the summer.

Speaker 3

实际上,我妻子贝基开车比我多,因为她比我更喜欢开车。

And, actually, my wife, Becky, tends to do most of the driving because she likes it more than me.

Speaker 3

所以我们基本上能看到一种季节性的变化。

So we kinda get, like, a seasonal look.

Speaker 3

几乎每到五月,我们就会检查一次进展。

It's almost like every May, we check-in.

Speaker 3

一直以来,进步都是持续不断的。

There was just always continuous progress.

Speaker 3

今年,当我们启用12.3版本时,感觉就像自从我第一次拥有那辆特斯拉以来,过去十年的所有进步都浓缩在了这一个版本里。

This year, it's like when we turned on 12.3, it was like all of the progress over the last ten years was in that one release from the first time I had that Tesla.

Speaker 3

当我从12.3升级到12.5时,又出现了同样的情况。

And then we had that again when I went 12.3 to 12.5.

Speaker 3

他们目前的算力大概处于GPT-2级别。

And they're at probably a GPT-two level of compute.

Speaker 3

我认为他们将迅速达到GPT-4.5级别的算力,这意味着按照这些数量级,你将很快获得百倍的提升。

Think they're going to go really fast to GPT-4.5 compute, which means you're going to get, using these orders of magnitude, you're going to get a 100x improvement really fast.

Speaker 3

所以我认为,那些一直持怀疑态度的人。

So I think there's all these people who have been skeptical.

Speaker 3

他们都等着彻底丢脸。

They're all in for abject humiliation.

Speaker 3

他们就是如此。

They just are.

Speaker 3

而且与GPT-2不同,只有特斯拉能访问基于行驶里程的视觉训练数据集。

And then unlike GPT-two, only Tesla has access to a visual trading dataset that is based on miles driven.

Speaker 3

我们可以争论这个数据集比第二大训练数据集微博大了100倍、1000倍还是10000倍。

We could argue whether it's a 100 x, a thousand x, 10,000 x bigger than the second biggest trading dataset, which is Weibo.

Speaker 3

所以人们会问:他们怎么赚钱呢?

So it's like people, oh, how are they gonna make money?

Speaker 3

但在自动驾驶领域,从我的角度看,他们欠的是YouTube、Meta的所有平台、开放互联网和X,而其他人却像用雅虎一样在尝试做这件事。

Well, it's like in this case, in the world of self driving, from my perspective, it's like they owed YouTube, they owed all of Meta's properties and the open Internet and X, and then other people are, like, trying to do it using Yahoo.

Speaker 3

是的。

Yeah.

Speaker 3

用雅虎。

Using Yahoo.

Speaker 3

祝你好运。

Like, good luck.

Speaker 3

谁会赢呢?

Like, who's gonna win?

Speaker 3

当然,这种情况可能会改变。

Now, obviously, that could change.

Speaker 3

保持谦逊很重要。

Important to have humility.

Speaker 3

可能会出现算法上的突破,降低对这个交易数据集的依赖。

There may be an algorithmic breakthrough that reduces the importance of that trading data set.

Speaker 3

毫无疑问,微博会试图用蛮力解决,不惜投入任何所需的资金来获取数据以竞争。

And for sure, Weibo is gonna try and brute force it and just throw whatever amount of dollars they need to get the data to compete.

Speaker 3

他们采用了不同的方法,使用激光雷达。

And they have a different approach using LiDAR.

Speaker 3

特斯拉没有。

Tesla doesn't.

Speaker 3

我们走着瞧。

We'll see.

Speaker 3

我觉得这并不是板上钉钉的事。

Like, I don't think it's a foregone conclusion.

Speaker 3

关于未来,没有什么是确定的。

Nothing about the future is certain.

Speaker 3

但只要我看看十二点五在AI四硬件上的表现,再想想它是在多么少的算力下训练出来的,而他们在奥斯汀正在用现有技术搭建的超级集群,我认为我们会直接跳过12.5到GPT-2。

But just if I look at how amazing twelve dot five is on AI four hardware and think about the tiny amount of compute that that was trained on and the mega cluster that they are standing up in Austin using known techniques, We're gonna skip, I think, 12 dot five to GPT-two.

Speaker 3

我们会迅速跳到GPT-4。

We're gonna skip really quickly to GPT-four.

Speaker 3

我知道Weibo肯定会加强它。

And then look, know, I'm sure Weibo will reinforce it.

Speaker 3

可能会有算法上的突破,让其他人也能参与进来。

There may be algorithmic breakthroughs such that there are other people.

Speaker 3

我们走着瞧。

We'll see.

Speaker 3

但另一件大事是使用大语言模型来实现全自动驾驶。

But then the other big thing is just using an LLM for FSD.

Speaker 3

X 上最好的关注者之一是范博士。

One of the best followers at X is Doctor.

Speaker 3

他是英伟达机器人部门的负责人。

Jim Fan, who's NVIDIA's head of robotics.

Speaker 3

他发布了很多帖子,谈到他和埃隆在 X 上的互动非常有趣。

And he's had a lot of posts about how there's a fascinating shape between him and Elon on X.

Speaker 3

人工智能在 X 上发生的程度真是惊人。

It is amazing the extent to which AI happens on X.

Speaker 3

谷歌的 Jax 团队和 Meta 的 PyTorch 团队爆发了一场激烈的争执,甚至闹到了 MAMAF。

The Jax team at Google and the PyTorch team on Meta got into this bitter fight, and it went to MAMAF.

Speaker 3

哪个框架更适合 MAMAF?

Which framework was better for MAMAF?

Speaker 3

最终,两个实验室的负责人不得不在 X 上公开出面和解。

Eventually, the heads of each lab had to step in publicly on X and make peace.

Speaker 3

但天啊,真的。

But like, wow.

Speaker 3

你知道吗,仅仅关注这场争论,你就能学到很多东西。

You know, you learn so much just following that fight.

Speaker 3

每个AI研究员都在X上活跃。

And like every AI researcher is active on x.

Speaker 3

AI在X上发生。

AI happens on x.

Speaker 3

这是一个绝佳的平台来使用它。

And it's such a great forum for for using it.

Speaker 3

但吉姆·法耶特与埃隆有过一次有趣的交流,吉姆·法耶特谈到LLM如何大幅改善FSD。

But Jim Fayette had this fascinating exchange with Elon where Jim Fayette talked about how LLMs could massively improve FSD.

Speaker 3

埃隆回复说:是的。

And Elon replied, yes.

Speaker 3

只有两种数据来源能够无限扩展:合成数据和真实世界视频。

The only two data sources that will scale infinitely are synthetic data and real world video.

Speaker 3

我觉得这很有趣。

And I thought that was interesting.

Speaker 3

然后这就引出了我认为的、我刚才描述的特斯拉自动驾驶未来观点可能出错的最大风险。

And then that goes to, I think, maybe the biggest risk in which this view that I just described of Tesla's autonomous future is wrong.

Speaker 3

那就是,合成视频数据是否能像合成数据那样被使用。

It's just if synthetic video data can be used in the same way that synthetic data can be.

Speaker 3

我们知道合成文本数据是有效的。

We know that synthetic written data works.

Speaker 3

但我们不知道合成视频数据是否有效。

We don't know if synthetic video data works.

Speaker 3

没人知道。

Nobody knows.

Speaker 3

显然,监管机构的要求非常高。

And obviously, there's a very high bar for regulator.

Speaker 3

我认为大概是,我忘了具体数字,全球每年有五万或十万人都死于车祸。

I think it's something like, I forget whatever it is, like fifty thousand or a hundred thousand people die in car crashes every year globally.

Speaker 3

甚至可能达到一百万。

It might even be a million.

Speaker 3

显然,我们可以用人工智能大幅降低这一数字,但人们对人工智能造成的交通事故的容忍度远低于对人类造成的交通事故。

Obviously, we could take that down dramatically using AI, but we're much less willing to tolerate traffic fatal accidents from AIs than humans.

Speaker 3

就是这样。

That is what it is.

Speaker 3

所以,这将会受到严格监管。

So, you know, it's gonna be heavily regulated.

Speaker 3

但范博士提出,LLibs之所以能真正帮助实现全自动驾驶,原因如下。

But doctor Jim Fan posited that the reason LLibs are gonna be able to really help with FSD is because of the following.

Speaker 3

这是我对于一些从事这些问题研究的人的看法,以我相对较低的智商来理解的方式。

This is the way my relative to some of the people working on these problems, my comparatively low IQ brain conceptualizes it.

Speaker 3

任何基于真实世界数据训练过的系统,都知道在那种真实情境下,一个优秀的人类司机该怎么做。

Anything that's been trained on real world data just knows what to do, what a really good human driver would do in that real world situation.

Speaker 3

如果遇到全新情况,它可能就不知道该如何应对。

If there's a novel situation, it may not know what to do.

Speaker 3

而从我的角度来看,大语言模型正是在这里能真正发挥作用的地方。

And that's where, from my perspective, the LLM can really help.

Speaker 3

因为GPT-4的一个涌现特性是——我们或许可以争论这究竟是真正的涌现特性,还是仅仅属于上下文学习——它具备所谓的世界模型。

Because one of the emergent properties of GPT-four, and we could debate whether or not it actually is an emergent property or just in context learning, It has what's called a world model.

Speaker 3

这意味着,我相信你了解这一点:如果你问GPT-3,‘如果你把一瓶香槟倒过来放,再在上面放一个涂满肥皂的篮球,会发生什么?’

And that means, I'm sure you know this, but if you ask GPT-three, hey, what happens if you stand a champagne bottle upside down and put like a basketball covered in soap on top of it?

Speaker 3

GPT-3完全不知道答案。

GPT-three, no idea.

Speaker 3

GPT-4通常能正确回答这类问题。

GPT-four will often get questions like that right.

Speaker 3

我其实应该去试试看它是否能正确回答这个问题。

I should actually see if he gets that exact question right.

Speaker 3

一个三岁的小孩会说:‘这会掉下来的。’

A three year old human will say, That's going to fall.

Speaker 3

香槟瓶会摔碎的。

The champagne bottle is going to shatter.

关于 Bayt 播客

Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。

继续浏览更多播客