The Twenty Minute VC (20VC): Venture Capital | Startup Funding | The Pitch - 20VC:英伟达 vs Groq:训练与推理的未来 | Meta、谷歌和微软的数据中心投资:谁将胜出 | 数据、算力、模型:AI的核心瓶颈及价值分布,专访Groq创始人乔纳森·罗斯 封面

20VC:英伟达 vs Groq:训练与推理的未来 | Meta、谷歌和微软的数据中心投资:谁将胜出 | 数据、算力、模型:AI的核心瓶颈及价值分布,专访Groq创始人乔纳森·罗斯

20VC: NVIDIA vs Groq: The Future of Training vs Inference | Meta, Google, and Microsoft's Data Center Investments: Who Wins | Data, Compute, Models: The Core Bottlenecks in AI & Where Value Will Distribute with Jonathan Ross, Founder @ Groq

本集简介

乔纳森·罗斯是Groq的创始人兼首席执行官,该公司创造了全球首个语言处理单元(LPU™)。在创立Groq之前,乔纳森以20%项目的形式启动了后来成为谷歌张量处理单元(TPU)的项目,设计并实现了第一代TPU芯片的核心组件。随后,他加入谷歌X的快速评估团队——即著名的“登月工厂”的初始阶段,在此期间他构思并孵化了谷歌母公司Alphabet的新业务单元。 在本期节目中,我们讨论: 04:20 乔纳森·罗斯访谈开始 04:59 扩展定律与AI模型训练 06:22 合成数据与模型效率 12:01 推理与训练成本:为何NVIDIA在推理中失利 17:06 AI推理的未来:效率与成本 18:15 芯片供应与扩展担忧 20:57 AI计算中的能效问题 25:40 为何数据中心的大部分资金将打水漂 31:05 Meta、谷歌和微软的数据中心投资 41:11 AI经济中的价值分配 42:10 初创企业成功的阶段 43:17 AI投资泡沫 45:00 风险投资中的凯恩斯选美博弈 48:40 NVIDIA在AI生态系统中的角色 53:39 中国的AI战略及其全球影响 57:51 欧洲在AI革命中的潜力 01:10:14 未来预测与AI对社会的影响

双语字幕

仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。

Speaker 0

我们并没有筹集到15亿。

We did not raise 1,500,000,000.0.

Speaker 0

那是收入。

That's revenue.

Speaker 0

这实际上是OpenAI收入的大约30%。

That's actually about 30% of the revenue of OpenAI.

Speaker 0

你的工作不是追随浪潮。

Your job is not to follow the wave.

Speaker 0

你的工作是为浪潮做好布局。

Your job is to get positioned for the wave.

Speaker 0

你几乎可以说,我们是给英伟达带来最好影响的公司之一,因为他们可以生产出原本计划制造的每一颗GPU,并以高利润率用于训练,成本分摊到部署中;而我们则接手低利润率、高体量的推理业务,这样他们就无需在两种利润率之间做取舍。

You can almost say we're one of the best things that ever happened to NVIDIA because they can make every single GPU that they were gonna make, and they can sell it for training, high margin, gets amortized across deployment, and we'll take the low margin, high volume inference business off their hands, and they won't have to sell either margin.

Speaker 0

我们的增长速度超过了指数级。

We are growing faster than exponential.

Speaker 0

当你增长得比指数还快时,任何利润数额都变得无关紧要。

And when you are growing faster than exponential, there is no amount of profit that you can make that matters.

Speaker 0

重要的是在市场中站稳脚跟并变得相关。

What matters is getting a toehold in the market and becoming relevant.

Speaker 1

这里是20VC的哈利·斯蒂宾斯。

This is 20 VC with me, Harry Stebbings.

Speaker 1

今天,我们要介绍一家刚刚实现了15亿收入的公司。

And today, we feature a company that has just booked 1,500,000,000.0 in revenue.

Speaker 1

他们非常简单地表示,NVIDIA应该掌控AI的训练市场,而他们会掌控推理市场。

They say very simply, NVIDIA should own the training market for AI, and they will own the inference market.

Speaker 1

很简单。

Simple.

Speaker 1

这是上周在巴黎录制的一场非凡对话,嘉宾是Grok的创始人兼首席执行官乔纳森·罗斯,Grok是全球首个语言处理单元LPU的创造者。

This is an exceptional discussion recorded in Paris last week with Jonathan Ross, founder and CEO of Grok, the creator of the world's first language processing unit, LPU.

Speaker 1

在创立Grok之前,乔纳森曾主导了谷歌的张量处理单元TPU项目,并在谷歌实现了第一代TPU芯片的核心架构。

And prior to Grok, Jonathan began Google's tensor processing unit, TPU, and implemented the core elements of the first generation TPU chip at Google.

Speaker 1

但在我们深入讨论之前,将一张餐巾纸上的想法转变为十亿美元的初创企业,需要无数小时的协作与团队合作。

But before we dive in today, turning your back of a nap kin idea into a billion dollar startup requires countless hours of collaboration and teamwork.

Speaker 1

组建一支在价值观和工作流程上都高度一致的团队可能非常困难,但这正是Coda的设计初衷。

It can be really difficult to build a team that's aligned on everything from values to workflow, but that's exactly what Coda was made to do.

Speaker 1

Coda是一个一体化的协作工作空间,最初只是一个餐巾纸上的草图。

Coda is an all in one collaborative workspace that started as a napkin sketch.

Speaker 1

如今,自Coda在五年前推出测试版以来,已帮助全球五万个团队实现步调一致。

Now just five years since launching in beta, Coda has helped 50,000 teams all over the world get on the same page.

Speaker 1

在20VC,我们用Coda为内容规划和节目准备建立了清晰的结构,效果显著。

Now at twenty v c, we've used Coda to bring structure to our content planning and episode prep, and it's made a huge difference.

Speaker 1

我们不再需要在不同工具之间来回切换,而是可以把嘉宾调研、日程安排和笔记全部集中在一个地方,节省了大量时间。

Instead of bouncing between different tools, we can keep everything from guest research to scheduling and notes all in one place, which saves us so much time.

Speaker 1

使用Coda,你既能获得文档的灵活性、表格的结构性,又能享受专为企业打造的应用程序的强大功能,再加上AI的智能加持,让它更加出色。

With Kodi, you get the flexibility of docs, the structure of spreadsheets, and the power of applications all built for enterprise, and it's got the intelligence of AI, which makes it even more awesome.

Speaker 1

如果你是初创团队,希望提升协同效率和敏捷性,Coda能帮助你以惊人的速度从规划转向执行。

If you're a start up team looking to increase alignment and agility, Coda can help you move from planning to execution in record time.

Speaker 1

想亲自体验一下,请立即访问coda.io/20vc,即可免费获得六个月的团队版服务。

To try it for yourself, go to coda.io/20vc today and get six free months of the team plan for start ups.

Speaker 1

前往 coda.io/20vc 免费注册,即可获得六个月的团队计划免费试用。

That's coda.io/20vc to get started for free and get six free months of the team plan.

Speaker 1

现在你的团队已经对齐并开始协作,让我们来处理那些混乱的报销单吧。

Now that your team is aligned and collaborating, let's tackle those messy expense reports.

Speaker 1

你知道的,那些在你钱包里像兔子一样不断增多的收据,还有没完没了的邮件在问‘你能批准这个吗?’

You know, those receipts that seem to multiply like rabbits in your wallet, the endless email chains asking, can you approve this?

Speaker 1

别提月底的恐慌了,当你意识到必须核对所有账目时,简直要崩溃。

Don't even get me started on the month end panic when you realize you have to reconcile it all.

Speaker 1

Pleo 提供智能公司卡,包括实体卡、虚拟卡和特定供应商卡。

Well, Pleo offers smart company cards, physical, virtual, and vendor specific.

Speaker 1

这样团队可以购买所需物品,同时财务部门保持控制,自动化报销流程,无缝处理发票,并在一个平台上轻松管理退款。

So teams can buy what they need while finance stays in control, automate your expense reports, process invoices seamlessly, and manage reimbursements effortlessly all in one platform.

Speaker 1

通过与 Xero、QuickBooks 和 NetSuite 等工具的集成,Pleo 能无缝融入你的工作流程,节省时间并全面掌握每个实体、付款和订阅情况。

With integrations to tools like Xero, QuickBooks, and NetSuite, Pleo fits right into your workflow, saving time and giving you full visibility over every entity, payment, and subscription.

Speaker 1

加入已使用 Pleo 简化财务流程的三万七千多家公司吧?

Join over 37,000 companies already using Pleo to streamline their finances?

Speaker 1

今天就试试 Pleo。

Try Pleo today.

Speaker 1

这就像魔法,但兔子更少。

It's like magic, but with fewer rabbits.

Speaker 1

了解更多,请访问 pleo.io/20vc。

Find out more at pleo.io/20vc.

Speaker 1

别忘了与客户建立信任。

Don't forget to secure trust with your customers.

Speaker 1

但信任不仅仅是赢得的。

Trust isn't just earned though.

Speaker 1

它是被要求的。

It's demanded.

Speaker 1

因此,包括 Atlassian、Cora 和 Factory 在内的超过 9,000 家公司都依赖 Vanta 来自动化他们的安全合规流程。

That's why over 9,000 companies, including Atlassian, Cora, and Factory rely on Vanta to automate their security compliance.

Speaker 1

Vanta 帮助企业轻松获得 SOC 2 和 ISO 27001 等认证,将数月繁琐的工作转化为快速而简洁的流程。

So Vanta helps businesses achieve certifications like SOC two and ISO 27,001, turning months of tedious work into this beautifully fast and straightforward process.

Speaker 1

他们的平台可自动化超过35种合规框架。

Their platform automates compliance across over 35 frameworks.

Speaker 1

它集中管理工作流程,并主动应对风险,同时通过自动化和人工智能为你节省时间。

It centralizes workflows, and it proactively manages risk, all while saving you time with automation and AI.

Speaker 1

无论你是刚刚起步还是正在扩展你的安全项目,Vanta都会帮你连接审计师和专家,快速做好审计准备,并与客户建立信任。

So whether you're just starting or scaling your security program, Vantas connects you with auditors and experts to get audit ready quickly and build trust with your customers.

Speaker 1

通过访问 vanta.com/20vc,即可享受首年1000美元优惠。

Get $1,000 off your first year by visiting vanta.com/20vc.

Speaker 1

那就是 vanta.com/20vc。

That's vanta.com/20vc.

Speaker 0

您已到达目的地。

You have now arrived at your destination.

Speaker 1

乔纳森,非常感谢你同意在巴黎接受这次访谈。

Jonathan, thank you so much for agreeing to do this in Paris.

Speaker 1

顺便说一句,你看起来棒极了。

You look fantastic, by the way.

Speaker 1

我感觉自己穿得太随便了,但你看起来很棒。

I feel so underdressed, but you look great.

Speaker 1

谢谢。

Thank you.

Speaker 0

如果你愿意,我可以把领带摘下来,但我再也没法系上了。

I I could take the tie off if you want, but I'll never be able to tie it again.

Speaker 0

我不知道怎么系领带,不行。

I don't know how to tie No.

Speaker 0

真的,我的首席幕僚得帮我系领带。

Literally, my chief of staff has to tie it for me.

Speaker 0

这简直是个难题,因为他是在给自己系领带。

It's and it's like a struggle because, like, he's putting it on himself.

Speaker 0

他在系领带。

He's tying it.

Speaker 0

我其实是最近才买这套西装的。

I literally only bought this suit recently.

Speaker 0

我的意思是,你看起来棒极了。

Well, mean, you look fantastic.

Speaker 0

我不,我觉得我

I don't I think I

Speaker 1

我没有西装,所以你比我强。

have a suit, so you're one up on me.

Speaker 1

我想把节目分成两部分。

I wanna split the show into two parts there.

Speaker 1

我想先聊聊我们当前所处的环境,然后再具体谈谈你所在的Grok。

I wanna talk about the landscape where we're at, and then I wanna dive specifically into Grok where you're at.

Speaker 1

你宣布了一项重大的新合作,但我认为大家对刚才我们讨论的内容有点误解。

You've announced a massive new deal that I think everyone's slightly misunderstanding what we're just talking about.

Speaker 1

我想先从我们目前所处的位置说起。

I just wanna start on where we're at.

Speaker 1

关于规模定律,大家都说我们已经触及了规模定律的极限,但像DeepSeek等公司却似乎在持续取得指数级的创新。

In terms of, like, scaling laws, everyone says we are at the limits of scaling laws, and then there seems to be exponential innovation happening with the likes of DeepSeek and others.

Speaker 0

在扩展定律的极限方面,我们现在处于什么阶段?

Where are we at in terms of the limits of scaling laws?

Speaker 0

扩展定律是OpenAI发布的一篇论文,它基本上指出,模型的参数越多,就能越好地吸收信息。

Scaling laws is a paper that was published by OpenAI, and what it does is it effectively says the more parameters your model has, basically the better it can absorb information.

Speaker 0

你会看到他们绘制的这些曲线,非常惊人。

You'll see these curves that they draw, and they're amazing.

Speaker 0

如果可以的话,你最好展示一下。

You should show it if you can.

Speaker 0

但实际上,这些曲线呈现出一种渐近式的衰减:你不断变得更好,但当你增加线性数量的训练数据时,提升却是对数级的。

But effectively, you have these sort of asymptotic drop offs where you keep getting better and better, but you get a logarithmic improvement when you put a linear number of tokens in.

Speaker 0

这就是为什么你会看到有人用15万亿个标记进行训练的原因。

This is why you see people doing 15,000,000,000,000 tokens of training and whatnot.

Speaker 0

但这些做法被误解了,因为人们假设所有数据的质量都是一样的。

But they're misunderstood because the assumption is that all of the data is the same quality.

Speaker 0

所以最终,你可能会用你孩子的数据来训练模型。

So eventually you're gonna be training your kid.

Speaker 0

你会说,跟我一起想象一下,一加一等于多少?

And you're gonna say, and play along with me here, what's one plus one?

Speaker 0

二。

Two.

Speaker 0

二乘以三等于多少?

What's two times three?

Speaker 0

六。

Six.

Speaker 0

双曲正切函数平方的二阶导数是多少?

What's the second derivative of the square of the hyperbolic tangent?

Speaker 0

但这就是我们训练这些模型的方式。

But that's how we train these models.

Speaker 0

我们先给他们一些非常简单的问题,然后再给他们一些特别难的问题。

We give them really simple problems to solve and then we give them these really hard ones.

Speaker 0

我们并没有真正循序渐进地训练它们。

We don't really train them up.

Speaker 0

我们并不是聪明地做这件事。

We don't do it smart.

Speaker 0

所以有些人会用互联网上的低质数据进行训练,然后把高质量的数据留到后期,以提升模型表现。

So what some people do is they will train on the dregs of the internet and then they'll save some high quality data for the end to make them better.

Speaker 0

但你可以这么做,这也是大家感到困惑的地方:就像AlphaGo Zero那样,它自己生成数据并进行训练。

But what you can do, and this is where I think everyone's getting confused, is it's sort of like with AlphaGoZero, where it generated its own data and trained.

Speaker 0

你能否让一个大语言模型生成合成数据?

Could You have an LLM generate synthetic data.

Speaker 0

当它生成合成数据时,这些数据质量更高,然后你再用这些合成数据进行训练。

And when it generates the synthetic data, the data's better, you then train on that synthetic data.

Speaker 0

所以你做的是,用合成数据来训练,为什么合成数据更好?

So what you do is you train Why is synthetic

Speaker 1

合成数据为什么比真实数据更好?

data better than real data?

Speaker 0

因为模型更聪明。

Because the model is smarter.

Speaker 0

Reddit很不错,但未必比得上和某个领域的博士直接交流的质量。

Reddit is great, but not necessarily as high quality as talking to someone with a PhD in a topic.

Speaker 0

所以,就像更专业的专家更有知识和能力一样,如果你有一个更好的模型,它生成的数据也会更好。

And so just like with more expert people who are more knowledgeable and more capable, if you have a better model, it generates better data.

Speaker 0

因此,你先训练模型,模型变强后,生成更优质的数据,并产生一系列数据,然后剔除所有错误的部分。

So you train the model, it gets better, you produce better data, and you produce a range of data here, and you get rid of all the parts that are wrong.

Speaker 0

这样一来,剩下的就是最精华的部分,比原始模型还要好一点,因为你进行了筛选。

So now it's the best part, so it's a little better than the model is because you're pruning it.

Speaker 0

你可以离线做这件事,对吧?

You get to do this offline, right?

Speaker 0

然后你训练模型,模型表现提升,再重复这个过程,保留更好的数据,再次训练,就这样不断进步。

And then you train the model, and the model comes up here, and then you do this again, and then you keep the better data, you train it again, you just keep moving up.

Speaker 0

当你这么做时,实际的扩展规律并不会呈现这些渐近线,而是会——

When you do that, the actual scaling laws don't look like these asymptotics, they actually But

Speaker 1

但必须要有

there has to be

Speaker 0

效率的上限。

a ceiling on efficiency.

Speaker 0

不是吗?

No?

Speaker 0

真的有吗?

Does there?

Speaker 0

所以存在一个数学极限。

So there's a mathematical limit.

Speaker 0

如果你学过计算机科学,你可能听说过一种叫做大O复杂度的东西。

If you study computer science, you've probably heard of something called Big O complexity.

Speaker 0

大O复杂度是指,当我解决一个问题时,观察我的解决方式,使用不同的算法可能需要更多步骤。

Big O complexity is, if I am solving a problem and I look at how I solve it, I might need to take more steps if I solve it with one algorithm versus another.

Speaker 0

比如快速排序和冒泡排序。

So for example, quick sort versus bubble sort.

Speaker 0

快速排序需要 n log n 步。

Quick sort, I need n log n steps.

Speaker 0

冒泡排序需要 n 的平方步。

Bubble sort, I need n squared.

Speaker 0

有什么区别?

What's the difference?

Speaker 0

如果我要排序 1000 个数字,n log n 就是 10,000 步。

If I'm sorting 1,000 numbers, n log n, that's 10,000 steps.

Speaker 0

但用 n 的平方,那就是一百万步,因为要么是 10 乘以一千,要么是一千乘以一千。

But with n squared, that's a million steps because it's either 10 times a thousand or a thousand times a thousand.

Speaker 0

这些大语言模型在处理大数乘法时遇到困难的原因之一,是因为乘法不是线性的。

One of the reasons that these LLMs struggle to multiply large numbers is because multiplying is not linear.

Speaker 0

这些大语言模型可以不加思考地完成任何线性操作。

These LLMs can do anything linear without needing to think.

Speaker 0

但就像在纸上一样,你需要写下所有中间步骤,这些大语言模型也需要中间空间来完成这些步骤,才能进行计算。

But just like on a piece of paper, how you need to write out all those intermediate steps, these LLMs need that intermediate space those steps in order to compute these things.

Speaker 0

这是一种数学上的要求。

It's a mathematical requirement.

Speaker 0

只要你训练模型足够充分,它就能识别任意大的数字并直接完成乘法运算。

There's nothing you cannot train a model enough so that it'll see any arbitrarily large number and just be able to multiply it.

Speaker 0

但你可以选择更大规模的数字组合让它去记忆,这样它就能用更少的步骤完成计算。

But you can choose bigger and bigger groupings of numbers for it to memorize, in which case it can do it in fewer steps.

Speaker 0

随着你不断用更多数据训练模型,它会接触到越来越多的示例。

And effectively, as you are training the model on more and more data, it's seeing more and more examples.

Speaker 0

因此,它现在对更多具体情境已经有了现成的答案,不再需要进行太多推理。

So now it just has the answer for more specific situations, so it doesn't need to do as much reasoning.

Speaker 0

但它仍然需要对其中一些问题进行推理。

But it still needs to do reasoning for some of these problems.

Speaker 1

这对下一步意味着什么?

What does that mean for the next step?

Speaker 1

如果我们没有效率上限,这个问题意味着什么?

If we have no efficiency ceiling, what does that issue mean?

Speaker 0

两者都需要。

You need both.

Speaker 0

模型的训练使其更加直观。

Training of the model makes it more intuitive.

Speaker 0

这意味着它能够几乎直接想出答案,像一种更流畅的思维流。

It means that it can sort of just come up with the answer like that, more stream of consciousness.

Speaker 0

推理部分是不同的。

The reasoning part is different.

Speaker 0

推理是位于顶层的算法,即大O复杂度部分。

The reasoning is the algorithm on top, the big O complexity portion.

Speaker 0

这就像是系统一和系统二的思维,或者说是快思考与慢思考,就像丹尼尔·卡尼曼的书里说的那样。

So it's system one, system two thinking, or thinking fast, thinking slow, like Daniel Kahneman's book.

Speaker 0

当你将两者结合时,当你让它变得更直观,你就会以这种方式变得更好,对吧?

When you pair them together, when you make it more intuitive, you get better this way, right?

Speaker 0

但当你开始加入系统二的部分时,你就会得到这种效果。

But when you start adding in the system two portion, you start to get this.

Speaker 0

你可能会发现它的输出量非常小,但当你这样做的时候,就会得到这种多项式增长——这是术语,但你可以理解为,当结合更优的训练以及所谓的测试时计算、运行时计算时,模型的性能呈几何级提升。

You you hear that the volume is very little, but when you do this and so you get this polylinear is the term, but you could think of it as geometrically increasing improvement in the model when you combine it with that improved training, but also the improved, what they call, test time computer, run time compute.

Speaker 1

为了让我理解清楚,当我们讨论瓶颈时,如果我们使用合成数据来驱动训练,

Just so I understand, so when we think about bottlenecks, if we have synthetic data that powers the training,

Speaker 0

它会变得更加直观。

it gets more intuitive.

Speaker 0

它能更快地得出答案。

It gets to the answer more quickly.

Speaker 0

就像国际象棋大师一眼就能看出正确的走法一样。

Sort of like a grandmaster in chess just seeing the right moves.

Speaker 1

当然,合成数据在供应端是不受限制的。

Sure, but synthetic data is not constrained in terms of its supply side.

Speaker 1

如果我们考虑其他瓶颈,还有硬件、能效和算法极限。

If we think about the other bottlenecks, there is hardware, there is energy efficiency, there's algorithmic limits.

Speaker 1

但什么是

What is the But

Speaker 0

如果我的工作是提高乘法能力,而我告诉你我希望你能更直观、更少步骤地完成,那么为了让你能计算三位数而不是两位数的乘法,你需要十倍的数据和十倍的示例。

if I'm telling if if your job is to get better at multiplying numbers, and I tell you that I want you to be able to do it with fewer steps, more intuitively, For you to be able to multiply three digit numbers versus two digit, you need 10x the data, and you need 10x the examples.

Speaker 0

因此,当你在直觉方面变得更好时,你需要更多的例子来进行训练。

And so as you get better on the intuitive part, you need more examples to train on.

Speaker 0

那么什么是

And so what is

Speaker 1

瓶颈所在呢?

the bottleneck then?

Speaker 1

是硬件质量吗?

Is it the hardware quality?

Speaker 1

是算力吗?

Is it compute?

Speaker 1

是算法吗?

Is it algorithms?

Speaker 1

是算力,是数据,是

It is the compute, it is the data, it is

Speaker 0

算法。

the algorithms.

Speaker 0

这三者都是。

It's all three of them.

Speaker 0

但人们误解了瓶颈这个概念。

But people misunderstand the concept of a bottleneck.

Speaker 0

算力已经不再是一个严格的瓶颈,而更像是一种软性限制之类的。

Compute has been more of a less of a bottleneck and more of a, you know, soft neck or something.

Speaker 0

对吧?

Right?

Speaker 0

当你提供更多的算力时,你甚至可以克服数据不足和算法改进缓慢的问题。

Where when you provide even more compute, you can sort of overpower the lack of data, the lack of improvement in algorithms.

Speaker 0

所以这不是一个硬性瓶颈,而是一个软性瓶颈。

So it's not a hard bottleneck, it's a soft bottleneck.

Speaker 0

但理想情况下,你应该同时提升这三个方面。

But ideally, you would improve all three.

Speaker 0

你会获得更好的数据。

You would be getting better data.

Speaker 0

你会获得更好的算法。

You would be getting better algorithms.

Speaker 0

算法的改进一定会到来。

And the algorithm improvements are gonna be there.

Speaker 0

数据的改进也一定会到来。

The the data improvements are gonna be there.

Speaker 0

但计算资源始终是最容易调动的杠杆,因为它非常灵活。

But compute has always been the easiest lever because it's so fungible.

Speaker 0

如果我给你更多计算资源,效果就会更好。

If I just give you more compute, works better.

Speaker 0

DeepSeek

Has DeepSeek

Speaker 1

没有向你展示其实我们不需要那么多计算资源,用更少的资源也能做得更好吗?

not shown you that actually we don't need the compute and you can do more with less?

Speaker 0

并不完全如此。

Not exactly.

Speaker 0

那上面有一个算法上的改进。

There was an algorithmic improvement on that.

Speaker 0

而这个算法改进看似很荒谬——他们只是把答案写在一个框里,然后就知道该找什么了,而不是必须让人去检查之类的,对吧?

And the algorithmic improvement, seemingly silly thing, where they just wrote the answer in a box and then they knew what to look for rather than having to have a human being check it or something like that, right?

Speaker 0

这非常简单。

It was very simple.

Speaker 0

但这是一个算法上的改进,它使得生成随后用于训练的数据变得更加容易。

But that was an algorithmic improvement and it made it easier to generate the data that was then trained on.

Speaker 1

我认为人们对计算、数据,尤其是你提到的合成数据和算法,存在一些误解。

I think there's misconceptions around compute data, especially kind of synthetic data, as you said there, algorithms.

Speaker 1

当你想到人们在人工智能,特别是推理方面最大的误解时,你认为是什么?

When you think about the biggest misconceptions that people have around AI and specifically kind of inference, what do you think they are?

Speaker 0

当我们刚开始时,第一个误解——现在人们不再这么认为了——是训练比推理更昂贵。

When we started, the first misconception, which people don't hold anymore, is that training was more expensive than inference.

Speaker 0

在谷歌,每次我们训练一个新模型时,用于推理的计算量最终都会是训练的10到20倍。

At Google, any time we would train a new model we would end up using 10 to 20 times as much compute on the inference as the training.

Speaker 0

推理一直是我们需要的关键基础设施。

Inference was always the the critical infrastructure piece that we needed.

Speaker 0

但一旦跨过那一步,现在每个人都明白推理很重要了。

But then after getting past that, now everyone understands inference is important.

Speaker 0

我认为其中一个关键点是

I think one of the Do

Speaker 1

你觉得他们真的明白了吗?

you think they fully do?

Speaker 1

因为当你看DeepSeek之后英伟达的股价,它下跌了15%。

Because when you look at Nvidia's stock price post DeepSeek, it'd be down 15%.

Speaker 1

如果你真正理解了推理的价值,它就不该下跌15%。

If you understood the value of inference, it shouldn't be down 15%.

Speaker 0

嗯,还有杰文斯悖论之类的那些问题。

Well, and Jevan's paradox and all that, yeah.

Speaker 0

我不认同英伟达的股价因为那个原因就该下跌。

I I I don't agree that Nvidia stock should've gone down for that.

Speaker 0

我认为这是大多数人的一种误解。

I think that was a misunderstanding on most people's part.

Speaker 0

但这也表明,我认为还揭示了更多东西。

But it also shows I think that shows more.

Speaker 0

每个人都说英伟达的股价不可能再涨了。

Everyone keeps saying NVIDIA stock can't possibly go higher.

Speaker 0

他们一直在找借口,说:‘好了,到此为止了。’

And they were looking for an excuse for, oh, now that's it.

Speaker 0

所以我们错了,现在该卖了。

That's why we were wrong and we need to sell now.

Speaker 0

但这与实际情况毫无关系,这仅仅是市场中一种流行度竞赛,与价值衡量机制无关。

But that has nothing to do with the that's just a sort of popularity contest side of the market that had nothing to with the weighing machine of

Speaker 1

市场。

the market.

Speaker 1

那么,创始人应该继续构建并坚持下去吗?他们应该假设规模定律会持续下去吗?

So should founders build and stay, should they build with the assumption that scaling laws will continue?

Speaker 1

他们应该基于当今的技术来构建吗?

Should they build with what we have today?

Speaker 1

你如何建议他们做出选择?

How do you advise them on that?

Speaker 0

我会建议你基于事物不断改善的方式来构建。

I would advise you to build based on things getting better.

Speaker 0

我还会更关注那些重大的质变。

I would also focus a little more on the sort of big quantum steps.

Speaker 0

我喜欢的一个类比是,看看信息时代,我们经历了印刷术、电话、电报、互联网和智能手机。

The analogy that I like is if you look at the information age, we went through the printing press, we had the telephone, we had the telegram, we had the internet, and we had smartphones.

Speaker 0

对吧?

Right?

Speaker 0

如果你在互联网刚出现的时候就创建了Uber,它根本行不通,因为你叫了车,去了目的地,但你怎么回家?

And if you had built Uber back when we had internet, it wouldn't have worked because you'd book a ride, you'd go somewhere, how do you get home?

Speaker 0

我们现在也处在类似的阶段。

And we're in the same sort of space now.

Speaker 0

所以这些模型会幻觉。

So the models hallucinate.

Speaker 0

因此,建立一家医疗诊断公司会很困难。

So it would be hard to build a medical diagnosis company.

Speaker 0

建立一家法律公司也会很困难。

It would be hard to build a legal company.

Speaker 0

然而,如果你正在做这件事,而算法改进降低了幻觉率,那你就会处于绝佳的位置。

However, if you are doing that and the algorithmic enhancements happen that get the hallucination rate down, you are perfectly positioned.

Speaker 0

就像摇滚乐一样。

Just like rock.

Speaker 0

我们在找到产品市场契合点之前,已经熬了七年。

We were around for seven years before we had product market fit.

Speaker 0

我们的赌注是大规模推理。

Our bet was scaled inference.

Speaker 0

我们相信推理将成为瓶颈,我们需要运行非常庞大复杂的模型。

That inference was gonna be the bottleneck that we were gonna need to run really big heavy models.

Speaker 0

每个人都以为你会用一张PCIe卡来运行推理,因为训练才是复杂的一部分。

Everyone was assuming you'd have a single PCIe card running inference because training was the complicated part.

Speaker 0

对吧?

Right?

Speaker 0

但事实上,我们提前做出了正确的判断,因此完全占据了有利位置。

But the reality was we made the right bet ahead of time, and then we were perfectly positioned.

Speaker 0

你的任务不是追随浪潮。

Your job is not to follow the wave.

Speaker 0

你的任务是为浪潮做好布局。

Your job is to get positioned for the wave.

Speaker 0

而这是最难做到的,因为每个人都在试图劝你回到岸边。

And that's the hardest thing to do because everyone is trying to talk you into coming onshore again.

Speaker 0

几乎所有人都告诉我们,别做LLM。

Almost everyone was telling us, don't do LLMs.

Speaker 0

它们会对你们造成灾难。

They're gonna be terrible for you.

Speaker 0

我们觉得,这正是我们当初打造这一切的目的。

We're like, this is literally what we built for.

Speaker 0

你有没有怀疑过自己?

Did you ever doubt yourself?

Speaker 0

七年是相当漫长的等待时间。

Seven years is an incredibly long wait time.

Speaker 0

确实有过怀疑,但从未停歇过。

There was doubt, but there was never a pause.

Speaker 0

原因在于,早在启动TPU之前,我就担心人工智能会成为让某些人获得过度控制力和影响力的工具。

And the reason was, even back before starting the TPU, I was concerned that AI was going to be a technology that would allow some people to have outsized control, outsized influence.

Speaker 0

如果你任由这种情况发生在可能并非最合适的手中,那么你有多富有根本无关紧要。

And if you allow that to just happen in potentially not the best hands, it doesn't really matter how rich you are.

Speaker 0

什么都不重要了。

It doesn't matter nothing matters.

Speaker 0

这是最重要的技术。

It's the most important technology.

Speaker 0

所以无论有多艰难,都不重要。

So it didn't matter how hard it got.

Speaker 0

我们别无选择,只能成功。

There was no choice but to be successful.

Speaker 0

我们的目标是在人工智能时代维护人类的自主性。

And our goal is to preserve human agency in the age of AI.

Speaker 0

如果我们做不到这一点,我们就失败了。

If we don't do that, we have failed.

Speaker 0

有没有怀疑其实都不重要。

It wouldn't matter whether there was doubt or not.

Speaker 0

而且,是的,怀疑非常多。

And, yes, there was plenty of doubt.

Speaker 0

我们曾经一度濒临资金耗尽。

There was a point where we were so close to running out of money.

Speaker 0

我们做了一件我们称之为‘Grok债券’的事情。

We did this thing that we called grok bonds.

Speaker 0

所以,你知道的,二战时期的战争债券?

So, you know, war bonds from World War two?

Speaker 1

当然。

Of course.

Speaker 1

但对那些不了解的人,战争债券是什么?

But for anyone that doesn't, what is a war bond?

Speaker 0

二战时期的战争债券是通过发行债券来筹资的。

So war bond World War two was funded with bonds.

Speaker 0

美国政府当时有这些海报。

The US government, they had these posters.

Speaker 0

就是支持你的军队之类的。

It was like fund your troops and and whatever.

Speaker 0

你会购买这些债券,然后它们会给你回报。

And you'd buy them and and they would pay you a return.

Speaker 0

这些资金用于支持战争努力。

And that funded the war effort.

Speaker 0

我们曾经一度几乎要没钱了。

We were very close to running out of money at one point.

Speaker 0

我们没有假装自己很强,而是对员工坦诚自己的脆弱。

Rather than trying to pretend to be strong, we were vulnerable with our employees.

Speaker 0

我们告诉他们:我们快没钱了。

We said, we're gonna run out of money.

Speaker 0

我们需要你们用薪资换取股权。

We need you to trade salary for equity.

Speaker 0

我们真的拿来了战争债券的图片,把上面的内容改成了‘Grok债券’。

We literally took pictures of the war bonds, and we put grok bonds on it instead.

Speaker 0

我们在全员大会上说了这件事。

And we had an all hands where we said this.

Speaker 0

我们担心所有人都会离开。

And we were worried everyone was going to leave.

Speaker 0

但结果没有人离开,大约80%的员工参与了,其中一半的人选择将薪资降至法律规定的最低标准。

Instead of leaving, about 80% of the employees participated, 50%, I think, went to the statutory minimum salary by law.

Speaker 0

当我们终于筹集到3亿美元融资的第一笔资金时,银行里剩下的钱甚至比我们通过Grok债券节省的还要少。

When we finally raised the first bit of our $300,000,000 round, we had so little money in the bank left that it was less money than we saved doing grok bonds.

Speaker 0

所以如果我们没有这么做,我们真的就会把钱花光。

So had we not done that, we would have literally run out of money.

Speaker 0

那段时间真的非常艰难。

So there were some really hard times.

Speaker 0

我知道每个创始人都会经历这些,但从外面看,人们很难理解。

And I know every founder has these, and from the outside it's so hard to understand.

Speaker 0

就像在看一部电视剧。

It's like watching a TV show.

Speaker 0

你不在其中,但当你身处其中时,一切都会变得强烈十倍到一百倍。

You're not in it, but when you are there, everything is 10 to a 100 times more intense.

Speaker 0

因为人们离开了他们的工作。

Because people left their jobs.

Speaker 0

他们放弃了他们的职业生涯。

They left their careers.

Speaker 0

他们的家人正把希望寄托在这上面。

Their families are banking on this.

Speaker 0

你必须做出决定。

You have to make decisions.

Speaker 0

如果我们出去要求每个人都做Grok债券,然后每个人都辞职了,会发生什么?

What would have happened if we went out there and asked everyone to do grok bonds and everyone quit?

Speaker 0

那么股东们会说,你有这么多人依赖着你。

Then the shareholders would have been like, you have all of these people depending on you.

Speaker 0

但如果你展现出这种脆弱性,人们往往会选择支持你。

But if you lean towards that vulnerability, people are often gonna go with you on it.

Speaker 0

在一个推理如此关键、比训练重要二十倍的世界里,会是什么样子?

So what is a world where inference is so crucial and 20 times more important than training?

Speaker 0

那样的世界会是什么模样?

What does that world look like?

Speaker 0

我认为最简单的理解方式是把LPU或GPU比作一名员工。

I think the simplest way to understand it is equate an LPU or a GPU to an employee.

Speaker 0

如果你有足够的LPUs或GPU,就可以像使用员工一样开展工作。

If you have enough of them, the LPUs or GPUs, you can do work just like with an employee.

Speaker 0

这有点不同,因为它们不会辞职去另找一份工作。

It's a little different in the sense that they can't quit and take another job.

Speaker 0

你不需要重新训练。

You don't have to retrain.

Speaker 0

一旦模型达到某种能力水平,它就始终至少保持那个能力。

Once you get a model to a certain capability, it'll always be at least that capability.

Speaker 0

对吧?

Right?

Speaker 0

因此,你能从中获得一致性。

So you get the consistency out of it.

Speaker 0

但现在想象你是一家初创公司,与其去招聘100个人,不如只招聘10个人,再购买相当于90名员工算力的计算资源。

But now imagine that you're a startup, and rather than having to go out and hire a 100 people, you hire 10, and you buy the amount of compute equivalent to 90 employees' worth.

Speaker 0

这是一种截然不同的世界观,因为现在可以使用资本支出(CapEx)或某些运营支出(OpEx)来替代单纯的人力。

That's a very different way of thinking about the world because now CapEx or in in some cases different types of OpEx can be used instead of just employees.

Speaker 0

在推理方面,为了让你了解我们的扩展规模,我们在2024年初时生产中大约有六千四百颗芯片。

And in terms of inference, just to give you a sense of our scaling, we started 2024 with about six forty chips in production.

Speaker 0

到年底时已超过四万颗。

We ended with over 40,000.

Speaker 0

今年我们目标是达到两百万颗以上。

This year we want to be at over 2,000,000.

Speaker 0

明年这个数字会大得多、大得多、大得多。

And next year the number is much, much, much larger.

Speaker 1

我们是否面临芯片供应的瓶颈?

We seeing constraints on chip supply?

Speaker 1

这真是一个难以置信的扩展故事。

Mean that is an unbelievable scaling story.

Speaker 0

是的,为了实现明年我们的目标数字——虽然我不能公开具体数据——我们需要用尽我们所用晶圆厂的几乎全部产能。

Yeah, so for us to hit our numbers next year, which I'm not sharing publicly, we're gonna need almost all of the capacity of the fab that we're using.

Speaker 0

目前最大的问题在于七纳米工艺。

The the biggest issue so seven powers.

Speaker 0

我们很喜欢七种力量,对吧?

We love seven powers, right?

Speaker 0

汉密尔顿·赫尔默斯?

Hamilton Helmers?

Speaker 0

是的。

Yeah.

Speaker 0

好的。

Okay.

Speaker 0

你通常不会认为科技公司拥有垄断性资源,但英伟达确实拥有这样的资源。

You don't normally think of tech companies as having a cornered resource, but Nvidia has a cornered resource.

Speaker 0

它们是一个买方垄断,也就是垄断的反面,是HBM的唯一买家。

They're a monopsony, the opposite of a monopoly, a single buyer for HBM.

Speaker 0

还有互连基板,也就是Co-OS。

And the interposer, the co OS.

Speaker 0

那么,HBM是什么?

So what is HBM?

Speaker 0

HBM是高带宽内存。

So HBM is high bandwidth memory.

Speaker 0

GPU是

And GPUs are And

Speaker 1

谁生产HBM?

who produces HBM?

Speaker 1

抱歉问了这么简单的问题。

I'm sorry for the dumb questions.

Speaker 0

世界上只有三家公司做这个,分别是SK海力士、三星和美光。

There's three companies in the world that do this, SK Hynix, Samsung, and Micron.

Speaker 0

这是一种专用内存。

It's a specialty memory.

Speaker 0

它只用于高端服务器,因此产量有限。

It's only used in high end servers, so there's a limited quantity that's built.

Speaker 0

扩大产能非常昂贵。

It's very expensive to ramp up.

Speaker 0

这是一种在技术上非常具有挑战性的内存类型,比其他类型更难制造。

It's a very technically challenging type of memory to build, more so than others.

Speaker 0

因此供应量非常有限。

And so there's a very limited supply.

Speaker 0

而GPU的计算速度如此之快,如果你使用普通内存,就像用鸡尾酒吸管喝水一样。

And GPUs are so fast computationally that if you were using regular memory it'd be like drinking out of a martini straw.

Speaker 0

那会花上无穷无尽的时间。

It would just take forever.

Speaker 0

这就是为什么人们更倾向于在GPU而非CPU上进行推理,尤其是训练,因为内存带宽太有限了。

This is why you see people preferring to do even inference, but especially training on GPUs rather than CPUs because the memory bandwidth is too limited.

Speaker 0

而CPU很少使用HBM。

And CPUs rarely use HBM.

Speaker 0

它们主要使用普通内存。

They're mostly regular memory.

Speaker 0

当我们开始Grok项目时,我们注意到一个现象:人人都知道摩尔定律。

The observation that we had when we started Grok, everyone knows Moore's Law.

Speaker 0

每十八到二十四个月,就像时钟一样精确,晶体管数量翻倍意味着算力翻倍。

Every eighteen to twenty four months, like clockwork, double the transistors means double the compute.

Speaker 0

但我们发现,人工智能的进步速度更快。

But we noticed that AI was getting better faster.

Speaker 0

这显然不是算法的原因,因为算法的提升是断断续续的。

And it it clearly wasn't the algorithms because algorithms have sort of discontinuous jump.

Speaker 0

也不像是数据的原因,因为数据并没有多出太多。

It also, didn't seem to be the data because there wasn't that much more data.

Speaker 0

而晶体管数量每十八到二十四个月才翻一番。

And the transistors were only doubling every eighteen to twenty four months.

Speaker 0

那么,所有这些能力究竟从何而来?

So where was all of this capability coming from?

Speaker 0

结果发现,芯片数量也是每十八到二十四个月翻一番。

Turns out the number of chips was also doubling every eighteen to twenty four months.

Speaker 0

所以,不是两倍,而是四倍。

So rather than two x, it was four x.

Speaker 0

我们提出的问题是,如果你实际上拥有无限数量的芯片,是否会在架构上采取不同的做法?

The question we asked was, if you're effectively gonna have an unlimited number of chips, do you do something architecturally different?

Speaker 0

答案是肯定的。

And the answer is absolutely.

Speaker 0

因此,我们不再使用外部内存,而是直接使用大量芯片,将模型的所有参数始终保留在芯片中运行。

So rather than using external memory, we just use a large number of chips and keep all of the parameters of the model in the chips live.

Speaker 0

我们建立了一个流水线,让计算像流水线一样依次通过。

And we just have this pipeline where the computation flows through it, sort of like an assembly line.

Speaker 0

想象一下,如果你要建造一座工厂,但工厂的规模只有所需装配线的百分之一。

So imagine if you were trying to build a factory and the factory was only one one hundredth of the size needed for the assembly line.

Speaker 0

于是你得让一批汽车通过这百分之一的区域,拆解后,再 setup 下一个百分之一的装配线。

So you'd run a bunch of cars through one one hundredth, tear it down, set up the next one one hundredth assembly line.

Speaker 0

你一遍又一遍地重复这个过程。

You just do this over and over again.

Speaker 0

这就是GPU的工作方式。

That's the way GPU works.

Speaker 0

LPUs,非常不同。

LPUs, very different.

Speaker 0

我们实际上是让计算流经大量芯片。

We actually just have the the computation flow through a whole bunch of chips.

Speaker 0

所以我们不是使用八块芯片,而是为一个模型使用600块或3000块芯片。

So rather than using eight chips, we'll use 600 or 3,000 for a model.

Speaker 0

这如何影响能效?

How does that change energy efficiency?

Speaker 0

能效提升了大约三倍,原因是

It improves at about three x, and the reason is

Speaker 1

它是如何提升的?

How does it improve it

Speaker 0

当你使用更多芯片时?

when you use more?

Speaker 0

每个令牌。

Per token.

展开剩余字幕(还有 480 条)
Speaker 0

所以占用空间更大。

So the footprint is higher.

Speaker 0

把它想象成工厂和后院车库之间的区别。

Think of it as the difference between a factory or a backyard sort of garage.

Speaker 0

后院的车库效率不会那么高。

The backyard garage is not gonna be as efficient.

Speaker 0

然而,它的能耗更低。

However, it has a lower energy footprint.

Speaker 0

或者另一个例子:如果你想把一吨煤从城市这一端运到另一端,是用摩托车运,还是用货运列车运,哪种更高效?

Or another example would be if you're trying to transport a ton of coal from one side of the city to the other, and you did it on mopeds, or you did it with freight trains, which one would be more efficient?

Speaker 0

摩托车每次运输耗能更少,但需要更多次运输,因此总体耗能更多。

The moped would use less energy per trip, but it would need more trips and therefore would use more energy overall.

Speaker 0

事实上,这是大多数人误解的地方。

In fact, this is one of the things most people misunderstand.

Speaker 0

他们以为边缘计算的能耗更低。

They think that edge computing is lower energy.

Speaker 0

实际上,边缘计算比在数据中心计算能效更低。

Actually, edge computing is less energy efficient than computing in the data center.

Speaker 0

当你在数据中心进行计算时,这有点像货运列车。

When you're computing in the data center, it's a little bit like that freight train.

Speaker 0

你可以同时处理大量任务。

You're actually getting to do a whole bunch of jobs simultaneously.

Speaker 0

因此,我们无需从外部内存读取数据,也就省下了这部分能耗。

So the fact that we don't have to read from that external memory means that we don't have to spend the energy doing that.

Speaker 0

即使使用GPU,你也能进行批量处理。

Even with GPUs you get to batch.

Speaker 0

但回到为什么它能效低的问题上,在芯片中消耗的能量主要来自这些物理导线。

But going back to why it's so energy efficient, the amount of energy used in a chip, there's these physical wires.

Speaker 0

这些物理导线有宽度,当你观察其宽度和长度时,需要给导线充电以设置为1,然后再放电以设置为0。

And the physical wires have a width, and when you look at the width and you look at the length, you charge that wire up to set it to a one, and then you discharge it to set it to a zero.

Speaker 0

这就像给电容器充电和放电,都需要消耗能量。

It's sort of like charging a capacitor and discharging a capacitor using energy.

Speaker 0

导线越长,需要的电荷就越多。

The longer that wire, the more charge.

Speaker 0

当这里有HBM,另一个芯片在那边时,每次传输一位数据,你都必须在芯片之间充电并放电。

When you have HBM here and another chip here, you're actually having to charge a wire between the chips and then discharge it every time you send a bit.

Speaker 0

这段距离很长,而且导线比芯片内部的导线更粗。

That's a long distance to travel, but also the wires are wider than the wires that are inside the chips.

Speaker 0

因此你消耗了多得多的能量。

So you just use a lot more energy.

Speaker 0

当我们把内存保留在芯片内部时,数据只传输很短的距离,使用细得多的导线,因此能耗低得多。

When we keep that memory in the chip, it's only traveling a little distance using much thinner wires, and therefore it uses a lot less energy.

Speaker 1

那么,我们是否能看到LPUs和GPUs混合使用的场景?这两种技术的使用比例会是怎样的?

So do we see a world of LPU and GPU usage in combin like how how does that distribution look between LPU usage and GPU usage?

Speaker 1

这有

There's a

Speaker 0

几个方面。

couple of things.

Speaker 0

首先是,训练应该在GPU上进行。

The first is training should be done on GPUs.

Speaker 0

NVIDIA会卖出他们生产的每一颗用于训练的GPU。

NVIDIA will sell every single GPU they make for training.

Speaker 0

目前,他们约40%的市场份额来自推理任务。

Right now, about 40% of their market is inference.

Speaker 0

如果我们部署大量成本低得多的推理芯片,你会看到同样数量的GPU会被售出,但对训练的需求会增加,因为推理越多,所需的训练也越多,反之亦然。

If we were to deploy a lot of much lower cost inference chips, what you would see is that same number of GPUs would be sold, but the demand for training would increase because the more inference you have, the more training you need and vice versa.

Speaker 0

另一个应用场景是,我们的速度实际上比GPU快得多。

The other use case is we're actually so crazy fast compared to GPUs.

Speaker 0

我们已经做了一些实验,将模型的一部分放在我们的LPU上运行,其余部分仍在GPU上运行。

We've actually experimented a little bit with taking some portions of the model and running it on our LPUs and letting the rest run on GPU.

Speaker 0

这实际上加快了速度,并让GPU的使用更经济。

And it actually speeds up and makes the GPU more economical.

Speaker 0

由于人们已经部署了大量GPU,我们考虑的一个应用场景是销售一些我们的LPU来实现加速提升。

So since people already have a bunch of GPUs they've deployed, one use case we've contemplated is selling some of our LPUs to sort of nitro boost

Speaker 1

那些GPU。

those GPUs.

Speaker 0

好吧,这正是我想问的,

Well, this is my question, is that,

Speaker 1

你知道,人们早就买好了GPU,是的。

you know, people have bought GPUs so far ahead of time Yeah.

Speaker 1

等你拿到的时候,它们已经被部署安装了,几乎已经过时了。

That by the time you get them, they're deployed and installed, they're almost out of date.

Speaker 0

实际上,我们已经和一些客户沟通过,他们提前一年多就下了订单。

Actually, we've we've spoken to some customers that put orders in over a year in advance.

Speaker 0

他们提前一年付款,但至今还没收到货。

They paid a year in advance and still haven't gotten them.

Speaker 0

我们在沙特阿拉伯最近的一次部署,从签约到在本国生产环境中首次生成token,只用了51天。

The recent deployment we did in in Saudi Arabia, fifty one days from contract to the first tokens being served in production in country.

Speaker 0

你们是怎么做到的

How are you able to do

Speaker 1

这么快?

it so quickly?

Speaker 1

五十一天令人震惊。

Fifty one days is astonishing.

Speaker 0

是的。

Yeah.

Speaker 0

部分原因是我们在架构上更加简洁。

Part of it is architecturally, things are much simpler for us.

Speaker 0

我们不需要一大堆其他硬件组件。

We don't have a bunch of other hardware components.

Speaker 0

我们实际上不使用交换机来连接我们的芯片。

We actually don't use switches to communicate between our chips.

Speaker 0

我们直接将芯片插到芯片上。

We just plug our chips into our chips.

Speaker 0

我们的芯片就是交换机。

Our chips are the switch.

Speaker 0

我们没有

We don't have

Speaker 1

所有这些网络调优工作。

all of this network tuning.

Speaker 1

考虑到能效和可预测性,为什么英伟达在LPU上不更积极一些?

Given the energy efficiency, given the predictability, why is NVIDIA not being more proactive on LPUs?

Speaker 0

你凭什么认为他们不想在这件事上更积极呢?

What makes you think that they don't want to be more proactive on it?

Speaker 0

他们从不谈论它。

They don't talk about it.

Speaker 0

那他们为什么要谈论它呢?

Well, why would they talk about it?

Speaker 0

那就像在试图展现实力而非脆弱时,故意谈论自己根本没有的东西。

That would be like talking about something you don't have when you're trying to project strength rather than vulnerability.

Speaker 1

想想看,如果你想要保护股东价值,想要维持华尔街对你占据主导地位、遥遥领先的形象

Well, think if you wanted to protect shareholder value and you wanted to protect a Wall Street image of dominance and being ahead of

Speaker 0

如果你在打这场游戏,至少你会说,当然,我们也在研发LPU。

the game, you'd at least say, oh, we are of course working on LPUs as well.

Speaker 0

但在他们拥有LPU之前,这样做实际上会暴露他们有所缺失。

But then, until they had LPUs, they would effectively be exposing that there's something missing.

Speaker 0

比如,看看最近的GTC大会,他们宣布最新一代GPU比上一代快了30倍。

Like, if you look at the last GTC, there was an announcement that the latest GPUs were 30 x faster than the previous generation.

Speaker 0

当你看它是如何实现的时,会看到一条曲线大致是这样的,然后就在这里戛然而止。

When you look at how it was done, there was this curve that looked kinda like this, and then it basically ended here.

Speaker 0

接着又出现了另一条曲线,大致是这样的。

And then there was another curve that was kinda like this.

Speaker 0

而这30倍的提升,是从这条曲线的末端到那条曲线的差距。

Now that 30 x was from the end of this curve to this curve.

Speaker 0

如果你把比较点移到这里,提升就不到30倍了。

If you moved it here, it would have been less than 30 x.

Speaker 0

如果你把比较点移到这里,提升就会变成无限大。

If you moved it here, it would have been infinite.

Speaker 0

所以他们的芯片比前一代快得无限,但这样说听起来不太合理。

So their chip is infinitely faster than the previous one, but that wouldn't have sounded reasonable.

Speaker 0

对吧?

Right?

Speaker 0

这个市场一直有夸大宣传的历史,因为很难真正接触到芯片。

There's a history in this market of specsmanship because it's so hard to, like, get access to chip.

Speaker 0

这是一个关于企业销售的教训。

And this is a lesson on enterprise sales.

Speaker 0

在企业销售中,人们依赖夸大宣传,因为夸大宣传是什么?

In enterprise sales, people rely on specsmanship because Specmanship is what?

Speaker 0

就是我的规格比你的规格更好。

It's well, my specs are better than your specs.

Speaker 0

我的芯片比你的芯片更快。

My chip is faster than your chip.

Speaker 0

我每秒获得的teraFLOPS比你多。

I get more teraflops per second than you do.

Speaker 0

但谁在乎呢?

But who cares?

Speaker 0

告诉我每美元能获得多少令牌,告诉我每瓦能获得多少令牌。

Tell me what the the tokens per dollar is, and tell me what the tokens per watt is.

Speaker 0

其他一切都无关紧要。

Nothing else really matters.

Speaker 0

但人们会找出所有这些其他奇怪的指标来证明自己更优秀。

But people will find all of these other weird things to measure that they might be better on.

Speaker 0

就像我会卖给你一辆转速更高的车。

Sort of like, I'll sell you a car with better RPMs.

Speaker 0

转速并不重要。

RPMs don't matter.

Speaker 0

重要的是每加仑能跑多少英里,也许还有你能达到的速度,不过限速让这些也变得无关紧要。

What matters is miles per gallon and maybe the speed that you can drive at, although speed limits kinda render them, you know, moot.

Speaker 0

对吧?

Right?

Speaker 0

在企业销售中,曾经有一段时间,你会推销肥皂。

In the case of enterprise sales, there there was a time when you would market soap.

Speaker 0

广告牌上会写着:我们的肥皂比其他品牌的肥皂泡沫更多。

The billboards would say, our soap has more bubbles than this other brand's soap.

Speaker 0

谁在乎呢?

Who cares?

Speaker 0

他们后来意识到,不如在广告牌上放一些使用肥皂后非常开心的人,这样人们可能会将这种快乐与产品联系起来。

And what they figured out was, let's put really happy people up on a billboard after they use the soap, and then maybe people associate that happiness.

Speaker 0

对吧?

Right?

Speaker 0

生活方式营销。

Lifestyle marketing.

Speaker 0

当然。

Sure.

Speaker 0

不知为何,企业界至今仍未吸取这个教训。

For some reason, enterprise still hasn't learned this lesson.

Speaker 0

现在还是这样,我们的气泡更多。

It's still, we have more bubbles.

Speaker 0

我们的TeraOps更多。

We have more TeraOps.

Speaker 0

我们有更多各种各样的东西。

We have more whatever.

Speaker 0

这些都是人们根本不在乎的东西。

Things that people just literally don't care about.

Speaker 1

所以你觉得……

So you think

Speaker 0

NVIDIA说‘我们快了30倍’,这并不是好的营销。

NVIDIA is, hey, we're 30 times faster is not good marketing.

Speaker 0

它之所以有效,是因为人们已经习惯了,但我们的回应是发布了一篇新闻稿,上面写着:Grock,仍然更快。

It worked because it's what people are used to, but our counter was we we did a press release to that that said, Grock, still faster.

Speaker 0

就这样。

That was it.

Speaker 0

人们为此疯狂了。

And people went gaga over it.

Speaker 0

对吧?

Right?

Speaker 0

因为我们就只是。

Because it was just we we are.

Speaker 0

我们仍然更快。

We're still faster.

Speaker 0

所以谁在乎呢?

So who cares?

Speaker 1

你觉得华尔街理解这种方式吗?

Do you think Wall Street understands that way?

Speaker 0

我觉得他们开始明白了。

Think they're starting to.

Speaker 0

是的。

Yeah.

Speaker 0

但再说一遍,我不认为这里存在真正的竞争。

But again, I don't think there's real competition here.

Speaker 0

我认为,如果你在竞争,那说明你已经犯了严重的错误。

I think if you are competing, you have done something seriously wrong.

Speaker 0

如果你在竞争,那就意味着你还没有找到一个未被解决的客户问题。

If you're competing, it means that you haven't found an unsolved customer problem.

Speaker 0

因为如果你在竞争,说明别人已经解决了这个问题。

Because if you're competing, someone else has already solved the problem.

Speaker 0

那你为什么还要花时间在这上面?

So why are you spending time on it?

Speaker 0

你不把英伟达视为竞争对手。

You don't view Nvidia as a competitor.

Speaker 0

对。

No.

Speaker 0

他们不提供快速令牌,也不提供低成本令牌。

They don't offer fast tokens and they don't offer low cost tokens.

Speaker 0

这是一个非常不同的产品。

It's a very different product.

Speaker 0

但他们非常擅长训练。

But what they do very, very well is training.

Speaker 0

他们在这一方面做得比任何人都好,而且优势如此之大,这已经是一个被解决的问题。

Do They it better than anyone else, and by such a wide degree, it's a solved problem.

Speaker 0

我们为什么要费心去解决一个已经被解决的问题呢?

Why would we bother trying to solve a problem that's already been solved?

Speaker 0

所以你的想法是,让出训练市场给他们,我们来掌控推理市场。

So you're like, seed the training market to them, we'll own the inference market.

Speaker 0

是的。

Yeah.

Speaker 0

而他们说,我们也想要推理市场。

And they're saying, we also want the inference market.

Speaker 0

当然。

Of course.

Speaker 0

事情一向都是这样发展的。

The way it always works.

Speaker 0

那我们现在该怎么办?

So what do we do now?

Speaker 0

所以我们现在是在推理市场中竞争。

So now we are competing in the inference market.

Speaker 0

但我们真的在竞争吗?

But are we?

Speaker 0

是的。

Yeah.

Speaker 0

我们并没有听到有人说要改买GPU而不是用你们的产品。

We don't really have people saying we're gonna buy GPUs instead of you.

Speaker 0

我们确实听到有人说要同时购买两者。

We do have people saying we're gonna buy both.

Speaker 0

这种情况确实会发生。

That happens.

Speaker 0

但我们并不在意。

But we don't care.

Speaker 0

我给一个人展示了演示,他说:我们是不是干脆别再买GPU了?

I showed a demo to someone, and he's like, should we just not buy any more GPUs?

Speaker 0

我说:不,别这样。

I'm like, no.

Speaker 0

你能搞到多少GPU就买多少。

You should buy every single GPU you can get your hands on.

Speaker 0

他困惑地看着我,我说:你打算怎么训练?

And he's looking at me very perplexed, and I'm like, how are you gonna do training?

Speaker 0

我们不做训练。

We don't do training.

Speaker 0

去买GPU吧。

Buy the GPUs.

Speaker 0

尽可能多买一些,因为我希望你们的模型能运行在我们的平台上

Get every single one you can because I want your models running on us to be

Speaker 1

确实如此,但对于推理来说,他们

really Totally, but for inference, they

Speaker 0

不再需要购买NVIDIA了。

don't need to buy NVIDIA anymore.

Speaker 0

他们不需要为推理购买GPU,但如果你能搞到,我的意思是,它们有点贵,但如果你已经习惯了,为什么不呢?

They don't need to buy GPUs for inference, but if you can get them, I mean, they're a little expensive, but if you're used to it, why not?

Speaker 0

仍然有很多人出售大型机。

Plenty people still sell mainframes.

Speaker 0

如果你想要更低的成本和更快的速度,那就选择LPU。

If you want lower cost and faster, then you want an LPU.

Speaker 1

成本低多少?

How much lower cost is this?

Speaker 0

超过五倍更低。

More than five x lower.

Speaker 1

超过五倍更低?

More than five x lower?

Speaker 0

仅最新显卡的内存成本就超过了我们每块芯片部署的全部资本支出。

Just the memory alone in the latest GPUs costs more than our fully loaded CapEx per chip deployed.

Speaker 0

此外,我们之前谈到了能效问题。

And on top of that, so we talked about the energy efficiency.

Speaker 0

因此,我们每个词元消耗的能量大约只有三分之一。

So we use about a third of the energy per token.

Speaker 0

在三年周期内,我们的运营支出占总成本的三分之一,主要是能源和数据中心租金,而资本支出占三分之二。

Over a three year period, one third of our cost is the OpEx, which is mostly energy and data center rent, and two thirds is the CapEx.

Speaker 0

这意味着,由于我们的能耗只有三分之一,运行显卡生成相同数量词元的费用,就等于我们总的总成本。

Which means that since we're one third of the energy, the cost to run that GPU to produce the same number of tokens for inference is the same as our total cost.

Speaker 0

仅显卡的运营支出,就等于我们的资本支出加上运营支出的总和。

Just the OpEx for the GPU is the same as our CapEx plus our OpEx.

Speaker 1

真的很抱歉。

I'm really sorry.

Speaker 1

我问的都是最蠢的问题,但我还是想继续问下去。

I'm I'm asking the most stupid questions, but I'm just rolling with it.

Speaker 1

我们现在在巴黎,快到一天的尾声了。

We're in Paris and it's nearly the end of the day.

Speaker 0

那为什么他们40%的收入来自推理呢?

Why is 40% of their revenue inference then?

Speaker 0

那你为什么没有拿下更多这部分业务呢?

Why have you not taken so much more of that?

Speaker 0

2024年初,我们只有640颗芯片。

At the beginning of 2024, we only had 640 chips.

Speaker 0

到年底,我们已经有40,000颗了。

At the end, we had 40,000.

Speaker 0

我们还没有达到那个规模。

We're not at that scale yet.

Speaker 0

你必须提供高质量。

You have to provide quality.

Speaker 0

你必须提供低成本。

You have to provide low cost.

Speaker 0

你必须提供速度,但也必须提供容量。

You have to provide speed, but you also have to provide capacity.

Speaker 0

因此,这就是不使用HBM最为关键的部分所在。

And so this is where that that most important part of not using HBM came in.

Speaker 0

这意味着我们实际上没有规模限制。

It means that we effectively have no scale limits.

Speaker 0

因此,GPU本身是使用与你的手机相同的工艺制造的。

So the GPU itself is actually manufactured using the same process that you use for your mobile phone.

Speaker 0

所以你手机里的硅芯片,和GPU用的是同样的硅芯片。

So the same silicon that's in your mobile phone is the same silicon for the GPU.

Speaker 0

事实上,他们先制造手机芯片,因为手机芯片更小,良率更高。

In fact, they build the mobile phones first, the mobile phone chips first, because they're smaller, so they yield better.

Speaker 0

英伟达实际上是在苹果之后才拿到的。

NVIDIA actually gets it after Apple.

Speaker 0

区别就在于内存。

The difference is that memory.

Speaker 0

这就是唯一的区别,但这种内存制造起来非常困难,且产能受限。

That's the only difference, but that memory is the hard part to manufacture that's that's limited in scale.

Speaker 0

因此,通过避免使用它,我们实际上在扩展规模上几乎没有限制。

So by us avoiding that, we effectively have almost no limit on how much we can scale up.

Speaker 0

这对推理非常重要。

And that's important for inference.

Speaker 0

英伟达的利润率是多少?

What is NVIDIA's margin?

Speaker 0

70%到80%。

70 to 80%.

Speaker 0

70%到80%。

70 to 80%.

Speaker 0

所以他们可以砍掉70%到80%的成本,从而变得极具竞争力。

So they can take 70 to 80% off and be radically more Yeah.

Speaker 1

相比你们,成本效益要高得多。

Comparatively cost effective compared to you.

Speaker 0

你可以摧毁他们的利润率。

You could destroy their margin.

Speaker 0

但同样地,你几乎可以说,我们是给英伟达带来过最好事情之一,因为他们可以制造出原本计划生产的每一款GPU,并以高利润率出售用于训练,而这些成本会分摊到整个部署中。

But in that same vein, you could almost say we're one of the best things that's ever happened to NVIDIA because they can make every single GPU that they were gonna make, and they can sell it for training, high margin, right, gets amortized across the deployment.

Speaker 0

我们接手低利润率、高容量的推理业务,让他们不必玷污自己的利润率。

You know, we'll take the low margin, high volume inference business off their hands, and they won't have to sully their margin.

Speaker 0

什么是低利润率?

What's low margin?

Speaker 0

取决于具体交易。

Depending on the deal.

Speaker 0

我们在后端确实能拿到一些,但前期大约是20%。

We do get some on the the backside, but upfront it's about 20%.

Speaker 0

大约20%?

About 20%?

Speaker 1

是的。

Yeah.

Speaker 1

好的。

Okay.

Speaker 1

所以他们的占80%,你们占20%,但你们的20%还要再看。

So theirs is 80, yours is 20, but then you're looking at a 20

Speaker 0

但我们之后还能从它那里获得更多收益,所以我们承担了一些风险。

But then we get more later off of it, so we take some of the risk.

Speaker 0

你什么意思?

What do you mean

Speaker 1

你说之后还能获得更多?抱歉,没听清。

you get more later, sorry?

Speaker 0

我们做的这些交易中,合作伙伴会承担费用,因为我们自己不会投入资本支出。

So the deals that we do, the partner will off because we don't deploy our we don't spend money for our own capex.

Speaker 0

合作伙伴会为我们垫付部署所需的资金。

The partner will put up the money for us to deploy.

Speaker 0

我们会以不错的内部收益率偿还,但收益会分成,大部分归合作伙伴。

We pay back with a decent IRR, but we split, and most of it goes to the partner.

Speaker 0

一旦我们达到内部收益率,情况就会反过来。

And then once we hit the IRR, it flips the other way.

Speaker 0

所以是别人为我们垫付资本支出。

So others are putting the CapEx up for us.

Speaker 1

那么最后看起来是什么样子呢?

What does it look like at the end then?

Speaker 0

嗯,最后的情况和其他商业模式不一样。

Well, at the end, it's not like other business models.

Speaker 0

所以我们不仅仅在芯片上进行了创新。

So we we didn't just innovate on the chip.

Speaker 0

我们还在商业模式上进行了创新。

We also innovated on the business model.

Speaker 0

我们的盈利上限取决于我们能部署多少,而不是我们拥有多少资金,因为资金是由合作伙伴提供的。

And we're limited in how much money we can make based on how much we can deploy, not how much money we have because the partners are putting that money up.

Speaker 0

所以当我考虑我们能做什么时,一切都关乎我们能扩张到多大规模。

So when I'm looking at what we can do, it's all about how much we can scale.

Speaker 0

什么是

What are

Speaker 1

你们部署的限制是什么?

the limits to your deployment?

Speaker 1

仅仅是芯片的限制吗?

Is it purely chip constraints?

Speaker 0

主要是。

Mostly.

Speaker 0

所以你在问关于AI的误解。

So you're asking about misconceptions in AI.

Speaker 0

我认为其中一个关于能源。

I think one of them is about power.

Speaker 0

确实,市场上在芯片拥有者和电力拥有者之间存在不匹配,但部分原因是因为你需要一个中间的数据中心,而目前数据中心数量不足。

It is true that there is a mismatch in the market between people with chips and people with power, but that's partially because you need a data center in the middle and there aren't enough data centers.

Speaker 0

建造这些并不是世界上最难的事情。

Those aren't the hardest thing in the world to build.

Speaker 0

它们并不容易,但也不是最难的。

They're not easy, but they're not the hardest thing.

Speaker 0

建立电力供应更困难。

It's harder to build up the power.

Speaker 0

然而,由于这种不匹配,大型超大规模企业四处宣称:我需要一吉瓦的电力。

However, because of that mismatch, you have big hyperscalers going around and saying, I need a gigawatt of power.

Speaker 0

他们会向多达60家潜在的数据中心建设方提出这个需求。

And they'll say this to to 60 different potential data center builders.

Speaker 0

然后突然间,你听到了回声。

Then all of sudden, hear this echo.

Speaker 0

嗯,我听说这里有一个吉瓦,那里也有一个吉瓦,另一个地方还有一个吉瓦。

Well, I heard that, you know, there's a there's a gigawatt here and a gigawatt here and a gigawatt here.

Speaker 0

突然之间,需求变成了60吉瓦,这正是从最初那一吉瓦引发的回响。

And all of sudden, there's, like, 60 gigawatts of demand, and it's this echo from that first gigawatt.

Speaker 0

事实上,我知道目前有大约20吉瓦的电力正被计划用于数据中心。

The thing is I am aware of about 20 gigawatts of power that people want to make available for data centers now.

Speaker 0

目前,全球大约有15吉瓦的数据中心,是当前容量的两倍多。

Right now, there's about 15 gigawatts of data centers worldwide, so more than double the current capacity.

Speaker 0

我担心的是,人们现在正在增加电力供应。

Concern that I have is that people are now building up more power.

Speaker 0

在未来三到四年里,人们会发现:我建了这么多电力,却没人使用。

And what's going to happen in the next three to four years is people are going to be like, I built up all this power and no one's using it.

Speaker 0

这简直就是完全的浪费,我们再也不会这么做了。

This was like a complete waste and we're never going do this again.

Speaker 0

那么,接下来会发生什么?还记得芯片每十八到二十四个月翻一番的情况吗?

Then what's going to happen, remember that doubling of chips every eighteen to twenty four months?

Speaker 0

三到四年的时间里,你把这15吉瓦翻两番,那就会变成多少?120吉瓦?

Well, three to four years, you double that 15 gigawatts twice, and now you're talking about, what, 120 gigawatts?

Speaker 0

根本没那么多电力可用。

There isn't that much power available.

Speaker 0

再翻一番,那就变成240吉瓦了。

And then another one after that, now you're at two forty.

Speaker 0

因此,由于这种不匹配和当前存在的沟通不畅,我们目前会稍微过度建设。

And so what's gonna happen is we're gonna overbuild slightly right now just because of that mismatch and the miscommunication that's going on right now.

Speaker 0

然后我们会放缓建设步伐,并逐步停止这种行为。

And then we're gonna dampen our building and we're gonna close down on that.

Speaker 0

接着,我们才会真正需要这些电力。

And then we're gonna have the real need for the power.

Speaker 0

这正是我目前最大的担忧,因为这种电力将在三到四年内成为难以突破的瓶颈。

That's my big concern right now because that power will become a hard bottleneck in three to four years.

Speaker 0

当我们正迈向一个推理需求是训练20倍的世界时,为什么还会出现数据或数据中心供过于求的情况呢?

Why will we have that data over or data center oversupply when we are moving into a

Speaker 1

因为推理的需求将是训练的20倍。

world of inference which will be 20 x larger than training?

Speaker 0

数据中心的问题在于,每个人都认为数据中心是房地产。

So the problem with data centers is everyone thinks that data centers are real estate.

Speaker 0

而且很多人确实从事房地产。

And a lot of people do real estate.

Speaker 0

数据中心不是房地产。

Data centers are not real estate.

Speaker 0

现在行业里一个常见的笑话是,有人会说:‘我给你提供100兆瓦的容量,三个月内搞定。’

The common joke in the industry now is someone says, I'm going to have, you know, 100 megawatts of capacity for you, and I'm going to have it in three months.

Speaker 0

你愿意签吗?

You willing to sign?

Speaker 0

然后你问一个问题:‘你的正常运行时间是多少?’

And then you ask a question like, what's your uptime?

Speaker 0

他们却说:‘不知道啊,就看电网怎么样吧。’

And they're like, I don't know, whatever the the power grid is.

Speaker 0

你会说:‘等等,什么?’

You're like, wait, what?

Speaker 0

你的发电机在哪?

Where are your generators?

Speaker 0

哦,还没下单呢。

Oh, haven't ordered those.

Speaker 0

我现在就下单。

I'll order them now.

Speaker 0

你知道现在发电机的交货周期是90个月吗?

You know that there's a ninety month lead time on generators right now?

Speaker 0

真的吗?

Oh, really?

Speaker 0

然后下一个‘90’。

And then the next '90.

Speaker 0

09/00。

09/00.

Speaker 0

接下来的问题是,你的水从哪里来?

And then the next question is, where are you getting the water from?

Speaker 0

等等。

Wait.

Speaker 0

数据中心需要水?

Data centers need water?

Speaker 0

我以为那只是些芯片。

I thought it was a bunch of chips.

Speaker 0

所以有很多人根本不知道自己在做什么,因为他们以为这是房地产。

So there's a bunch of people who have no idea what they're doing going into it because they think it's real estate.

Speaker 0

这些人现在正在建造过多的数据中心,但他们其实并没有真正建造出来。

Those people are now building an oversupply of data centers, but they're not really building them.

Speaker 0

所以他们建造的是虚假的数据中心,人们以为那是真的。

So they're they're fake data centers that people think are real.

Speaker 0

这些数据中心会怎么样?

What happens to those data centers?

Speaker 0

因为它们不会被利用,对吧?

Because they're not gonna be utilized, are they?

Speaker 0

亚马逊不会为一个根本不存在的数据中心付费,嗯,亚马逊不会上这种当。

Amazon is not gonna pay for a data center that doesn't Well, Amazon doesn't fall for this.

Speaker 0

亚马逊有非常优秀的人才。

Amazon has really good people.

Speaker 0

无论买家是谁,都不会为一个

Whoever the the buyer is is not gonna pay for a

Speaker 1

没有水源或没有电力的数据中心付款。

data center that's got no water or got no power.

Speaker 0

是的。

Yeah.

Speaker 0

所以,在你看来,这些项目大多数都永远不会建成,只是浪费了吗?

And so is it just wasted from your Most of these projects will never be developed.

Speaker 1

我们能建得足够快吗?

Will we build them fast enough?

Speaker 1

你提到过数据方面的供过于求。

You said about the data like the oversupply.

Speaker 1

建造一个数据中心确实需要时间。

It take it does take time to build a data center.

Speaker 0

是的。

Yeah.

Speaker 0

所以这几乎是可接受的。

So it's almost okay.

Speaker 0

如果你训练一个模型,你真的希望在六个月内摊销成本。

If you train a model, you really wanna amortize it for about six months.

Speaker 0

如果你部署芯片,你真的希望在三到五年内摊销成本。

If you deploy chips, you really wanna amortize it for three to five years.

Speaker 0

在三年这一端更多一些,其他人则更偏向五年。

One more on the three year side, others are more on the five year side.

Speaker 0

如果你建造一个数据中心,你可能要考虑十年到十五年。

If you build a data center, you're probably talking ten to fifteen years.

Speaker 0

在发电厂,你可能要考虑十五年、二十年。

In a power plant, you're talking like fifteen, twenty years.

Speaker 0

我们这个行业面临的问题在于,融资方式与实际需求之间存在这种不匹配。

The the problem we have in the industry is not on this and there's this mismatch between the sort of financing and the needs here.

Speaker 0

所以,有人想训练一个模型,他们只会做六个月。

So you have someone who wants to train a model, and they're gonna be doing this for six months.

Speaker 0

他们不明白为什么人们希望芯片有三到五年的承诺。

They don't understand why people want three to five year commitments on the chips.

Speaker 0

而部署芯片的人也不明白,为什么有人希望数据中心有十五到二十年的承诺。

And then the people deploying the chips don't understand why someone wants a fifteen to twenty year commitment on the data center.

Speaker 0

对吧?

Right?

Speaker 0

现在数据中心的期限已经是七年了。

It's at seven years now on the data centers.

Speaker 0

然后建造数据中心的人就需要长期承诺,这是一份七十八年的承诺。

And then the people building the data centers then need a long It's a seventy eighth commitment.

Speaker 0

是的。

Yeah.

Speaker 0

他们要的就是这种东西。

That's the kind of thing they're asking for.

Speaker 0

所以整个生态系统中都存在这种完全的不匹配。

So you've got this complete mismatch throughout the ecosystem.

Speaker 0

但有趣的是,尽管他们都希望零风险,并且希望另一方拥有主权财富基金级别的信用评级和长期承诺,但回报周期越长,基础设施就越通用。

But the funny part about it is while they all want to take zero risk and have a committed sovereign wealth level sort of credit rating on the other side of it with long commits, the longer the payoff time, the more generic the infrastructure is.

Speaker 0

一个模型有非常特定的用途,但像LP和GPU这样的加速器除了用于生成式AI或大语言模型外,还能用于其他用途。

A model has a pretty specific use, but accelerators like LPs and GPUs can be used for other things besides generative AI or LLMs.

Speaker 0

数据中心除了用于加速器之外,还能用于其他用途。

The data center can be used for other things besides accelerators.

Speaker 0

电力可以用于任何事情。

The power can be used for anything.

Speaker 0

当他们在这里寻求最低风险时,这恰恰是风险最低的地方,因为如果我们不把它用于AI,我们就可以用它来为所有电动汽车供电。

While they're looking for the least risk over here, it's the place where there is the least risk because if we don't use it for for AI, we'll use it to power all of the electric cars.

Speaker 0

这是不是一种

Is this a

Speaker 1

由于只有 incumbents 能够匹配数据中心提供商所需的期限,所以他们能胜出的情况吗?

case where incumbents win because they're one of the only ones who are able to match the durations required by data center providers?

Speaker 0

没错,这正是我们与沙特阿美以及沙特阿拉伯的新实体合作的原因,因为他们拥有长期资助这种项目的能力。

Well, and and this is why we partnered with Aramco and this new entity in Saudi Arabia because they have an enormous ability to fund this over the long term.

Speaker 0

他们有着非常长远的视角。

They have a very long term perspective.

Speaker 0

他们拥有极佳的信用评级。

They have an amazing credit rating.

Speaker 0

当你说到他们

When you say they

Speaker 1

有能力为它提供资金时,这里存在一个误解,人们以为这是15亿美元的融资轮。

have an ability to fund it, and this is where the misconception was, people think it's a funding round of a billion and a half.

Speaker 1

这并不是10亿美元的融资轮。

It's not a funding round of a billion.

Speaker 0

不,他们没有筹集15亿美元,那是收入。

No, did not raise 1,500,000,000.0, that's revenue.

Speaker 0

这实际上大约是OpenAI收入的30%。

That's actually about 30% of the revenue of OpenAI.

Speaker 1

你能给我详细讲讲这个交易的结构吗?

Can you just walk me through how that deal is structured?

Speaker 0

是的

Yeah.

Speaker 0

我们去年刚开始的时候,部署了19000片我们的芯片。

We started off last year, right, and we got to 19,000 of our chips deployed.

Speaker 0

我们大约在51天内完成了这一点。

Did We that in about fifty one days.

Speaker 0

问题是,今年我们能做些什么?

The question was, what can we do this year?

Speaker 0

他们已经去各地收集了大量电力,这笔交易的结构是:他们会为我们提供资本支出,以便我们在这些数据中心部署我们的芯片,而我们则根据所获得的收入进行偿还。

So they've gone off, they've collected up a bunch of power in the country, and the deal is structured so that they will put up the capex for us to deploy our our chips in that data center or those data centers, and we pay back based on the money that we make.

Speaker 0

这与债务略有不同,因为他们参与了收益的上行部分。

It's a little bit different than debt in that they participate in the upside.

Speaker 0

其性质是相似的。

It's similar in nature.

Speaker 0

这是收入,因为我们实际上提前实现了盈利。

It is revenue because we actually make profit upfront.

Speaker 0

这如何改变了你能做的事情?

How does that change what you can do?

Speaker 0

好吧,我们现在不再受资本限制了。

Well, we are not limited by capital anymore.

Speaker 0

关于Grok,有一个误解。

There is one misconception around grok.

Speaker 0

有一篇论文声称,我们不可能在定价最低的同时实现盈利。

There was a paper that was written that said that we couldn't be profitable while being the lowest price.

Speaker 0

我们可以收取更高的费用。

We could charge more.

Speaker 0

但实际上,我们目前的贡献利润率非常积极。

But actually, we have a very positive contribution margin right now.

Speaker 0

据我们所知,我们是唯一真正通过运行这些开源模型赚钱的公司。

As far as we know, we're the only ones that are actually making money running these open source models.

Speaker 0

因为对于开源模型来说,大家都在用风投资金竞争,试图像优步那样抢占市场份额。

Because with the open source models, everyone's sort of competing with VC dollars trying to take market share Uber style.

Speaker 0

对吧?

Right?

Speaker 0

与此同时,我们坐在这里想,我们可以一直这样做下去,因为我们正在赚钱。

And meanwhile, we're sitting here going, we could do this all day long because we're making money.

Speaker 0

我们甚至能够偿还内部收益率,并让我们的合作伙伴赚钱。

We're able to even pay off an IRR and and and make our partners money.

Speaker 0

模型的另一部分是。

There's another part of the model.

Speaker 0

我们还在与一些专有模型提供商合作。

We're also working with some proprietary model providers.

Speaker 0

我们实际上在周日的Leap活动上展示了第一个,那是我们与Play.ai合作的语音模型。

We actually showed off the first one at Leap on Sunday where we did a voice model with Play dot ai.

Speaker 0

这个也是收入分成。

That one is also a rev share.

Speaker 0

但关键是,他们能从这个模式中赚钱,而行业中的大多数其他人都因为模型的同质化而亏钱。

But the thing is, they get to make money off of that, whereas most others in the industry are losing money because of the commoditization of the models.

Speaker 1

那么,随着你们垄断力量的减弱,价格会变得更便宜吗?还是随着垄断增强,价格反而会上涨?

So do you have cheaper pricing over time as you bluntly have less monopoly power, or do you have higher prices as your monopoly increases?

Speaker 0

我们希望利润率保持稳定,但希望价格下降,因为这样就能触发杰文斯悖论,生活就会变得美好,因为我们将实现规模化。

We want the margin to stay about the same, but we want the prices to go down because then we get into Jevan's paradox, and and life gets great because we're gonna scale.

Speaker 0

我们的重点是实现规模化。

Our focus is on getting to scale.

Speaker 0

为了在人工智能时代保留人类代理,我们需要成为全球最重要的计算提供商之一。

To preserve human agents in the age of AI, we need to be one of the most important compute providers in the world.

Speaker 0

我们的目标是到2027年,提供至少占全球一半的AI推理计算能力。

Our goal by the 2027 is to be providing at least half of the world's AI inference compute.

Speaker 0

我们认为,如果我们没有受到所有限制,甚至可能超过两倍。

We think we could be further than two x, given that we don't have all the constraints.

Speaker 0

但为了实现这一目标,我们必须大力扩张,并通过额外收费,让人们没有任何理由不使用我们的平台运行他们的模型,也不使用我们平台上的模型。

But in order to get there, we do need to be very aggressively building out, and we need to give people no excuse for not running their models on us and and using the models that are on us by charging extra.

Speaker 0

我不断提醒团队,因为有时候你必须反复强调:我们的增长速度超越了指数级。

And and what I keep telling the team over and over again, because you have to remind them sometimes, is we're growing faster than exponential.

Speaker 0

当你以超越指数的速度增长时,任何利润数额都变得无关紧要。

And when you are growing faster than exponential, there's no amount of profit that you can make that matters.

Speaker 0

重要的是在市场中站稳脚跟并变得重要。

What matters is getting a toehold in the market and becoming relevant.

Speaker 1

什么可能阻止这一点?

What could prevent that?

Speaker 0

我们曾经担心有人会定价低于我们,但后来我们意识到这并不是问题,因为投入这个领域的资金太多了,人们更愿意通过使用我们的服务来减少亏损。

We used to be worried that someone would try and price below us, and then we realized that wasn't a concern because there's so much money going into this that people are going to want to lose less money by running on us.

Speaker 0

这不再是问题了,但在我们意识到这一点之前,这曾是最大的担忧。

That isn't a concern, but that was the big one early on until we realized that.

Speaker 0

当我们看到扎克伯格投资650亿美元建设数据中心时,这究竟意味着什么?

When we see Zuck investing $65,000,000,000 in data centers, what does that actually mean?

Speaker 0

这意味着他正在内部化原本必须支付给之前提到的供应商的数据中心利润,而Facebook正在做全栈布局。

That means he's internalizing all of the margins that he would have had to spend on data centers with the providers that we mentioned earlier, and Facebook is doing full stack.

Speaker 0

Meta每年投入650亿美元。

Meta's doing 65,000,000,000 a year.

Speaker 0

我认为谷歌说70或75,萨提亚说微软在做80,然后你还有Stargate。

I think Google said 70 or 75, and Satya said Microsoft's doing 80, and then you've also got Stargate.

Speaker 0

是的。

Yeah.

Speaker 0

这些是惊人的巨额资金。

These are crazy sums of money.

Speaker 0

这些全部都是用于数据中心建设吗?

And this is all for data center build out?

Speaker 0

不是。

No.

Speaker 0

它还包括了里面配套的东西。

It also includes the stuff that goes in.

Speaker 0

包括芯片、系统,所有东西。

Includes the chips as well, the systems, everything.

Speaker 0

我们从未见过这么多钱。

We've never seen money like this.

Speaker 0

不是吗?

No?

Speaker 0

不是。

No.

Speaker 0

以前从未有过这样的情况,但以前也从未有过如此明确的情况,即最终会带来巨大的价值。

There's never been anything like this, but there's never been a case where it was so clear that there was gonna be value at the end.

Speaker 0

如果你早知道搜索会如此成功的话。

If you knew how successful search was gonna be.

Speaker 0

对吧?

Right?

Speaker 0

记住,谷歌之所以尽可能长时间保持私有,是因为他们担心微软会发现搜索能赚多少钱,然后试图复制。

Remember, Google stayed private as long as they did because they were afraid that Microsoft would figure out how much money search was making and then would try and replicate.

Speaker 0

他们一上市,Bing就出现了。

And the moment that they went public, bing.

Speaker 0

他们对这一点的判断非常准确。

They they called that perfectly.

Speaker 0

每个人都清楚人工智能能赚多少钱,所以大家都在争相进入这个领域。

Everyone knows how much money there is in AI, so everyone's going after it.

Speaker 1

你认为这些价值是分散在众多参与者之间,还是集中在一两个玩家手中?

Do you think that value is distributed amongst many players or concentrated towards one or two?

Speaker 1

我完全同意

I completely agree with

Speaker 0

你在明确价值分配方面的观点,但这种价值是相对均匀地分布,还是高度集中?

you in terms of the clear value when assigned, but is it distributed to some levels evenly or concentrated?

Speaker 0

这符合幂律分布。

It's a power law.

Speaker 0

经济中的价值越大,就越有可能出现某个实体远远领先,从而彻底主导市场的风险。

The more value there is in the economy, the more risk there is of a single entity being so far on one end that they just dominate.

Speaker 0

你在七巨头身上就能看到这种现象。

And you see this with the mag seven.

Speaker 0

人们预测,经济规模越大,经济结果的波动就会越剧烈。

And it's predicted just the bigger the economy gets, the more you will have big swings in in the economic outcomes.

Speaker 0

目前,超大规模云服务商的市值都差不多。

Right now, the hyperscalers are all sort of even in their market caps.

Speaker 0

这很奇怪。

It's strange.

Speaker 0

你本以为其中一家会表现得特别突出,远远领先。

You would expect one of them to just be killing it and taking it much further.

Speaker 0

所以我搞不懂为什么它们如此接近。

And so I don't understand why they're so closely grouped.

Speaker 0

那么当我们思考这种分布时,我们该如何改变它呢?

So when we think about that distribution, how do we think about changing that then?

Speaker 0

比如,显然要成为

Like, obviously, with

Speaker 1

Grok,你希望成为镁光灯下的七巨头之一。

the Grok, you want to be one of the MAG seven.

Speaker 1

你希望成为世界上最重要的公司之一。

You want to be one of the most important companies in the world.

Speaker 0

你如何看待这一点?

How do you see that?

Speaker 0

你达成目标的方式和保持地位的方式是两件截然不同的事。

So the way that you get there and the way that you stay there are two very different things.

Speaker 0

初创公司中存在着一种生命周期循环。

There's sort of a circle of life that happens in startups.

Speaker 0

第一阶段是解决一个未被解决的问题。

The first circle the first stage is solve an unsolved problem.

Speaker 0

这就是你走红的原因。

That's how you go viral.

Speaker 0

这就是你取得成功的途径。

That's how you do well.

Speaker 0

第二阶段是营销阶段,此时其他人开始模仿你的做法,因为他们自己想不出新点子。

The second stage is the marketing stage, which is now other people are trying to copy what you've done because they can't think of something themselves.

Speaker 0

现在你必须在广告和营销等方面与他们展开竞争。

And now you have to fight it out in in advertising and marketing and whatnot.

Speaker 0

你看到消费品公司常常卡在这一阶段,它们的关注点更多变成了产品在货架上的位置,而不是其他任何东西。

You see CPG companies often get stuck there, It and becomes more about where on the shelf they are than anything else.

Speaker 0

然后是最后阶段:七大力量。

Then the final stage is the seven powers.

Speaker 0

一旦你找到了其中一些优势,并真正开始提升它们,你就拥有了系统性的优势。

It's once you've found some of those and you've really started improving it and you have sort of systemic advantages.

Speaker 0

然后,当有人解决了某个未被解决的客户问题时,整个生命周期就继续循环下去。

And then what happens is someone solves an unsolved customer problem, and the whole cycle of life continues.

Speaker 0

对吧?

Right?

Speaker 0

现在谷歌必须重新开始,因为大语言模型比搜索更优秀。

Now Google has to redo this because LLMs are better than search.

Speaker 0

所以,你要成为七巨头之一,首先就要解决那个未被解决的问题。

So the way that you start off to become a mag seven is you solve that unsolved problem.

Speaker 0

而保持这一地位的方式,首先是找到这七大力量中的一项或多项,但你还要随时准备应对被颠覆,并继续反击、解决客户问题。

The way that you stay there is first you find one of those seven powers or or multiple, but then you have to be ready for when you get disrupted to continue fighting back and and solving customer problems.

Speaker 1

我们提到了这里投入的巨额资金。

We mentioned the different huge amounts of money that's being spent here.

Speaker 1

这是一个良好的泡沫吗?它为未来十年到二十年打下基础,最终资本真正变得有生产力,尽管在账面上看起来并非如此?

Is this a good bubble that lays the foundations for an incredible next ten to twenty years, where, finally, the capital actually turns out to be productive but not seemingly so on paper?

Speaker 1

还是说,实际上大量资金被浪费在了贬值的资产上?

Or is it where actually just a huge amount of money is incinerated on depreciating assets?

Speaker 0

我可以保证,会有巨额资金被浪费。

I can guarantee you that a huge amount of money will be incinerated.

Speaker 0

我也打赌,总体而言,赚到的钱会比投入的更多。

I also bet that in total, more money will be made than will be put in.

Speaker 0

这就是问题所在。

This is the problem.

Speaker 0

你必须从整体上或单个投资的角度来看待它。

You you have to look at it either in aggregate or individual bets.

Speaker 0

当所有人都在市场中进行投资时,有些人会亏钱,因为并非每家公司都能成功。

When everyone is making investments in the market, some people are gonna lose money because not every company is gonna be successful.

Speaker 0

你总是会看到,当有一些真正的技术进步或新事物出现时。

What you always see is when there is some real tech improvements or things coming.

Speaker 0

那些早期的、人们大力投资的、极其成功的东西,然后所有人都想加入进来。

You've got the things that were early, that people are investing in heavily, that are super successful, and then everyone else wants to get in on it.

Speaker 0

于是,从AI芯片和AI模型,变成了现在满大街的AIT恤,接下来你就会看到AI导热膏。

And, you know, it goes from you have AI chips and AI models to now you've got AI t shirts, and next thing you know, you've got AI thermal grease.

Speaker 0

人们开始把AI应用到一切事物上。

People just start applying AI to everything.

Speaker 0

转眼间,你就会看到AI公寓。

Next thing you know, you'll have an AI condo.

Speaker 0

所以关键在于分辨什么是真实的,什么是虚假的。

So the the trick is discerning what is real and what isn't.

Speaker 0

每当有真正的东西出现时,总会有这些令人讨厌的骗子涌入,这很不幸。

You're always gonna have all of these really obnoxious charlatans coming in whenever there's something real, and that's unfortunate.

Speaker 0

但一旦人们开始理解这项技术,明白什么是真实的、什么是虚假的,这些骗子最终就会被淘汰。

But eventually they get cleared away once people start to understand the technology and what's real and what isn't.

Speaker 0

工作就是开始进行教育。

The job is to start educating.

Speaker 0

人们越有知识,就越不会投资于AI保暖膏。

And the more educated people are, the less they'll invest in AI thermal grease.

Speaker 1

什么是最有可能导致巨额资金打水漂的最大单笔投资?

What is the largest individual bet that will lead to the largest incineration of cash?

Speaker 0

我不会点名任何人,但我认为这会在每一个学科中都发生。

I'm not gonna call anyone out in particular, but I actually think it will happen across every single discipline.

Speaker 0

你知道凯恩斯的选美比赛吗?

The are you aware of the Keynesian beauty contest?

Speaker 0

不知道。

No.

Speaker 0

约翰·梅纳德·凯恩斯,那位经济学家?

John Maynard Keynes, The Economist?

Speaker 0

这能解释你关于风险投资需要知道的一切。

This will explain everything you need to know about VC.

Speaker 1

所以我有点紧张,但你继续说。

So I'm nervous, but keep going.

Speaker 0

想象一本满是模特的杂志,就是那种长得很好看的真人模特。

So take a magazine full of models, human models, like, you know, good looking models.

Speaker 0

让一群风投人士坐在房间里。

Have a whole bunch of VCs in the room.

Speaker 0

他们可以下注,猜谁是最美的模特。

They're allowed to make bets on who the the most beautiful model is.

Speaker 0

最后,押注金额最多的人获胜。

In the end, whoever has the most money on them is the winner.

Speaker 0

你押在某个模特身上的比例,决定了你分得所有资金的比例。

And based on the proportion that you put on that particular model's face, you get the share of all of the money.

Speaker 0

如果你押的不是最终获得最多资金支持的那一位,那你就会输掉你的钱,归那些押对的人所有。

If you put money on one that isn't the most beautiful by dollars, then you lose your money to the the people who bet on that one.

Speaker 0

这正是软银所下的赌注,他们相信自己能赢得这场凯恩斯式选美比赛。

That was sort of the the bet that SoftBank was making, which was they could win the Keynesian beauty contest.

Speaker 0

我就要再多投点钱,然后赢。

I'm just gonna put more money in and I'm gonna win.

Speaker 0

当你拥有真正的技术优势而非营销时,这就成问题了。

That is problematic when you have true technological advantages as opposed to marketing.

Speaker 0

当你在解决客户问题时,这就像一台秤。

When you're solving customer problems, it's a weighing machine.

Speaker 0

一旦客户问题被解决,你就进入了这种营销的流行度竞赛。

Once the customer problem has been solved, you then get into this sort of popularity contest of marketing.

Speaker 0

这次出现了一件不同寻常的事,我认为这在风险投资史上从未发生过:你看到一些人筹集了数十亿美元,而他们的竞争对手也筹集了数十亿美元。

Now something unusual has happened this time around, which I don't think has ever happened in in VC before, which is you see people raising billions of dollars who have competitors who've raised billions of dollars.

Speaker 0

通常情况下,凯恩斯式选美比赛会有明确的赢家。

It usually, there is a clear winner in the Keynesian beauty contest.

Speaker 0

你不会遇到这种争斗,比如,我得再投点钱。

You don't have this, like, fight where, you know, it's it's sort of like, well, I gotta put a little more money in.

Speaker 0

我得再投点,你知道的,投一百亿美元。

I gotta put a little more I gotta, you know, put 10,000,000,000 in.

Speaker 0

不。

No.

Speaker 0

我要投200亿美元。

I'm gonna put 20,000,000,000.

Speaker 0

我要投5000亿美元,因为凯恩斯式选美比赛已经完全失控了。

I'm gonna put 500,000,000,000 in because the Keynesian beauty contest has gone completely amok.

Speaker 0

这以前从未发生过,所以现在人们甚至不知道该如何应对,因为以前如果有人融了十亿美元,你会觉得,哦,他们是赢家。

This has never happened before, and so now people don't even understand how to react because it used to be if someone had raised a billion dollars, you're like, oh, they're the winner.

Speaker 0

现在的情况是,有三四个竞争对手都融了十亿美元。

Now it's like there's three or four competitors who have a billion dollars.

Speaker 1

那谁赢谁输呢?

So who wins and who loses?

Speaker 1

比如,马西会烧掉有史以来最多的现金吗?

Like, is Massey gonna incinerate the largest amount of cash ever?

Speaker 1

我认为

I think

Speaker 0

凯恩斯选美竞赛在这里不再适用,因为现在有大量资金被分散出去,我认为真正有最好产品的人才会成为赢家,因为每个人都能获得资本。

the Keynesian beauty contest no longer applies here because there's so much money available being spread out, and I think you're gonna see that the people who have the best products are actually gonna be the winners because everyone can be capitalized.

Speaker 0

但赢家也会因为这种情况而面临问题。

But there will be problems for the winners because of this.

Speaker 0

问题在于,你原本打算雇佣的员工,却被别人提供了荒谬的高薪。

The problems are gonna be of the sort you had this employee that you were gonna hire, and someone offered them a ridiculous amount of money.

Speaker 0

是的。

Yeah.

Speaker 0

你现在经常能看到这种情况。

You see this all the time now.

Speaker 0

他们本可以去支持赢家,但现在却去支持了一个本不该存在、或同样可能获胜的竞争对手,导致人才被分散了。

And they could have gone and contributed to the winner, but now they're contributing to a competitor that shouldn't even exist or is equally likely to win, and now you're splitting the talent.

Speaker 1

当你面对如此高的薪资时,你还会怎么做?

What do you also do when you have such high salaries?

Speaker 1

我们已经看到一些公司给初级到中级员工开出一百万、两百万的薪资,而且他们

We've seen 1,000,000, 2,000,000 for kind of junior to mid level in some of these companies, And they

Speaker 0

他们实际上生活在很棒的地方,过着精彩的生活。

are living an amazing life actually in great places.

Speaker 0

你以为他们在广东为深度求索或其他中国竞品公司工作时,过着这样的精彩生活吗?

You think they're living that amazing life in Guangdong when they're working for DeepSeek or any other Chinese alternative?

Speaker 0

我认为他们实际上拿的钱少得多,每天拼命工作二十小时,连康普茶都喝不上,却只拿两百万年薪。

I think they're actually getting paid much less working their fucking ass off twenty hours a day and not getting kombucha and being paid 2,000,000 a year.

Speaker 0

公平吗?

Fair?

Speaker 0

不仅公平,我们有一项政策:从不提供最高薪资,因为我们希望人们选择我们,而不是选择薪水。

Not only fair, we have a policy that we never offer the highest because we want people to choose us, not choose the salary.

Speaker 0

如果我们卷入了薪资竞价战,那下次再有人开出更高的薪水,他们就走了。

If we win in a bidding war, then that means the next time someone comes along with a higher salary, that's it.

Speaker 0

他们只会去拿那份其他的工作。

They're just gonna go take that other job.

Speaker 0

根本没有什么忠诚度。

There's no loyalty.

Speaker 0

他们不相信这个使命。

They don't believe in the mission.

Speaker 0

相反,我们关注的是:我们要打造这个产品。

Instead, we focus on, look, we're gonna build this.

Speaker 0

这是你的机会。

This is your opportunity.

Speaker 0

你将有机会与杰出的人共事。

You're gonna get to work with amazing people.

Speaker 0

花点时间与团队相处。

Spend some time with the team.

Speaker 0

这些是你想共事的人吗?

Are these the people you wanna be working with?

Speaker 0

因为坦白说,你会赚到很多钱。

Because frankly, you're gonna make so much cash.

Speaker 0

这并不重要。

It doesn't matter.

Speaker 0

但要押注于股权和结果。

But bet on the equity, the outcome.

Speaker 0

帮助我们让这件事变得有价值。

Help us make this thing valuable.

Speaker 0

那些认同这一点的人更容易管理,因为他们以使命为导向。

And people who buy into that, they're so much easier to manage because they're mission oriented.

Speaker 0

他们都想做同样的事。

They all wanna do the same thing.

Speaker 0

他们不是因为想喝康普茶才来,也不会因为拿铁机坏了就抱怨。

They're not there because they want the kombucha, they're not going to complain because the cappuccino machine is broken.

Speaker 0

他们会直接去隔壁买咖啡。

They'll just go and buy their coffee next door.

Speaker 0

你会吗

Will you

Speaker 1

NVIDIA 会进入模型空气领域吗?

and NVIDIA move into the model air?

Speaker 1

每个人都说模型提供商正在成为应用提供商,或者基础设施提供商正在成为模型提供商。

Everyone talks about model providers becoming application providers, or infrastructure providers become model providers.

Speaker 0

我们决定不自己训练模型。

We have decided that we're not going to train our own models.

Speaker 0

我们只会为特定情况做一点微调,但我们不想参与竞争。

We'll do a little fine tuning for specific cases or whatnot, but we don't want to compete.

Speaker 0

这一点非常重要,因为人们把他们的模型和权重都放在我们这里。

That's really important because people are putting their models with their weights on us.

Speaker 0

他们不希望我们从中学习并把它们用于自己的利益。

And they don't want us to learn from and take that stuff for our own benefit.

Speaker 0

当你与超大规模云服务商合作时,就会遇到这个问题,因为你知道他们也在做你正在做的一切。

This is the problem you have when you work with a hyperscaler because you know they're also doing everything that you are doing.

Speaker 0

所以我们决定:模型提供商,你们负责做模型,我们不做这个。

So we've decided model providers, you make the model, we don't do that.

Speaker 0

还有用户和查询的数据层面。

There's also the data side of the users and the queries.

Speaker 0

所以我们能做但没有做的另一件事是记录查询,这样如果我们想训练的话就有数据了。

So the other thing that we could do that we do not do is log the queries and then we've got data if we wanted to train.

Speaker 0

我们不进行训练。

We don't train.

Speaker 0

我们没有理由保留这些数据。

We have no reason to hold the data.

Speaker 0

我们只在DRAM中临时存储数据。

We only temporarily store things in the DRAM.

Speaker 0

没有持久化存储。

There's no persistent storage.

Speaker 0

如果断电了,所有数据都会消失。

If the power went out, everything's gone.

Speaker 0

而且DRAM容量有限,所以我们无法长时间保存数据。

And DRAM is limited, so we can't hold things for a long time.

Speaker 0

所以你知道我们不会拥有你的数据。

So you know that we don't have your data.

关于 Bayt 播客

Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。

继续浏览更多播客