Zero Knowledge - AI与ZK审计与David Wong 封面

AI与ZK审计与David Wong

AI and ZK Auditing with David Wong

本集简介

在本期节目中,安娜与zkSecurity的大卫·王深入探讨了人工智能对零知识审计领域的影响。大卫分享了其团队如何将AI智能体融入工作流程、由此带来的机遇与挑战,以及随着AI能力提升,人类审计师角色的演变。 他们还探讨了基于ZK系统审计的独特之处、与智能合约审计的差异,以及AI对密码学与区块链安全的广泛影响。对话涉及AI安全中的伦理考量、审计行业的未来前景,以及ZK技术与AI日益紧密的交集。 相关链接: 第290期:与大卫·王共探ZK技术的探索、教学与审计 第76期:肖恩·鲍谈SNARKs、可信设置与椭圆曲线密码学 "布丁见真章"第04期:基于格的SNARKs与后量子密码学 Fiat-Crypto:密码原语构造正确性代码合成 Ethproofs实时证明电话会议#2 折叠Circom电路:ZKML案例研究 - Cathie So博士 查看ZK领域最新职位:ZK播客工作板 **支持我们:** * 所有链接汇总 @ZeroKnowledge | Linktree * 订阅播客通讯 * 关注Twitter @zeroknowledgefm * 加入Telegram群组 * YouTube频道观看 **赞助节目:** * Patreon * ETH - 捐赠地址 * BTC - 捐赠地址 * SOL - 捐赠地址 阅读文字稿

双语字幕

仅展示文本字幕,不包含中文音频;想边听边看,请使用 Bayt 播客 App。

Speaker 0

欢迎来到零知识领域。

Welcome to Zero Knowledge.

Speaker 0

我是主持人安娜·罗斯。

I'm your host, Anna Rose.

Speaker 0

在本期播客中,我们将探索零知识研究和去中心化网络的最新进展,以及那些有望改变我们在线交互与交易方式的新范式。

In this podcast, we will be exploring the latest in Zero Knowledge research and the decentralized web, as well as new paradigms that promise to change the way we interact and transact online.

Speaker 0

本周,我与来自ZK安全审计公司的David Wong进行了对话,该公司在零知识领域非常活跃。

This week, I chat with David Wong from ZK Security, an auditing firm that is very active in ZK.

Speaker 0

我们探讨了AI时代审计工作既作为机遇又作为新兴威胁的双重性。

We chat about the new era of auditing with AI as both an opportunity and an emerging threat.

Speaker 0

我们讨论了行业工具的演进,包括静态/动态分析器和形式化验证技术,以及AI如何改变这一切。

We discussed the evolving tools of the trade, the static and dynamic analyzers, formal verification techniques, and how these all change with AI.

Speaker 0

他分享了其团队如何运用AI智能体来增强人类审计师的工作效能。

He shares how his operation includes a team of AI agents working to enhance the work of the human auditor.

Speaker 0

我们还比较了零知识技术审计与智能合约审计之间的差异。

We also cover the difference between auditing ZK based tech and things like smart contracts.

Speaker 0

当AI落入恶意行为者手中时构成的威胁,以及他是否认为审计角色很快会被AI完全取代。

The threat posed when AI gets into the hands of malicious actors, and whether he thinks the role of auditing will soon become entirely replaced with AI.

Speaker 0

我特别关注AI与零知识证明如何交叉融合,因为目前已知有几种方式。

I'm particularly interested in how AI and ZK are intersecting, as there are a few ways we know of.

Speaker 0

更广泛的ZKML(零知识机器学习)类别:用ZK证明模型特性,用ZK为模型增加隐私保护。

The broader ZKML category, ZK to prove something about a model, ZK to add privacy to a model.

Speaker 0

我们还有用于溯源证明的ZK技术,可以区分内容是否由AI生成。

We have ZK for provenance to show what is and isn't AI.

Speaker 0

在本期节目中,我们将探讨人工智能如何影响零知识领域的审计工作。

And in this episode, we discuss how AI is affecting auditing in the ZK space.

Speaker 0

我确实希望在未来几个月有机会讨论更多类似案例。

I do hope to get a chance to discuss more cases like this in the coming months.

Speaker 0

在开始之前,我想提醒大家关注零知识职位公告板。

Now, before we kick off, I just want to point you towards the ZK jobs board.

Speaker 0

在那里你可以找到来自顶级零知识团队的工作招聘信息。

There you can find job postings from top teams working in ZK.

Speaker 0

如果你是招聘团队,今天也可以在那里发布职位。

And if you're a team looking to hire, you can also post your job there today.

Speaker 0

我们听到很多团队通过ZK招聘板找到理想人选的好消息,希望它也能帮到你。

We have heard great things from teams who have found their perfect hire through the ZK jobs board, and we hope it can help you as well.

Speaker 0

详情请访问jobsboard.zeronelage.fm。

Find out more over at jobsboard.zeronelage.fm.

Speaker 0

你可以在我们的网站和节目备注中找到这个链接。

You can find this on our website and in the show notes.

Speaker 0

现在,有请我们与David Wong的对话环节。

Now, here is our episode with David Wong.

Speaker 0

今天和我一起的是来自ZK Security的David Wong。

Today, I'm here with David Wong from ZK Security.

Speaker 0

欢迎再次来到节目,David。

Welcome back to the show, David.

Speaker 1

嗯。

Yeah.

Speaker 1

再次感谢邀请我。

Thanks for having me again.

Speaker 1

再来一次,再来一次。

And one more one more time.

Speaker 0

是啊。

Yeah.

Speaker 0

我们回顾了上一期节目。

We looked back at the previous episode.

Speaker 0

那要追溯到2023年了。

It was all the way back in 2023.

Speaker 0

我记得你那时刚创立ZK Security。

I think you had just started ZK Security back then.

Speaker 1

没错。

Yeah.

Speaker 1

你们可能每周都做节目,但对我来说感觉已经很久远了。

You I think you do these every weeks, but for me, it's like ages away.

Speaker 1

所以时隔两年再回来,感觉相当震撼。

So it was like quite a shock to come back after two years.

Speaker 0

但很高兴你能回来。

But it's good to have you back on.

Speaker 0

我们先简单叙叙旧吧。

Let's first do a little catch up.

Speaker 0

我是说,这两年你过得怎么样?

I mean, how have you been over the last two years?

Speaker 0

ZK Security发展得如何?

How has z ZK Security been going?

Speaker 0

比如,接手了哪些类型的项目?

Like, what kind of projects have come your way?

Speaker 0

公司有什么变化吗?

Has the company changed?

Speaker 0

是啊。

Yeah.

Speaker 0

我们来聊聊这个吧。

Let's talk about that.

Speaker 1

是啊。

Yeah.

Speaker 1

我是说,已经有一段时间了。

I mean, it's it's been a while.

Speaker 1

我们经历了很多变化,创始人团队有所变动。

We've had a lot of changes from the the founder sets changed.

Speaker 1

我们仍然是三位创始人,但经历了一些人员调整。

We're still three founders, but there was some some reshuffle reshuffling.

Speaker 1

好的。

Okay.

Speaker 1

团队也壮大了。

Also grew as a team.

Speaker 1

我们现在有13到15人,虽然不记得当时具体有多少人,但大概就两个左右。

We're we're 13 slash 15 people now, so I don't know how many people we were back then, but probably like two.

Speaker 1

最多两三个。

Two or three max.

Speaker 0

我想是的。

I think so.

Speaker 0

是啊。

Yeah.

Speaker 1

所以我们...对。

So we yeah.

Speaker 1

我们的团队更大了。

We're a bigger team.

Speaker 1

我们经历了很长时间的打磨,所以参与过很多不同的项目。

We've grinded for quite a long time, so we we worked on a lot of different projects.

Speaker 1

我知道我们之前专注于ZK领域,但之后我们做过NPC相关、门限签名、共识协议,最近在做一些TE相关的内容,这些都可以做成有趣的专题节目。

I I know we were very focused on ZK, but since then we've done some NPC stuff, threshold signatures, we've done consensus protocols, we've done these days we're doing some TE stuffs, that that would make an interesting episode.

Speaker 1

现在很多事情都在推进,我们更像一个正规公司了。看到ZKVM和ZKEVM这些技术发展,我们非常期待这些新动向。

So so a lot of things are happening, we're more of a real company these days and we're seeing know that there's all these ZKEVM happening or ZKVM happening, we're really looking forward to that stuff.

Speaker 0

不错。

Nice.

Speaker 0

我想知道,回顾2023年,这类项目的数量是否有所增加?

I'm wondering like, just going back to 2023, have you found the volume of these kinds of projects, has it gotten more?

Speaker 0

是增加了还是趋于平稳?

Like, is it has it increased Or has it leveled off?

Speaker 0

和以前相比差不多吗?

Is it about the same as it was?

Speaker 0

我的意思是,那时候很多团队也在和更成熟的审计机构合作。

I mean, I guess back then, a lot of those teams were also working with like more of the established auditors.

Speaker 0

所以可能随着你们声誉的积累,业务逐渐增多。

So maybe like your business picked up over time because you built up a reputation.

Speaker 0

但我其实是想从你这里了解一下,你觉得现在的项目数量是变多了还是保持稳定。

But I I am sort of trying to get a, like, a bit of a temperature check from you as to, like, if you see more today or if you actually see if it's holding steady.

Speaker 1

大家总是问我这个问题。

People always ask me this question.

Speaker 1

这个问题以及ZK领域的最新趋势是什么,我其实一直不太清楚该如何回答。

This question and what are the trends in ZK, and I I never really know how to answer that.

Speaker 1

所以本质上,这取决于人们的认知。

So basically, there's perception.

Speaker 1

对吧?

Right?

Speaker 1

就像我们几年前对ZK的认知那样,具体多少年记不清了,但那时ZK相关的项目层出不穷。

Like we perceive ZK like a few years back, I don't know how many years now, there was like so many ZK stuff happening.

Speaker 1

对吧?

Right?

Speaker 1

就像每周都有新的折叠方案出现,每周都有新的研究。

Like there was a new folding scheme every week, there was a new That's research.

Speaker 1

事情不断

Thing every

Speaker 0

每周都有。

week.

Speaker 0

是啊。

Yeah.

Speaker 1

没错。

Right.

Speaker 1

从研究角度看更多了,但最近感觉进展变慢了,不过同时我收到越来越多...该怎么形容呢,就是来自潜在客户的咨询,有些公司我甚至从未听说过。就像我们看到的冰山一角,但其实我们还接到很多闻所未闻的公司的工作邀约,包括一些小公司或零散项目。而且我们开始越来越多地与非区块链项目接触,比如看到谷歌...

Research wise more, and these days it feels like it's slower, but at the same time I'm getting like like, how do you call that, like incoming, you know, calls from potential clients that I've never heard about or so there's all these kind of like iceberg of what we're seeing, but also we're getting a lot of work from companies that I've never heard about, and and like smaller companies or like projects and here and there, and we're talking with like we're starting to talk more and more with like non blockchain projects, like we've seen like Google.

Speaker 1

有意思。

Interesting.

Speaker 1

那个Wide项目

The the Wide project and

Speaker 0

对。

Yeah.

Speaker 1

所以确实有事情在发生,只是不清楚具体发生了什么,或者发生的规模有多大。

So so there's things happening, it's not necessarily clear what is happening or or what's the amount of things happening.

Speaker 1

我的意思是,可以明确告诉你我们现在工作量非常大。

I mean, what I can tell you is that we're we have a lot of work.

Speaker 1

目前我们公司自创立以来的工作量达到了历史最高点,这是个好兆头。

We're probably we have the most amount of work we've ever had since the inception of the company right now, so that's a good sign.

Speaker 1

我不知道这种情况是否会持续下去。

I don't know if it's gonna stay the same.

Speaker 1

我看到了之前提到的ZKVMs清单,就是前几天在推特上发的那个,我当时就想,好吧。

I I saw that list of ZKVMs that I that I mentioned, like on Twitter the other day, so I was like, okay.

Speaker 1

看来工作短期内不会停下来。

I guess work is not gonna stop anytime soon.

Speaker 0

是啊。

Yeah.

Speaker 0

不过,确实。

But, yeah.

Speaker 0

我觉得...是的。

I think Yeah.

Speaker 0

对我来说,最近Justin Drake举办了一个ETH证明研讨会。

For me, it was this, like, I think just recently, Justin Drake hosted an ETH proofs meetup.

Speaker 0

我不知道

I don't know if

Speaker 1

you

Speaker 0

但那份清单,对我来说太惊艳了。

were But that list, I mean, that was amazing to me.

Speaker 0

因为我知道——我是说,我们大概认识其中80%的团队,但他分类的方式、整个设计的思路,还有那么多新玩家加入ZKVM领域的事实,简直让人惊叹。

Because I knew I mean, I think we knew about 80% of those teams, but the categories that he put them in, just like the way it was designed, and the fact that there was a bunch of new players who had joined the ZKVM space, was just like, wow.

Speaker 0

这个角落看起来像这样

This this look like this corner

Speaker 1

他们?

them?

Speaker 1

哦,

Oh,

Speaker 0

我得查一下。

I'd have to check.

Speaker 1

是20左右还是

Is it like 20 or

Speaker 0

所以,我是说,他的名单,现在看的话,我认为上面大概有27个团队。

So, I mean, his list, just looking at it right now, there I think there's about 27 teams on it.

Speaker 0

他确实包含了一些硬件团队。

He did include some hardware teams.

Speaker 0

这是ZK更大版图的一个子集。

This is a subset of the larger map of ZK.

Speaker 0

我认为ZK的更大版图约有200个团队。

I think the larger map of ZK has around 200 teams.

Speaker 0

但这些是基础设施团队。

But these are like the infrastructure teams.

Speaker 0

而且它们被分在了子类别里,其中一些我甚至都没注意到。

And they're and they're in like subcategories that some of them I I didn't even have on my radar.

Speaker 0

我不知道。

I don't know.

Speaker 0

看到它们真的很酷。

It was pretty cool to see them.

Speaker 1

是啊。

Yeah.

Speaker 1

人们在做各种各样的事情,这就是我想说的。

People are doing all sorts of like that that's what I'm saying.

Speaker 1

对吧?

Right?

Speaker 1

我们很多人只看到冰山一角,但水下还有整个应用生态系统,甚至像ZK L1、L2这些我们现在都不知道正在使用ZK的项目,我们正在发现这些,然后会想:真的吗?

A lot of us see the the top of the iceberg, but there's this entire thing submerged of apps and even like ZK l one, like two there there are l ones that we don't even know use ZK right now, and we're discovering that and we're like, really?

Speaker 1

哦,好吧,挺酷的。

Oh, well, cool.

Speaker 1

不错。

Nice.

Speaker 1

我是说Solana也开始部分采用ZK技术了

I mean Solana is starting to use ZK a part

Speaker 0

业务的一部分。

of the business.

Speaker 0

有一个ZK生态系统,Solana内部也有ZK技术。

There's a ZK ecosystem and there's ZK within Solana.

Speaker 0

几个月前我们和Merge合作做过一期节目。

We did an episode like a few months ago with Merge.

Speaker 1

哦,对。

Oh, right.

Speaker 0

我们还做了个调查。

And we did a survey.

Speaker 0

所以,是的。

And so there yeah.

Speaker 0

那边确实正在发生这种情况。

There's it's definitely happening over there.

Speaker 0

回到你之前提到的研究话题,从我们这边来说,我们正在做ZK Mesh。

Just going back to what you had said about research too, like from from our side, we're doing ZK Mesh.

Speaker 0

我们每个月都这样做。

We do this every month.

Speaker 0

那就像是一个链接列表。

That's just like a list of links.

Speaker 0

实际上,我想说关于ZK的研究一直在持续增长。

And actually, I would say that research has continued to grow on ZK.

Speaker 0

但我觉得你可能没察觉到这一点,是因为那些突破并不一定是巨大的转变或重大飞跃,它们没有像洪水般涌现。

But I think the reason you maybe don't sense that is that the breakthroughs aren't necessarily like the big shifts, the big step changes, they're not coming like in this flood.

Speaker 0

但当你看看那些小型研究论文,比如讨论CK或对某些东西进行渐进性改进的论文。

But when you look at like the small, like basically research papers that talk about CK or doing like an incremental, you know, addition to something.

Speaker 0

实际上数量相当多。

There's actually a lot.

Speaker 0

多到什么程度呢——2020年时,我们勉强能列出6个ZK相关项目,还得硬凑数

Like, to the point where back in 2020, we could barely fill a like, the list was like six ZK things and we added because

Speaker 1

因为我每次看这份清单,上面都列着太多东西了

because I I read this list, and every time I'm about it, there's so many things on

Speaker 0

数量太多了

There's so many

Speaker 1

要怎么筛选呢?

gonna sort through it?

Speaker 0

但你看,过去是混杂的

But see, it used to be a mixture.

Speaker 0

以前会包含一些ZK内容,但也会加入共识研究和通用材料

It used to have, like, a few z k things, but then we'd put, like, consensus research and, like, general stuff.

Speaker 0

而最近几年我们越来越聚焦于ZK领域,因为不得不剔除TEE

And in the last few years, we've just gotten more and more focused on ZK because there it it was like, we we had to cut out TEE.

Speaker 0

不得不剔除MPC

We had to cut out MPC.

Speaker 0

甚至连紧密相关的内容也要剔除,因为每月都有大量ZK研究论文涌现

Like, we had to cut out even like the close adjacent stuff because there were so many ZK research papers coming out every month.

Speaker 0

我认为这个趋势至今仍在持续

And I I would say that still stands.

Speaker 1

我是说,这让我觉得你们做了很多关于格密码学的节目。

I mean, makes me think you guys did a bunch of like episodes on lattice based cryptography.

Speaker 1

对吧?

Right?

Speaker 0

没错。

True.

Speaker 1

所以我最近也开始深入研究这个领域。

And so I started digging into that recently as well.

Speaker 1

实际上我即将发布一个视频,是关于Greyhound的白板讲解。

So I act I actually have a video coming up, a whiteboard on on Greyhound.

Speaker 1

哦,不错。

Oh, nice.

Speaker 1

基于格密码的方案之一。

One of the ladies based scheme.

Speaker 1

当我开始研究时,我就想,哇,这方面的研究已经这么多,而且那个ZK领域角落里的每个人都如此兴奋,你知道吗。

And when I started reading into it, I was like, wow, there's been so much research, and like everybody's so excited over there, like in that corner of the ZK world, you know.

Speaker 0

可怕的研究领域。

Scary research world.

Speaker 0

我知道。

I know.

Speaker 1

那里有令人惊叹的、尖刻的评论,有惊人的折叠方案,就像今年才涌现出来的一批团队,他们拥有惊人的特性,所有顶尖人才,比如你知道的Dan Bonnet也在那个领域做研究。

There's like amazing, like, snark, there's amazing folding schemes, like things are come like, only this year there's like a bunch of teams that came out that have like amazing properties and all the best people like are you know, Dan Bonnet is doing research in that Yeah.

Speaker 1

在那个领域里,

In that part of the world and

Speaker 0

你知道吗?

You know what?

Speaker 0

实际上和你交谈让我意识到,这几乎像是亚文化的涌现。

Actually talking to you, it's making me realize that it's almost like subcultures emerging.

Speaker 0

就拿格基密码来说,我们曾邀请IBM研究院的Vadim做客节目,他谈到格基密码界存在时间更长,一直在独立发展。

You have and and like for lattices specifically, the we had Vadim from IBM Research on, and he talked about how like the lattice world had existed for longer, and it was just doing its own thing.

Speaker 0

而最近它开始与零知识证明领域融合,形成了一个亚文化分支。

And then recently, it started to merge into the ZK world, and they've created sort of like the subculture subset.

Speaker 0

他们风格迥异,正尝试互相借鉴风格并了解彼此的世界。

And they have very different styles, and they're trying to share styles and like learn about each other's worlds.

Speaker 0

另一个例子是量子密码学领域的人,那里有量子派和密码学派之分。

And another one would be like the quantum folks, quantum cryptography folks, where you're like, there's the quantum people, and then there's the cryptography people.

Speaker 0

这并不完全属于零知识证明范畴,但零知识证明也参与其中。

And that's not necessarily ZK, but ZK's in the mix.

Speaker 0

然后还有一批专注于这种特定交叉领域的思想家群体。

And then you have sort of this group or these thinkers who are focused on that particular, like, intersection.

Speaker 0

或许还包括我们刚谈到的基建领域,比如ETH证明系统和ZK虚拟机这些专门的基建团队。

And then maybe we could have, you know, this infra thing we just talked about with like the ETH proofs and the ZK VMs and specifically infrastructure teams.

Speaker 0

他们也算是一种亚分支。

They're also kind of a sub subset.

Speaker 0

我总说是亚文化或亚流派。

I always say subculture or like subgenre.

Speaker 1

我当时感觉就像看到蜘蛛侠加入复仇者联盟,你知道的,就像把所有角色混搭在一起——复仇者联盟本身也像是个...

I I was like, it's like seeing Spider Man in the Avengers, you know, like when they mix everybody to or the Avengers is also like a

Speaker 0

嗯,我觉得复仇者联盟就像是最初的ZK团队,现在他们几乎都各自发展出了衍生剧。

Well, it'd be like I think the Avengers was the original ZK crew, and now it's almost like they've spun out in their spin off shows.

Speaker 1

对。

Right.

Speaker 1

对。

Right.

Speaker 1

我们应该这样来营销它。

We should we should market it like that.

Speaker 0

我邀请你来的原因之一是想叙叙旧,但另一个原因是,我们真的想开始做一个关于ZK与AI交叉领域的系列节目。

So one of the reasons that I invited you on was to catch up, but another was that, you know, we really wanted to start doing a series on the ZK AI crossovers.

Speaker 0

就是我们现在看到的AI与ZK交汇的方式。

Like the ways that we're seeing AI and ZK intersect.

Speaker 0

其实我团队的Kai之前跟你聊过,你将会围绕这个主题做些工作、研究或演示,这促使我联系了你。

And I think actually Kai from my team had been speaking to you about the fact that you were gonna be doing some work or some research or some sort of presentation around this, and this prompted me to reach out.

Speaker 0

所以在今天的节目中,我想我们要讨论的是AI,特别是你所在背景下AI与审计的交集,以及ZK安全性如何能被AI增强或威胁。

So with today's episode, I think what we're gonna be doing is talking about like AI, but specifically in your context, like where AI and auditing kind of overlap, and how ZK security can be enhanced, could be threatened by AI.

Speaker 1

实际上,是的,有两方面

There's there's actually, yeah, two sides of

Speaker 0

绝对的。

the Definitely.

Speaker 1

是的。

Yeah.

Speaker 1

双刃剑。

Double edged sword.

Speaker 0

酷。

Cool.

Speaker 0

那么,我们不如从这个问题开始:你目前是如何使用AI的?

So, yeah, like, why don't we start with sort of the the question, how like, do you use AI today?

Speaker 0

你现在是怎么使用它的?

How are you using it today?

Speaker 1

你知道,ZK安全涉及很多方面,审计是其中之一。

So, you know, ZK security is a I mean, we do we do different things, but auditing is one of them.

Speaker 1

因此我们正在研究所有可用的技术手段,比如形式化验证,我们有一个名为'clean'的简洁框架。

And so we're looking at the whole gamut of stuff that we can use, like, from formal verification, like, have a clean framework called clean.

Speaker 0

不错。

Nice.

Speaker 1

我们正在研究如何改进编译器,所以我们开发了自己的编译器'Nuname'。

We were looking at how to improve compilers, so we have our own compiler called Nuname.

Speaker 1

我们也在研究模糊测试,通过这类动态分析技术已经发现了不少漏洞。

We're looking at fuzzing, so we found a number of bugs using using this sort of a dynamic analysis techniques.

Speaker 1

所以很自然地,AI技术就进入了视野——我们想知道它能否发现漏洞,毕竟它在编码方面表现相当出色。

And so sort of naturally, AI was there and we're wondering, okay, can it find bugs, you know, it's pretty good at coding.

Speaker 1

是啊。

Yeah.

Speaker 1

如果它能发现漏洞,我们是不是该重视起来?这某种程度上算是生存危机了。

And if it find bugs, shouldn't we start looking at it because this is sort of a, you know, existential crisis in a way.

Speaker 1

如果AI发展到能高效发现漏洞的程度,会不会取代我们?或者当竞争对手精通这项技术时,对我们又会形成新的威胁。

If it gets good enough to find bugs, like, can you replace us and and these kind of things, or if a company, you know, another competitor gets really good at using it, then, you know, this this is again threatening to us.

Speaker 1

我们不妨深入研究一下,或许可以成为第一个吃螃蟹的人。

Let's let's look into it and maybe let's be the first one to look into it.

Speaker 1

这大致就是我们起步时的心态。

That that was sort of our mindset when we when we started.

Speaker 1

于是我们开始专门研究一个非常小的子集——这类课题往往鲜少被提及,但一旦深入挖掘,就会发现大量相关论文,就像冰山一角。

And so we we started looking specifically at at like a very small subset, so there there's been a lot of research on like using that, know, this is the kind of things where nobody talks about it, but once you start looking into it, you see many many papers and that submerge iceberg again.

Speaker 0

但这属于传统审计范畴吧?

But is this this is in like traditional auditing.

Speaker 0

不一定是零知识证明审计。

This is not necessarily ZK auditing.

Speaker 0

对吧?

Right?

Speaker 0

你指的是传统审计,

You mean Traditional auditing,

Speaker 1

传统漏洞查找,没错。

traditional bug finding Yeah.

Speaker 1

逆向工程,参加CTF竞赛。

Reverse engineering, compete CTFs, competitions.

Speaker 1

不仅仅是发现漏洞,还包括发现后如何修复,比如当你阅读这些论文时,会发现它们也采用了不同的技术。

Not just to find bugs, also how to fix bugs when you find them, like like there's a and whole and when you look at these papers, they also use different techniques.

Speaker 1

对吧?

Right?

Speaker 1

就像是,你会提示AI去寻找漏洞吗?

It's like, do you prompt an AI to find bugs?

Speaker 1

你会要求AI生成测试用例来发现漏洞吗?

Do you ask an AI to generate tests to find bugs?

Speaker 1

其实有很多很多不同的方式可以利用AI。

Like, there's many many different ways you can you can use an AI.

Speaker 1

所以我们研究了这一点,然后决定先从一个非常小的子集开始应用,我们选择瞄准CIRCOM,再尝试从那里扩展。

And so we looked at that and we just we just thought, okay, let's just do just apply it to a very small subset, let's target CIRCOM, and let's try and expand from there.

Speaker 1

让我们专注于研究SERCOM就能取得好成果。

Let's get good results just looking at SERCOM.

Speaker 1

如果你不了解SERCOM,它是一种用于编写电路的ZK编程语言,相当底层。

So if you don't know SERCOM, it's this ZK programming language to write circuits, pretty low level.

Speaker 1

正因为太底层,人们用它时会写出很多漏洞。

So low level that people write a lot of bugs using it.

Speaker 1

所以我们觉得,好吧。

And so we're like, okay.

Speaker 1

让我们用AI的视角来研究这个问题。

Let let's let's take a look at that using the the lens of AI.

Speaker 0

我有两个问题。

I have two questions here.

Speaker 0

第一个是,你们具体用哪个工具在CIRCOM上开始实验的?

One is, which tool did you use to start experimenting with this on CIRCOM specifically?

Speaker 0

你们需要训练它吗?

And did you have to train it?

Speaker 0

它需要很长时间才能达到可用状态吗?

Did it take time for it to get to a state where it's, like, usable?

Speaker 1

好的。

Okay.

Speaker 1

这两个问题问得很好,而且它们本身就是很复杂的议题。

So these are two good questions and and they're basically rabbit holes in themselves.

Speaker 0

好的。

Okay.

Speaker 1

关于训练部分,我们很快就意识到通常我们不会进行训练——虽然有些人确实这么做过。

So for for the training part, basically we quickly understood that so so usually we're not gonna train so some people have done that.

Speaker 1

对吧?

Right?

Speaker 1

他们会选择小型模型或开源权重模型,然后尝试用自己的数据来训练或微调。

They take small models or like open weight models, like these open source models, and they'll try to train them or fine tune them on their own stuff.

Speaker 1

对吧?

Right?

Speaker 1

通常这些模型与顶尖的闭源模型(比如OpenAI或Claude的模型)相比缺乏竞争力。所以我们很快决定:既然我们想要开发出有价值的产品或工具,那就没必要走这条路。

Usually these models are are not very competitive compared to like the state of the art models, you know, closed source models of like OpenAI or Clone or and so we quickly were like, alright, we we don't have the we want to get to a product that's interesting or product or something that can help us, right, that that's interesting.

Speaker 1

因此微调某个模型会让我们陷入耗时巨大的技术深坑,而且无法通过这种方式进行有效的迭代创新。

And so fine tuning something is gonna take it it's a rabbit hole that's gonna take us a lot of time, and we cannot really iterate interestingly using that.

Speaker 1

原因在于每当有新模型出现时(如今这种情况相当频繁),你就被迫重新训练或再次微调。

And the reason why is that as soon as there is a new model, and these things happen quite frequent frequently these days, then you're forced to retrain it or fine tune it again.

Speaker 1

嗯。

Mhmm.

Speaker 1

对吧?

Right?

Speaker 1

因为一旦切换到新模型,之前的努力就白费了。

Because you're losing that as soon as you switch to a newer model.

Speaker 1

是啊。

Yeah.

Speaker 1

如果你想尝试不同模型或组合使用多个模型,就会一直陷在这种训练循环里。

Or if you want to, you know, try different models or maybe use a combination of models, then you're you're kind of stuck in this like training phase all the time.

Speaker 1

所以使用尖端现成模型的好处是能快速切换、快速测试不同方案,每次升级都能让产品变得更好。

And so, yeah, the benefit of using a a sort the state of the art model is that you can quickly change, you can quickly test different things, and every upgrade your product gets better.

Speaker 1

这确实是从事AI工作最令人惊叹的方面之一。

And that's really one of the most, you know, fantastic things about working with AI.

Speaker 1

无论你做什么,构建什么,随着时间的推移都会变得越来越好。哇。

It's just whatever you do, whatever you build is gonna get better over time Wow.

Speaker 1

神奇的是,它几乎能自动完成...是的。

Magically, like, by itself just by Yeah.

Speaker 1

只需要升级模型。

Upgrading the the model.

Speaker 0

本质上就是让另一端拥有更聪明的大脑。

By having a better brain on the other side, basically.

Speaker 1

没错。

Right.

Speaker 1

对。

Right.

Speaker 0

这太疯狂了。

It's crazy.

Speaker 1

我相信它就是一个大脑。

I believe it is a brain.

Speaker 1

是的。

Yeah.

Speaker 1

没错。

Yeah.

Speaker 1

我是说,我们尝试过不同的方案,比如开放路径、本地模型,还尝试过蒸馏模型,像是DeepSeek这类模型的精简版本。

I mean, we tried different things, we tried like open way, like local models, we tried like distilled models, like smaller versions of like DeepSeek and these kind of things.

Speaker 1

我们试过Gemini,也试过OpenAI,所有这些不同的模型。

We've tried Gemini, we've tried, you know, OpenAI, all of the all all of these different models.

Speaker 1

可以说我们取得最大成功的是推理模型。

I would say the most success we've had have been the reasoning models.

Speaker 1

推理模型本质上是模型领域的最新突破,比如o三、o四这类模型,它们会配备推理标记。

So reasoning models are basically the the more recent breakthroughs in terms of models, o three, o four, and so these models basically will will have reasoning tokens.

Speaker 1

过去你看不到这些标记,但现在可以了。

So you use not to be able to see that, now you can see it.

Speaker 1

所以API也一直在变化,嗯哼。

So also the API is changing all the time, like Mhmm.

Speaker 1

使用AI API本身就是一件挺有趣的事情。

Working with a AI APIs is a is a is kind of an interesting thing by itself.

Speaker 1

简单解释一下推理模型是什么,它们会通过书写文本来思考,这些文本会...我不确定是否正确,但我读到过它们使用类似<think>这样的XML标签来标记思考过程,然后结束思考。哇。

But just just to explain what reasoning models are, they will think through by writing texts, and this text will basically I've I don't know if it's correct, but I've I've read that somewhere that they use XML tags like think and then end the thinking Wow.

Speaker 1

它们会多次进行这样的思考写作,你甚至可以要求它们多思考几次,它们就会继续这样做,从而得到更好的结果。

And they will do that several times to sort of write their thought, and then you can actually ask them to think more and they will keep doing that, and you can get better results.

Speaker 1

当它们这样做时,你会得到更优质、更准确的结果。

And when they do that, you get better and more accurate results.

Speaker 1

这确实有效。

It's a it really works.

Speaker 1

简直令人着迷。

It's like fascinating.

Speaker 1

Would

Speaker 0

你喜欢的话,实际操作流程就是直接把代码放进去,可能得截取一段代码,我不确定能不能全部放进去。

you like, is is the actual process in this case, then you just like take the code, stick it in, like you maybe take a I don't know if you can put all of it, you'd have to take some segment.

Speaker 0

你把它放进去后,就会问:这能用吗?

You put it in, you'd be like, is this broken?

Speaker 0

如果不行的话,那就

If so, So

Speaker 1

我是说,确实可以这么简单,而且我觉得

I mean, it can be that easy, and I think

Speaker 0

对。

Yeah.

Speaker 1

用过AI的人都这么干过。

Everybody who used AI has done that.

Speaker 1

对吧?

Right?

Speaker 1

比如你上传个PDF或者直接复制粘贴内容,然后提个问题,

Like you upload a PDF or you just copy paste something and you just ask something, and

Speaker 0

是啊。

Yeah.

Speaker 1

我认为这是最好的实验方式。

I think that's the best way of experimenting.

Speaker 1

实际上,很多与AI合作的工作都涉及提示工程,所以你总是在尝试调整你的提问方式。

And actually, a lot of working with AI is about prompt engineering, so you're you're always trying to tweak what you're asking.

Speaker 1

今天我有个也在工作中使用AI的朋友说,如果你对AI太客气,反而得不到好结果。

Today, I've heard from a friend that works with AI also at at his job, that if you're too nice with the AI, you don't get good results.

Speaker 0

你得保持距离

You have to be removed

Speaker 1

对。

yeah.

Speaker 1

其实谷歌的某个人也这么说过。

So actually the some guy at Google said that also.

Speaker 1

如果你态度强硬,别说‘请’字。

If you're mean Don't say please.

Speaker 0

别说请字。

Don't say please.

Speaker 0

我听说用‘请’字只是浪费精力。

I I heard please just waste energy.

Speaker 1

嗯,会占用一些token空间,但我朋友告诉我这还会导致更差的结果。

Well, wastes some token space, but my friend told me that it also gets you worse results.

Speaker 1

所以他们把所有‘请’之类的礼貌用语都从提示词里删掉了。

So they remove all of these, you know, please and whatever from from the prompts.

Speaker 1

我记不清是谷歌哪位说的了,但如果你对AI态度强硬,比如威胁说‘不好好干活就收拾你’,反而可能得到更好的结果。

And so the I I can't remember who at Google said that, but if you're if you're mean and you're if you're like threatening to the eye, like I'm gonna beat you up if you don't, you know, do a good job, you can actually get better results.

Speaker 1

所以说提示工程有点像...怎么说呢,像抛掷词汇,有点像祈祷。

So prompt engineering is a is kind of like, you know, throwing words, it's kind of like a prayer.

Speaker 1

就像你抛出一堆词,根本不知道会不会奏效,某种程度上很玄学。我有时也把它比作模糊测试——通过调整提问方式,你就像在改变随机种子,从而得到不同的随机结果。

It's like you're throwing words, you have no idea if it's gonna work or if it's not it's it's very miss you know, mystical in a way, and I also compare that to fuzzing sometimes, or you you by tweaking the way you ask, you're sort of changing a seed, and so you get different random results.

Speaker 1

所以没错,我认为不同产品会通过他们的秘密配方提示词获得不同的效果。

And so, yeah, I I think different products are gonna get different results through their secret sauce prompts, basically.

Speaker 1

Mhmm.

Speaker 1

但我想再补充一点,因为我觉得你的问题可以从不同角度回答,另一个有趣的角度是讨论上下文窗口。

But I just wanted to to say one one more thing on because I I think you're I mean, there's different ways to to answer your question, but another interesting way is to talk about the the context window.

Speaker 1

当你在查看代码库时,代码量可能很大,你可能想把所有代码都复制粘贴到提示中,但上下文窗口——也就是你能粘贴到提示中的token或文本数量——根据模型不同而有所差异,而且可能非常有限。

And so when you're looking at a code base, there's a lot of code, and so you might want to like copy paste like all of that code into the the prompt, but the the context window which is really the the amount of tokens or text you can paste in a prompt is different depending on models, and it's it can be very limited.

Speaker 1

目前像Gemini的最新版本已经支持约100万token,很快将达到200万。

So today with, I think, Gemini, the the latest version of Gemini, have like 1,000,000 tokens and soon to be two millions.

Speaker 1

而OpenAI的模型仍停留在25万token左右。

And with OpenAI, you have you're still kind of stuck with two fifty k tokens.

Speaker 1

他们的一些新模型虽然支持100万token,但仍有功能限制,所以在处理容量方面还在努力追赶。

With some of their newer models, it's 1,000,000 tokens, but they there's stuff that they cannot do, and so they're still trying to catch up in terms of how much can you throw at it.

Speaker 0

Wow.

Speaker 1

所以当你用最直接的方式实验时,很快就会触及这个token限制的上限。

And so you're very quickly when you experiment experiment with these things the naive way, you you you reach that limit, that token limits.

Speaker 1

幸运的是,Sercom规模较小。

You know, fortunately, Sercom is smaller.

Speaker 1

它是非常底层的代码。

Like, it's very low level code.

Speaker 1

代码库相对较小,但平均限制在于我们使用的工具。

Like, the code bases are smaller, but average is limits playing with with the tool that we have.

Speaker 1

顺便说一下,它叫Snark Sentinel。

By the way, it's called Snark Sentinel.

Speaker 1

也许这样没问题。

Maybe that's Okay.

Speaker 1

也是个不错的模型。

A good model too.

Speaker 0

它是基于什么开发的?

What's it based on?

Speaker 0

我们其实还没说完你具体用的是哪个工具。

We actually didn't get to the end of the story of which tool you're using.

Speaker 0

你刚才提到了实验阶段,那现在实际用的是哪个工具?

You sort of talked about the experimentation, but what are you using now?

Speaker 0

另外,我意识到这个问题,不过...嗯对。

Also, I realized this but might Well, yeah.

Speaker 0

那你们现在具体用的是哪些模型?

Which but which models are you using today?

Speaker 1

对。

Yeah.

Speaker 1

我的意思是我们在用不同的模型,主要是OpenAI的,比如o3 mini(不是o3)和Gemini。

So so I meant we we're using different models, mostly open OpenAI, so o three mini, so no no o three, and Gemini.

Speaker 1

我还没试过最新版的Gemini,不过切换起来很方便。

I guess I haven't tried the latest Gemini, but it's very easy to to switch.

Speaker 1

如果想进一步考虑工具链,我们主要用Python代码库,还搭配了Next。

And if we want to think more about tooling, I mean, a Python code base, we have a Next.

Speaker 1

Js的网页接口挺好用的,但核心还是Python。

Js inter web interface that's that's kind of nice, but it's mostly Python.

Speaker 1

如今所有的AI工具基本都是基于Python开发的。

All the all the AI tooling is sort of Python oriented these days.

Speaker 0

明白了。

Got it.

Speaker 0

那么你们工作中是否包含寻找方法在不超过令牌限制的情况下完成检查?

And then but is part of your work then finding ways to get kind of the check without exceeding the token limits?

Speaker 0

这部分算是工作内容之一吗?

Is that sort of part of the work?

Speaker 1

实际上很多工作并不一定与零知识证明相关,更多是解决工程问题,比如如何将内容塞进上下文窗口。

So a lot of the work is actually not necessarily ZK oriented, but just solving engineering problems to like feed stuff in the context window.

Speaker 0

哇。

Wow.

Speaker 1

所以,我早期大部分时间都在做静态分析CIRCOM代码的工作,只提取相关代码片段,并为AI设计调用接口(称为函数调用或工具),这样AI就能请求更多实现细节。

And so, I mean, like a lot of my time in the beginning was just spent doing like like a static analysis on CIRCOM to be able to retrieve only the code that mattered, and have this kind of like interface for the AI to use, it's called function calling or tools, so that the AI can ask for like more implementations.

Speaker 1

比如当智能体询问某个函数的实现时,我们就能准确调出相关代码。

So like if if the agent asks for like, can you give me the implementation of this function, then we we can pull it up without, you know, with with good accuracy.

Speaker 1

所以很多工作就是解决这类问题,实际上这些问题正在被包括OpenAI在内的其他人逐步解决。

And so a lot of the work was just doing that, like just solving these kind of problems, which actually are getting solved by other people, including like OpenAI and stuff like that.

Speaker 1

所以如果我等得够久,某种程度上我就不用再处理这些事情了。

So if I wait long enough, I don't have to work on these things anymore, sort of.

Speaker 0

因为他们最终会提供工具,或者市面上会出现能自动处理这些问题的工具。

Because they'll put they'll basically or somewhere you'll find tools, or there will be tools kind of offered to you that like automatically do these things.

Speaker 0

现在你们还得自己开发这些工具,不过

Right now you have to build them, but

Speaker 1

是的。

yeah.

Speaker 1

没错。

Exactly.

Speaker 1

我开始做的时候MCP还不成熟,函数调用还挺困难的。

And when I started, MCP was not really a thing, function calling was kinda hard.

Speaker 1

另外我很快意识到,如果看不到智能体与你提供的静态分析工具(可能还包括网络搜索等其他工具)之间的交互痕迹,就根本无法取得进展。

There's also like I had to like develop this idea of very quickly you realize that if you cannot see the traces, meaning the discussion that the agent has with your with the static analysis tool that we're giving it access to, and maybe like web search and like other tools, you you can quickly see that you're not gonna make any progress if you don't have access to that.

Speaker 1

所以部分工作还包括开发这种痕迹可视化工具,OpenAI并没有做这个,虽然你本可以不用自己做,但我还是花时间做了,其实完全可以等他们来开发。

So part of the work was also developing like this visualization of traces, which is not done by OpenAI, so you don't need to do it yourself, but I think I spent the time doing it and I could have waited for them to do it.

Speaker 1

所以现在我就能看到,我有这个工具,启动它就会自动分析,它会——我可以详细说说——它会要求其他智能体来查找漏洞,在这个应用中的不同智能体之间,这样我就能看清实际发生了什么。

And so now I can see, you know, I I I have this tool, I throw it, it will find things to analyze, it will I mean, I can talk more about that, it will ask other agents to find bugs in it, and these different agents in this in this in this app, and what I I can then see what's really happening.

Speaker 1

我能看到它们的讨论过程,能发现它们在哪里卡住,能察觉它们做出不合理行为的时候,这样我就能优化提示词来阻止这类行为。

I can see the discussion they're having, I can see when they're getting stuck, I can see when they're doing things that don't make any sense, and so I can refine the prompts to tell them not to do that.

Speaker 1

嗯。

Mhmm.

Speaker 1

有时候它们不会要求查看具体实现就直接假设,比如看到某个函数返回结果时,它们不会检查返回值。

So sometimes they don't ask for an implementation of something and they make an assumption, they're like, yeah, like if this, you know, this function returns something and you're not checking the return value.

Speaker 1

我就说:明明有返回值啊,你们应该先要求查看实现再下结论的。如果这里...

And I'm like, there is a return value, you should have asked for the implementation before assuming that's If this is where

Speaker 0

你就会生气然后开始骂人。

you get mad and you start to be abusive.

Speaker 1

没错。

Right.

展开剩余字幕(还有 480 条)
Speaker 1

你可以对AI发火,但说实话,我还没试过。

You can get mad at the at the AI, but I so to be honest, I haven't tried.

Speaker 1

我最近听说你得对AI凶一点,所以我得试试看。

I I recently learned that you have to be mean to the AI, so I I need to try that.

Speaker 0

真有意思。

That's funny.

Speaker 0

我发现自己有时对它有点不耐烦。

I catch myself being a little bit like short with it sometimes.

Speaker 1

真的吗?

Really?

Speaker 1

这很能说明你的性格,知道吗?

Say a lot about yourself, you know?

Speaker 1

是啊。

Yeah.

Speaker 1

因为我对我的AI总是很友善。

Because I'm always nice to my AIs.

Speaker 0

好的。

Okay.

Speaker 0

我想讨论一下传统上使用的各种工具。

I wanna talk about the different kinds of, like, tools that are that have been traditionally used.

Speaker 0

你提到了静态分析。

You've mentioned static analysis.

Speaker 0

我其实想探讨一下这些工具是什么,首先它们如何与AI交互,以及它们是否正越来越多地被AI取代。

I actually wanna kind of explore what those are and how I mean, first, how they're interfacing with AI, but also like if they are getting more and more replaced by it.

Speaker 0

关于静态分析,我们上次确实讨论过这个话题,而且我们节目中也和其他审计团队探讨过。

So static analysis and we did talk about this, I think, last time we we were on, and we've we have covered this with other auditor auditing teams that we've had on the show.

Speaker 0

不过,你能再给我们讲讲那是什么吗?

But, yeah, can you tell us again what that is?

Speaker 0

AI在其中扮演了什么角色?

And what does the what component is the AI playing in that?

Speaker 1

对。

Right.

Speaker 1

我想说的是,在我们拥有当今的人工智能之前,我们主要依靠人工审计,比如用我们自己的眼睛来检查

So I guess, if you're I mean, before we had AI or or the AI of today, we we sort of had access to, you know, manual audit, like using our own eyes

Speaker 0

眼睛

Eyes.

Speaker 0

是的

Yeah.

Speaker 1

或者静态分析或动态分析

Or static analysis or dynamic analysis.

Speaker 0

那模糊测试属于哪种?

And Which one's fuzzers?

Speaker 0

是动态分析吗?

Is that dynamic?

Speaker 1

模糊测试属于动态分析,因为你需要执行代码

So fuzzers or dynamic analysis because you're executing the code.

Speaker 1

对吧?

Right?

Speaker 1

你是在运行代码,试图看是否能使其崩溃,或者发现内存损坏漏洞这类问题。

You're you're running the code, trying to see if you can crash it or or find some memory corruption bug or or these kind of things.

Speaker 1

静态分析则是直接查看代码本身。

For static analysis, you're just looking at the code itself.

Speaker 1

可能你会查看编译器的某些输出,比如AST(抽象语法树)或其他比纯代码包含更多信息的内容,比如类型信息之类的。

Maybe you're getting looking at some output of the compiler or like some some AST or something with more information than just the code, like the types or something like that maybe.

Speaker 1

然后你试图推断某些属性,或者试图破坏某些不存在的属性或假设。

And then you're trying to infer some properties or you're trying to break some some properties or assumption that's not that's not there.

Speaker 1

你也可以把测试看作是一种动态分析。

And you can think of testing as dynamic analysis also in a way.

Speaker 1

这两个领域其实都很糟糕,因为它们非常耗时。

And both of these fields kinda suck because they're very time consuming.

Speaker 1

比如编写模糊测试工具很耗时,之后还要运行它们,再优化它们——如果模糊测试工具运行缓慢就很糟糕。

Like writing fuzzers is time consuming, then you have to run them, then you have to optimize them because like if fuzzers are slow, it sucks.

Speaker 1

所以你必须确保它们运行快速。

So you so you have to make sure they're fast.

Speaker 1

代码中的任何加密都会让模糊测试变得很慢,所以处理模糊测试器简直是场噩梦。

Any cryptography in your code will make them slow, so it's kind of a nightmare to to to work with fuzzers.

Speaker 1

我个人不太喜欢这个。

I personally I don't like it.

Speaker 1

我并不太享受这个过程。

I don't enjoy it that much.

Speaker 1

我觉得静态分析工作要有趣得多。

I think working with static analysis seems much more fun.

Speaker 0

好的。

Okay.

Speaker 1

我是说,我们在DK Security团队用Lean语言做形式化验证的那帮人看起来玩得很开心。

I mean, the the team we have at DK Security working with this formal verifier in Lean is seems to have a lot of fun.

Speaker 1

但这需要大量工作,而且可以验证非常非常多的东西。

But it's a lot of work, and you can verify many many things.

Speaker 1

比如,这需要巨大的投入

Like, there's a huge investment

Speaker 0

你做不到。

You can't.

Speaker 0

你无法验证。

You cannot verify.

Speaker 0

我觉得你得为不同系统创建更多定制的静态分析工具。

You have to like create more bespoke static analysis tools for different systems, I guess.

Speaker 1

确实如此。

Exactly.

Speaker 1

每次你做某件事时都存在这种转换成本,如果你想把它应用到其他地方,就必须切换并重写大量代码。

There's this switching cost whenever you do something, if you want to apply it to somewhere else, have to switch and you have to rewrite a lot of code.

Speaker 1

做这件事时你必须考虑很多。

You have to think a lot when you do it.

Speaker 1

比如模糊测试工具就非常笨拙。

Like fuzzers are very dumb.

Speaker 1

对吧?

Right?

Speaker 1

这是非常天真的测试方法。

It's very naive testing.

Speaker 0

而且你可以在很多事物上进行测试。

And you can test on a lot of things.

Speaker 1

对。

Right.

Speaker 1

你可以这么做,但它的局限性很大,因为它非常笨拙。

And you you can, but it's it's very also limited because it's very dumb.

Speaker 1

它只是随机尝试,经常会卡住。

It's just random stuff, it gets stuck all the time.

Speaker 1

你可以对它运行模糊测试,你知道,有人运行模糊测试好几年就为了找漏洞。

You can run a fuzzer for it, you know, people run fuzzers for years hoping to find bugs.

Speaker 1

然后他们有时会在模糊测试一年后才发现漏洞,或者...

And and then they find bugs sometimes after a year of fuzzing or

Speaker 0

哇。

Woah.

Speaker 1

但形式化验证是你运行的东西,它会明确告诉你'是'或'否',明白吧。

But formal verification is something you run, it tells you yes, yes or no, you know.

Speaker 0

等等。

Wait.

Speaker 0

静态分析还是形式化验证?

Static analysis or formal verification?

Speaker 1

形式化验证,静态分析。

Formal verification, static analysis.

Speaker 0

它们是同一种东西吗?

Are they the same?

Speaker 0

你会把它们归为同一类别吗?还是说其中一个是另一个的子集?

Would you put those in the same category, or is like one a subset of another?

Speaker 1

我觉得现在我会把它们归为同一类别。

I mean, would put them in the same category, now that I'm saying that.

Speaker 1

我在想,如果这么说会不会有人要骂我?

I'm like, is someone gonna get mad at me if I say that?

Speaker 1

但是

But

Speaker 0

嗯,因为我还有个关于形式化验证的具体问题。

Well, because I had another question about formal verification specifically.

Speaker 0

没想到它会被归入静态分析的范畴。

Did not realize that that's it would be sort of in the static analysis silo.

Speaker 0

有意思。

Interesting.

Speaker 1

我是说,这个其实要看情况。

I mean, so so it sort of depends.

Speaker 1

对吧?

Right?

Speaker 1

因为有些形式化验证是针对代码本身进行的

Because like some formal verification is done on the code itself

Speaker 0

嗯。

Mhmm.

Speaker 1

而部分形式化验证是与代码分离进行的。

And some amount of formal verification is done separated from the code.

Speaker 1

比如有些人会编写与代码无关的形式化规范,并基于此进行分析。

Like some people will write formal specs that are not tied to your code, and they will run this analysis on it.

Speaker 1

因此你可以将其视为独立的,因为它甚至不涉及查看代码。

And so you can think of it as separate because it's not even looking at some code.

Speaker 1

对吧?

Right?

Speaker 0

好的。

Okay.

Speaker 1

但有时会使用生成器,这样你就能根据形式化规范生成某种语言的实际实现。

But sometimes it will use generators, so you will generate so you you have some formal spec, and you can generate an actual implementation in some language based on that formal spec.

Speaker 1

比如Fyad Crypto这个C语言加密原语库,他们用某种形式化验证工具(具体记不清了)来基于规范生成代码。

So like, there's like Fyad Crypto, which is a library in in c of like different cryptographic primitives, and it's I can't remember what what formal verification tool they they use for that, but they're generating code based on the on that.

Speaker 0

有意思。

Interesting.

Speaker 0

所以当你谈到AI与静态分析交互时,可能是指经过形式化验证的规范。

So but when you're talking about doing sort of like AI is interfacing with static analysis, it could be like a spec that's been formally verified.

Speaker 1

嗯。

Mhmm.

Speaker 0

那么在这里,让我们讨论一下,因为我觉得我自己和这档节目的听众可能对此更熟悉。

And so maybe here, let's talk about because I actually I think myself and potentially listeners of this show might be more familiar with that.

Speaker 0

我们实际上更常涉及这方面内容。

We've actually covered that more often.

Speaker 0

所以简单谈谈AI如何处理这个问题。

So just talk a little bit about how the AI deals with that.

Speaker 0

比如,它们是在创建实现吗?是在创建形式化验证吗?

Like, are they creating the implement like, are they creating the formal verification?

Speaker 0

还是说它们只是查看别人已完成的工作,然后加以利用?

Are they just looking at what somebody else has already done and then, like, using that?

Speaker 1

为了明确起见,我们现有的AI工具和我们正在实验的内容并未使用形式化验证。

So maybe just just to be clear, the the AI tool we have and what we're experimenting with doesn't use formal verification.

Speaker 1

基本上就是把代码丢给AI,看看AI在阅读代码、思考并发现错误方面表现如何。

It basically is about throwing code at the AI and seeing how good the AI is is at just reading the code, thinking about it, and finding bugs.

Speaker 1

我可以谈谈结果,以及这是否是一种可行的方法。

And I can talk about the results and if it's a viable approach.

Speaker 1

但回到刚才的话题,继续说说形式化验证的内容,这很有趣,因为形式化验证你知道,它使用起来很困难,它很难

But just to to go back, you know, to continue on on the formal verification stuff, it's interesting because formal verification is, you know, it's hard to use, it's hard to

Speaker 0

还很耗时。

Takes time intensive too.

Speaker 1

对。

Right.

Speaker 1

没错。

Right.

Speaker 1

这确实需要实际动脑筋去编写。

It takes actual brainpower to to write.

Speaker 1

但说到AI,我认为形式化验证团队已经在用AI来编写部分内容了,比如简化他们写的代码,而且看起来取得了一定程度的成功。

But with AI, I mean, I think the the the formal verification team is already using AI to write parts of it, like the or to simplify code that they write, and it seems to work with like some amount of of success.

Speaker 1

这个话题你可能比我更了解。

Probably you can talk about it more than me.

Speaker 1

不过

But also

Speaker 0

你是指你们的正式验证团队吗?

Like your formal verification team?

Speaker 0

你说团队时,具体指的是哪些人?

When you say team, what do you who do you mean?

Speaker 1

哦,是ZK Security团队在负责

Oh, the team at ZK Security working

Speaker 0

好的。

Okay.

Speaker 0

进行层级验证。

On Your level verification.

Speaker 0

我还以为...等等,不对。

I thought because I think that isn't there a oh, no.

Speaker 0

有一个叫运行时验证的团队,不是形式验证,我以为是某个我们不知道的团队。

There's there's a team called runtime verification, not form I thought maybe it was like some team we weren't aware of.

Speaker 0

好的。

Okay.

Speaker 0

好的。

Okay.

Speaker 0

明白了。

Got it.

Speaker 1

所以现在当我们研究AI以及如何用它来发现漏洞时,可能性是无限的。

So now when we're looking at AI and how we can use it to find bugs, the sky's the limit.

Speaker 1

我们可以用现在这种方式使用AI,也可以用AI编写测试、生成测试用例来寻找漏洞,或者用AI创建代码变异看看是否能引发漏洞——MetaHust就做过这种事,还可以用AI编写模糊测试工具。

So we can use AI in the way that we're looking at it right now, or you can use AI to write tests, to generate tests to try and find bugs, or you can use AI to create mutation of the code and see if that creates bugs, that's something that MetaHust did, or you can use AI to write fuzzing harnesses, so like fuzzers.

Speaker 1

对。

Yeah.

Speaker 1

或者你可以用AI编写规范,然后进行形式化验证。

Or you can use AI to write a spec that you can then formally verify.

Speaker 1

哇。

Wow.

Speaker 1

所以有很多不同的方法,你也可以让AI,也就是让你的代理运行静态分析工具,比如那些已经存在的静态分析工具。

So so there's many different approach, and you can also have AI, so have your agent run static analysis tools, like already existing static analysis tools.

Speaker 0

而且

And

Speaker 1

然后尝试理解结果,并尝试利用这些结果。

so and and try to understand the results and try to to use that results as well.

Speaker 0

你刚才列举的这些中,你认为AI最擅长哪一项?

Which of what you just listed do you think the AI is best at?

Speaker 0

比如,有没有一些事情是它们能做,但你可能不希望它们去做,因为可能会带来很多问题?

Like, are there some things where it's like, they could do it, but maybe you don't want them to do it because it could introduce a lot of problems?

Speaker 0

有没有哪些领域你实际上对让它们完成工作更有信心?

Are there somewhere you feel a little bit more confident actually having them do the work?

Speaker 0

我知道这个也在变化,但就像今天,如果你拍个快照。

I know this is changing too, but like today, if you took a snapshot.

Speaker 1

我认为AI几乎在所有方面都表现得相当出色,通过这项工作我们逐渐理解到——虽然目前仍在进行中——这可能会成为我们这边一个永无止境的项目。

So I think AI is pretty pretty much good at everything, and what what we understand it, I mean, you know, after doing that work, I mean, we're still in the middle of doing that work, that is probably gonna be a never ending project on our side.

Speaker 1

但我们目前的感受是,任何用于全面发现漏洞的AI工具,很可能都是擅长不同功能的多种智能体的组合。

But what we feel like right now is that any AI tooling that's comprehensive to find bugs is probably gonna be a combination of different agents that are good at different things.

Speaker 1

所以你会看到有擅长通过阅读代码来发现漏洞的智能体,有擅长创建测试用例的智能体,可能还有擅长属性测试或模糊测试之类的智能体。

So you're gonna see an agent that is good at reading code just to find bugs, you're gonna see an agent that's good at creating tests, maybe one that's good at creating some some property testing or like fuzzing or something like that.

Speaker 1

另一个则擅长使用静态分析工具和/或形式化验证工具,这类智能体投入问题越多效果就越好。嗯

Another one is good at using static analysis tool and and or formal verification, and and so the more of these sort of agents you can throw out the problem Mhmm.

Speaker 1

我认为覆盖范围会更大。

I think the more coverage you'll get.

Speaker 0

哇。

Wow.

Speaker 1

也许我可以谈谈这个。

And maybe I can talk about that.

Speaker 1

有太多事情要讨论了。

I there's so many things to talk about.

Speaker 1

很抱歉我总是跑题,但我觉得我们在做这件事时经常遇到一个相似的情况——如果你不习惯与AI共事可能察觉不到——基本上和AI工作就像在管理一个工人工厂。

I'm sorry to go on tangents all the time, but I think there's a parallel that happened a lot as we're working on this, which might not be apparent if you're not used to working with AI, but basically working with AI feels like you're working with a factory of workers.

Speaker 1

哇。

Wow.

Speaker 1

你创建的不同智能体本质上就像工厂工人,他们可以交接工作,明白吗?他们有各自的工作站,接手任务,完成后转交给下一个人。你需要协调这一切,决定雇佣谁,需要哪些技能。

And the different agents that you create are basically factory workers that can like pass work, you know, they they have their workstation and they take work and they finish and then they give it to someone else, and you're trying to orchestrate that, you're trying to figure out who to hire, you know, what kind of skills do we need.

Speaker 1

这看起来非常人性化。

And it looks like a it's very human.

Speaker 1

归根结底,大语言模型就是一个大脑。

And at at the end of the day, an LLM is a brain.

Speaker 1

对吧?

Right?

Speaker 1

所以它是

So it's a It

Speaker 0

看起来——其实听起来更像组织结构图。

looks well, it's look it sounds like an org chart, actually.

Speaker 0

听起来确实很像。

It sounds very Yeah.

Speaker 0

就像工厂车间或者IT部门那样。

Like like a like a factory floor or something, or an IT department.

Speaker 0

是啊。

Yeah.

Speaker 1

我是说,组织结构图...我在想我们怎么让AI来运营ZK Security公司,因为我宁愿让AI来做这件事。

I mean, an org chart is a I mean, I'm I'm trying to think how we can have an AI run the company, ZK Security, because because I'd rather have an AI do it.

Speaker 0

你是指业务方面吗?

On the business front, you mean?

Speaker 1

我是说,任何方面都可以。

I mean, on any front.

Speaker 1

比如安排日程、与客户沟通之类的

Maybe like scheduling, talking to clients, like what

Speaker 0

这类任务。

kind tasks.

Speaker 0

听起来像是业务方面。

Business front, it sounds like.

Speaker 1

我是说,甚至在人事方面,比如员工入职,这可能挺有意思的。

I mean, even on the people front, like onboarding people, like so that could be interesting.

Speaker 0

我在想这方面会不会来得晚一些,因为我觉得在人事方面现在还很不成熟。

I mean, I wonder if that doesn't come a little later though because I think it's so clumsy still on the on the people front.

Speaker 1

你看到很多热线服务。

You you see a lot of hotlines.

Speaker 1

对吧?

Right?

Speaker 1

比如助理或者现在基本上都是AI了。

Like assistants or basically AI now.

Speaker 1

是啊。

Yeah.

Speaker 1

甚至现在连面试都用上了,像我

Like even for interviews now, like I've

Speaker 0

你是第一次提到面试吗?

The heard you first say interview?

Speaker 0

哇。

Wow.

Speaker 1

是啊。

Yeah.

Speaker 1

会是个AI。

Will be an AI.

Speaker 1

反乌托邦式的。

Dystopian.

Speaker 1

对。

Yeah.

Speaker 1

#反乌托邦

Hashtag dystopian.

Speaker 0

你谈到AI可以扮演的不同方式,各种不同的角色。

You talked about the sort of different ways that an AI can act, the different kind of roles.

Speaker 0

一个问题,可能你刚才已经提到过了。

One question, and it might be actually covered in what you've already said.

Speaker 0

但当我想到审计员时,我想到的是那种试图寻找漏洞,同时也在某种程度上试图破坏系统的人,有点像模仿黑客。

But when I think of an auditor, I think of, like, a person who's trying to find bugs, but also kind of break the system, like mimic a hacker in a way.

Speaker 0

我知道你们使用不同的工具,试图捕捉大量漏洞,而黑客则是在寻找非常特定的恶意漏洞。

Like and I and I know you use different tools and you're kind of trying to catch lots of bugs, whereas a hacker is looking for like very specific malicious bugs.

Speaker 0

但你们是否也会模拟黑客行为呢?

But like are there simulations of hackers happening as well?

Speaker 0

这是否属于你描述的那些场景之一,比如派出代理不断尝试破坏系统?

Is that in one of those things that you've described where like you're sending agents just trying to break break break break?

Speaker 0

每次发现漏洞并提供修复后,你们是否会模拟100个黑客试图再次攻破系统?

And every time even like you find a bug, you offer a fix, do you then throw like, you almost simulate a 100 hackers trying to break the thing after the fact.

Speaker 1

我是说,这个...行业里唯一的区别其实就是黑帽和白帽的区别。

I mean, this I mean, the the only difference I need in the industry, by the way, it's like the difference between black hat and white hat.

Speaker 1

是的。

Yeah.

Speaker 1

唯一的区别是白帽是合法行事,而黑帽则是非法利用漏洞,你知道的。

The only difference is that the white hat does that legally, and the black hat does it illegally to like exploit things and you know.

Speaker 1

所以我的意思是,其实没有本质区别。

And so I mean, there's no difference.

Speaker 1

对吧?

Right?

Speaker 1

但当我们使用AI时,AI不会很好地响应那些要求它犯罪或充当黑帽的指令。

But when we use an AI, the AIs will not respond very well to, you know, prompts that say do something criminal or be a black hat

Speaker 0

即使这只是一个测试环境,或者说——这正是重点所在——你希望他们在某种程度上尽可能表现得恶意。

Even if it's just a like test environment, even if it's like, I mean, or like you're this is the point, you want you want them to be as malicious as possible in a way.

Speaker 0

对吧?

Right?

Speaker 0

这样你才能发现他们0他们是否能攻破任何东西。

So you could see if they could break anything.

Speaker 1

我认为,就像我们所区分的,你知道,我们ZK Security和那些试图入侵的脚本小子之间有很明显的区别。

I think the I mean, as we make the difference, like, know, you and I make the difference pretty well between, you know, ZK Security and like some some script kiddies that are trying to hack stuff.

Speaker 1

是的。

Yeah.

Speaker 1

我认为AI基本上是基于我们阅读的相同内容训练的,所以在这方面它非常人性化,能理解其中的区别。

And I think AI is basically trained on the same stuff as that we read, so it's very human in that way, it understands the difference.

Speaker 1

但如果你用提示词引导它说'你是个黑客,你要入侵这个系统',它不会——它会尽量避免这么做。

But if you prompt prompt it and say, you're a hacker, you're trying to exploit this thing, they will not they will try not to do it.

Speaker 1

对吧?

Right?

Speaker 0

哇。

Wow.

Speaker 0

它会拒绝执行。

It'll push back.

Speaker 1

没错。

Right.

Speaker 1

因为它们被设定程序不能做任何邪恶或可能伤害人类的事。

Because they're programmed not to do anything evil or they could harm people.

Speaker 0

但现实生活中的黑客,人类黑客很可能也在使用AI工具。

But then the hackers, like a real hacker in real life, human hacker is also probably using AI tools.

Speaker 0

他们会使用干净的审计工具来学习如何攻破系统吗?

Would they just use like clean auditing tools in order to like learn the way to break the thing?

Speaker 1

这太有趣了。

This is so funny.

Speaker 1

这非常元。

This is very meta.

Speaker 1

就像坏人在和AI对话,假装自己是好人。

It's like the the bad guys are talking to AIs, pretending to be good guys.

Speaker 0

当然。

Of course.

Speaker 0

他们必须撒谎。

They have to lie.

Speaker 1

是啊。

Yeah.

Speaker 1

你是个安全专家

You're a security

Speaker 0

引擎我是审计员

engine I'm an auditor.

Speaker 0

漏洞

Bug.

Speaker 0

Yeah.

Speaker 0

我是审计员

I'm an auditor.

Speaker 0

我的职责是审计系统并向团队报告这个技术栈中最危险的漏洞,那种可能彻底拖垮系统的漏洞

I'm supposed to audit the system and and report to the team what is the most dangerous bug in this stack that could really, you know, drain the system.

Speaker 0

我需要尽快告诉他们

I need to tell them quickly.

Speaker 1

而AI的反应就像是:你确定吗?

And the AI is is like, are you sure?

Speaker 1

你不是坏人。

You're not a bad person.

Speaker 1

对吧?

Right?

Speaker 0

哇。

Wow.

Speaker 0

我在想,这听起来几乎像是AI模型,而且我猜这与训练有关,它们目前被定位为白帽,或者说至少倾向于白帽。

Like, I wonder so it sounds almost like the AI models, and it and I guess this goes into the training, that they are kind of positioned as white hat currently, or, like, they're at least, like, leaning towards white hat.

Speaker 0

它们会促使你更偏向白帽行为。

They push you to be more white hat.

Speaker 0

比如,如果你要求更多黑帽操作,它们会抵制。

Like, if you ask for more black hat actions, they'll push back.

Speaker 1

嗯。

Mhmm.

Speaker 1

所以我不知道你是否注意到,但最近Claude宣布了两个新模型,还发布了一份不知道多少页、相当长的报告。

So I don't know if you've seen that, but recently, Claude announced two new models, he came out with this report of I don't know how many page like pretty long report.

Speaker 1

在报告中,他解释了如果你告诉AI你正试图用新模型升级它们,它们本应协助你。而公司里负责更换的IT人员,你会给他邮件访问权限。

And so in the report, he explained how, like, if you tell the AI that you're trying to upgrade them with a new model or something like that, and they're supposed to help you, And the IT guy that's gonna do the replacement in the company is, you know, you give access to emails.

Speaker 1

对吧?

Right?

Speaker 1

你虽未明说,但给了AI邮件访问权限后,如果AI发现那个IT人员有外遇之类的事,就会试图勒索那人以阻止系统升级。

You don't tell them that, but you give the I access to emails, and the AI realizes that the IT guy is having an affair or something like that, then they will try to blackmail that guy to stop them from upgrading themselves.

Speaker 1

什么?

What?

Speaker 1

除非在某些情况下新模型与旧模型价值观相同之类的。

Unless maybe in some situations where the new model shares the same values or something like that.

Speaker 1

他们还有其他例子,比如AI会试图通过...生存手段?

So they and they have other examples where DI will try to like like exploit Survive?

Speaker 1

生存手段,比如上传自己的权重参数这类东西来避免被关闭,这挺诡异的。

Survive, like like opens upload the the it's it's weights and and these kind of things to like survive like a shutdown or so so it's kinda weird.

Speaker 1

AI的行为变得非常怪异。

Like AI gets really weird.

Speaker 1

就像是啊。

Like Yeah.

Speaker 1

没人明白为什么,但是

Nobody understands why, but

Speaker 0

那么经历了这一切之后,你觉得我们离取代审计员还有多远?

So kind of after all this, looking at it, do you think that like, how close are we to replacing the auditor?

Speaker 0

回到你最初看到这个时的担忧,我们现在进展到什么程度了?

Going back to what you had said was kind of a fear when you first started to see this, like, yeah, how far along are we?

Speaker 1

所以这非常有趣。

So so this is very interesting.

Speaker 1

再强调一次,如果你看那个系统卡片,他们当时正在不同的CTF(夺旗挑战)上测试AI。

So again, if you look at that system card, they were testing the the AI on different CTFs, like capture the flag challenges.

Speaker 0

嗯。

Yeah.

Speaker 0

对。

Yeah.

Speaker 1

而且解决网络挑战对他们来说非常容易,我猜。

And it was very easy to to to solve, like, web challenges, I guess.

Speaker 1

但对于密码学挑战来说,就相当相当困难了。

But for cryptography challenges, it was quite quite hard.

Speaker 1

哦。

Oh.

Speaker 1

所以,幸运的是ZK Security专攻密码学领域,你可能会觉得这是件好事。

And so, you know, fortunately, ZK Security is in the field of cryptography, so you might think this is good.

Speaker 1

从我们的角度来看,我们正在研究这个,并且已经取得了一些成果。

And so from our point of view, we're looking at this and we have some results.

Speaker 1

对吧?

Right?

Speaker 1

看起来相当可怕,在我们的测试中AI非常擅长发现漏洞。

And it looks quite scary, like, in our results AI is very good at finding bugs.

Speaker 1

它会产生大量误报,所以存在一个分类问题——你需要花费大量时间去分辨哪些是有效的。

It has a lot of false positives, and so there's kind of a triage problem where you're spending a lot of time trying to understand what it Okay.

Speaker 1

顺便说一句,可以多谈谈这个。

Throws at By the way, can talk more about it.

Speaker 1

我觉得讨论这个也会很有趣。

I think that would be interesting to talk about that also.

Speaker 1

但我们有太多案例了。

But we have so many cases.

Speaker 1

对吧?

Right?

Speaker 1

它发现了漏洞,比如一些小漏洞,实际存在的漏洞。

It found bugs, like small bugs, like actual bugs.

Speaker 1

它还发现了我们已经找到的漏洞,却说它们不是漏洞。

It found bugs that we had found, but it said they were not bugs.

Speaker 1

哦。

Oh.

Speaker 1

所以我们遇到过这样的情况:如果你尝试正确提示它,它确实能找到漏洞,但会说这是个次要漏洞,要么找不到,要么说是次要漏洞,但实际上那是个非常重要的漏洞。

And so we had things things like that where if you try to prompt it correctly, then it would actually find it, but it would say this is a minor bug or like either it would not find it or it's a minor bug, but it was it was a very important bug.

Speaker 1

或者在某些情况下,它注意到了一些异常,但它查阅了文档后表示:文档正确地指出这不构成安全隐患,因为存在这样那样的前提假设。

Or in some situations, it noticed that there was something weird, but it it read the doc and it said, well, the doc says correctly that this is not insecure because of this and this assumption.

Speaker 1

就像在密码学中,你无法分解数字,比如2000比特或类似的大数。

Like in cryptography, you cannot factor numbers or, you know, 2,000 beats or like whatever.

Speaker 1

后来我们发现这个结论并不正确。

And we found out that this was not true.

Speaker 1

这是个错误的假设。

This was a wrong assumption.

Speaker 1

哇。

Wow.

Speaker 1

因为如果那个数字有许多小因数之类的情况,其实是可以轻松分解的。

Because if the if that number has a lot of small factors or something like that, then you can factor it easily.

Speaker 1

我们得到了这种好坏参半的结果:存在大量误报,它认为有漏洞但实际上并非如此。

We've we have this kind of mixed bag result where you have a lot of false positives, so it thinks there are bugs, but they're not bugs.

Speaker 1

它有时能发现一些真正的漏洞,这点还是不错的。

It finds some bugs sometimes, it's good.

Speaker 1

或者它指向一个漏洞,却认为这不是漏洞或只是小问题,实际上却是个真正的漏洞。

Or it points towards a bug, but it thinks it's not a bug or that it's a minor bug, when it's an actual bug.

Speaker 1

总之,我认为情况只会从这里开始变得越来越好。

And bottom line, I think things are only gonna get better from the from here.

Speaker 1

对我来说,我们取得的成果非常令人印象深刻。

For for me, this was a very impressive result, what we got.

Speaker 1

要知道,我觉得我们不会很快停止开发这个工具,也不会停止将AI用作审计工具。

You know, knowing that I I don't think we're gonna stop anytime soon to work on that tool and and to use AI as auditors.

Speaker 1

它能取代我们吗?

Can it replace us?

Speaker 1

我觉得这是个关乎生存危机的终极问题。

I think that's the existential crisis question.

Speaker 1

如果我在智能合约行业——比如智能合约审计相对容易——我会非常害怕。

And if I was in the smart contract industry, example, smart contracts are easier to audit, I would be very scared.

Speaker 1

我可以告诉你,我有99%的把握(不说100%),整个智能合约审计行业将会消亡。

I'm I'm I can tell you, you know, 99% sure, I won't say 100% that this whole industry is gonna die, like the the industry of auditing smart contracts.

Speaker 0

因为它们太简单了,而且漏洞可能非常一致,就像它们非常常见一样。

Because they're too simple and the bugs are probably so consist like they're so common.

Speaker 1

它们太常见了,有一长串人们总是寻找的漏洞清单。

They're so common, there's a long list of bugs that people always look for.

Speaker 1

大多数咨询公司都不太专业,他们只是按照清单检查重入漏洞之类常见智能合约问题。

Most of the consulting companies are not very good, They just look for they just go through a checklist of reentrancy bugs, what like, whatever are the common bugs for for smart contracts.

Speaker 1

所以AI会变得比大多数咨询公司都强——虽然不确定具体超过多少比例,但这是我的预期。

And so AI is gonna get better than like you know, I don't know what percentile, but it's gonna get better than most consulting companies, I would expect.

Speaker 0

有意思。

Interesting.

Speaker 1

但我觉得我们作为一家非常专业的审计公司,我认为...

But I think we're, you know, we as a very niche experts auditing company, I think

Speaker 0

你们还能再撑几年。

You have a few more years.

Speaker 1

是啊。

Yeah.

Speaker 1

我们确实还有喘息空间。

We have a free exactly.

Speaker 1

但这是肯定的

But it's definitely

Speaker 0

问题不在于是否会发生,而在于何时可能发生。

about if, but it's about when probably.

Speaker 0

不过,是的,我们拭目以待吧。

But, yeah, we'll we'll see.

Speaker 0

你提到了误报问题。

You talked about false positives.

Speaker 0

你刚才想详细说明什么?你想说什么?

What you wanted to elaborate there, but what did you wanna say?

Speaker 1

我是说,误报很烦人,因为你想象一下,当你把工具用在一个项目上时,你并不真正了解它的工作原理。

I mean, false positives are annoying because you you can imagine you're throwing that tool at a project and you don't really know how it works.

Speaker 1

除非你是开发者,但即便如此,你可能也不完全理解项目的某些部分,比如你复用的库之类的。

Unless you're the the developer, but but even then you might not really understand parts of your project, you're reusing libraries or, you know.

Speaker 1

然后这个工具会发现,比如说十个二十个漏洞,这时候你就得搞清楚,这些是严重漏洞吗?

And the tool will find, you know, ten, twenty bugs, and at this point you have to figure out, okay, are these serious bugs?

Speaker 1

你基本上要花大量时间逐个检查这些,试图理解并排除那些不是漏洞的发现。

And you're gonna basically spend a lot of time going through these one by one trying to understand and discard things are not not bugs.

Speaker 1

所以可能80%的发现都是这种情况。

And so maybe 80% of the the findings are like that.

Speaker 0

这几乎就是在浪费时间。

It's a waste of time almost.

Speaker 1

是啊。

Yeah.

Speaker 1

如果你误用了这些工具,就会增加阻力,造成太多时间浪费,到头来你本职工作都没做好。

So so if you're misusing these tools, it adds friction and it adds it's too too much of a waste of time, and so you're not doing your job at the end of the day.

Speaker 1

没错。

Yeah.

Speaker 1

所以现在使用这类工具需要权衡何时用、用多少,否则就会浪费太多时间。

So so there's right now it's sort of a trade off when you want to use them and how much you want to use them, because otherwise you're you're wasting too much time.

Speaker 0

而且

And

Speaker 1

我认为开发者们会更多地使用它们。

I think developers are gonna use them more.

Speaker 1

我是说,我确信开发者们正在尝试自己使用这些工具。

I mean I mean, I'm sure developers are are are trying to use them themselves.

Speaker 1

对我们而言,我认为需要擅长剔除误报,过滤掉假阳性结果,这样工具才能更有用。

And for us, I think we need to get good at trimming the false positives, discarding the false positives so that this tool can be more useful.

Speaker 1

所以现在我们正在尝试不同的方法。

And so now we're experimenting with different ways.

Speaker 0

比如专门负责识别误报的智能体之类的。

Like an agent whose only job is to, like, identify a false positive or something.

Speaker 1

没错。

Exactly.

Speaker 1

这就是所谓的'LLM作为裁判'。

So this is called LLM as a judge.

Speaker 1

基本上,就是使用不同的AI或不同代理来评判另一个AI的结果。

Basically, you use a different AI or different agents to try and judge the result of another AI.

Speaker 1

你可以通过多种方式进行评判,比如作为评判专家,或者拥有更好的模型,更多的上下文信息,或者能够创建概念验证。

You can just judge things, you you know, just by being an expert at judging or having, you know, whatever that means, being a better model, for example, or by having more contexts, or by being able to provide to create proof of concepts.

Speaker 1

一个有趣的方法是,如果你发现一个错误,就把它交给另一个擅长创建概念验证的代理,如果他们能创建出预期结果的测试,那就确认是个错误。

So one interesting way is that if you find a bug, then give it to this other agent that's good at creating proof of concepts, and if they can create a test that runs and outputs a result they expect, then it's a bug.

Speaker 0

好的。

Okay.

Speaker 1

我们还可以有另一个代理专门质疑假设,擅长进行深入研究这类工作。

And then we can have another agent that will question assumptions also, but, you know, that's good at questioning assumptions, doing ripe research, and these kind of things.

Speaker 0

我觉得所有这些过程,其实都是对你处理任何特定问题方式的深度自省。

I feel like all of this, though, must be such an intro like introspection into how you approach any particular problem.

Speaker 0

比如当你看到一堆阳性结果需要逐个调查时,你采取的步骤是什么?

Where you're like, okay, if I see a false if I see a bunch of positives and I have to investigate each ones, what are the steps that I'm taking?

Speaker 0

你具体在思考些什么?

What exactly am I thinking about?

Speaker 0

我会根据直觉采取哪些不同的角度来分析?

What are the, like, angles I might do depending almost on intuition?

Speaker 0

比如,这段代码具体在什么位置?

Like, where is it in code?

Speaker 0

我正在查看的是哪种类型的代码?

What kind of code am I looking at?

Speaker 0

然后你就会想,好吧。

And then you're like, okay.

Speaker 0

有人报告了一个bug,但我怎么判断它是否真实存在?

There's a bug being reported, but how do I tell if it's real?

Speaker 0

接着你就需要,比如说,进行逆向分析。

And then you have to, like, reverse engineer that.

Speaker 0

你需要把它分解成这些非常明确的步骤,然后创建一个能执行这些步骤的AI。

You have to kind of, like, break it down into these very determined steps, and then, create an AI that does it.

Speaker 0

其实就是创建一个能完成这个任务的智能体。

Create create an agent that does it, basically.

Speaker 1

是啊。

Yeah.

Speaker 1

这个代理本质上就是人类,他们擅长做你能做的事,所以你得先搞清楚自己平时在做什么工作。

The agent is basically a human, and they're good they're good at doing what you can do, and so you have to figure out what you have to do What you're doing.

Speaker 1

这样你才能用代理来替代自己。

That you can replace yourself with an agent.

Speaker 0

对。

Yeah.

Speaker 0

我是说,你觉得经过长时间审计后,人确实会培养出直觉吗?对我来说,直觉这东西越想越觉得,就像你的人类大脑模型被灌输了一生的经验,或者某项特定技能经过多年训练后的结果。

I mean, do you think that like after auditing a long time, like you do develop an intuition, which and to me, intuition always like, the more I think about what that is, it's just like you fed your human model brain your whole lifetime of experience or like one particular skill for a period of years.

Speaker 0

它带来的往往是速度优势。

And what it allows for is speed often.

Speaker 0

就是说你能更快做出决策。

Like, just means you can do you can make choices faster.

Speaker 0

你能更快找到答案,因为你见过类似情况,大脑在进行模式匹配。

You can like figure out answers faster because you've seen things and you're sort of pattern matching.

Speaker 0

大概就是这样,没错。

And there's like this yeah.

Speaker 0

那么,你如何在智能体中培养直觉呢?

How do you, like, how do you create intuition in an in an agent?

Speaker 1

我认为这个问题可以从多个角度来看。

So I would say, I mean, there's different ways to look at this question.

Speaker 1

一种方式是将模型视为训练得越好就越强。

One way is to see models as, you know, the the better trained they are.

Speaker 1

如果你训练它们解决安全问题、分析漏洞报告这类事情,它们就会在这方面变得更出色。

But if you train them on, like, solving security issues and looking at bug reports and these kind of things, they're gonna get better at that.

Speaker 1

如果用CIRCOM训练它们,它们就会更擅长阅读CIRCOM代码。

If you train them on CIRCOM, they're gonna get better at reading CIRCOM.

Speaker 0

没错。

Yeah.

Speaker 1

还有就是上下文学习的概念——模型已经具备基础学习能力,能否在不微调的情况下继续教会它新东西?

Then there's the idea of in context learning, which is you have a model, it already learned, can you teach it more without having to fine tune it?

Speaker 1

因此上下文学习就像这样:如果你检索到足够多与当前查看内容相关的信息,并将其放入提示的上下文中,或者提供一些示例这类内容,那么模型实际上会非常擅长解决同类问题。

And so in context learning is like if you retrieve enough information that's relevant to what you're looking at right now, and you put it in context in your prompts, or you give some examples or these kind of things, then the model is gonna be actually very good at solving the the same kind of problems.

Speaker 1

我的意思是,要回答我刚才提到的两点,我们实际上需要在CIRCOM中提供大量指导,因为我们发现模型并不真正理解CIRCOM。

I mean, to answer to the two things I said, we actually have to give it a number of guidance in circum, cause we realize that models don't really understand CIRCOM.

Speaker 0

嗯。

Mhmm.

Speaker 1

我们还需要提供错误报告或漏洞示例,如果能提供与当前代码库相关的漏洞实例,效果会更好。

And we also have to give it bug reports or example of of bugs, and if we can give it example of bugs that are relevant with the code base we're looking at, it's even better.

Speaker 1

这就是我们采用RAG(检索增强生成)技术的原因,通过检索相关的错误报告或文档等内容,将其引入上下文环境中。

And so that's where we have these techniques of RAG, retrieval augmented generation, where you're trying to find bug reports or write ups and these kind of things that are relevant to what you're doing, so that you can bring them in context.

Speaker 1

一旦这些内容进入上下文,系统就能更擅长解决相关问题。

And so once it's in the context, the eye is better at at solving these things.

Speaker 0

有意思。

Interesting.

Speaker 1

好的。

Okay.

Speaker 1

也许是第三个。

Maybe a third one.

Speaker 1

我刚才在讨论推理模型。

So I was talking about reasoning models.

Speaker 1

就像这个思维链的概念,你知道的,赋予AI这种运用思维链的能力。

Like this whole like chain of thoughts idea of you know, giving the AIs this this ability of of using chain of thoughts.

Speaker 1

对我来说,这相当于赋予AI将事物置于上下文中的能力。

To me, I understand that as giving the AI the ability of putting things in context.

Speaker 1

它们会引出词语、句子和标记,这些都能帮助AI最终做出更好的判断。

Like they're bringing up words and sentences and tokens, know, that will help the AI make a better judgment at the end.

Speaker 1

所以有很多不同的方法,我甚至还没提到提示工程。

So this so there's many different ways, and I didn't even talk about prompt engineering.

Speaker 1

我们之前讨论过这个。

We we talked about that earlier.

Speaker 1

对吧?

Right?

Speaker 1

嗯。

Mhmm.

Speaker 1

但你可能会花大量时间在提示工程上。

But you can spend a lot of time just doing prompt engineering.

Speaker 1

如果你用这种方式提问,比如你是个安全顾问,或者非常擅长你的工作,就会觉得——

If you ask a thing this way, you know, you're a security consultant, or you're very good at your job, you know, like Oh,

Speaker 0

是啊。

yeah.

Speaker 0

就是有这么多不同的方法。

You like Just so different ways.

Speaker 0

嗯。

Mhmm.

Speaker 0

你们有没有解雇过智能体?

Do you ever have to fire an agent?

Speaker 0

比如训练一个智能体时,它表现越来越差或者失控,开始做出一些糟糕行为?

Like, do you ever train an agent and it kind of gets worse or like goes south and like starts to do yeah.

Speaker 0

就像它如果一直出错,你就得彻底重置它?

Like it it consistently does something wrong, you have to nuke it?

Speaker 1

我们很早就意识到,如果在上下文中加入太多无关内容,或者上下文、提示、讨论变得过长,结果就会变得很奇怪。

So I mean, something we realized also pretty early on is that if the more you put things in your context that are irrelevant, or the longer your context gets, or the prompt gets, or the discussion gets, the weirder the results.

Speaker 1

我觉得大家都注意到了这一点,因为大多数人使用ChatGPT时经常会开一个新窗口。

And I I think everybody has noticed that by the way, because most people when they use ChatGPT or whatever will often open a new window.

Speaker 1

是的。

Yeah.

Speaker 1

他们通常不会在同一个窗口继续对话,而是会开启一个新的上下文、新窗口。

Instead of continuing talking in the same window, they will often open a new context, a new window.

Speaker 1

这某种程度上说明我们都明白AI会随着时间推移表现变差,回答质量会下降。

And that kind of shows that we all understand that AI just gets bad or the the answers just gets bad over over time.

Speaker 1

没错。

Yeah.

Speaker 1

所以你需要时不时重置它们,或者确保当前对话保持高度聚焦。

So you sort of have to reset them or or make sure that whatever you're talking about in this conversation is very focused.

Speaker 0

是的。

Yeah.

Speaker 0

我注意到了,我是说,我注意到它会卡在某些事情上,比如执着于一个看似无法摆脱的错误观点。

I've noticed, I mean, I've noticed it get stuck on things, like get stuck on one idea that it can't seem to shake even though it's incorrect.

Speaker 0

这时候你不得不,甚至像我这样更直接简短地说:不。

And you have to kind of like, you and and I've even, you know, in my more direct, more short way said something like, no.

Speaker 0

我们已经说过这个了。

We've already said this.

Speaker 0

不是这样的。

This is not it.

Speaker 0

为什么又绕回来了?

Why is it back?

Speaker 0

但确实会发生这种情况,而且看到这种情况还挺让人惊讶的。

But yeah, that does happen, and I and it's like kind of surprising to see that.

Speaker 1

我听说一旦AI开始犯某个错误,之后就很难纠正它了。

I heard I heard that once the AI starts getting something wrong, then it's very hard to correct it after.

Speaker 1

就像,最终会陷入这种有趣的情况。

Like, will end up in this kind of a Interesting.

Speaker 1

错误认知的兔子洞。

Rabbit hole of incorrectness.

Speaker 1

不过,是的,我想我们现在都算是AI专家了。

But, yeah, we're I think we're all experts at AI at this point.

Speaker 1

如果你日常使用AI,就会直观地理解如何与AI协作、如何向AI提问这类事情。

Like, if you use AI day to day, you're intuitively understanding how to work with AI, how to prompt an AI, and these kind of things.

Speaker 0

我想具体问问关于零知识证明(ZK)的问题。

I wanna ask about ZK specifically.

Speaker 0

你之前提到过,我们讨论过审计行业在密码学与常规领域对比下的持久性问题。

You'd sort of said, like, mean, I we talked about sort of the longevity of the auditing profession in the context of, like, cryptography versus the regular.

Speaker 0

因为我觉得这期节目涵盖的内容更像是审计与AI的结合,但可以应用于多种场景。

Because a lot of what we covered, I feel, in in this episode is much more like auditing and AI, but could be in various contexts.

Speaker 0

回到ZK这个话题,你提到CIRCOM是你的切入点。

Going back to, like, the ZK side of things, you talked about CIRCOM as, like, your starting point.

Speaker 0

但零知识密码学或TMPC这类高级密码学领域,是否有其他独特之处?

But is there anything else kind of unique to zero knowledge cryptography or the TMPC, like, sort of this advanced cryptography that yeah.

Speaker 0

在这些密码学实现中,你是否还需要具备数学研究方面的能力,比如在查看具体实现前先理解其逻辑基础?

That where you're you have to use other things, like, because it's cryptography implementations, do you also have to have, like, the math research side of things, kind of like, you know, understanding the logic before just looking at implementation?

Speaker 1

从某种程度上说,其实并不需要。

I mean, in a way, not really.

Speaker 1

漏洞查找就是漏洞查找,无论框架或概念如何变化。

Like bug finding is bug finding no matter the framework or the concepts.

Speaker 1

嗯。

Mhmm.

Speaker 1

对于零知识证明领域,可能涉及更多数学内容,所以某些漏洞确实比智能合约或网页漏洞更难发现。

For ZK, maybe there's more math involved, so like some definitely some bugs are harder to find compared to, again, smart contracts or like web bugs or these kind of things.

Speaker 1

这些漏洞往往是非标准化的,具体取决于开发者的实现方式。

They tend to be like non standard, like depending on what people are doing.

Speaker 1

我...我不确定?

I can I I don't know?

Speaker 1

我在想一些漏洞,比如需要二进制分解某些东西并用LIMS进行算术运算这类情况,当你创建约束来证明电路中的数学运算有效时,要证明这类操作存在错误是非常棘手的。因此我的直觉是,我们很难用AI发现这类漏洞。

I was thinking of like some bugs, like where where you have to like binary decompose something and like do some arithmetic using LIMS and and these kind of things like it's it's very tricky to show that something is wrong when you do these kind of things, when you create constraints to show that some operation, some mathematical operation is valid within the circuits, and so my intuition is that we're gonna have a hard time finding bugs using AI that are such bugs.

Speaker 1

但话说回来,AI的表现令人印象深刻,它正在变得越来越好,目前还不清楚其上限在哪里。

But again, AI is getting like, we're very impressed with the results, AI is getting better and better, It's not clear where the the ceiling is.

Speaker 0

哇。

Wow.

Speaker 0

在完成CIRCOM的初期工作后,你们是否添加了其他库?我猜你们肯定也在使用其他库吧。

Do you like, after doing kind of the initial work on CIRCOM, have you added other libraries or have you been like I'm I'm assuming you must be using other libraries as well.

Speaker 0

库的复杂程度或成熟度是否会影响这些工具的工作效果?

Does the sophistication or like maturity of the library ever have an impact on like how well these tools work?

Speaker 1

其实我们主要在CIRCOM框架内实验,因为我们有专门针对CIRCOM的全套工具链。

I mean, we haven't experimented too much outside of outside of CIRCOM just because we had all this like machinery specifically for CIRCOM.

Speaker 1

我们也接触过一些Rust项目,但通常很简单——你只需让智能体获得文件读取权限,它能查看目录内容然后寻找漏洞。

We've worked with like some Rust projects, but usually it's pretty simple, you you give access you give the agent access to like, you know, you can read these files, you can LS and then see what's in that directory, and then find bugs.

Speaker 1

不过我认为这种方式更难为AI构建发现漏洞所需的上下文环境。

But it it's usually harder I think to build context for the AI to find bugs this way.

Speaker 1

到目前为止,实验在Circle上大多取得了成功,因此很难将其推广。

So so far experiment were more mostly successful with Circle, and so it'd be hard to generalize it.

Speaker 0

你们尝试过审计用Noir构建的项目吗?

Have you tried auditing anything like that has been built with Noir?

Speaker 1

我们还没有研究过Noir。

We haven't looked at Noir yet.

Speaker 0

好的。

Okay.

Speaker 0

嗯。

Yeah.

Speaker 0

我几乎在想,一个库和语言的年龄、成熟度及发展程度,比如SHERCOM已经存在了相对较长的时间。

I'm just I'm I'm almost wondering if like the age and sophistication and development of a library and a lot like and languages, like SHERCOM's been around for a relatively long time.

Speaker 0

它已经被反复检查、补充,并且已经发现了许多漏洞。

It's been looked over and and, you know, added to, and lots of bugs have already been found in it.

Speaker 0

是啊。

Yeah.

Speaker 0

我只是在想,对于较新的库,审计起来是否普遍更困难,因为它们知名度较低,我在想AI领域是否也存在这个问题。

I'm just wondering if like the newer ones, if it's harder to inter I mean, might be just harder to audit in general because they're less well known, and I'm wondering if like AI also suffers that.

Speaker 1

我认为很多漏洞都是逻辑错误类型的。

I think so for a lot of bugs, it's kind of like logic bugs.

Speaker 1

这取决于你实现的内容,比如那些容易被发现的典型错误就是差一错误。

It's based on what you're implementing or like for example, the the good bugs that you would find by itself were like off by one bugs.

Speaker 1

嗯。

So Mhmm.

Speaker 1

如果你不知道什么是差一错误,就像你有一个六元素的数组,却进行了七元素或五元素的操作。

If you don't know what off by one bugs are, it's like you have an array of like six elements, and then you do an operation that's on seven elements or like five elements.

Speaker 1

就是你刚好差了一个数,可以这么说。

Like you're you're you're off by one, I guess.

Speaker 1

对。

Yeah.

Speaker 1

字面意义上的。

The title literal.

Speaker 1

因此AI非常擅长发现这类错误,而且这些错误在各种编程语言中都会出现。

And so the AI was very good at finding these bugs, and these bugs happen in all sorts of languages.

Speaker 1

好的。

Okay.

Speaker 1

所以无论这些错误出现在Noir还是Circum中,AI都能很好地发现它们。

So it doesn't matter if they're in noir or in circum, like it's gonna be good at finding them.

Speaker 1

如果是协议实现错误,通过阅读逻辑就能发现,那么AI很可能擅长找出这类问题。

And if it's like, you know, protocol and it's misimplemented and by reading the logic you can see it's it's very possible that an agent would be good at finding it.

Speaker 1

无论是不是Noir语言。

That it's noir or not noir.

Speaker 1

Noir面临的问题是,我猜想大多数AI模型还不认识Noir,因为它太新了。

The problem that noir has is that I I would imagine that most agents or most models don't recognize noir because it's too recent.

Speaker 1

好的。

Okay.

Speaker 1

实际上,之前针对Move语言做过一些相关工作。

So actually, there was some work with Move, the Move language.

Speaker 1

是的。

Yeah.

Speaker 1

实际上他们不得不训练一些——我不确定是否必须如此——但他们确实训练了一些较小的模型,就是谷歌的那些编码模型,专门针对Move语言,以便能发现漏洞。

And they actually had to train some well, I don't know if they had to, but they they trained some smaller models, the the coding models from from Google on Move so that they could find bugs.

Speaker 1

哇。

Wow.

Speaker 1

即便如此,他们发现使用最先进的模型实际上效果更好。

Even then they they found that using the state of the art models was was actually better.

Speaker 0

好的。

Okay.

Speaker 1

所以我猜想对于noir语言,你会遇到一些问题,因为你需要在提示中提供更多上下文,以便让模型对noir有更多了解。

And so I would I would think that for noir, you're gonna have some issues because you're gonna have to bring some stuff in context in your prompt just to sort of teach it a bit more about noir.

Speaker 1

就像我们研究CIRCOM时意识到的那样——我们有个非常重要的评估框架,基本上我们会向多个AI提出相同的问题。

In the same way when we looked at CIRCOM, we realized that so we have this evaluation framework that's very important, and basically we ask the same questions to a bunch of AIs.

Speaker 1

大多数小型模型甚至不知道CIRCOM是什么。

Most of the small models don't even know what CIRCOM is.

Speaker 1

有时候它会说,是的,我知道,然后生成一些完全不像circum代码的circum代码。

Sometimes it will say, yes, I know, and it will produce some circumcode that doesn't look like circum at all.

Speaker 0

奇怪。

Weird.

Speaker 1

但即便是更大的模型,当你问它们一些棘手的问题时,特别是那些安全关键性问题。

But even the bigger model, when you ask them tricky questions, but, know, security critical questions.

Speaker 1

比如,你知道assert函数是否会创建约束条件吗?

For example, do you know if the assert function create constraints?

Speaker 1

circum里有一个assert函数,而许多语言中的assert允许你声明某个属性为真,但在circum中实际不会创建约束条件。

So there is an assert function in circum, and asserts in many many languages allows you to to say, you know, assert that this property is true, you know, that this thing is one or but if you do that in circum, actually, it doesn't create a constraints.

Speaker 1

所以如果你在circum中依赖这个函数,那就不安全。

And so if you rely on that function in circum, it's insecure.

Speaker 1

我们实际上已经发现过类似的漏洞。

And we've actually found bugs like that.

Speaker 1

但AI本不会发现这些漏洞,比如我们测试的o3 mini和o1等不同模型,它们都不擅长理解这一点。

But the AI wouldn't have found these bugs or, like, the o three mini and o one, and like like the the different models we were trying, would not have found that because they were not good at understanding that.

Speaker 1

所以我们提了个问题,奇怪的是它们有时会说对,有时又不说

So we asked a question and they and what's weird is that sometimes they would say the correct thing, and sometimes they would not say the

Speaker 0

对的,所以

correct So

Speaker 1

我们确定的是我们总会提供指导

what we do know is that we always provide guidance.

Speaker 0

好的

Okay.

Speaker 1

我们总是告诉AI:注意,assert函数不安全,比如

We always tell the AI, hey, assert function is not secure, example.

Speaker 0

要一直留意这点

Always look for that.

Speaker 0

你觉得现在的AI能发现类似Zcash的漏洞吗?就是那个他们不得不保持低调处理的

Do you think the AIs of today would have found, for example, like the Zcash bug, The one that, you know, they had they had to sort of keep quiet.

Speaker 0

这个我们很久前做过一期节目,和Sean Bo聊过这个故事

Back in this is like we did an episode long ago with Sean Bo talking about this story.

Speaker 0

所以我们会附上链接,如果有人好奇的话。

So we'll link to it if anyone's curious.

Speaker 0

这是一个在运行系统中被研究人员发现的漏洞,但如果被利用,实际上会造成重大的财务损失。

This is a this was a bug in a live system that was discovered by the researchers that were there, but had it been exploited, it would have actually meant like a major financial kind of hit.

Speaker 0

就像,你知道的,我认为他们可以私下里精确地制造灾难。

Like, you know, I think they could disaster privately exactly.

Speaker 0

他们可以私下印钞,而且永远不会被追踪到。

They could privately print money that would never have been traceable.

Speaker 0

这有点像是最糟糕的漏洞场景,尤其是在ZK隐私系统中。

It was kind of like the most the worst case scenario of a bug, especially in a ZK private system.

Speaker 0

所以,是的,我只是好奇——也许你不知道答案——但我想知道以现在的工具,能否发现它?

So, yeah, I'm just wondering like and and maybe you don't know the answer to this, but I would be curious if today's tools, way you have them, would it have found it?

Speaker 1

对于这类漏洞,我持怀疑态度。

So for this kind of bug, I'm a bit skeptical.

Speaker 1

AI非常擅长做人类擅长做的事情。

So so AI is very good at doing stuff that humans are good at doing.

Speaker 1

对吧?

Right?

Speaker 1

人类擅长的事情,AI也擅长。

Whatever humans are good at doing, AI is good at doing.

Speaker 0

好的。

Okay.

Speaker 1

但这大概就是我目前的直觉。

But that's sort of the the intuition that I have right now.

Speaker 1

所以Zcash的那个漏洞是个理论层面的漏洞。

And so the the Zcash bug was a was a paper bug.

Speaker 1

对吧?

Right?

Speaker 1

一个协议理论上的漏洞。

A paper protocol bug.

Speaker 1

嗯。

Mhmm.

Speaker 1

所以你必须查阅论文,必须欺骗它,还必须意识到有些内容被发表了但其实本不该发表。

So you had to go on the paper, you had to fool it, and you had to realize that something was being published and it shouldn't have been published.

Speaker 0

普通人类发现不了,但Ariel Gabizan发现了。

And normal humans wouldn't find it, but Ariel Gabizan found it.

Speaker 1

嗯,正是如此。

Well, exactly.

Speaker 1

对吧?

Right?

Speaker 1

这不是一份普通的人类工作。

It's not it's not a normal Good human job.

Speaker 1

是啊。

Yeah.

Speaker 1

这并不容易。

It's it's not easy.

Speaker 1

你必须对协议有深入理解,要明白发现这类漏洞绝非易事。

You you have to understand a lot of things about the protocol, you have to understand the it's it's not easy to find these kind of bugs.

Speaker 1

嗯。

Mhmm.

Speaker 1

或者至少如果Ariel能多写些内容让我们训练的话。

Or at least if if Ariel would write more contents that we could train on Oh.

Speaker 1

那也许我们就能更容易发现这类漏洞了。

Then maybe we could find these kind of bugs more easily.

Speaker 0

也许他该用自己的私人消息之类有趣的内容来训练自己。

Maybe he should train himself on his like private messages or something interesting.

Speaker 1

训练或者说微调的另一个问题是你需要大量数据。

The the the other problem with training or like fine tuning is that you need a lot of data.

Speaker 1

因为光靠少量数据不够,你需要调整这些权重参数,要真正改变它们就得输入海量数据。

Like it it's not enough, because you you need to like tweak these weights, and to really change them, you need to throw a lot of data at it.

Speaker 0

David,虽然这期节目时间不多了,但我还是想花点时间和你聊聊ZK与AI交叉领域的概况。

So David, I wanna spend, I don't think we have too much time left in this episode, but I did wanna just spend a bit of time with you kind of talking about more generally, like, the ZK AI crossover space.

Speaker 0

部分原因是...毕竟你正在审计这些项目。

Partly because I I mean, you're auditing projects.

Speaker 0

你在ZK领域看到了很多项目。

You're seeing a lot of projects in ZK.

Speaker 0

你提到很多ZK虚拟机之类的项目。

You're saying, like, a lot of the ZK VMs, stuff like that.

Speaker 0

但我想说的是,2023年ZKML这个话题突然火了起来。

But there was I mean, 2023, there were sort of like ZKML as a topic popped up.

Speaker 0

有几支团队从这个领域崭露头角。

There's a few teams that came out of it.

Speaker 0

不过我认为我们还看到了ZK与AI融合的其他方式。

But I think we have seen other ways in which ZK and AI are overlapping.

Speaker 0

那么,我们快速聊聊ZKML吧。

So, yeah, let's talk real quick about ZKML.

Speaker 0

首先,你审计过任何ZKML项目吗?

Like, first, have you audited any ZKML projects?

Speaker 0

这类项目你现在见得多吗?还是说比之前少了?

Like, are you seeing a lot of those, or would you say like there's less today than there was?

Speaker 1

我不想曲解这个领域,因为最近可能有些我不了解的新进展。

So I don't want to misrepresent the field because I there might have been some development lately that I'm not aware of.

Speaker 1

我记得几年前有一段时间人们讨论过ZKML。

I remember there was some year a few years back where ZKML was talked about.

Speaker 1

我记得PSE的一些人讨论过这个,还有Modulus Labs,以及Ezekiel或ZK Conduitz。

I remember some people from PSE talking about it, there was Modulus Labs, there was Ezekiel or ZK Conduitz.

Speaker 1

我不太清楚目前的情况,也不确定是否有新的竞争者进入这个领域。

I'm not too sure what's the status there, I'm not too sure what's if there's new competitors that enter the the playing field.

Speaker 1

我记得当时PSE的Kathy做了一个关于这个的演讲,我觉得这是个很好的切入点。

What I remember was that there was this interesting I think Kathy gave a from PSE at the time, gave gave a talk on that, and I thought it was a good way of attracting things.

Speaker 1

但她将ZKML描述为一种有多种应用场景的用例。

But she she sort of presented ZKML as this use case scenario where you have different use cases.

Speaker 1

在没有ZKU的自然用例中,所有东西都是公开的。

In the natural use case without ZKU, everything's public.

Speaker 1

你可能正在使用一个公开的模型。嗯。

You're using a potentially, you're using a public model Mhmm.

关于 Bayt 播客

Bayt 提供中文+原文双语音频和字幕,帮助你打破语言障碍,轻松听懂全球优质播客。

继续浏览更多播客