I want to tell you a story about artificial intelligence and farmers.
我想给你讲一个 关于人工智能与农民的故事。
Now, what a strange combination, right?
这听起来是个奇怪的组合,对吧?
Two topics could not sound more different from each other.
这两个话题听起来简直完全不相干。
But did you know that modern farming actually involves a lot of technology?
但你知道现代农业 其实与科技息息相关吗?
So computer vision is used to predict crop yields.
比如,计算机视觉 用于预测作物产量。
And artificial intelligence is used to find, identify and get rid of insects.
人工智能用于发现、 识别和消灭昆虫。
Predictive analytics helps figure out extreme weather conditions like drought or hurricanes.
预测分析有助于找出干旱,飓风等 极端天气状况。
But this technology is also alienating to farmers.
但是这项技术 也疏远了农民。
And this all came to a head in 2017 with the tractor company John Deere when they introduced smart tractors.
在 2017 年, 这一切到达了顶峰。拖拉机公司约翰迪尔 推出了智能拖拉机,
So before then, if a farmer's tractor broke, they could just repair it themselves or take it to a mechanic.
在此之前,如果农民的拖拉机坏了, 他们可以自己修理或者交给机械师。
Well, the company actually made it illegal for farmers to fix their own equipment.
虽然,该公司实际上 将农民私自维修设备 视为非法。
You had to use a licensed technician and farmers would have to wait for weeks while their crops rot and pests took over.
农民必须聘请有执照的技术人员。农民将不得不等上好几个星期, 直到他们的农作物腐烂, 虫害已然占据上风的时候。
So they took matters into their own hands.
因此,他们想把事情掌握在自己手中。
Some of them learned to program, and they worked with hackers to create patches to repair their own systems.
他们中的一些人学会了编程, 他们与黑客合作开发了 补丁来修复自己的系统。
In 2022, at one of the largest hacker conferences in the world, DEFCON,
在 2022 年, 在世界上最大的 黑客会议之一 DEFCON 上
a hacker named Sick Codes and his team showed everybody how to break into a John Deere tractor, showing that, first of all,
一位名叫 Sick Codes 的黑客 和他的团队向所有人 展示了如何闯入约翰迪尔拖拉机,这首先表明了,
the technology was vulnerable, but also that you can and should own your own equipment.
该技术很脆弱, 但是,这也代表 你可以研发自己的设备了。
To be clear, this is illegal, but there are people trying to change that.
当然,这是非法的, 但有人试图改变这种状况。
Now that movement is called the "right to repair." The right to repair goes something like this.
现在,该运动被称为 “维修权”。维修权是这样的。
If you own a piece of technology, it could be a tractor, a smart toothbrush, a washing machine, you should have the right to repair it if it breaks.
如果你拥有一项技术, 它可能是拖拉机、智能牙刷、 洗衣机,如果它坏了, 你应该 有权修理它。
So why am I telling you this story?
那我为什么要告诉你这个故事?
The right to repair needs to extend to artificial intelligence.
维修权需要扩展 到人工智能。
Now it seems like every week there is a new and mind-blowing innovation in AI.
现在,人工智能似乎每周 都有一项令人兴奋的新创新。
But did you know that public confidence is actually declining?
但是你知道公众的信心 实际上在下降吗?
A recent Pew poll showed that more Americans are concerned than they are excited about the technology.
皮尤最近的一项民意调查 显示,美国人对这项技术的 担忧多于兴奋
This is echoed throughout the world.
这在全世界都得到了共鸣。
The World Risk Poll shows that respondents from Central and South America and Africa all said that they felt AI would lead to more harm than good for their people.
世界风险民意调查显示 来自中南美洲和非洲的受访者 都表示,他们认为人工智能 对人民带来的弊大于利。
As a social scientist and an AI developer, this frustrates me.
作为一名社会科学家和 人工智能开发人员, 这让我感到沮丧。
I'm a tech optimist because I truly believe this technology can lead to good.
我是一个技术乐观主义者, 因为我真的相信 这项技术可以带来好处。
So what's the disconnect?
那么,脱节是如何发生的?
Well, I've talked to hundreds of people over the last few years.
事实上,在过去的几年里, 我和数百人谈过话。
Architects and scientists, journalists and photographers, ride-share drivers and doctors, and they all say the same thing.
建筑师和科学家、 记者和摄影师、 拼车司机和医生, 他们都有同样的想法。
People feel like an afterthought.
他们都有同样的后知后觉。
They all know that their data is harvested often without their permission to create these sophisticated systems.
他们都知道,他们的数据 通常是在未经允许的情况下收集的, 用于创建这些复杂的系统。
They know that these systems are determining their life opportunities.
他们知道这些系统 正在改变他们的生活。
They also know that nobody ever bothered to ask them how the system should be built,
他们也知道从来没有人过问他们的意见 来如何构建系统,
and they certainly have no idea where to go if something goes wrong.
他们也无从得知如果 系统出现问题该去哪里解决。
We may not own AI systems, but they are slowly dominating our lives.
我们可能不拥有人工智能系统, 但它们正在慢慢主导我们的生活。
We need a better feedback loop between the people who are making these systems,
我们需要建立一个良好的反馈循环
and the people who are best determined to tell us how these AI systems should interact in their world.
在系统研发者 与大众之间构建一个桥梁 来帮助AI系统更好地融入大众的生活
One step towards this is a process called red teaming.
而“红队合作” ,是我们 迈向这个目标的一步。
Now, red teaming is a practice that was started in the military, and it's used in cybersecurity.
红队合作是一种源于军队的做法, 现在,它被用于网络安全。
In a traditional red-teaming exercise, external experts are brought in to break into a system, sort of like what Sick Codes did with tractors, but legal.
在传统的红队活动中, 外部专家被邀请进入一个系统, 有点像 Sick Codes 对拖拉机所做的那样,但合法。
So red teaming acts as a way of testing your defenses and when you can figure out where something will go wrong, you can figure out how to fix it.
因此,红队合作是测试 你的防御能力的一种方式, 当你弄清楚 哪里会出现问题时, 你就能想出解决问题的方法。
But when AI systems go rogue, it's more than just a hacker breaking in.
但是,当人工智能系统变成盗贼时。不仅仅是黑客的入侵。
The model could malfunction or misrepresent reality.
该模型可能会出现故障或歪曲现实。
So, for example, not too long ago, we saw an AI system attempting diversity by showing historically inaccurate photos.
因此,举例来说,不久前, 我们看到一个人工智能系统试图 通过显示历史上不准确 的照片来实现多样性。
Anybody with a basic understanding of Western history could have told you that neither the Founding Fathers nor Nazi-era soldiers would have been Black.
任何对西方历史有基本了解的人 都可以告诉你无论是 开国元勋还是纳粹时代的士兵, 都不会是黑人。
In that case, who qualifies as an expert?
在这种情况下, 谁有资格成为专家?
You.
你。
I'm working with thousands of people all around the world on large and small red-teaming exercises,
我正在与世界各地成千上万的人 一起进行各规模的红队练习,
and through them we found and fixed mistakes in AI models.
通过不断的练习,我们发现 并修复了AI 模型中的缺陷。
We also work with some of the biggest tech companies in the world: OpenAI, Meta, Anthropic, Google.
我们还与世界上一些最大的 科技公司合作:OpenAI、Meta、Anthropic、谷歌。
And through this, we've made models work better for more people.
通过这种方式, 我们使模型更好地 适用于更多的人。
Here's a bit of what we've learned.
以下是我们学到的一些东西。
We partnered with the Royal Society in London to do a scientific, mis- and disinformation event with disease scientists.
我们与伦敦皇家学会合作,与疾病科学家 一起举办了一次 关于科学与虚假信息的活动。
What these scientists found is that AI models actually had a lot of protections against COVID misinformation.
这些科学家发现, 人工智能模型实际上具有很多 针对 COVID 错误信息的保护措施。
But for other diseases like measles, mumps and the flu, the same protections didn't apply.
但同样保护措施并不适用于 麻疹、流行性 腮腺炎和流感等其他疾病,
We reported these changes, they're fixed and now we are all better protected against scientific mis- and disinformation.
我们报告了这些变化, 它们得到了改进, 而我们都得到了更好的保护, 让我们免受科学错误 和虚假信息的侵害。
We did a really similar exercise with architects at Autodesk University, and we asked them a simple question: Will AI put them out of a job?
我们对 Autodesk 大学的建筑师 做了一个非常相似的练习, 我们问了他们一个简单的问题:人工智能会让他们失业吗?
Or more specifically, could they imagine a modern AI system that would be able to design the specs of a modern art museum?
或者更具体地说, 他们能否想象一个 能够设计现代艺术博物馆的 现代人工智能系统?
The answer, resoundingly, was no.
答案很明显,是否定的。
Here's why, architects do more than just draw buildings.
这就是建筑师所做的 不仅仅是画建筑物的原因。
They have to understand physics and material science.
他们必须了解物理学 和材料科学。
They have to know building codes, and they have to do that while making something that evokes emotion.
他们必须了解建筑规范, 而且他们必须制作 引起大众共鸣的东西。
What the architects wanted was an AI system that interacted with them, that would give them feedback, maybe proactively offer design recommendations.
建筑师他们想要的是一个 可以与他们互动的人工智能系统, 可以向他们提供反馈, 也许可以主动提供设计建议。
And today's AI systems, not quite there yet.
而当今的人工智能系统, 还没有完全到来。
But those are technical problems.
但这些都是技术问题。
People building AI are incredibly smart, and maybe they could solve all that in a few years.
构建人工智能的人非常聪明, 也许他们可以在几年内解决 所有这些问题。
But that wasn't their biggest concern.
但这并不是他们最关心的问题。
Their biggest concern was trust.
他们最担心的是信任。
Now architects are liable if something goes wrong with their buildings.
现在,如果建筑物出现问题, 建筑师要承担责任。
They could lose their license, they could be fined, they could even go to prison.
他们可能会失去执照, 可能会被处以罚款, 甚至可能入狱。
And failures can happen in a million different ways.
失败可能 以一百万种不同的方式发生。
For example, exit doors that open the wrong way, leading to people being crushed in an evacuation crisis,
例如,出口门 开错了方向, 导致人们在疏散危机 中被压伤,
or broken glass raining down onto pedestrians in the street because the wind blows too hard and shatters windows.
或者 由于风吹得太厉害,窗户被打碎, 碎玻璃落到街上的行人身上。
So why would an architect trust an AI system with their job, with their literal freedom, if they couldn't go in and fix a mistake if they found it?
那么,建筑师要如何信任人工智能 来完成自己的工作 他们会失去自主权 无法干预或修正 AI 的在工作上失误
So we need to figure out these problems today, and I'll tell you why.
因此,我们今天 需要弄清楚这些问题,
The next wave of artificial intelligence systems, called agentic AI, is a true tipping point between whether or not we retain human agency,
下一波人工智能系统, 即代理人工智能, 是一个真正的转折点。
or whether or not AI systems make our decisions for us.
要么保留人类能动性, 要么人工智能系统来 完全为我们做出决策。
Imagine an AI agent as kind of like a personal assistant.
想象一下 AI 代理有点像私人助理。
So, for example, a medical agent might determine whether or not your family needs doctor's appointments, it might refill prescription medications,
因此,举例来说, 医疗代理人可能会决定你的家人 是否需要预约医生, 它可能会补充处方药,
or in case of an emergency, send medical records to the hospital.
或者在紧急情况下向医院 发送病历。
But AI agents can't and won't exist unless we have a true right to repair.
但是,人工智能代理 无法真正地诞生 如果我们没有真正的修复权 没有家长会将孩子的健康委托 给人工智能系统
What parent would trust their child's health to an AI system unless you could run some basic diagnostics?
除非你能进行一些基本的判断 哪位专业人士会相信人工智能 能做出专业的决策
What professional would trust an AI system with job decisions, unless you could retrain it the way you might a junior employee?
除非你能对人工智能系统 进行专业培训?
Now, a right to repair might look something like this.
现在,维修权 可能看起来像这样。
You could have a diagnostics board where you run basic tests that you design, and if something's wrong,
你可以有一个诊断板 来运行你设计的基本测试, 如果出现问题,
you could report it to the company and hear back when it's fixed.
你可以向公司报告, 并在修复后收到回复。
Or you could work with third parties like ethical hackers who make patches for systems like we do today.
或者你可以与第三方合作, 比如道德黑客, 他们像我们今天一样 为系统制作补丁。
You can download them and use them to improve your system the way you want it to be improved.
你可以下载它们并使用它们 来按照你想要的方式改进系统。
Or you could be like these intrepid farmers and learn to program and fine-tune your own systems.
或者你可以像这些干劲十足的 农民一样,学会编程 和微调自己的系统。
We won't achieve the promised benefits of artificial intelligence unless we figure out how to bring people into the development process.
我们需要想出如何让人们 参与开发过程, 否则我们将无法实现 人工智能所承诺的好处。
I've dedicated my career to responsible AI, and in that field we ask the question, what can companies build to ensure that people trust AI?
我的职业生涯一直 致力于负责任的人工智能, 在这个领域我们问一个问题, 企业可以建立 什么来确保人们信任人工智能?
Now, through these red-teaming exercises, and by talking to you, I've come to realize that we've been asking the wrong question all along.
现在,通过这些红队练习, 通过与你们的交谈, 我意识到我们一直在问错误的问题。
What we should have been asking is what tools can we build so people can make AI beneficial for them?
我们应该问的是,我们可以开发 哪些工具,让人们能够让 人工智能为他们带来好处?
Technologists can't do it alone.
技术人员无法独自完成这项工作。
We can only do it with you.
我们只能和你一起做。
Thank you.
谢谢。
(Applause)
(掌声)
|
|