Keeping AI Well Behaved: How Do We Engineer An Artificial System That Has Values?

让人工智能完美运行:我们怎么给人工智能系统赋予价值观?

In Brief
摘要:

It is estimated that some 10 million self-driving cars will be on the road by the close of 2020, which raises questions about how an AI will respond in lethal situations.
预计到2020年为止,全球将会有超过100万辆人工智能驱动的无人车上路,同时也引出了人工智能在发生致命事故时会怎样反应的问题。

In this exclusive interview, FLI researcher David Parkes discusses his work to engineer a value-aligned AI system.
在这篇专访中,FLI的研究员大卫.帕吉斯讨论了他构建带有价值观的人工智能系统的一些工作。


Teaching AI to make Decisions
教人工智能进行讨论

Imagine you’re sitting in a self-driving car that’s about to make a left turn into on-coming traffic. One small AI system in the car will be responsible for making the vehicle turn, one system might speed it up or hit the brakes, other systems will have sensors that detect obstacles, and yet another system may be in communication with other vehicles on the road. Each system has its own goals — starting or stopping, turning or traveling straight, recognizing potential problems, etc. — but they also have to all work together toward one common goal: turning into traffic without causing an accident.

想象一下你正坐在一辆自我驱动的无人车中,即将左转面对迎面而来的车流。车内的人工智能系统将会负责控制转向的过程,一个系统也许会加速或者是刹车,另一些系统将会通过传感器探测障碍物,还有一些系统也许会和路上的其他车辆进行通信。每一个系统都有其自己的目标 — 启动或者停止,转向或者直接行驶,分辨出潜在的问题等等。— 但是他们都需要被整合到一起朝向同一个目标:在不发生事故的情况下转换为正常的交通。

Harvard professor and Future of Life researcher, David Parkes, is trying to solve just this type of problem. Parkes told FLI, “The particular question I’m asking is: If we have a system of AIs, how can we construct rewards for individual AIs, such that the combined system is well behaved?”

哈佛教授,未来生活研究员,大卫.帕吉斯,正在试图解决这一类型的问题。帕吉斯告诉FLI,“实际上我正要解决的问题是:如果我们有一个人工智能系统,我们如何才能为其中每一个独立的系统构建奖励模式,以便于整合的系统可以完美的运行?”

Essentially, an AI within a system of AIs—like that in the car example above—needs to learn how to meet its own objective, as well as how to compromise so that it’s actions will help satisfy the group objective. On top of that, the system of AIs needs to consider the preferences of society. For example, it needs to determine if the safety of the passenger in the car or a pedestrian in the crosswalk is a higher priority than turning left.

尤其是,一个人包含很多子系统的工智能系统 — 比如上面无人车的例子中说的 — 需要学会怎样才能完成自己的目标,同时还要学会妥协以便于其子系统的行为能对完成宏观的目标提供帮助。在这些问题之上,包含多个子系统的人工智能需要将社会的价值偏好考虑进来。比如,当其需要决定在转弯的过程中究竟是车内乘客的安全更为重要还是人行道上行人的安全更重要。

Because environments like a busy street are so complicated, an engineer can’t just program an AI to act in some way to always achieve its objectives. AIs need to learn proper behavior based on a rewards system. “Each AI has a reward for its action and the action of the other AI,” Parkes explained. With the world constantly changing, the rewards have to evolve, and the AIs need to keep up not only with how their own goals change, but also with the evolving objectives of the system as a whole.

因为像忙碌的街道这样的环境太复杂,工程师无法通过简单的编程告诉人工智能去如何通过某种特定的行为模式去反应,同时又总能完成其目标。人工智能需要通过基于奖励的系统去学习合适的行为。“每一个人工智能的行为都有自己的激励模式,同时其他系统也有。”


Making an Evolving AI
构建一个自我进化的智能系统

The idea of a rewards-based learning system is something most people can likely relate to. Who doesn’t remember the excitement of a gold star or a smiley face on a test? And any dog owner has experienced how much more likely their pet is to perform a trick when it realizes it will get a treat. A reward for an AI is similar.

基于奖励的学习系统的想法其实也和大多数人的经历相关。谁不记得在考试之后得到金色小星星或者一个笑脸时的兴奋呢?而且每一个有宠物狗的主人都有这样的经验:当宠物狗发现某种行为可以获得奖励的时候,它们是多么愿意执行某种特定的行为。对于人工智能的奖励也是类似的道理。

A technique often used in designing artificial intelligence is reinforcement learning. With reinforcement learning, when the AI takes some action, it receives either positive or negative feedback. And it then tries to optimize its actions to receive more positive rewards. However, the reward can’t just be programmed into the AI. The AI has to interact with its environment to learn which actions will be considered good, bad, or neutral. Again, the idea is similar to a dog learning that tricks can earn it treats or praise, but misbehaving could result in punishment.

在设计人工智能系统的时候有一种常用的技术 — 强化学习。在强化学习的作用下,当人工智能采取行动的时候,其会获得积极的或者消极的反馈。而且其随后就会通过最优化其行为去获得更为积极的奖励。然而,这奖励不能直接被编程到AI的系统中。这些人工智能必须要和周围的环境进行互动,去学习哪些行为被认为是好的,坏的或者中性的。同样,这个想法和训练宠物狗的过程类似:执行某种行为获得奖励或者赞美,做错了则可能会引发惩罚。

More than this, Parkes wants to understand how to distribute rewards to subcomponents (the individual AIs) in order to achieve good system-wide behavior. How often should there be positive (or negative) reinforcement, and in reaction to which types of actions?

不仅如此,帕吉斯想要理解如何通过将奖励分发给子系统(独立的人工智能)以便于能够产生系统层面的合理行为。正向强化或者负向强化的频率要怎么控制,而且应该基于哪种行为去强化呢?


"Rather than programming a reward specifically into the AI, Parkes shapes the way rewards flow from the environment to the AI in order to promote desirable behaviors as the AI interacts with the world around it."“相比于将特定的奖励机制编程到人工智能,帕吉斯重塑了环境对人工智能施加奖励的途径,这样一来就可以在系统和环境互动的过程中提升目标行为的预期。”

For example, if you were to play a video game without any points or lives or levels or other indicators of success or failure, you might run around the world killing or fighting aliens and monsters, and you might eventually beat the game, but you wouldn’t know which specific actions led you to win. Instead, games are designed to provide regular feedback and reinforcement so that you know when you make progress and what steps you need to take next. To train an AI, Parkes has to determine which smaller actions will merit feedback so that the AI can move toward a larger, overarching goal.

例如,如果你正在玩一款视频游戏,但是没有分数,没有命数限制或者其他的成功或者失败的提示,你也许会满世界跑打外星人,杀怪玩,也许最后游戏被你通关了,但是你不知道具体什么样的行为可以让你获得胜利。相反,游戏通常被设计成频繁的进行反馈和强化,这样你可以知道,你正在获得一定的进展而且你知道下一步你要做什么。为了训练一个人工智能系统,帕吉斯必须决定哪些小的行为会产生反馈以至于人工智能能够朝向更加宏观的目标。

But this is all for just one AI. How do these techniques apply to two or more AIs?

但是这只是对于单个人工智能而言。这些技术怎样被应用于两个或者更多的人工智能混合的系统呢?


Gaming the System
用博弈训练整个系统

Much of Parkes’ work involves game theory. Game theory helps researchers understand what types of rewards will elicit collaboration among otherwise self-interested players, or in this case, rational AIs. Once an AI figures out how to maximize its own reward, what will entice it to act in accordance with another AI?

帕吉斯很大一部分工作包含博弈论。博弈论能帮助研究者理解哪些奖励的类型会引出“自私的”局中人的合作行为,在当前的情况中,是理性的人工智能系统。一旦人工智能搞清楚如何去让自身的奖励最大化,什么会影响其系统和其他人工智能系统的协同呢?

To answer this question, Parkes turns to an economic theory called mechanism design.

为了解答这些问题,帕吉斯转向了对一种叫机制设计的的经济学理论的研究。

Mechanism design theory is a Nobel-prize winning theory that allows researchers to determine how a system with multiple parts can achieve an overarching goal. It is a kind of “inverse game theory.” How can rules of interaction – ways to distribute rewards, for instance – be designed so individual AIs will act in favor of system-wide and societal preferences? Among other things, mechanism design theory has been applied to problems in auctions, e-commerce, regulations, environmental policy, and now, artificial intelligence.

机制设计理论是一个获得过诺奖的理论,这个理论允许研究者去决定一个多部分组成的系统去如何获取宏观的目标。这是一种“反向博弈论”。人工智能系统之间的互动规则 — 在这里是发放奖励的方法或者途径 — 如何被设计用来推动系统层面或者社会层面的偏好?在这些问题中,机制设计理论已经被应用到很多实际的领域中,电子商务,法律法规,环境政策,现在,人工智能。

The difference between Parkes’ work with AIs and mechanism design theory is that the latter requires some sort of mechanism or manager overseeing the entire system. In the case of an automated car or a drone, the AIs within have to work together to achieve group goals, without a mechanism making final decisions. As the environment changes, the external rewards will change. And as the AIs within the system realize they want to make some sort of change to maximize their rewards, they’ll have to communicate with each other, shifting the goals for the entire autonomous system.

帕吉斯人工智能系统的工作和机制设计理论的不同在于后者要求某种机制或者管理者来监督整个系统。而在自动驾驶汽车或者无人车的例子中,其中包含的人工智能系统需要一起协作区完成整个团体的目标,并不需要一个机制去做最终的决策。当环境改变的时候,内部的奖励也会改变。并且当这些人工智能系统意识到他们需要做一些改变,以减少其奖励,他们会相互进行通信,将目标和整个自动驾驶系统做整合协调。

Parkes summarized his work for FLI, saying, “The work that I’m doing as part of the FLI grant program is all about aligning incentives so that when autonomous AIs decide how to act, they act in a way that’s not only good for the AI system, but also good for society more broadly.”

帕吉斯对FLI总结了其工作,他说,“我正在做的事情,作为FLI整个项目的一部分,其实就是协调激励以便于自动的人工智能系统可以决定如何去行动,而且不仅仅是基于为了自身系统的利益,更要基于广义上社会的利益。”

Parkes is also involved with the One Hundred Year Study on Artificial Intelligence, and he explained his “research with FLI has informed a broader perspective on thinking about the role that AI can play in an urban context in the near future.” As he considers the future, he asks, “What can we see, for example, from the early trajectory of research and development on autonomous vehicles and robots in the home, about where the hard problems will be in regard to the engineering of value-aligned systems?”

帕吉斯同时也参与了“人工智能百年研究”项目,他这样解释自己的工作“在FLI的研究为自己提供了一个更加宽广的角度,以便思考在未来的城市生活中人工智能可以扮演怎样的角色。”他同时也对未来有所顾虑,他问道,“我能可以看到什么,比如,早期的研究路径和自动驾驶汽车以及家庭机器人,包括在有价值倾向的人工智能系统工程中,困难的问题又会在什么地方出现呢?”

让人工智能完美运行:我们怎么给人工智能系统赋予价值观?相关推荐

  1. 人工智能ai下海百度云_人工智能系统存在偏差,云可用性使情况更糟

    人工智能ai下海百度云 根据Gartner Research的数据,从现在到2022年, 由于数据,算法或负责管理这些项目的团队的偏见 ,有85%的AI项目将交付错误的结果 . 此外,盖洛普民意调查机 ...

  2. 破解人工智能系统的四种攻击方法!

    来源:未来科技前沿 没有人喜欢早上起床,但现在人工智能驱动的算法可以设置我们的闹钟.管理我们家中的温度设置以及选择适合我们心情的播放列表,贪睡按钮的使用越来越少.人工智能安全辅助系统使我们的车辆更安全 ...

  3. 一种通用强人工智能系统的实现

    官网链接 面向对象思想与人工智能 引言 人工智能系统,从概念的诞生起到现在(二十一世纪二十年代),已经有了一定的进步和发展,但是依然面对很多难以克服的困难. 缺乏知识表达 现在的人工智能系统,没有一个 ...

  4. Yann LeCun:掌舵Facebook人工智能 | 完美人物志

     极客Yann LeCun:掌舵Facebook人工智能 | 完美人物志                             Facebook人工智能研究部门(FAIR)负责人Yann LeC ...

  5. 人工智能(8)---一文读懂人工智能产业链:基础技术、人工智能技术及人工智能应用

    一文读懂人工智能产业链:基础技术.人工智能技术及人工智能应用 概要:针对人工智能产业链,主要有三个核心:基础技术.人工智能技术及人工智能应用,本文将从主要从这三个方面进行梳理 人工智能(Artific ...

  6. 人工智能ai思维_人工智能系统如何学习创造性思维

    人工智能ai思维 In such a hot situation of artificial intelligence, there have been many examples of using ...

  7. 中国人工智能行业运行环境分析与投资战略建议报告2022-2028年版

    [报告目录]:     第一章 人工智能的基本介绍 1.1 人工智能的基本概述 1.1.1 人工智能的内涵 1.1.2 人工智能的分类 1.1.3 人工智能的特征 1.1.4 人工智能关键环节 1.1 ...

  8. csgo怎么通过人机验证_让你的MacBook Pro完美运行CSGO!(图解版)

    背景:学生党一枚,使用机器为MacBook Pro 2017底配版,累计游戏时间大约300小时. 先说结论:mbp是可以在保证画面质量的情况下流畅地进行游戏(但个人建议在没有外置散热设备的情况下不要长 ...

  9. Matlab R2016a中如何完美运行“mex-setup”

    接上一篇 "Matlab R2016a中运行'mex -setup',错误使用 mex 未找到支持的编译器或 SDK"的解决办法 上篇博文提到: (4)使用MATLAB将MW_MI ...

最新文章

  1. django model filter_Django框架使用流程(四)
  2. 在Centos6.5下安装AR8161网卡驱动
  3. 我总结的js性能优化的小知识
  4. 深入浅出设计模式原则之依赖倒置原则
  5. Eclipse配置CAS client
  6. android mock测试资源,Android 单元测试 --Mock 及 Mockito
  7. different styles of button in footer area
  8. 30万手表推荐_今年六十岁生日,儿子说要送只30万的手表,请问有哪些推荐?...
  9. IoTDI2019论文列表
  10. CentOS x86_64系统手动释放内存
  11. 大数据_Flink_流式处理_简介_Flink是什么---Flink工作笔记0001
  12. kettle升级jetty10实验(未完成jaas认证)
  13. oracle中导入,ORACLE数据导入
  14. 6U VPX 超高速信号采集处理板(XC7K325T + 4 片DSP TMS320C6678)
  15. 编译原理 上下文无关文法
  16. Sentaurus TCAD 2013 在RedHat7.0 Linux系统的安装教程
  17. VS Code 快捷打开(localhost)PHP页面
  18. 阴阳师魂十挂机实现(后台句柄截屏+鼠标点击)
  19. 数据表为什么又叫透明表?
  20. Linux 中有效用户组和初始用户组

热门文章

  1. 网页 插件 html,HTML 插件
  2. 七、手写实现决策树算法
  3. 华为修改优先级命令_华为LTE重选参数命令简介
  4. 5. 有 2*n 的一个长方形方格,用一个1*2 的骨牌铺满方格
  5. mindspore 1.3.0版本GPU环境下源码编译前的准备工作——依赖环境的安装
  6. C语言程序设计博客作业07
  7. 计算机游戏中屏幕上显示的,电脑在玩全屏游戏的时候显示屏老是出现无信号
  8. python 字符串结束符_python字符串以反斜杠结尾
  9. 计算机师徒结对方案,师徒结对计划
  10. 在观察研究中减少数据偏倚:倾向评分(一)