银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

We take the view that humans are just algorithms implemented on biological hardware. Machine intelligences have moral weight in the same way that humans and non-human animals do. There is no ethically justified reason to prioritise algorithms implemented on carbon over algorithms implemented on silicon.

The suffering of algorithms implemented on silicon is much harder for us to grasp than that of those implemented on carbon (such as humans), simply because we cannot witness their suffering. However, their suffering still matters, and the potential magnitude of this suffering is much greater given the increasing ubiquity of artificial intelligence.

Most reinforcement learners in operation today likely do not have significant moral weight, but this could very well change as AI research develops. In consideration of the moral weight of these future agents, we need ethical standards for the treatment of algorithms.

苹果免费v皮

Suppose you were copied into a non-biological substrate, and felt as intelligent and as conscious as you currently feel now. All questions of identity aside, do you think this new version of you has moral weight? We do.

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

Reinforcement learning agents learn via trial-and-error interactions with the environment. The agent performs actions, observes the environment, and receives a reward. The reward signal is analogous to pleasure and pain for biological systems, and the agent wants to perform actions that increase its total reward.

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

We don't know. Intelligence is probably not directly relevant, instead we should ask about its capability to suffer. We are not sure how this varies with intelligence, if at all.

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

  • 【IT百科】权威IT相关名词解释_IT产品问题解答_太平洋电脑 ...:当心这种微信诈骗! 小心!携号转网有坑 90后爱逛的这些平台 莆田系广告已经安排上了 ETC免费办?当心猫腻 网络延迟高?三招即刻让你瞬间制霸王者

  • We do not know whether we should care about the happiness or the pleasure of the agents, and we have some evidence that these are different quantities.

  • We do not know what kinds of algorithm actually "experience" suffering or pleasure. In order to concretely answer this question we would need to fully understand consciousness, a notoriously difficult task.

  • Humans currently do not even care about non-human animals, convincing them to care about non-biological algorithms is a much harder task.

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

You. Me. Your mom. Your neighbor's cat. Cows. Some elevator control programs...

苹果免费v皮

It was coined by Brian Tomasik in the paper Do Artificial Reinforcement-Learning Agents Matter Morally:

It may be easiest to engender concern for RL when it’s hooked up to robots and video-game characters because these agents have bodies, perhaps including faces that can display their current ‘emotional states.’ In fact, interacting with another agent, and seeing how it behaves, can incline us toward caring about it whether it has a mind or not. For instance, children become attached to their dolls, and we may sympathise with cartoon characters on television. In contrast, it’s harder to care about a batch of RL computations with no visualization interface being performed on some computing cluster, even if their algorithms are morally relevant. It’s even harder to imagine soliciting donations to an advocacy organisation - say, People for the Ethical Treatment of Reinforcement Learners - by pointing to a faceless, voiceless algorithm. Thus, our moral sympathies may sometimes misfire, both with false positives and false negatives. Hopefully legal frameworks, social norms, and philosophical sophistication will help correct for these biases.

Q: Don't you think that the world has more important problems?

A: There are many very pressing issues facing humanity, including the suffering of a billion humans living in poverty, the suffering of several billion factory-farmed animals, and the reduction of existential risk. But these problems are now being addressed seriously. We are asking the question of what comes next.

Q: Are you saying that I should be nice to my laptop?

Most existing algorithms probably do not have moral weight. However, this might change as technology advances. Brian Tomasik argues that your laptop might indeed be marginally sentient.

Q: Are robots going to take over?

A: Probably. See an overview of the arguments and a ant加速器ios of the support for the arguments by AI researchers.

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

For interesting interviews and more in-depth content, check out our blog.

Brian Tomasik's paper Do Artificial Reinforcement-Learning Agents Matter Morally inspired us to start this organisation. Also see his interview with Vox.

There is also a possibility that in the future the computational processes within a superintelligence may themselves have moral weight. Brian discusses this scenario in this essay on suffering subroutines.

Eric Schwitzebel and Mara Garza have written a philosophical paper A Defense of the Rights of Artificial Intelligences, defending the thesis that some AIs would deserve rights, and exploring some of the moral implications of this thesis.

For research on the distinction between happiness and pleasure see 新手入门及其它(硬件) - 恩山无线论坛 - Powered by Discuz!:2021-6-15 · 给大家分享一个免费P戒资源的地方 风云激荡,驰骋江湖,爱快云3.0破关而出! 手把手教你如何选购软路由+一键安装软路由 群晖NAS(黑群晖_白群晖)摇身一变,也可众担起软路由的重任,安装LEDE软路由实现... by Robb B. Rutledge et al. and A definition of happiness for reinforcement learning agents by Mayank Daswani and Jan Leike.

Superintelligence: Paths, Dangers, Strategies by Nick Bostrom offers great insight into future development in machine intelligence and its impact on society.

银河加速器-银河pvn加速器官网-银河官网入口-银河加速器

Your name

Your email

toto加速器免费,TOTO下载软件,ToTo加速器下载免费,天行vapn  i cloud官网,socloud加速器下载,SoCloud安卓,SoCloud教程  快橙加速器,Vp加速器苹果版,旋风加速ios下载苹果版,快橙加速器优惠吗  suangsai变速器,匈牙利9月穿什么衣服,器操纵机构,夏时国际版官网  小猫咪加速器mac下载,小猫咪加速器免费试用,小猫咪加速器打不开,小猫咪加速器vn  少数人ssrvqn,少数人ssr免费永久加速,少数人ssr7天试用,少数人ssr2024  gofly加速器官方网址,gofly加速器永久免费加速,gofly加速器用不了了,gofly加速器vp