
Science fiction author Isaac Asimov famously predicted that weāll one day have to program robots with a set of laws that protect us from our mechanical creations. But before we get there, we need rules to ensure that, at the most fundamental level, weāre developing AI responsibly and safely. At a recent gathering, a group of experts did just that, coming up with 23 principles to steer the development of AI in a positive directionāand to ensure it doesnāt destroy us.
The new guidelines, dubbed the 23 Asilomar AI Principles, touch upon issues pertaining to research, ethics, and foresightāfrom research strategies and data rights to transparency issues and the risks of artificial superintelligence. Previous attempts to establish AI guidelines, including efforts by the IEEE Standards Association, Stanford Universityās AI100 Standing Committee, and even the White House, were either too narrow in scope, or far too generalized. The Asilomar principles, on the other hand, pooled together much of the current thinking on the matter to develop a kind of best practices rulebook as it pertains to the development of AI. The principles arenāt yet enforceable, but are meant to influence the way research is done moving forward.
Artificial intelligence is at the dawn of a golden era, as witnessed by the emergence of digital personal assistants like Siri, Alexa, and Cortana, self-driving vehicles, and algorithms that exceed human capacities in meaningful ways (in the latest development, an AI defeated the worldās best poker players). But unlike many other tech sectors, this area of research isnāt bound by formal safety regulations or standards, leading to concerns that AI could eventually go off the rails and become a burden instead of a benefit. Common fears include AI replacing human workers, disempowering us, and becoming a threat to our very existence.
To address these and other issues, the Future of Life Institute recently brought together dozens of experts to come up with a set of core principles to nudge AI development in positive directions and steer us clear of apocalypse scenarios. Attendees of the conference came from diverse backgrounds, including engineers, programmers, roboticists, physicists, economists, philosophers, ethicists, and legal scholars.
Advertisement
The 23 Asilomar AI Principles have since been endorsed by nearly 2,300 people, including 880 robotics and AI researchers. Notable supporters include physicist Stephen Hawking, SpaceX CEO Elon Musk, futurist Ray Kurzweil, and Skype co-founder Jaan Tallinn, among many others.

āOur hope wasāand will be going forwardāto involve a very diverse set of stakeholders, but also to ground the process in the expert knowledge of the AI developers who really understand what systems are being developed, what they can do, and where they may be going,ā noted participant Anthony Aguirre, a physicist at the University of California, Santa Cruz.
Advertisement
Discussions were often contentious during the conference, but a high level of consensus eventually emerged. To set a high bar, the FLI organizers only accepted a principle if at least 90 percent of the attendees agreed with it.
The principles were organized into three sectionsāresearch issues, ethics and values, and longer-term issues. Under research, principles included the need to create ābeneficial intelligenceā as opposed to āundirected intelligenceā (more on this in just a bit), and an admonition for AI developers to maintain a healthy dialogue with policy-makers. Ethics and values included the need to ensure safety at all stages of AI development, the instilling of human values into a machine mind, the avoidance of an AI arms race, and the need to maintain human control. Long-term considerations included risk assessments, control measures, and adherence to the so-called ācapability cautionāāa warning that we should never underestimate the potential power of advanced AI.
Advertisement
Patrick Lin, a conference attendee and the the director of the Ethics + Emerging Sciences Group at California Polytechnic State University, says the Asilomar AI principles sprung from āa perfect storm of influencesā he hadnāt encountered before. āThis was a standard-setting exercise in a field that has no cohesive identity, making the exercise much more difficult,ā he told Gizmodo.
While perfect consensus was not expected, everyone at the conference could agree that AI is going to be impactful in ways weāve never seen before. āTo think AI merely automates human decisions is like thinking electricity is just a replacement for candles,ā Lin said. āGiven the massive potential for disruption and harm, as well as benefits, it makes sense to establish some norms early on to help steer research in a good direction, as opposed to letting a market economy thatās fixated mostly on efficiency and profit to shape AI.ā

Advertisement
Some of the items on the list are a bit of a no-brainer, such as the need to avoid an AI arms race and to ensure safety. But other items, though simply stated, are considerably more complex. Take the injunction to develop ābeneficial intelligence.ā Itās vague, but Aguirre says thatsā a strength at this early juncture.
āWe hope people will interpret it in many ways,ā he told Gizmodo. āThe crucial aspect of this principle, as I see it, is that benefit is part of the equation. That is, AI developers are actively thinking about what the ramifications of the tools they are developing might be, rather than just making the tools better, stronger, and faster.ā
Another complicated issue is the one about maintaining human control. According to this principle,
Humans should choose how and whether to delegate decisions to AI systems, to accomplish human-chosen objectives.
Advertisement
Trouble is, AI is becoming increasingly better and more efficient at many tasks, making us more inclined to offload those responsibilities. Today, for example, algorithms have taken over as our stock traders. Thereās concern that battlefield robots will eventually be ātaken out of the loop,āand given the autonomy to kill enemy combatants of their own accord. Aguirre says we will cede some decisions to AI systems, but we should do so with care.
āI already listen to Google Maps, for example, which helps me decide what route to take through traffic,ā said Aguirre. āBut recently I was caught in a mudslide-caused traffic jam for an hour with zero movement, while Google doggedly informed me it was a 48 minute trip. I eventually ignored the instructions and drove home a very roundabout way that saved me hours.ā He says itās important that we keep our eye on what decisions we are delegating and which ones we really want to keep for ourselves.
Advertisement
Many of the principles on the new list fall into the āeasier said than doneā category. For instance, thereās an injunction for AI developers to be open and transparent about their research. Given the high degree of competition in the tech sector, that seems like a pipe dream. Aguirre agrees itās an important and difficult issue.
āOne positive sign is the partnership on AI, envisaged in part as a group that can create collaborative dynamics where the benefit of participating counter the competitive ācostā of sharing information,ā he said. āWhile big advances may come from scrappy little startups with small resources, thus far much of the real advance has been in powerhouses like Google/Deepmind, IBM, Facebook, etc.ā
In terms of getting developers to comply with these principles, Lin says people tend to have a hard time understanding the value of norms, leading to calls for the āfangs of law.ā Failure to disclose security faults in an AI, for example, could lead to stiff fines and penalties.
Advertisement
āEven without the teeth of legal punishment, norms are still powerful in shaping behavior, whether itās at home, or in the classroom, or in international relations,ā Lin said. āFor instance, the US is not a signatory to the Ottawa Treaty that bans anti-personnel landmines. But we still abide by it anyway, because breaking such a strong norm has severe political and therefore economic consequences. The global condemnation that would follow is enough of a deterrent (so far).ā
But this might not always be the case, requiring strictāand even international-levelāoversight. Itās not impossible to imagine a company like Facebook or Google developing a superintelligent AI that can simultaneously boosts profits to unprecedented levels and pose a catastrophic threat to humanity. With profits soaring, a big company might not care about violating ānormsā and having to face global condemnation. Alarmingly, we have no provision for this sort of thing.
Another principle that will be hard, if not impossible, to implement is the one stating that:
AI systems should be designed and operated so as to be compatible with ideals of human dignity, rights, freedoms, and cultural diversity.
Advertisement
This concept, known as āvalue alignment,ā is an often-cited strategy to keep AI safe, but thereās no universally agreed-upon way to define a āhuman value.ā
Lin agrees that value alignment is a particularly thorny problem. āAt Asilomar, there was a lot of talk about aligning AI with human values, but little discussion of what these human values are,ā he said. āTechnologists understandably are uncomfortable confronting this prior question, but itās an essential one. If you have a perfectly aligned AI [i.e. an AI perfectly aligned with human values] but calibrated to the wrong human values, then you just unleashed a terrible technology or even weapon.ā As a classic (but extreme) example, an AI could be told that making lots of paperclips is goodāand it proceeds to turn the entire planet into paperclips. As Lin fears, a seemingly innocuous value could pose a serious threat.
Advertisement
Finally, thereās the issue of artificial intelligence as an existential threat. On that topic, the Asilomar guidelines speak to the importance of overseeing any machine capable of recursive self-improvement (i.e. the ability of an AI to continually improve its own source code, leading to a potential ārun awayā effect, or program itself with a set of features that run contrary to human values). The challenge there will be in implementing the required safeguards, and ensuring that the machine stays āin boundsā when it comes to certain modifications. This is not going to be easy.
A final important item on the list is the ācapability caution.ā Given that thereās no consensus on how powerful AI might become, it would be wise to refrain from making grand proclamations about its maximum potential. In other words, we should never, ever, underestimate the power of artificial superintelligence.
āNo current AI system is going to āgo rogueā and be dangerous, and itās important that people know that,ā Aguirre said. āAt the same time, if we envision a time when AIs exist that are as capable or more so than the smartest humans, it would be utterly naive to believe that the world will not fundamentally change. So how seriously we take AIās opportunities and risks has to scale with how capable it is, and having clear assessments and forecastsāwithout the press, industry or research hype that often accompanies advancesāwould be a good starting point.ā
Advertisement
For now, this list of principles is just thatāa list of principles. Thereās no provision to have these guidelines enforced or updated, nor is there any call for institutional or governmental oversight. As it stands, developers can violate these best-of-breed recommendations, and face no consequences.
So will these 32 principles keep us safe and protect us from an AI apocalypse? Probably not, but the purpose of this exercise was to pool together the best ideas on the matter in hopes of inspiring something more concrete. As imperfect and idealistic as these principles may be, weāre better off having them.
Advertisement