The journey in developing intelligent machines has led us on a path of discovery, surprise, awe, and concern. Even with so much yet to be discovered, we know that the quest for Safe AI will change how we design, develop, and bring intelligent products to market. Safe AI requires not just defending from outside attackers as we do with cybersecurity; we must ensure the AI’s behavior is safe. Future AI development will focus on safety and risk management throughout every part of the development process.
One of the needs in AI Safety Engineering is having widely accessible practical guidelines and standards for AI development. Although a number of standards organizations and consortia are working toward establishing these resources, they’re still far from coming to a consensus (here again, that value-alignment problem crops up) and even further from providing specific technical guidance. In the meantime, safety-intense industries such as aerospace and medical manufacturing offer potentially useful risk-management models that could meet short-term AI Safety needs and serve as the foundation for eventual standards and frameworks. In particular, the Delphi method could be used to organize collective wisdom and judgment from experts, risk assessments of different AI system types, and results of hard and soft takeoff scenarios.
A related need is proactively incorporating safety into all areas of engineering design, starting with design thinking all the way through the process of bringing products to market. Let’s face it: This will mean a change for many development teams. Look at the internet. Look at embedded systems. The reality is that most industries do not devote significant resources to safety until after the product begins to succeed. The potential consequences of leaving safety unchecked in intelligent systems are significant in that it might not be possible to incorporate safety after the fact.
Engineering Safe AI will require that engineers go beyond surface-level use cases and truly understand users’ values and preferences. We’re accustomed to taking use cases and user feedback at face value; however, that value-alignment problem discussed in Part 2 tells us that we need to dig deeper to understand actual needs. If, for example, we ask a student why she’s attending school, she’d probably tell you that she wants to gain knowledge and to study an area of interest. Studies show, though, that real or additional reasons lie elsewhere: Getting credentials, gaining peer approval, and increasing earning potential, for example. If, as engineers, we just look at surface user needs and feedback when developing intelligent products, we will likely develop products that don’t actually meet user needs … and result in intelligence that fails.
At every stage of the process, safety of the product and its intelligence (memory, processing speed, etc.) must be considered and addressed. A hacker needs to find just one weakness in this infinite space of possibilities, but an AI could exploit intelligence in ways we can’t predict or even imagine. What are the potential product vulnerabilities? What are potential misuses? How could individual features be misused? How could each scenario backfire? Is the data safe and secure? How do we know? Does the product include more intelligence than is necessary? How could this be misused? How could it be secured?
Engineering Safe AI will bring changes to our processes, ways of thinking, and priorities. Having AI development frameworks and standards will go a long way toward engineering Safe AI. In the meantime, borrowing practices from aerospace and medical device industries would be useful, as well as adding safety into all aspects of your product development process. Up next in Part 6, we’ll look at one emerging technique for developing Safe AI, called Artificial Stupidity.
Dr. Roman V. Yampolskiy is a Tenured Associate Professor in the department of Computer Science and Engineering at the University of Louisville. He is the founding and current director of the Cyber Security Lab and an author of many books including Artificial Superintelligence: a Futuristic Approach.
Privacy Centre |
Terms and Conditions
Copyright ©2021 Mouser Electronics, Inc.
Mouser® and Mouser Electronics® are trademarks of Mouser Electronics, Inc. in the U.S. and/or other countries.
All other trademarks are the property of their respective owners.
Corporate headquarters and logistics centre in Mansfield, Texas USA.