Opinion
Artificial Intelligence and Machine Learning Opinion

Dilemma of the Artificial Intelligence Regulatory Landscape

Navigating the complex and varying AI regulatory landscape.

Posted
red wavelike bands surround a striped lighthouse, illustration

When legal regulations get ahead of technological developments, the progress of society may be constrained. Conversely, when technological developments run ahead of legal regulations, unregulated new technologies may harm society, defying technological development’s fundamental purpose.

This is exactly what has happened in the world in the past decade, as technological developments have far outpaced legal regulations. Worse, traditional legal frameworks focus on the relationship between people, whereas we must develop a legal framework to regulate relations between people and intelligent machines in the current era. Integrating AI technologies into human society imposes unique legal challenges without any precedence.

For the first time in history, AI has the potential to generate solutions superior to what a human would expect.8 However, today AI is not bound by ethics, and decisions made by AI may be considered out of line with standards and ethics generally accepted by human society.1 The unforeseeable superiority and the problem of ethical control may create chaos in the AI regulatory landscape. If AI is not adequately regulated, such problems will further exacerbate unforeseen social issues the current legal framework is unprepared for.

As an autonomous driving start-up founded in 2016, we launched commercial autonomous driving operations in the U.S., Europe, Japan, and China to find the most promising market.4 It has been a treacherous journey filled with regulatory roadblocks and land mines. Such a problem is not unique to autonomous driving but universal to all AI applications. Here, we share our firsthand experiences tackling regulatory requirements and attempting to reduce the friction between innovative technologies and regulatory requirements to advance human society with properly regulated AI technologies.

Back to Top

Our Global Commercial Deployment Experiences

Our business model is to provide turnkey solutions to customers. The solution includes hardware such as drive-by-wire chassis and computing units, a full stack of autonomous driving software such as perception and localization, and a user interface.10

In the past four years, when we performed commercial deployments in different regions worldwide, we encountered a broad spectrum of regulatory rules across countries or regions within a country. This lack of regulation standardization has become unhealthy as companies in the AI sector, mostly start-ups, spend a significant amount of budget, time, and effort dealing with different regulatory measures. Our firsthand experiences in working with authorities worldwide are described here.

Deployments in China. In 2017, autonomous driving test permits in China were virtually nonexistent. Only a few governmental documents stated high-level goals and principles in some pilot cities such as Shanghai. The absence of test permit certification pushed start-ups in this field to test their vehicles wildly on highways, where the traffic was more predictable and less complicated, or in rural areas where the population is scarce, both posing fatal risks to the general public.

To minimize compliance and safety risks, we chose to carry out autonomous driving projects in restricted areas, such as factories and school campuses. Since these areas are private, the test pod could go on the roads once approved by the property management. In addition, we set the speed limit at 20mph to cope with safety risks and designed several redundant perception systems to avoid accidents.7


For the first time in history, AI has the potential to generate solutions that are superior to what a human would expect.


Having accumulated 1.5 months of testing data and different driving scenarios, our company was invited to join an autonomous driving pilot program in Shenzhen, China’s Silicon Valley. We were selected to present the latest autonomous driving technology and global trends in this field in a seminar hosted by the local government, where we seized the opportunity to emphasize the urgency and importance of regulated test drives. After six months, our voice was heard, and we could file our first test permit application. In this case, the whole regulatory process lagged behind the actual technological development, pushing companies to play wild first and file applications later. Furthermore, if a company goes to another city for a new autonomous driving project, the whole process will start from zero.

Deployments in European Union. In late 2018, we were invited to carry out autonomous driving operations in a European city. It took us more than seven months to complete the application process and launch commercial deployment. The first three months focused on technical specifications such as communication bandwidth requirements and visual perception response time. The latter four months focused on functional specifications in more than 40 scenarios, from recognizing an unexpected object in the test field to preparing a backup test plan on rainy days in the dry city. Filing all requested reports, we were granted a restrictive operation permit that lasted a month. In Europe we were strictly not allowed to perform any experiments on the roads, whether public or private, without a valid permit. In addition, there was no room for negotiation between local authorities and us.

Regardless of the actual operation environment, we had to spend extended amount of time and effort to prepare for all scenarios and go through the regulatory process. This imposes a significant financial burden on AI start-ups, which are already constrained on budget and race against time for growth. In a way, the complex and varying regulatory processes across different regions subtly give advantage to well-established and resourceful technology firm over resource-constrained start-ups.

Deployments in Japan. With autonomous driving cases launched in China and Europe, we entered the Japanese market and directly worked with Japan’s Ministry of Land, Infrastructure, Transport, and Tourism (MLIT) to deploy a driverless vehicle.3 The experience in Japan was a combination of China and Europe. We could negotiate with authorities to start from the most straightforward scenarios and reduce paperwork. Meanwhile, the authority closely reviewed technical specifications and risk-mitigation plans in those scenarios.

The application and endorsement process is top-down. We had to work with the national government to gain approval and then with local governments to carry out the operation. We first ran the test operation in a park in Fukuoka under the supervision of MLIT. The operation data was then used to go through a basic safety check to assess the risk of our autonomous driving operations. After the procedure was completed, we were allowed to continue our operation at the new site in a historic park in Nara. The whole operation scale-up process took more than 12 months, and it was worthwhile since we did not have to repeat the initial certification process once we received endorsements from MLIT.


Autonomous driving start-ups must deal with various compliance issues from day one, which imposes a significant financial burden on autonomous driving start-ups.


Deployments in the U.S. In contrast to Japan, we had to take a bottom-up approach for deployments in the U.S. The local governments are responsible for inner-city transportation, and these local governments have the authority to grant autonomous driving operation permits. Since the state and federal governments had limited guidance on the deployment of autonomous driving services, we had to deal with each city individually to expand our operation, which is an inefficient and costly process given the regulatory differences among cities. In our cases in the U.S., the test data and safety assurance plans in California did not give us privileges when we applied for the test permit in Indiana. Although such an arrangement may be adequate to mitigate people’s safety risk in different states with different traffic densities,9 there could be room to improve process standardization and transparency, which would save AI start-ups’ resources and, in turn, help advance the technology.

Back to Top

The Dilemma

The accompanying figure summarizes the basic process of autonomous driving deployment in the four areas discussed previously. The deployment process consisted of four essential parts: Application filing to initiate the conversations, national-level approval, local-level approval, and commercial deployment. There were no explicit instructions to follow, and it took at least eight months to reach commercial deployment after technical readiness. In addition, the broad spectrum and the opacity of regulatory requirements consumed a significant amount of the company’s resources, costing 42% of the annual budget to deal with regulatory issues, compared with the average 13% of the budget in software companies.6

uf1.jpg
Figure. Process of getting a test permit for autonomous driving in different regions.

Traditionally, software companies may take an incremental approach to comply with regulations. They first launch a minimally viable product, collect feedback from pilot users, and then reiterate several times until the product rolls out to the general public. When more compliance issues arise, they are addressed on the fly with software updates.

Conversely, autonomous driving start-ups must deal with various compliance issues from day one, which imposes a significant financial burden on autonomous driving start-ups. Dealing with regulatory uncertainty is a ubiquitous problem among many AI companies, as exemplified by Microsoft’s retracting its facial recognition products.2 We believe this dilemma is caused by a lack of standardization in the AI product deployment process and a lack of understanding of AI technologies in the regulatory bodies.

Without a properly regulated environment, AI companies cannot thrive and transform their latest technology into successful products. When regulatory requirements are absent, players in the field may go wild and put both the company’s future and public safety at stake, jeopardizing technological advancement in the long run. When the requirements are too tight, AI companies, especially those start-ups that do not have sufficient budget for compliance, would be exhausted with filing repetitive paperwork to different offices.

Back to Top

Recommendations for AI Start-Ups

Our lessons can help AI companies, especially start-ups with limited resources, to navigate through various regulatory barriers. While the specific questions being asked by regulators varied from time to time, we have found the key to settling concerns is to clearly convey the message that potential benefits outweigh relevant risks. These recommendations have been proven effective in our own deployment projects across multiple countries.


Without a properly regulated environment, AI companies cannot thrive and transform their latest technology into successful products.


First, cross-function communication and information sharing are critical to unifying the team to deal with external complexities, as complex regulatory requirements often cause chaos within an organization. For instance, as we carried out deployment projects, our field engineers often needed to work directly with regulators to demonstrate technical progress, fix deployment problems, and address their concerns. Engineers with cross-functional experiences tend to outperform those without, as the former ones are capable of communicating effectively with regulators who have limited technical background.

Second, show strong evidence to support the benefits of the project. Most regulators welcome innovations, and they are eager to integrate innovative technologies into their economies. They look for strong evidence that the technology indeed works, and they look for evidence from credible sources, such as publications from leading journals in the field or reports released by leading research institutes. Hence, we suggest AI start-up companies collaborate with academic partners to perform scientific research to verify the project’s benefits. Industry-academia collaboration is essential to foster the AI industry more than ever before.

Third, present a concrete risk mitigation plan to regulators. Most regulatory actions are designed to mitigate risks. The local authority will often carry the blame if an innovative project goes wrong. Therefore, regulators would try their best to mitigate risks. Presenting a clear risk-mitigation plan and success stories of deployment case studies elsewhere will undoubtedly improve the regulator’s confidence in the project. While in our own experiences, presenting the cases in China and Europe helped us obtain approval for operation in Japan. For instance, development and awareness of technical standards, such as Safety of the Intended Functionality (SOTIF) for Autonomous Driving,7 is an effective method to improve product creditability and to boost regulators’ confidence.

Back to Top

Conclusion

Our firsthand experiences of dealing with a broad spectrum of regulatory requirements pinpoint the dilemma of the AI regulatory landscape today. We have provided some practical recommendations for AI start-ups to work efficiently and effectively within the regulatory framework to address this problem. We hope these recommendations make AI start-ups more adaptable to the regulatory landscape and hope for standardization in the AI regulatory process, which will make the AI industry much more efficient and effective.

 

    1. Awad, E. et al. The moral machine experiment. Nature 563, 7729 (2018), 59–64.

    2. Hill, K. Microsoft plans to eliminate face analysis tools in push for 'responsible A.I.' The New York Times (June 21, 2022); https://bit.ly/43P31ai

    3. Japan Smart Island Pilot Project; https://bit.ly/43JdtjE

    4. Liu, S. Critical business decision making for technology startups: A PerceptIn case study. IEEE Engineering Management Review 48, 4 (2020), 32–36.

    5. Liu, S. and Gaudiot, J.L. Autonomous vehicles like self-driving technologies should start small, go slow. IEEE Spectrum 57, 3 (2020), 36–49.

    6. Poneman Institute. The True Cost of Compliance with Data Protection Regulations; https://bit.ly/44JxU1c

    7. Shaoshan, L. Safety of the intended functionality (SOTIF) for autonomous driving. Forbes (Sept. 23, 2022); https://bit.ly/43JNUPt

    8. Scherer, M.U. Regulating artificial intelligence systems: Risks, challenges, competencies, and strategies. Harvard Journal of Law and Technology 29, 2 (Spring 2016); https://bit.ly/3YdCqSY

    9. Widen, W.H. and Koopman, P. Autonomous vehicle regulation and trust. UCLA Journal of Law and Technology (Spring 2022); https://bit.ly/44IONcp

    10. Yu, B. et al. Building the computing system for autonomous micromobility vehicles: Design constraints and architectural optimizations. In Proceedings of the 2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO) (Oct. 2020), 1067–1081.

Join the Discussion (0)

Become a Member or Sign In to Post a Comment

The Latest from CACM

Shape the Future of Computing

ACM encourages its members to take a direct hand in shaping the future of the association. There are more ways than ever to get involved.

Get Involved

Communications of the ACM (CACM) is now a fully Open Access publication.

By opening CACM to the world, we hope to increase engagement among the broader computer science community and encourage non-members to discover the rich resources ACM has to offer.

Learn More