International Trends in AI Safety and Governance - 東京カレッジ

International Trends in AI Safety and Governance

When:
2024.03.28 @ 10:00 – 12:00
2024-03-28T10:00:00+09:00
2024-03-28T12:00:00+09:00

Event Report

International Trends in AI Safety and Governance

 On March 28, 2024, the University of Tokyo's Institute for Future Initiatives and the University of Tokyo's Tokyo College held a public event titled “International Trends in AI Safety and Governance.” The event was held at the SMBC Academia Hall in the International Academic Building at the Hongo Campus of the University of Tokyo and was attended by 178 people online and about 30 people at the venue.

 While discussions on the safety of AI have been developing domestically and internationally with the spread of generative AI, there are various types of discussions on "safety" and their countermeasures. In addition, with the AI Safety Institute being established in the U.K., U.S., and Japan, it is important to understand the types of "safety" and countermeasures specific to Japan as a basis for future international collaboration. The event was attended by experts on AI governance from overseas, and discussions were held on international AI safety and governance trends.

Speakers and Panelists

Merve Hickok: President and Research Director at Center for AI & Digital Policy (CAIDP)

Cyrus Hodes: Lead, SAFE project at the Global Partnership on AI (GPAI)

Inma Martinez: Chair of the Multi-stakeholder Experts Group, Global Partnership on AI (GPAI)

Michael Sellitto: Head of Global Affairs at Anthropic

Yoichi Iida: Special Negotiator for Information and Communications International Strategy, International Strategy Bureau, Ministry of Internal Affairs and Communications

Hideaki Shiroyama: Professor, Institute for Future Initiatives, The University of Tokyo

Arisa Ema: Associate Professor, Tokyo College, University of Tokyo (Moderator)

(1) Opening remarks

 Professor Hideaki Shiroyama of the University of Tokyo's Institute for Future Initiatives first gave opening remarks. Focusing on the governance of emerging technologies, Professor Shiroyama explained how the Institute's Technology Governance Research Unit has contributed to international discussions on AI safety through its research on Risk Chain Models and participation in the GPAI. In light of recent rapid changes such as the explosive spread of AI and the G7 Hiroshima AI Process, he expressed his hope that this event will serve as a catalyst for discussion in the context of Japan, given the current need to organize issues related to the safety of AI and to develop a system to address them in a manner that is relevant to each site.

(2) Topics from presenters

 First, as an introduction to the discussion points from the panelists, Ms. Inma Martinez of the GPAI mentioned the GPAI's emphasis on equity and inclusion of vulnerable peoples as "AI for all," and the leadership role Japan has played in the GPAI discussion, including these issues. In addition, she introduced that the most recent GPAI activities have been conducted with an emphasis on consensus building toward the realization of common values, a feature of the G7 Hiroshima AI process.

 Ms. Martinez also explained that AI is not limited to automation, but will impact and transform all industrial sectors, and that while the "safety" of such AI can be interpreted in culturally diverse ways around the world, the "trustworthiness" of such AI is "technically functional," and that there is a consensus. It was then emphasized that while the GPAI seeks to build consensus, definitions on AI are no longer valid and should be in line with each country's culture and values and should not be monocultured.

 Next, Mr. Cyrus Hodes, also from GPAI, stated that GPAI is working with multi-stakeholders to guarantee the safety of generated AI.

 Mr. Hodes then noted that one of the risks of generative AI is that as AI systems become more sophisticated, which is bringing risks of misalignment, control and robustness of these systems and where tools addressing these raising issues will become increasingly important (such as audits, evaluations, cybersecurity red-teaming) and where an infrastructure for such alignment needs to be established, he expressed hope for collaboration with the AI Safety Institute. In addition, he mentioned that he expects Japan to cooperate in mapping the various set of tools developed by the global community and contribute to international coordination on AI safety.

 Ms. Merve Hickok of CAIDP then spoke, first explaining that the Center is tasked with providing recommendations on AI policy to governments and international organizations, and training of future AI policy leaders. She then introduced the current state of AI policy in the U.S., which is consistent across Administrations, and the development of binding presidential executive orders for government agencies and voluntary guidelines that can also be used in the private sector. She also explained that the bipartisan agreement on the need for AI regulation in the U.S. is a reflection of the failure to regulate harmful impact of social media. She noted the AI Safety Institute was established in the U.S., and that which ministry is in charge of this type of organization indicates what the nation is focusing on, she stated that in the U.S., unlike in UK, the definition of "safety" is broad and includes the economy and current risks of AI, and therefore, the Department of Commerce is in charge of this type of organization. In addition, she introduced recent initiatives such as the AI Safety Summit by the UK, upcoming AI Summit in France, and the Mini-Virtual Summit in South Korea.

 Ms. Hickok emphasized the importance of "interoperability" to avoid governance fragmentation. However, she also warned about reducing the protections to a minimum number of common elements in the discussion of AI and human rights. She underlined the importance of international collaboration with multi-stakeholder participation, and advancing the elements of the Hiroshima AI process.

 Finally, Mr. Michael Sellitto of Anthropic introduced the company's Responsible Scaling Policy, introduced that under the Responsible Scaling Policy, an AI Safety Levels (ASL), analogous to the biosafety level, is set and safety and security measures are taken according to the degree of risk. He also said that calls for a moratorium on AI development should not be based on abstract risks, but should be considered only when there is concrete evidence that safety or security measures may be insufficient.

 Mr. Sellitto also praised the international code of conduct developed during the Hiroshima AI process as a highly effective framework, and expressed hope that the public and private sectors will work together to monitor commitments and thus increase confidence in the code.

(3) Panel Discussion

 Following the introduction of the above issues, Mr. Iida, Special Negotiator for International Information and Communications Strategy, International Strategy Bureau, Ministry of Internal Affairs and Communications, and Prof. Shiroyama joined a panel discussion moderated by Associate Prof. Ema on the topic of “What is expected of Japan in AI governance.”

 First, Mr. Iida expressed his appreciation for the substantial presentations, including the introduction of overseas case studies, as well as his compliments for the ambitious efforts by each stakeholder to address AI safety. Mr. Iida also stressed the importance of ensuring commonality and interoperability in the diversity of AI policies, while pointing out that even among developed countries, there is still diversity, especially in approaches, as indicated by the comments of each speaker. He also noted Anthropic's voluntary efforts and willingness for international collaboration, which he appreciated and felt reinforced by such efforts.

 Professor Shiroyama posed the question as a discussant, what is safety and why is it important? He then asked for further views on two points: what are the new risks posed by advanced and generative AI that differ from conventional AI, and what does the existence of bipartisan consensus and differences in competent ministries mean when comparing AI policies in different countries?

 In response to the issues raised by Professor Shiroyama, Mr. Sellitto first responded that while there are a wide range of concerns and risks surrounding AI, "safety" in the context of Anthropic's focus is to ensure that AI can be used in a reliable and safe manner.

 Ms. Martinez then noted that the 21st century is the first century in which safety has been brought to all industries but pointed out that "safety" is about preventing harm, not causing harm.

 In response, Ms. Merve pointed out that while the objective function of AI is the starting point for trust and performance evaluation, it is not possible to envision all use cases for general-purpose AI. This makes it harder to manage risk and performance.

 Mr. Hodes also noted that in the era of AGI, any task can be subject to improvement by AI, but values must be maintained by adjusting AI systems in such a society.

 In response to these issues, Mr. Iida explained how the Hiroshima Process was launched to discuss the risks of generative AI but was later added to cover infrastructure systems and advanced AI as well. He also recognized that in international discussions, "safety" and "trust" have been discussed simultaneously, and that discussions on the definition of safety have been avoided, and that a detailed definition is needed in the course of taking concrete measures in the future.

 Associate Professor Ema, the moderator of the session, also raised the point that discussions on safety should be framed not only in terms of the safety of AI itself, but also in terms of the safety realized by AI, such as its use in law enforcement agencies, and the trade-off relationship with other values.

 In response, Mr. Iida noted that both Professor Shiroyama's and Associate Professor Ema's points of view are extremely important, but also expressed the view that the gap between political and administrative actors in terms of minimizing risk while advancing technology-based innovation is not so large. Mr. Iida also reiterated the importance of a multi-stakeholder approach in the AI policy-making process.

 Mr. Hodes agreed with Mr. Iida, pointing to the composition of the U.S. and China as the two giants, and praised Japan's efforts, such as the establishment of the AI Safety Institute, and expressed hope that Japan would play a coordinating role.

 Ms. Merve, while noting the differences in authority among ministries, emphasized the importance of a multi-stakeholder approach because of the need for diverse competencies, and praised Japan’s work to drive commonalities across actors.

 Ms. Martinez noted that even in Europe, the development of regulations pertaining to the Internet has been slow but said that regulations on AI have been developed under a global consensus based on principles, values, and commonalities, taking into account Japan's recommendations.

 Mr. Sellitto noted that in the early stages of technology development, there can be concern that regulations will hinder innovation, but people will gradually learn what to regulate, and that Anthropic's ASL was also a practice of first developing and implementing commitments and then publishing the lessons learned from them, and he hopes that this will lead to the development of best practices that can inform regulations in the future.

(4) Questions from an audience

 In response to a question from an online participant about what is needed to ensure the safety and reliability of AI, given that Japan has been the target of cyber-attacks in recent years, Mr. Sellitto explained that while there are currently no clear guidelines for AI cyber security, he explained that cybersecurity standards are being formed. Ms. Martinez also expressed the view that there have been many cyberattacks targeting AI, we can learn from them to increase resilience.

(5) Summary and closing remarks

 In concluding the event, Professor Shiroyama summarized the discussions and pointed out the need to organize a common vocabulary and know-how for "safety," although it seems better not to dare to establish a detailed definition. He also suggested that the dichotomy of hard law/soft law for regulation of AI is too simplistic, and that the learning process needs to start with abstract principles and shared experiences.

 In addition to thanking the participants, Associate Professor Ema mentioned the need to adhere to an agile process for AI security and safety, and ultimately AI governance, in the face of rapid technological innovation.

 Finally, Prof. Takeo Hoshi, Deputy Director of Tokyo College at the University of Tokyo, gave closing remarks. Prof. Hoshi pointed out the importance of today's discussion, and expressed the pleasure for Tokyo College to host this event together with the Institute for Future Initiatives. Drawing on debates on regulatory attempts to prevent financial crises, which is one of his areas of expertise, he stated that financial crises have been happening despite the various efforts to build sound and safe financial systems. here seem to be no regulatory mechanisms that make the financial systems completely safe.  The lesson is that, in addition to trying to prevent crises, we need to be ready to respond. Prof. Hoshi concluded the event by noting the need to prepare for AI crises while promoting human-centered AI development, and expressed his hope that today's discussion would serve as a starting point for future discussions.

Finished
Zoom Webinar
Date(s) Thursday, 28 March 2024, 10:00-12:00 JST
Venue

Zoom Webinar (In-person registration here, Online registration here)

Registration Advance registration is required. *Registration will close when we reach maximum capacity. *The Zoom URL will be emailed to those who register for the event on March 27.
Language English (Japanese interpretation)
Abstract

With the proliferation of generative AI, discussions on the safety of AI are unfolding domestically and internationally. However, there are various types of discussions on “safety,” and corresponding measures. Moreover, what is considered “safe” or what threats and risks are emphasized can differ depending on the situation and context in each country. While AI Safety Institutes have been established in the UK, US, Japan and to be established in other countries, grasping the types of “safety” specific to Japan and its corresponding measures will become increasingly important as a foundation for future international cooperation.
In this event, we will host discussions with experts on overseas AI governance to delve into the trends in international AI safety and governance.

Program

10:00:

Opening Remarks

Hideaki Shiroyama (Institute for Future Initiatives, The University of Tokyo)

 

10:10:

Introduction from panelists

Introduction from each panelist on current important issues in AI governance

 

11:10: 

Panel Discussion : What are the expectations for Japan?

Merve Hickok (President and Research Director at Center for AI & Digital Policy (CAIDP))

Cyrus Hodes (Lead, SAFE project at the Global Partnership on AI)

Inma Martinez (Chair of the Multi-stakeholder Experts Group, Global Partnership on AI)

Michael Sellitto (Head of Global Affairs, Anthropic)

 

11:40:

Q&A

 

Moderator

EMA Arisa (Associate Professor, Tokyo College and Institute for Future Initiatives, The University of Tokyo)

 

Speaker Profile

Merve Hickok, President and Research Director at Center for AI & Digital Policy (CAIDP)
Merve Hickok is the President and Research Director at Center for AI and Digital Policy (CAIDP), advising governments and international organizations on AI policy and regulation. She is a globally renowned expert on AI policy, ethics and governance. Her contributions and perspective have featured in The New York Times, Guardian, CNN, Forbes, Bloomberg, Wired, Scientific American, The Atlantic, and Politico. Her work focuses on impact of AI systems on individuals, society, public and private organizations – with a particular focus on fundamental rights, democratic values, and social justice. Merve is also the founder of AIethicist.org. She is the Data Ethics lecturer at University of Michigan School of Information, and the Responsible Data and AI Advisor at Michigan Institute for Data Science.

 

Cyrus Hodes, Lead, SAFE project at the Global Partnership on AI
Cyrus Hodes is a co-founder of Stability AI a leading generative AI platform, which he exited to launch infinitio.ai (AIGC Chain), the first foundation model of AI generated content on a blockchain. He is a General Partner at 1infinity Ventures, a global fund investing in responsible AI ventures. Cyrus leads the Safety and Assurance of Generative AI (SAFE) project at the Global Partnership on AI, responding to the G7 Hiroshima AI Process. He previously co-founded and chaired the AI Initiative at The Future Society—a think tank incubated at Harvard Kennedy School. Cyrus is a member of the OECD Expert Group on AI and a Board member of Intelmatix (Saudi’s largest AI company). Educated at Sciences Po Paris, M.A. Paris II University, M.P.A. Harvard.

 

Inma Martinez, Chair of the Multi-stakeholder Experts Group, Global Partnership on AI
Inma Martinez is technology pioneer and AI scientist who advises leaders in business and government on technology as competitive advantage and contribution to societal progress. She was a pioneer of digital technologies and AI in the 2000s and has combined her career in innovation with advisory appointments at government agencies in the United Kingdom (UKTI and the Innovation Fund of the Department of Sport, Media and Culture), Spain (State Secretariat for Artificial Intelligence at the Ministry of Economy and Digital Transformation) as well as provided expert testimonies across various technology boards at the European Commission since 2002. She has collaborated with the United Nations Industrial Development Organisation (UNIDO) highlighting the implications of the 4IR for developing countries in the post-pandemic world and is a UNESCO Ambassador for Intercultural Values in AI. She is a guest lecturer at Imperial College Business School in London and a published author of scientific books and research papers on emerging technologies.

 

Michael Sellitto, Head of Global Affairs, Anthropic

Michael Sellitto is the Head of Global Affairs at Anthropic, an AI safety and research company. He is also an Adjunct Senior Fellow in the Technology and National Security Program at the Center for a New American Security, and a Member of the Council on Foreign Relations.

Prior to joining Anthropic, Michael was the founding Deputy Director of the Stanford Institute for Human-Centered Artificial Intelligence (HAI), which is dedicated to advancing AI research, education, policy, and practice to improve the human condition. As HAI’s first staff member, he was instrumental in designing and executing the Institute’s strategic plans and establishing HAI’s global reputation among policymakers.

Michael served in the White House as Director for Cybersecurity Policy on the National Security Council staff from 2015-2018. He led international engagement on cybersecurity policy and strategy, promoted international adoption of a framework for strategic stability in cyberspace, and advanced issues related to the digital economy and Internet governance. Before that, Michael served as Special Assistant to Deputy Secretaries of State William J. Burns and Antony Blinken, advising the Deputies on political, energy, security, and trade issues related to South and Central Asia and on worldwide cyber policy and counterterrorism strategy.

Organized by Institute for Future Initiatives, The University of Tokyo Tokyo College, The University of Tokyo
Contact E-mail: ifi_tg[at]ifi.u-tokyo.ac.jp (replace [at] with @)

Upcoming Events

Individualism in Japanese Life (Lecture by Prof. John LIE)

イベント予定講演会/Lecture

Tuesday, 26 November 2024, 13:00-14:30 JST

We have been repeatedly told that Japan is a "collectivist" or "group-oriented" society, in contradistinction to the United States and other Western countries, which are said to be "individualist." The argument strikes me as wrong, at best. After briefly rebutting the received view, I trace the genealogy of the mistaken idea and explain its cogency.

From Invisible to Visible Genders (Lecture by Prof. Tricia OKADA)

イベント予定講演会/Lecture

Friday, 6 December 2024, 15:00-16:30

This lecture will cover ethnographic research on Filipino trans women or transpinay before, during, and after migration in Japan from the 1980s to the early 2000s. Drawing from an intersectional invisibility (Purdie-Vaughns & Eibach, 2008) framework, it will relate the Filipino trans women’s migration experiences to the cases of current issues transgender migrants are facing. This talk will also explore how social media and films create spaces to show and negotiate the (in)visibility of genders.

Dealing with the Brussels Effect: How should Japanese companies prepare for the EU-AI Act?

イベント予定講演会/Lecture

Wednesday, 11 December 2024, 12:00-13:00 JST

This webinar will outline the overview of the EU-AI Act, the activities of four working groups involved in the formulation of the Code of Practice, and important points that Japanese companies and organizations should particularly pay attention to. 
We look forward to the participation of companies, research institutions, and development communities involved in the development, provision, and distribution of AI-related technologies as an opportunity to deepen understanding of the “Brussels Effect” brought about by EU regulatory trends and its impact on Japan. 

Environmental Problems in Developing Countries: What Role for Taxation? (Lecture by Ushioda Fellow Michael KEEN)

イベント予定講演会/Lecture

Wednesday, 11 December 2024 10:30-12:00 JST

Many low income countries face severe environmental problems. They also face an urgent need for tax revenue to finance social needs and economic development. Can environmental taxes provide a way to meet both objectives? Drawing on a recent book, this lecture will take stock of the most pressing of the many environmental challenges faced by low income countries—including in air quality, waste management, soil quality, deforestation, congestion, adaptation to climate change—and consider to what extent improved tax policy can simultaneously help address them and raise a significant amount of tax revenue.

A Conversation with the Ambassador of Georgia to Japan “Japan viewed from inside and outside”

イベント予定対話/Dialogue

Available from Friday, 13 December 2024 17:00 JST

H.E. Mr. Teimuraz LEZHAVA, Ambassador of Georgia to Japan, who has deep knowledge and insight into Japanese culture and business practices, and Professor SHIMAZU Naoko, a global historian who has studied Japan from an external perspective, will explore one of Tokyo College’s key research themes, “Japan viewed from the inside and outside.”
We invite you to join us for this insightful and engaging discussion.

The Salon ー Conversations with Prominent Professors at the University of Tokyo (Season 3)

イベント予定対話/Dialogue

Every Monday from December 2, 2024 (Available from 17:00 JST)

“The Salon” is a dialogue series featuring distinguished scholars in the humanities at the University of Tokyo that aims to transcend disciplinary boundaries. It is hosted by Professor Naoko Shimazu of Tokyo College and Professor John Lie of UC Berkeley, who is currently staying at Tokyo College.
We invite you to listen to an informal discussion between experts in different fields, as if you are sitting next to them.This is a chance to see a new side of our guests that you have never seen before.

Previous Events

From Competitors to Partners: Banks’ Venture Investments in Fintech (Lecture by Prof. Manju PURI)

イベント予定講演会/Lecture

Tuesday, 12 November 2024 10:30-12:00 JST

Prof. Manju Puri has hypothesized and found evidence that banks use venture investments in fintech startups as a strategic approach to navigate fintech competition. She first documented that banks’ venture investments have increasingly focused on fintech firms. She found that banks facing greater fintech competition are more likely to make venture investments in fintech startups. Banks target fintech firms that exhibit higher levels of asset complementarities with their own business. Finally, instrumental variable analyses showed that venture investments increase the likelihoods of operational collaborations and knowledge transfer between the investing bank and the fintech investee.

Towards Building Multicultural and Multilingual Safe Large Language Models

イベント予定講演会/Lecture

Monday, 11 November 2024, 10:00-11:00 JST

As generative AI becomes more widely used, it is crucial for AI models to accurately reflect cultural and linguistic risks in different regions. Identifying harmful content specific to each culture must be continuously updated. This requires collaboration between AI researchers, social scientists, policymakers, and practitioners to form a global community for ongoing discussions. This event will discuss frameworks to sustain such communities, welcoming those interested in AI safety and governance.

Economic Policies under Japan’s New Cabinet: Wish List and Prospects

イベント予定パネルディスカッション/Panel discussion共催/Joint Event

Friday, 8 November 2024 8:00 - 9:15 JST

Japan had two elections—LDP President (September 27) and House of Representatives (October 27)—that elected a new prime minister. During the two elections, many economic policy proposals were presented and debated. The webinar will discuss economic policies that are likely to be adopted and those that are unlikely to be adopted but desirable for the Japanese economy.

50 Ways to Kill a Robot (Lecture by Prof. Jennifer ROBERTSON)

イベント予定講演会/Lecture

Tuesday, 5 November 2024 10:30-12:00 JST

The idiom “50 Ways” in my title should not be taken as a literal metric. “Fifty” is a simply a metaphor for a number larger than a few. In his hit song, “Fifty Ways to Leave Your Lover” (1975), Paul Simon provides six ways to leave. In this paper, I will review a number of ways to “kill” a robot together with some of the ways that robots “die.” Death here is broadly defined as the permanent ending of vital processes. I will also review how “deceased” robots are handled. My cultural area focus is primarily Japan and the United States. Both humans and robots are, in several respects, electrical entities, and so in my concluding remarks, I address the question of what happens to each after the electricity is out.

Exploring Queer Fantasy Work in Idol Fandom Culture Across East and Southeast Asia (Lecture by Dr. Thomas BAUDINETTE)

イベント予定講演会/Lecture

Friday, 1 November 2024 14:00-15:30 JST

Within this presentation, Prof. Baudinette argues that the fundamentally transformative nature of fan subjectivity encourages the production of queer fantasies tied to idols that marginalized social subjects can utilize to critique the social structures which disadvantage them. He unpacks how LGBTQ+ fans across Asia transform idol fandom into a queer space where their fantasy work creates transnational solidarities grounded in the political project of queer emancipation. Through this discussion, he theorizes “queer fantasy work” as it is tied to idol fandom as an explicitly political force in contemporary Asian culture designed to actively produce a more egalitarian and hopeful world.

Queer Demography in Japan: Decentering Universalized Knowledge of Gender and Sexuality in the West (Lecture by Prof. HIRAMORI Daiki)

イベント予定講演会/Lecture

Thursday, 24 October 2024 15:00-16:30 JST

In this talk, Prof. Hiramori will present findings from his methodological studies to develop questions to measure sexual orientation and gender identity on population-based surveys in the Japanese context. He will also examine the issue of heterosexual respondents being misclassified as non-heterosexual and the difficulty of fully separating heterosexual and non-heterosexual people in survey data. He will conclude the talk by discussing the findings from his latest study that half of those who select “other” as their gender on surveys may be cisgender women, even though this category was meant to capture non-binary respondents.


TOP