How to counter China’s scary use of artificial intelligence data



Nowhere is the competition in developing artificial intelligence fiercer than in the accelerating rivalry between the United States and China. At stake in this competition is not just who leads in AI but who sets the rules for how it is used around the world.

China is forging a new model of digital authoritarianism at home and is actively exporting it abroad. It has launched a national-level AI development plan with the intent to be the global leader by 2030. And it is spending billions on AI deployment, training more AI scientists and aggressively courting experts from Silicon Valley.

The United States and other democracies must counter this rising tide of techno-authoritarianism by presenting an alternative vision for how AI should be used that is consistent with democratic values. But China’s authoritarian government has an advantage. It can move faster than democratic governments in establishing rules for AI governance, since it can simply dictate which uses are allowed or banned.

One risk is that China’s model for AI use will be adopted in other countries while democracies are still developing an approach more protective of human rights.

The Chinese Communist Party, for example, is integrating AI into surveillance cameras, security checkpoints and police cloud computing centers. As it does so, it can rely on world-class technology companies that work closely with the government. Lin Ji, vice president of iFlytek, one of China’s AI “national team” companies, told me that 50% of its $1 billion in annual revenue came from the Chinese government.

China is building a burgeoning panopticon, with more than 500 million surveillance cameras deployed nationwide by 2021 — accounting for more than half of the world’s surveillance cameras. Even more significant than government cash buoying the AI industry is the data collected, which AI companies can use to further train and refine their algorithms.

Facial recognition is being widely deployed in China, while a grassroots backlash in the U.S. has slowed deployment. Several U.S. cities and states have banned facial recognition for use by law enforcement. In 2020, Amazon and Microsoft placed a moratorium on selling facial-recognition technology to law enforcement, and IBM canceled its work in the field. These national differences are likely to give Chinese firms a major edge in development of facial-recognition technology.

The problem is not just that AI is being used for human rights abuses but that it can supercharge repression itself, arming the state with vast intelligent surveillance networks to monitor and control the population at a scale and degree of precision that would be impossible with human agents.

In the face of these AI threats, democratic governments and societies need to work to establish global norms for lawful, appropriate and ethical uses of technologies like facial recognition. One of the challenges in doing so is that there is not yet a democratic model for how facial recognition or other AI technologies ought to be employed.

The U.S. government needs to be more proactive in international standard-setting, working with domestic companies to ensure that international AI and data standards protect human rights and individual liberty.

AI can be used to bolster individual freedom or crush it. Russian President Vladimir Putin has said about AI: “Whoever becomes the leader in this sphere will become the ruler of the world.” The race to lead in AI and write the rules of the next century is underway, and with it, the future of global security.

Paul Scharre is vice president and director of studies at the Center for a New American Security. He is author of the forthcoming book “Four Battlegrounds: Power in the Age of Artificial Intelligence.” ©2023 Los Angeles Times. Distributed by Tribune Content Agency.

Similar Articles

Comments

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Advertismentspot_img

Instagram

Most Popular