【原版杂志阅读】 文章预读:Just how good can China get at generative AI?


Business | The Sino-American tech race
Just how good can China get at generative AI?

Its models may in time rival America’s. But that may not translate into an economic or military edge 

截屏2023-05-24 下午8.15.38

If you listen to the bombast in Beijing and Washington, America and China are engaged in an all-out contest for technological supremacy. “Fundamentally, we believe that a select few technologies are set to play an outsized importance over the coming decade,” declared Jake Sullivan, President Joe Biden’s national security adviser, last September. In February Xi Jinping, China’s paramount leader, echoed the sentiment, stating that “we urgently need to strengthen basic research and solve key technology problems” in order to “cope with international science and technology competition, achieve a high level of self-reliance and self-improvement”.


No technology seems to obsess policymakers on both sides of the Pacific more right now than artificial intelligence (ai). The rapid improvements in the abilities of “generative” ais like Chatgpt, which analyse the web’s worth of human text, images or sounds and can then create increasingly passable simulacrums, have only strengthened the obsession. If generative ai proves as transformational as its boosters claim, the technology could give those who wield it an economic and military edge in the 21st century’s chief geopolitical contest. Western and Chinese strategists already talk of an ai arms race. Can China win it?



On some measures of ai prowess, the autocracy pulled ahead some time ago (see chart). China surpassed America in the share of highly cited ai papers in 2019; in 2021, 26% of ai conference publications globally came from China, compared with America’s share of 17%. Nine of the world’s top ten institutions, by volume of ai publications, are Chinese. According to one popular benchmark, so are the top five labs working on computer vision, a type of ai particularly useful to a communist surveillance state.


Yet when it comes to “foundation models”, which give generative ais their wits, America is firmly in front (see charts 2 and 3). Chatgpt and the pioneering model behind it, the latest version of which is called gpt-4, are the brainchild of Openai, an American startup. A handful of other American firms, from small ones such as Anthropic or Stability ai to behemoths like Google, Meta and Microsoft (which part-owns Openai), have their own powerful systems. ernie, a Chinese rival to Chatgpt built by Baidu, China’s internet-search giant, is widely seen as less clever. Alibaba and Tencent, China’s mightiest tech titans, have yet to unveil their own generative ais.



This leads those in the know to conclude that China is two or three years behind America in building foundation models. There are three reasons for this underperformance. The first concerns data. A centralised autocracy should be able to marshal lots of it—the government was, for instance, able to hand over troves of surveillance information on Chinese citizens to firms such as SenseTime or Megvii that, with the help of China’s leading computer-vision labs, then used it to develop top-notch facial-recognition systems.


That advantage has proved less formidable in the context of generative ais, because foundation models are trained on the voluminous unstructured data of the web. American model-builders benefit from the fact that 56% of all websites are in English, whereas just 1.5% are written in Chinese, according to data from w3Techs, an internet-research site. As Yiqin Fu of Stanford University points out, the Chinese interact with the internet primarily through mobile super-apps like WeChat and Weibo. These are “walled gardens”, so much of their content is not indexed on search engines. This makes that content harder for ai models to suck up. Lack of data may explain why Wu Dao 2.0, a model unveiled in 2021 by the Beijing Academy of Artificial Intelligence, a state-backed outfit, failed to make a splash despite its possibly being computationally more complex than gpt-4.


The second reason for China’s lacklustre generative achievements has to do with hardware. Last year America imposed export controls on technology that might give China a leg-up in ai. These cover the powerful microprocessors used in the cloud-computing data centres where foundation models do their learning, and the chipmaking tools that could enable China to build such semiconductors on its own.


That hurt Chinese model-builders. An analysis of 26 big Chinese models by the Centre for the Governance of ai, a British think-tank, found that more than half depended on Nvidia, an American chip designer, for their processing power. Some reports suggest that smic, China’s biggest chipmaker, has produced prototypes just a generation or two behind tsmc, the Taiwanese industry leader that manufactures chips for Nvidia (see chart 4). But smic can probably mass-produce only chips which tsmc was churning out by the million three or four years ago.


Chinese ai firms are having trouble getting their hands on another American export: know-how. America remains a magnet for the world’s tech talent; two-thirds of ai experts in America who present papers at the main ai conference are foreign-born. Chinese engineers made up 27% of that select group in 2019. Many Chinese ai boffins studied or worked in America before bringing expertise back home. The covid-19 pandemic and rising Sino-American tensions are causing their numbers to dwindle. In the first half of 2022 America granted half as many visas to Chinese students as in the same period in 2019.


The triple shortage—of data, hardware and expertise—has been a hurdle for China. Whether it will hold Chinese ai ambitions back much longer is another matter.

Take data. In February local authorities in Beijing, where nearly a third of China’s ai firms are located, promised to release data from 115 state-affiliated organisations, giving model-builders 15,880 data sets to play with. The central government has previously signalled it wants to dismantle Chinese apps’ walled gardens, potentially liberating more data, says Kayla Blomquist, an American former diplomat in China now at Oxford University. The latest models are also able to transfer their machine learnings from one language to another. Openai says that gpt-4 performs remarkably well on tasks in Chinese despite scarce Chinese source material in its training data. Baidu’s ernie was trained on lots of English-language data, notes Jeffrey Ding of George Washington University.


In hardware, too, China is finding workarounds. The Financial Times reported in March that SenseTime, which is blacklisted by America, was using middlemen to skirt the export controls. Some Chinese ai firms are harnessing Nvidia’s processors through cloud servers in other countries. Alternatively, they can buy more of Nvidia’s less advanced wares—to keep serving the vast Chinese market, Nvidia has designed sanctions-compliant ones that are between 10% and 30% slower than top-of-the-range kit. These end up being costlier for the Chinese customers per unit of processing power. But they do the job.


China could partly alleviate the dearth of chips—and of brain power—with the help of “open-source” models. Such models’ inner workings can be downloaded by anyone and fine-tuned to a specific task. Those include the numbers, called “weights”, which define the structure of the model and which are derived from costly training runs. Researchers at Stanford used the weights from llama, Meta’s foundation model, to build one called Alpaca for less than $600, compared with perhaps $100m for training something like gpt-4. Alpaca performs just as well as the original version of Chatgpt on some tasks.

.


相关课程
免费试听
实用口语:5个pan out 小短句
实用口语:老外嘴里的put a sock in it 是什么意思?
精选外教自然拼读系列02
每日一词 consistency
rise raise 辨析
实用口语:two-way street
(高中英语) 多所名校都考过的一道从句典型题目
精选外教自然拼读系列01
实用口语:throw a curveball
(每日一词)infallible
铅笔英语