Development of Artificial Intelligence and Intelligent Computing

“IT有得聊” is a professional information and service platform under the Machinery Industry Press, dedicated to helping readers master more professional and practical knowledge and skills in the broad field of IT, quickly enhancing workplace competitiveness. Click the blue WeChat name to quickly follow us!

On the morning of April 26, after the closing of the ninth meeting of the 14th National People’s Congress Standing Committee, the tenth special lecture was held. Academician Sun Ninghui from the Chinese Academy of Engineering and researcher at the Institute of Computing Technology, Chinese Academy of Sciences, delivered a lecture titled “Development of Artificial Intelligence and Intelligent Computing”. Below is the lecture manuscript recently published by the National People’s Congress website.

Lecture Manuscript of the Tenth Special Lecture of the 14th National People’s Congress Standing Committee

Development of Artificial Intelligence and Intelligent Computing

Sun Ninghui

Development of Artificial Intelligence and Intelligent Computing

Chairman, Vice Chairmen, Secretary General, and all members:

In recent years, the field of artificial intelligence has been experiencing explosive development led by generative artificial intelligence large models. On November 30, 2022, OpenAI launched an AI conversational chatbot, ChatGPT, whose outstanding natural language generation capabilities attracted widespread attention worldwide, surpassing 100 million users in just two months. This sparked a wave of large models both domestically and internationally, with various models such as Gemini, Wenxin Yiyan, Copilot, LLaMA, SAM, and SORA emerging like mushrooms after rain, making 2022 known as the year of large models. The current information age is rapidly entering the development stage of intelligent computing, with breakthroughs in artificial intelligence technology emerging one after another, gradually empowering various industries and promoting artificial intelligence and data elements as typical representatives of new productive forces. General Secretary Xi Jinping pointed out that the new generation of artificial intelligence should be regarded as a driving force for promoting leapfrog development in science and technology, optimizing industrial upgrades, and enhancing overall productivity, striving to achieve high-quality development. Since the 18th National Congress of the Communist Party, the Central Committee with Comrade Xi Jinping at its core has attached great importance to the development of the intelligent economy, promoting the deep integration of artificial intelligence and the real economy, injecting strong momentum into high-quality development.

1. Overview of the Development of Computing Technology

The history of the development of computing technology can be roughly divided into four stages: the emergence of the abacus marked the entry of humanity into the first generation—mechanical computing; the second generation—electronic computing was marked by the emergence of electronic devices and electronic computers; the advent of the internet ushered us into the third generation—network computing; and currently, human society is entering the fourth stage—intelligent computing.

Early computing devices were manual auxiliary computing devices and semi-automatic computing devices. The history of human computing tools began with the Chinese abacus in 1200 AD, followed by the invention of Napier’s bones (1612) and the wheel-based adder (1642), leading to the birth of the first automatic calculator capable of performing four arithmetic operations—the stepped reckoner in 1672.

The mechanical computing era had already introduced some basic concepts of modern computers. Charles Babbage proposed the designs for the difference engine (1822) and the analytical engine (1834), supporting automatic mechanical computation. During this period, the concepts of programming and programs were basically formed. The concept of programming originated from the Jacquard loom, which controlled the printing pattern through punched cards, eventually evolving into the storage of all mathematical calculation steps in the form of computational instructions; the first programmer in human history was Ada, the daughter of poet Lord Byron, who wrote a set of computational instructions for Babbage’s difference engine to solve Bernoulli numbers. This set of instructions is also the first computer algorithm program in human history, separating hardware from software and introducing the concept of a program for the first time.

It was not until the first half of the 20th century that four scientific foundations of modern computing technology emerged: Boolean algebra (mathematics), Turing machines (computational models), von Neumann architecture (architecture), and transistors (devices). Among them, Boolean algebra is used to describe the underlying logic of programs and hardware such as CPUs; Turing machines are a universal computational model that transforms complex tasks into automated calculations without human intervention; the von Neumann architecture proposed three basic principles for constructing computers: using binary logic, storing and executing programs, and that computers consist of five basic units: arithmetic unit, control unit, memory, input devices, and output devices; transistors are semiconductor devices that form the basic logic circuits and storage circuits, serving as the “bricks” for building modern computers. Based on these scientific foundations, computing technology has developed rapidly, forming a large-scale industry.

From the birth of the world’s first electronic computer, ENIAC, in 1946 to today in the 21st century, five types of successful platform computing systems have been formed. Currently, various types of applications in various fields can be supported by these five types of platform computing devices. The first type is high-performance computing platforms, which solve scientific and engineering computing problems for core national departments; the second type is enterprise computing platforms, also known as servers, used for enterprise-level data management and transaction processing, with computing platforms of internet companies like Baidu, Alibaba, and Tencent belonging to this category; the third type is personal computer platforms, appearing in the form of desktop applications, allowing people to interact with personal computers through desktop applications; the fourth type is smartphones, characterized by mobility and portability, connecting to data centers via the internet, primarily focused on internet applications, distributedly deployed in data centers and mobile terminals; the fifth type is embedded computers, embedded in industrial equipment and military devices, ensuring the completion of specific tasks within a determined time through real-time control. These five types of devices cover almost all aspects of our information society, while the long-sought sixth type of platform computing system centered on intelligent computing applications has yet to be formed.

Modern computing technology can be roughly divided into three eras.

IT 1.0, also known as the electronic computing era (1950-1970), is characterized by being “machine-centered”. The basic architecture of computing technology was formed, and with the advancement of integrated circuit technology, the scale of basic computing units rapidly shrank, with transistor density, computing performance, and reliability continuously improving, leading to widespread application of computers in scientific engineering computing and enterprise data processing.

IT 2.0, also known as the network computing era (1980-2020), is centered around “people”. The internet connects the terminals used by people with backend data centers, and internet applications interact with people through intelligent terminals. Internet companies like Amazon proposed the idea of cloud computing, encapsulating backend computing power as a public service rented to third-party users, forming the cloud computing and big data industry.

IT 3.0, also known as the intelligent computing era, began in 2020, adding the concept of “things” compared to IT 2.0, that is, various edge devices in the physical world, which are digitized, networked, and intelligent, achieving a triadic integration of “people-machine-things”. In the intelligent computing era, in addition to the internet, there is also data infrastructure supporting various terminals to achieve the interconnection of all things through edge-cloud computing, with terminals, things, edge, and cloud all embedded with AI, providing intelligent services similar to large models like ChatGPT, ultimately achieving AI intelligence wherever there is computation. Intelligent computing has brought massive data, breakthroughs in artificial intelligence algorithms, and explosive demand for computing power.

2. Overview of the Development of Intelligent Computing

Intelligent computing includes artificial intelligence technology and its computing carriers,which has roughly gone through four stages: general computing devices, logic reasoning expert systems, deep learning computing systems, and large model computing systems.

The starting point of intelligent computing isgeneral automatic computing devices (1946). Scientists like Alan Turing and John von Neumann initially hoped to simulate the process of the human brain processing knowledge, inventing machines that think like the human brain. Although this was not achieved, it solved the problem of automation in computation. The emergence of general automatic computing devices also promoted the birth of the concept of artificial intelligence (AI) in 1956, and all subsequent developments in artificial intelligence technology have been built on new generation computing devices and stronger computing capabilities.

The second stage of intelligent computing development islogic reasoning expert systems (1990). Scientists from the symbolic AI school, such as Edward Albert Feigenbaum, aimed primarily at automating logical reasoning capabilities, proposing expert systems capable of performing logical reasoning with knowledge symbols. Human prior knowledge enters the computer in the form of knowledge symbols, enabling the computer to assist humans in making certain logical judgments and decisions in specific fields, but expert systems heavily rely on manually generated knowledge bases or rule bases. A typical representative of this type of expert system is Japan’s fifth-generation computer and China’s 863 Program-supported 306 intelligent computer theme. Japan adopted dedicated computing platforms and knowledge reasoning languages like Prolog to complete application-level reasoning tasks in logic expert systems; China took a different technical route, based on general computing platforms, transforming intelligent tasks into artificial intelligence algorithms, integrating both hardware and system software into general computing platforms, giving rise to a number of backbone enterprises such as Shuguang, Hanwang, and iFlytek.

The limitations of symbolic computing systems lie in their explosive computational time-space complexity, meaning that symbolic computing systems can only solve linear growth problems and cannot solve high-dimensional complex space problems, thus limiting the size of problems they can handle. Additionally, because symbolic computing systems are based on knowledge rules, we cannot exhaustively enumerate all common sense, greatly restricting their application range. With the arrival of the second AI winter, the first generation of intelligent computers gradually exited the historical stage.

Until around 2014, intelligent computing advanced to the third stage—deep learning computing systems. Represented by Geoffrey Hinton and others from the connectionist school, aiming for automation of learning capabilities, new AI algorithms such as deep learning were invented. Through the automatic learning of deep neural networks, the model’s statistical induction capabilities were significantly enhanced, achieving tremendous breakthroughs in applications such as pattern recognition, with recognition accuracy in certain scenarios even surpassing that of humans. Taking facial recognition as an example, the entire training process of the neural network is equivalent to a process of adjusting network parameters, inputting a large number of labeled facial image data into the neural network, and then adjusting parameters between networks to make the output probability of the neural network approach the real result. The greater the probability of the neural network outputting the real situation, the larger the parameters, thus encoding knowledge and rules into network parameters, allowing for learning various common sense as long as there is enough data, greatly enhancing generality. The applications of connectionist intelligence are more widespread, including speech recognition, facial recognition, and autonomous driving. In terms of computing carriers, the Institute of Computing Technology, Chinese Academy of Sciences proposed the world’s first deep learning processor architecture in 2013, and internationally renowned hardware manufacturer NVIDIA has continuously released several leading performance general-purpose GPU chips, all of which are typical representatives of deep learning computing systems.

The fourth stage of intelligent computing development islarge model computing systems (2020). Driven by the technology of large AI models, intelligent computing has reached new heights. In 2020, AI shifted from “small models + discriminative” to “large models + generative”, upgrading from traditional facial recognition, object detection, and text classification to today’s text generation, 3D digital human generation, image generation, speech generation, and video generation. A typical application of large language models in dialogue systems is OpenAI’s ChatGPT, which uses the pre-trained base large language model GPT-3, incorporating a training corpus of 300 billion words, equivalent to the total of all English text on the internet. Its basic principle is: by giving it an input, it predicts the next word to train the model, improving prediction accuracy through extensive training, ultimately allowing it to generate an answer when asked a question, engaging in real-time dialogue with humans. Based on the base large model, it is further fine-tuned with supervised instruction prompts, gradually teaching the model how to engage in multi-turn dialogue with humans; finally, through human-designed and automatically generated reward functions, reinforcement learning iterations are conducted to gradually align the large model with human values.

The characteristics of large models are defined by “large” in three layers: (1) large parameters, with GPT-3 having 170 billion parameters; (2) large training data, with ChatGPT using approximately 300 billion words and 570GB of training data; (3) large computing power requirements, with GPT-3 requiring tens of thousands of V100 GPUs for training. To meet the explosive demand for intelligent computing power from large models, new intelligent computing centers are being built on a large scale both domestically and internationally, and NVIDIA has also launched a large model intelligent computing system using 256 H100 chips and 150TB of massive GPU memory.

Large models have brought about three transformations. First is the scaling law in technology, where the accuracy of many AI models rapidly improves once the parameter scale exceeds a certain threshold, the reasons for which are still not very clear in the scientific community and are highly debated. The performance of AI models is logarithmically linearly related to three variables: model parameter scale, dataset size, and total computing power, thus increasing model scale can continuously improve model performance. Currently, the most advanced large model, GPT-4, has reached a parameter scale of trillions to tens of trillions and is still growing; second is the explosive growth in computing power demand in the industry, where training large models with hundreds of billions of parameters typically requires training on thousands to tens of thousands of GPU cards for 2-3 months, leading to a rapid increase in computing power demand driving related computing companies to develop at super high speed, with NVIDIA’s market value approaching $2 trillion, an unprecedented occurrence for chip companies; third is the impact on the labor market, with a report released by Peking University’s National Development Research Institute and Zhaopin Recruitment indicating that among the 20 most affected occupations, accounting, sales, and clerical positions rank at the top, while physical labor jobs that require interaction with people and providing services, such as human resources, administration, and logistics, are relatively safer.

Artificial intelligence’s technological frontier will develop in the following four directions. The first frontier direction ismultimodal large models. From the perspective of human beings, human intelligence is inherently multimodal, possessing eyes, ears, nose, tongue, body, and mouth (language). From the AI perspective, vision, hearing, etc., can also be modeled as sequences of tokens, and the same methods used for large language models can be employed for learning, further aligning with the semantics in language to achieve intelligent capabilities of multimodal alignment.

The second frontier direction isvideo generation large models. OpenAI released the video generation model SORA on February 15, 2024, significantly increasing video generation duration from a few seconds to one minute, with significant improvements in resolution, visual realism, and temporal consistency. The greatest significance of SORA is that it possesses the basic characteristics of a world model, which is the ability to observe and predict the world. A world model is built on understanding the basic physical knowledge of the world (e.g., water flows downhill) and then observing and predicting what event will happen next. Although SORA still faces many issues to become a world model, it can be considered that SORA has learned the ability of visual imagination and minute-level future prediction, which are fundamental characteristics of a world model.

The third frontier direction isembodied intelligence. Embodied intelligence refers to intelligent agents with bodies that support interaction with the physical world, such as robots and unmanned vehicles, processing various sensor data inputs through multimodal large models, generating motion instructions to drive the intelligent agents, replacing traditional rule-based or mathematical formula-based motion driving methods, achieving deep integration of the virtual and real worlds. Therefore, robots with embodied intelligence can integrate the three major schools of artificial intelligence: connectionism represented by neural networks, symbolism represented by knowledge engineering, and behaviorism related to control theory, which is expected to bring new technological breakthroughs.

The fourth frontier direction isAI4R (AI for Research) becoming the main paradigm for scientific discovery and technological invention. Currently, scientific discovery mainly relies on experiments and human brain wisdom, with humans making bold conjectures and careful verifications, where information technology plays only a supportive and validating role. Compared to humans, artificial intelligence has significant advantages in memory, high-dimensional complexity, full vision, reasoning depth, and conjecture. Whether AI can take the lead in some scientific discoveries and technological inventions, significantly improving the efficiency of human scientific discovery, such as actively discovering physical laws, predicting protein structures, designing high-performance chips, and efficiently synthesizing new drugs, is a key question. Because large AI models possess complete data and have a god-like perspective, they can look further ahead than humans. If AI can achieve a leap from inference to reasoning, AI models have the potential to possess the imagination and scientific conjecture capabilities akin to Einstein, greatly enhancing the efficiency of human scientific discovery and breaking human cognitive boundaries. This is the true disruption.

Finally,general artificial intelligence (Artificial General Intelligence, AGI) is a highly challenging and controversial topic. There was once a bet between a philosopher and a neuroscientist: whether researchers could reveal how the brain achieves consciousness in 25 years (i.e., by 2023). At that time, there were two schools of thought regarding consciousness: one called integrated information theory, which posited that consciousness is formed by the “structure” created by specific types of neurons in the brain, and the other called global workspace theory, which suggested that consciousness arises when information spreads through interconnected networks to brain regions. In 2023, adversarial experiments were conducted across six independent laboratories, and the results did not fully match either theory, with the philosopher winning and the neuroscientist losing. This bet illustrates that people always hope artificial intelligence can understand the mysteries of human cognition and the brain. From a physics perspective, physics has thoroughly understood the macroscopic world and has begun to understand the microscopic world starting from quantum physics. The intelligent world, like the physical world, is a research object of immense complexity. AI large models still study the macroscopic world through data-driven methods to improve machine intelligence levels, and directly seeking answers in the microscopic world of the nervous system is challenging. Since its inception, artificial intelligence has always carried humanity’s dreams and fantasies about intelligence and consciousness, inspiring continuous exploration.

3. Security Risks of Artificial Intelligence

The development of artificial intelligence has promoted technological progress in today’s world, but it has also brought many security risks that need to be addressed from both technical and regulatory perspectives.

First is the rampant spread of false information on the internet. Here are several scenarios: first, digital avatars. AI Yoon is the first official “candidate” synthesized using DeepFake technology, modeled after South Korean National Power Party candidate Yoon Suk-yeol, created by a local DeepFake technology company using 20 hours of audio and video clips of Yoon and over 3,000 sentences recorded specifically for researchers. In reality, the content expressed by AI Yoon was written by the campaign team, not the candidate himself.

Second, forged videos, especially those of leaders, can cause international disputes, disrupt election orders, or trigger sudden public opinion events, such as the forged announcement of the first moon landing failure by Nixon or the forged announcement of “surrender” by Ukrainian President Zelensky, leading to a decline in social trust in the news media industry.

Third, forged news, mainly generated through automated false news generation for illegal profit, using ChatGPT to generate trending news to earn traffic. As of June 30, 2023, there were 277 global websites generating fake news, severely disrupting social order.

Fourth, face and voice swapping used for fraud. For example, due to AI mimicking the voice of a corporate executive, an international company in Hong Kong was defrauded of $35 million.

Fifth, generating inappropriate images, especially targeting public figures, such as the production of pornographic videos of film stars, causing adverse social impacts. Therefore, there is an urgent need to develop detection technologies for the forgery of false information on the internet.

Secondly, AI large models face serious trust issues. These issues include: (1) factual errors that are “seriously misleading”; (2) narratives based on Western values, outputting political biases and erroneous statements; (3) susceptibility to inducement, outputting erroneous knowledge and harmful content; (4) exacerbated data security issues, with large models becoming significant traps for sensitive data, as ChatGPT incorporates user inputs into its training database to improve itself, allowing the US to utilize large models to access Chinese language data that may not be available to us, mastering “Chinese knowledge” that we may not even grasp. Therefore, there is an urgent need to develop security regulatory technologies for large models and establish our own trustworthy large models.

In addition to technical means, the security of artificial intelligence requires relevant legislative work. In 2021, the Ministry of Science and Technology issued the “Ethical Norms for the New Generation of Artificial Intelligence”; in August 2022, the National Information Security Standardization Technical Committee released the “Information Security Technology Machine Learning Algorithm Security Assessment Specification”; from 2022 to 2023, the Central Cyberspace Administration successively issued the “Regulations on Algorithm Recommendation Management for Internet Information Services”, “Regulations on Deep Synthesis Management for Internet Information Services”, and “Management Measures for Generative Artificial Intelligence Services”. European and American countries have also successively introduced regulations, such as the EU’s General Data Protection Regulation issued on May 25, 2018, the US’s Blueprint for the Artificial Intelligence Bill released on October 4, 2022, and the European Parliament’s passage of the EU Artificial Intelligence Act on March 13, 2024.

China should accelerate the introduction of the “Artificial Intelligence Law”, build an artificial intelligence governance system, ensure that the development and application of artificial intelligence follow common human values, promote harmonious and friendly human-machine interactions; create a policy environment conducive to the research, development, and application of artificial intelligence technology; establish reasonable disclosure and audit assessment mechanisms to understand the principles and decision-making processes of artificial intelligence mechanisms; clarify the security responsibilities and accountability mechanisms of artificial intelligence systems, traceable responsible entities, and remedies; and promote the formation of fair, reasonable, open, and inclusive international artificial intelligence governance rules.

4. Dilemmas in the Development of Intelligent Computing in China

The technology of artificial intelligence and the intelligent computing industry are at the focal point of Sino-US technological competition. Although China has made significant achievements in recent years, it still faces many development dilemmas, especially difficulties brought about by the US’s technology suppression policies.

The first dilemma is that the US has long been in a leading position in AI core capabilities, while China is in a tracking mode. China lags behind the US in the number of high-end AI talents, innovation in AI foundational algorithms, capabilities of foundational large models (large language models, text-to-image models, text-to-video models), training data for foundational large models, and training computing power for foundational large models, and this gap is expected to persist for a long time.

The second dilemma is the ban on the sale of high-end computing products and the long-term bottleneck in high-end chip processes. High-end intelligent computing chips such as A100, H100, and B200 are banned from being sold to China. Companies like Huawei, Loongson, Cambricon, Shuguang, and Haiguang have all been placed on the entity list, and their advanced chip manufacturing processes are restricted. The domestic processes capable of meeting large-scale production are 2-3 generations behind the international advanced level, and the performance of core computing chips lags 2-3 generations behind the international advanced level.

The third dilemma is the weak domestic intelligent computing ecosystem and insufficient penetration of AI development frameworks. NVIDIA’s CUDA ecosystem is complete and has formed a de facto monopoly. The domestic ecosystem is weak, specifically manifested in: (1) insufficient R&D personnel, with nearly 20,000 developers in the NVIDIA CUDA ecosystem, which is 20 times the total number of personnel in all domestic intelligent chip companies; (2) insufficient development tools, with CUDA having 550 SDKs, which is hundreds of times more than related domestic companies; (3) insufficient funding, with NVIDIA investing $5 billion annually, which is dozens of times more than related domestic companies; (4) the AI development framework TensorFlow occupies the industrial market, PyTorch occupies the research market, while domestic AI development frameworks like Baidu PaddlePaddle have only 1/10 the number of developers compared to foreign frameworks. More seriously, domestic enterprises are fragmented, unable to form a united front; although there are related products at each layer, such as intelligent applications, development frameworks, system software, and intelligent chips, there is no deep adaptation between layers, making it impossible to form a competitive technical system.

The fourth dilemma is that the cost and threshold for applying AI in industries remain high. Currently, AI applications in China are mainly concentrated in the internet industry and some defense fields. When promoting AI technology applications across various industries, especially when migrating from the internet industry to non-internet industries, a large amount of customization work is required, making migration difficult and the cost of single use high. Finally, there is also a significant shortage of talent in the AI field compared to actual demand in China.

5. Path Choices for the Development of Intelligent Computing in China

The path choices for the development of artificial intelligence are crucial for China, relating to the sustainability of development and the final international competitive landscape. Currently, the cost of using artificial intelligence is extremely high; for example, the Microsoft Copilot suite requires a monthly fee of $10, ChatGPT consumes 500,000 kilowatt-hours of electricity daily, and the price of NVIDIA’s B200 chip exceeds $30,000. Overall, China should develop affordable, safe, and trustworthy artificial intelligence technology, eliminate information poverty among its population, and benefit countries along the Belt and Road; empower various industries with low thresholds, maintain the competitiveness of China’s advantageous industries, and significantly narrow the gap for relatively lagging industries.

Choice 1: Should we unify the technical system to follow a closed-source path or an open-source path?

The intelligent computing industry is supported by a tightly coupled technical system, which is a technical whole closely linking materials, devices, processes, chips, complete machines, system software, and application software through a series of technical standards and intellectual property. China has three paths for developing an intelligent computing technology system:

First, to catch up and be compatible with the US-led A system. Most of China’s internet companies are following the GPGPU/CUDA compatible path, and many startups in the chip field are also trying to build ecosystems compatible with CUDA, which is a more realistic path. Due to the restrictions imposed by the US on China’s processes and chip bandwidth in terms of computing power, it is difficult for domestic ecosystems to form a unified front in algorithms, and the maturity of the ecosystem is severely limited. Additionally, the lack of high-quality Chinese data further exacerbates the gap between followers and leaders, which may sometimes even widen.

Second, to build a dedicated closed B system. In specialized fields such as military, meteorology, and justice, build enterprise closed ecosystems based on domestically produced mature processes, focusing more on specific vertical large models rather than foundational large models, and training large models using proprietary high-quality data from the field. This path is easier to form a complete and controllable technical system and ecosystem. Some large backbone enterprises in China are following this path, but its downside is that it is closed and cannot gather the majority of domestic forces, making it difficult to achieve globalization.

Third, to globally co-build an open-source C system. Break the ecological monopoly with open-source, lowering the threshold for enterprises to own core technologies, allowing every enterprise to create its own chips at low cost, forming a vast ocean of intelligent chips to meet ubiquitous intelligent demands. Use openness to form a unified technical system, where Chinese enterprises join forces with global powers to co-build a unified intelligent computing software stack based on international standards. Establish a shared mechanism for enterprise competition, sharing high-quality databases and open-source general foundational large models. In the global open-source ecosystem, Chinese enterprises have benefited greatly during the internet era, and in the intelligent era, Chinese enterprises should become major contributors to the RISC-V + AI open-source technology system, becoming the leading force in global open sharing.

Choice 2: Should we focus on algorithm models or new infrastructure?

Artificial intelligence technology must empower various industries, exhibiting a typical long-tail effect. 80% of China’s small and medium-sized enterprises need low-threshold, low-cost intelligent services. Therefore, China’s intelligent computing industry must be built on a new data space infrastructure, where the key is to achieve comprehensive infrastructure for intelligent elements, namely data, computing power, and algorithms. This work can be compared to the historical role of the US information superhighway plan in the early 20th century (i.e., information infrastructure construction) for the internet industry.

The core productivity of the information society is cyberspace. The evolution of cyberspace is from a computing space formed by one-dimensional machine connections, evolving into an information space formed by two-dimensional human-machine information connections, and then evolving into a data space formed by three-dimensional human-machine-thing connections. From the perspective of data space, the essence of artificial intelligence is the refinement of data, and large models are products of deep processing of all internet data. In the digital age, what is transmitted over the internet is information flow, which is a structured abstraction of data processed roughly by computing power; in the intelligent age, what is transmitted over the internet is intelligent flow, which is a model-based abstraction of data processed and refined deeply by computing power. A core feature of intelligent computing is to process massive data items in the computing power pool using numerical computing, data analysis, and artificial intelligence algorithms to obtain intelligent models, which are then embedded in various processes of the information world and the physical world.

The Chinese government has proactively laid out new infrastructure, seizing the initiative in global competition.First, data has become a national strategic information resource.Data has both resource element and value processing attributes, including production, acquisition, transmission, aggregation, circulation, trading, ownership, assets, and security. China should continue to strengthen the construction of national data hubs and data circulation infrastructure.

Second, AI large models are a type of algorithmic infrastructure in the data space. Based on general large models, build infrastructure for the research and application of large models, supporting enterprises in developing specialized large models for industries such as robotics, autonomous driving, wearable devices, smart homes, and intelligent security, covering long-tail applications.

Finally, the construction of a national integrated computing power network plays a pioneering role in promoting the infrastructure of computing power. The Chinese solution for computing power infrastructure should significantly reduce the cost and threshold of using computing power while providing high-throughput, high-quality intelligent services for the widest range of people. The Chinese solution for computing power infrastructure needs to achieve “two lows and one high”, that is, on the supply side, significantly reduce the total costs of computing power devices, computing power equipment, network connections, data acquisition, algorithm model invocation, power consumption, operation maintenance, and development deployment, making high-quality computing power services affordable for small and medium-sized enterprises, encouraging them to develop computing network applications; on the consumer side, significantly lower the threshold for the use of computing power for the general public, ensuring that public services are easy to access and use, like water and electricity, and that computing power services can be easily customized like writing a webpage. In terms of service efficiency, China’s computing power services should achieve low entropy and high throughput, where high throughput refers to the ability to provide high-concurrency services while maintaining a high satisfaction rate for end-to-end service response times; low entropy means ensuring system throughput does not sharply decline under high-concurrency loads with resource disorder. Ensuring “more computing power” is especially important for China.

Choice 3: Should AI+ focus on empowering the virtual economy or strengthen the real economy?

The effectiveness of “AI+” is a touchstone for the value of artificial intelligence. After the subprime mortgage crisis, the value added of the US manufacturing industry as a percentage of GDP dropped from 28% in 1950 to 11% in 2021, and the proportion of employment in the US manufacturing industry dropped from 35% in 1979 to 8% in 2022, indicating that the US tends to favor the virtual economy with higher returns, neglecting the real economy with high investment costs and low economic returns. China tends to develop the real economy alongside the virtual economy, placing greater emphasis on developing equipment manufacturing, new energy vehicles, photovoltaic power generation, lithium batteries, high-speed rail, and 5G as part of the real economy.

Correspondingly, AI in the US is mainly applied to the virtual economy and IT infrastructure tools, reflecting the trend of “de-physicalization” since 2007, with Silicon Valley continuously promoting virtual reality (VR), the metaverse, blockchain, Web 3.0, deep learning, and AI large models.

China’s advantage lies in the real economy, with the most complete range of global manufacturing industries and the most comprehensive systems, characterized by diverse scenarios and abundant private data. China should select several industries to increase investment, forming a paradigm that can be promoted across all industries with low thresholds, such as choosing the equipment manufacturing industry as a representative industry to continue its advantages and the pharmaceutical industry as a representative industry to rapidly narrow the gap. The technical challenge of empowering the real economy lies in the integration of AI algorithms with physical mechanisms.

The key to the success of artificial intelligence technology is whether it can significantly reduce the costs of an industry or product, thereby expanding the number of users and the scale of the industry by tenfold, producing transformative effects similar to those of the steam engine on the textile industry or the smartphone on the internet industry.

China should find a high-quality development path suitable for its own artificial intelligence to empower the real economy.

(The speaker is an academician of the Chinese Academy of Engineering, a researcher at the Institute of Computing Technology, Chinese Academy of Sciences, and the director of the academic committee)

Notes:

① Pattern recognition refers to the use of computational methods to classify samples into certain categories based on their characteristics, studying the automatic processing and interpretation of patterns through mathematical methods by computers, with major research directions including image processing and computer vision, speech language information processing, brain networks, and brain-like intelligence.

② Token refers to symbols used to represent words or phrases in the natural language processing process. Tokens can be single characters or sequences composed of multiple characters.

③ General artificial intelligence refers to a type of artificial intelligence that possesses intelligence comparable to or exceeding that of humans. General artificial intelligence can not only perceive, understand, learn, and reason like humans but also flexibly apply knowledge across different fields, learn quickly, and think creatively. The research goal of general artificial intelligence is to seek a unified theoretical framework to explain various intelligent phenomena.

④ Chip manufacturing processes refer to the processes used to manufacture CPUs or GPUs, specifically the size of transistor gate circuits, measured in nanometers. The most advanced processes currently in mass production are represented by TSMC’s 3nm. More advanced manufacturing processes allow CPUs and GPUs to integrate more transistors, providing more functions and higher performance while being smaller and cheaper.

⑤ CUDA is a parallel computing platform and programming model designed and developed by NVIDIA, which includes the CUDA instruction set architecture and the parallel computing engine within GPUs. Developers can use the C language to write programs for the CUDA architecture, and the programs can run at ultra-high performance on processors that support CUDA.

⑥ RISC-V (pronounced “risk-five”) is an open general-purpose instruction set architecture initiated by the University of California, Berkeley, which allows anyone to freely use the RISC-V instruction set to design, manufacture, and sell chips and software, unlike other paid instruction sets.

⑦ The long-tail effect refers to the phenomenon where products or services that were originally overlooked, with small sales but many varieties, accumulate total revenues that exceed those of mainstream products due to their sheer volume. This effect is particularly significant in the internet field.

⑧ High concurrency generally refers to the design ensuring that the system can process many requests in parallel simultaneously.

Recommended Hot Books

Development of Artificial Intelligence and Intelligent Computing

<Click the book cover for details>

▊《Attention Mechanism in Artificial Intelligence: Analysis of Systems, Models, and Algorithms

Fu Gang

  • Integrating years of frontline work experience from senior development engineers, this book discusses the background and development history of the attention mechanism, providing detailed theoretical analysis and focusing on the introduction of the attention mechanism in the two major AI directions of computer vision and natural language processing, while also expanding its applications in other intelligent fields. The content is presented in a step-by-step manner with detailed illustrations, allowing readers to quickly and deeply grasp various experiences and techniques.

Development of Artificial Intelligence and Intelligent Computing

<Click the book cover for details>

▊《Deep Learning and Artificial Intelligence Practical Guide》

Zhang Chongsheng

  • The content includes principle explanations, practical cases, source code, and datasets, helping technical personnel quickly master knowledge and skills in deep learning.

  • Comprehensively covering relevant technologies in deep learning, including the basics of deep learning programming, target classification and detection, image segmentation, and generative adversarial networks.

Development of Artificial Intelligence and Intelligent Computing

<Click the book cover for details>

▊《GPT: The Ultimate Guide to Building NLP Products Using OpenAI API》

[English] Sandra Kublik; [English] Shubham Saboo

Translated by Li Zhaojun

  • Led by the Vice President of OpenAI, this book is highly recommended by nine AI experts from both domestic and international backgrounds.

  • It serves as a perfect starting point for practitioners and developers who want to understand the GPT-3 language model and learn how to build applications on the OpenAI API.

  • It provides a comprehensive guide on how to easily use the OpenAI API, exploring methods to use the tool based on specific user needs, and showcasing successful business cases based on GPT-3.

Development of Artificial Intelligence and Intelligent Computing

<Click the book cover for details>

▊《Practical AI Large Models》

You Yang

  • This book provides a detailed introduction from basic concepts to practical techniques, offering a comprehensive interpretation of AI large models, presented in a step-by-step manner. The book includes QR codes for videos, allowing readers to immerse themselves and quickly and deeply grasp various experiences and techniques. It also includes rich additional resources: open-source tools and libraries, datasets, model case studies, and practical applications, as well as an online communication community, enabling readers to utilize these resources comprehensively for a richer learning experience and accelerate their learning and growth.

Source: WeChat public account “Network Law Frontline”

Editor: Zhang Shuqian Review: Cao Xinyu

Leave a Comment