Advertisement
Canada markets closed
  • S&P/TSX

    24,822.54
    +132.06 (+0.53%)
     
  • S&P 500

    5,864.67
    +23.20 (+0.40%)
     
  • DOW

    43,275.91
    +36.86 (+0.09%)
     
  • CAD/USD

    0.7246
    -0.0004 (-0.05%)
     
  • CRUDE OIL

    69.34
    -1.33 (-1.88%)
     
  • Bitcoin CAD

    94,454.29
    +69.29 (+0.07%)
     
  • XRP CAD

    0.75
    -0.00 (-0.33%)
     
  • GOLD FUTURES

    2,736.40
    +28.90 (+1.07%)
     
  • RUSSELL 2000

    2,276.09
    -4.76 (-0.21%)
     
  • 10-Yr Bond

    4.0730
    -0.0230 (-0.56%)
     
  • NASDAQ

    18,489.55
    +115.94 (+0.63%)
     
  • VOLATILITY

    18.03
    -1.08 (-5.65%)
     
  • FTSE

    8,358.25
    -26.88 (-0.32%)
     
  • NIKKEI 225

    38,981.75
    +70.56 (+0.18%)
     
  • CAD/EUR

    0.6666
    -0.0024 (-0.36%)
     

Tech war: China eyes supercomputers for building LLMs amid US sanctions on advanced chips

China must find an alternative approach to artificial intelligence (AI) development, in lieu of stacking up processors inside data centres, as US sanctions continue to bar the country's access to advanced semiconductors and chip-making equipment, according to industry experts on the mainland.

Leveraging supercomputing technology that China has developed over the past decade could help break the stranglehold of US-led restrictions on the mainland's AI industry, according to Zhang Yunquan, a researcher at the Institute of Computing Technology under the Chinese Academy of Sciences (CAS), who was quoted in a report on Monday by state-backed tabloid Global Times.

Supercomputing systems designed for training large language models (LLMs) - the technology underpinning generative AI services like ChatGPT - are crucial to replacing power-hungry, data-centre computing clusters, which typically employ from 10,000 to 100,000 graphics processing units (GPUs) for such training, Zhang said in a recent conference, according to the report.

Do you have questions about the biggest topics and trends from around the world? Get the answers with SCMP Knowledge, our new platform of curated content with explainers, FAQs, analyses and infographics brought to you by our award-winning team.

China's quest to establish a viable, advanced computing platform to train LLMs and develop AI applications shows the urgency of becoming technologically self-sufficient on the mainland, as its AI progress remains hindered by limited GPU choices amid US sanctions that have prevented top GPU firm Nvidia from supplying its most cutting-edge chips to the country.

More enterprises are using data centres - secure, temperature-controlled facilities that house large-capacity servers and data-storage systems - to host or manage computing infrastructure for their artificial intelligence projects. Photo: Shutterstock alt=More enterprises are using data centres - secure, temperature-controlled facilities that house large-capacity servers and data-storage systems - to host or manage computing infrastructure for their artificial intelligence projects. Photo: Shutterstock>

Nvidia is currently working on a version of its new flagship AI chips for the China market that would be compatible with current US export controls, according to a Reuters report last week that cited four sources familiar with the matter.

"I believe that [building] LLMs are not achieved by simply adding more chips," CAS academician Chen Runsheng said at the same conference, according to the Global Times report. "They must learn, like the human brain, to lower energy consumption, while improving their efficiency."

China - the biggest data centre market in the Asia-Pacific, with 3,956 megawatts of capacity - relies heavily on coal power. The country generated nearly two-thirds of its electricity from coal last year.

Chen called on China to work on fundamental research for intelligent computing of LLMs, combined with high-performance computing (HPC) technology, to achieve breakthroughs in computing power, the report said. HPC refers to the ability to process data and perform complex calculations at high speeds, which are accomplished by supercomputers containing thousands of compute nodes that work together to complete tasks.

Engineers work at the Wuhan Supercomputer Centre in central Hubei province on May 24, 2023. Photo: AFP alt=Engineers work at the Wuhan Supercomputer Centre in central Hubei province on May 24, 2023. Photo: AFP>

The batch of LLMs that have been developed on the mainland are based on models and algorithms developed by the US, without enough consideration of fundamental theories, according to Chen. "If we can make progress in fundamental theory, we will achieve groundbreaking and authentic innovation," Chen said.

At present, Chinese companies have been building computing clusters for training LLMs that have 10,000 GPUs or more, which include home-grown chips from Chinese GPU start-up Moore Threads Technology.

Chinese Big Tech firms, such as Tencent Holdings, have been making optimisations in their infrastructure, including speeding up network communications and boosting AI training efficiency. Tencent's Xingmai HPC network, for example, can support a single computing cluster with more than 100,000 GPUs, according to the company.

This article originally appeared in the South China Morning Post (SCMP), the most authoritative voice reporting on China and Asia for more than a century. For more SCMP stories, please explore the SCMP app or visit the SCMP's Facebook and Twitter pages. Copyright © 2024 South China Morning Post Publishers Ltd. All rights reserved.

Copyright (c) 2024. South China Morning Post Publishers Ltd. All rights reserved.