当前位置:主页 > 要闻 >

Infinigence Unveils Next-Gen AI Infrastructure Suite, Aims to Lead China’s A...

时间:2025-07-30 14:27:32

  
 

  Xia Lixue, Co-founder and CEO of Infinigence

  AsianFin -- Infinigence, an AI infrastructure startup backed by Tsinghua University, introduced a sweeping portfolio of performance-optimized computing platforms targeting the full spectrum of AI deployment at this year’s World Artificial Intelligence Conference .

  The company officially launched three flagship products under its integrated solution suite: Infinicloud, a global-scale AI cloud platform for clusters of up to 100,000 GPUs; InfiniCore, a high-performance intelligent computing platform designed for multi-thousand-GPU clusters; and InfiniEdge, a lean, edge computing solution optimized for terminal deployments with as few as one GPU.

  Together, the platforms represent what CEO Xia Lixue calls a “software-hardware co-designed infrastructure system for the AI 2.0 era.” Built for compatibility across heterogeneous computing environments, the Infinigence stack offers full lifecycle support—from model scheduling and performance optimization to large-scale application deployment.

  “We’re addressing a core bottleneck in China’s AI industry: fragmentation in compute infrastructure,” Xia said. “With InfiniCloud, InfiniCore, and InfiniEdge, we’re enabling AI developers to move seamlessly between different chips, architectures, and workloads—unlocking intelligent performance at scale.”

  In a fast-evolving AI landscape dominated by open-source large language models such as DeepSeek, GLM-4.5, and MiniMax M1, Chinese infra startups are racing to build the backbone that powers model deployment and inference.

  Early on July 29, Infinigence announced that InfiniCloud now supports Zhipu AI’s latest GLM-4.5 and GLM-4.5-air models, which currently rank third globally in performance. The move signals Infinigence’s ambition to anchor the growing synergy between Chinese model developers and domestic chipmakers.

  Xia likened the trio of newly launched platforms to “three bundled boxes” that can be matched to AI workloads of any scale. “From a single smartphone to clusters of 100,000 GPUs—our system is designed to ensure resource efficiency and intelligent elasticity,” he said.

  Infinigence’s platforms are already powering Shanghai ModelSpeed Space, the world’s largest AI incubator. The facility sees daily token call volumes exceed 10 billion, supports over 100 AI use cases, and reaches tens of millions of monthly active users across its applications.

  A key challenge for China’s AI infrastructure sector is hardware heterogeneity. With dozens of domestic chip vendors and proprietary architectures, developers often struggle to port models across systems.

  Xia emphasized that Infinigence has developed a “universal compute language” that bridges chips with disparate instruction sets. “We treat computing resources like supermarket goods—plug-and-play, interoperable, and composable,” he said.

  The company’s infrastructure has already achieved full-stack adaptation for more than a dozen domestic chips, delivering 50%–200% performance gains through algorithm and compiler optimization. It also supports unified scheduling and mixed-precision computing, enabling cost-performance ratios that beat many international offerings.

  “What’s missing in China’s ecosystem is a feedback loop,” Xia said. “In the U.S., NVIDIA and OpenAI form a tight cycle: model developers know what chips are coming, and chipmakers know what models are being built. We’re building that loop domestically.”

  Infinigence is also targeting AI democratization with a first-of-its-kind cross-regional federated reinforcement learning system. The system links idle GPU resources from different regional AIDC centers into a unified compute cluster—allowing SMEs to build and fine-tune domain-specific inference models using consumer-grade cards.

  To support this, Infinigence launched the “AIDC Joint Operations Innovation Ecosystem Initiative” in partnership with China’s three major telecom providers and 20+ AIDC institutions.

  Xia noted that while training still depends heavily on NVIDIA hardware, inference workloads are rapidly migrating to domestic accelerators. “Users often start with international chips on our platform, but we help them transition to Chinese cards—many of which now deliver strong commercial value,” he said.

  Infinigence has also rolled out a series of on-device and edge inference engines under its Infini-Ask line. These include:

  Infini-Megrez2.0, co-developed with the Shanghai Institute of Creative Intelligence, the world’s first on-device intrinsic model.

  Infini-Mizar2.0, built with Lenovo, which enables heterogeneous computing across AI PCs, boosting local model capacity from 7B to 30B parameters.

  A low-cost FPGA-based large model inference engine, jointly developed with Suzhou Yige Technology.

  Founded in May 2023, Infinigence has raised more than RMB 1 billion in just two years, including a record-setting RMB 500 million Series A round in 2024—the largest to date in China’s AI infrastructure sector.

  Its product portfolio now spans everything from model hosting and cloud management to edge optimization and model migration—serving clients across intelligent computing centers, model providers, and industrial sectors.

  The company’s broader mission, Xia said, is to balance scale, performance, and resource availability. “Our vision is to deliver ‘boundless intelligence and flawless computing’—wherever theres compute, we want Infinigence to be the intelligence that flows through it.”

  IEEE Fellow and Tsinghua professor Wang Yu, also a co-founder of Infinigence, argued that the future of China’s AI economy depends on interdisciplinary collaboration. “We need people who understand chips, models, commercialization, and investment,” Wang said. “Only then can we solve the ‘last mile’ problem—connecting AI research with real-world deployment.”

  As China looks to decouple from foreign hardware dependence while competing globally in next-gen AI, Infinigence is positioning itself as a vital enabler—fusing chip-level control with cloud-scale ambition.

  “Every AI system runs on two forces: models and compute,” Xia said. “They cannot evolve in silos—they must move forward in sync.”

热点推荐
1 美国现货比特币ETF连续9天净流入达21.2亿

美国现货比特币ETF在4月14日至4月24日期间,连续9天录得净流入,总额达到约21.2亿美元,显示出...

2 日本交易所集团考虑修改TOPIX指数纳入规

消息,日本交易所集团旗下的JPXI正在考虑修改TOPIX指数的纳入规则,拟将加密资产占总资产5...

3 稳定币总市值创3200亿美元历史新高,BN

消息,据Binance Research数据,稳定币总市值已创下约3200亿美元的历史新高。数据显示,自2025年...

4 新加坡警方反诈骗行动成功拦截286万新元

新加坡警方反诈骗中心与网络调查部门联合Coinbase、Coinhako、StraitsX、Upbit等平台,在3月16日至...

5 Satsuma Technology增持22.77 BTC现持有668.48 BT

消息,英国上市公司Satsuma Technology近日增持22.77 BTC,目前总持有量达到668.48 BTC,当前在比特币...

6 特朗普报告3月债券购买交易

消息,美国总统特朗普在最新的财务披露报告中透露,他在3月份进行了数十笔债券买卖交易,...

7 委内瑞拉反对派领导人提出以石油换取比

委内瑞拉反对派领导人Mara Corina Machado提出以出售该国石油换取比特币,并建立国家级比特币储...

8 Grayscale再次质押102,400枚ETH,价值约2.37亿

消息,据链上监测平台Lookonchain数据,Grayscale在10小时前再次质押了102,400枚ETH,按当前价格计算...

9 Meta签下AWS数十亿美元Graviton5大单,数千万

消息,Meta与AWS签署多年期合作协议,合同价值数十亿美元,部署数千万个AWS Graviton5 CPU核心。...

10 以太坊基金会与Lido、Mellow完成21269枚AWE

消息,吴说获悉,以太坊基金会表示,已与Lido和Mellow协调完成约21,269枚AWETH向WSTETH的双边兑换...

成都来彰科技 蜀ICP备2025134723号-1

资讯来源互联网,如有版权问题请联系管理员删除。