Software stack manages GPUs, Kubernetes, and AI workloads at scale TOKYO–(BUSINESS WIRE)–SoftBank Corp. (TOKYO:9434, President & CEO: Junichi Miyakawa, “SoftBankSoftware stack manages GPUs, Kubernetes, and AI workloads at scale TOKYO–(BUSINESS WIRE)–SoftBank Corp. (TOKYO:9434, President & CEO: Junichi Miyakawa, “SoftBank

SoftBank Corp. Announces “Infrinia AI Cloud OS,” a Software Stack for AI Data Centers

Software stack manages GPUs, Kubernetes, and AI workloads at scale

TOKYO–(BUSINESS WIRE)–SoftBank Corp. (TOKYO:9434, President & CEO: Junichi Miyakawa, “SoftBank”) announced that its Infrinia Team*1, which works on the development of next-generation AI infrastructure architecture and systems, has developed “Infrinia AI Cloud OS,” a software stack*2 designed for AI data centers.

By deploying “Infrinia AI Cloud OS,” AI data center operators can build Kubernetes*3 as a Service (KaaS) in a multi-tenant environment, and Inference as a Service (Inf-aaS) that provides Large Language Model inference capabilities via APIs, as part of their own GPU cloud services. In addition, the software stack is expected to reduce total cost of ownership (TCO) as well as operational burden compared with bespoke solutions or in-house development. This will enable the rapid delivery of GPU cloud services that efficiently and flexibly support the full AI lifecycle—from AI model training to inference.

SoftBank plans to deploy “Infrinia AI Cloud OS” initially within its own GPU cloud services. Furthermore, the Infrinia Team aims to expand deployment to overseas data centers and cloud environments with a view to global adoption.

Background of “Infrinia AI Cloud OS” Development

The demand for GPU-accelerated AI computing is expanding rapidly across the generative AI, autonomous robotics, simulation, drug discovery, and materials development fields. As a result, user needs and usage patterns for AI computing are becoming increasingly diverse and sophisticated, and requirements including the following have emerged:

  • Access to infrastructure that is fully managed by GPU cloud service providers, abstracted GPU bare-metal servers
  • Cost-optimized, highly abstracted inference services without concerning with GPU management
  • Advanced operations in which AI models are trained and optimized on centralized servers and deployed for inference at the edge

Building and operating GPU cloud services that meet these requirements requires highly specialized expertise and involves complex operational tasks, placing a significant burden on GPU cloud service providers.

To address these challenges, the Infrinia Team developed “Infrinia AI Cloud OS,” a software stack that maximizes GPU performance while enabling the easy and rapid deployment and operation of advanced GPU cloud services.

Key Features of “Infrinia AI Cloud OS”

Kubernetes as a Service

  • Reduces the operational burden of managing the physical infrastructure and the Kubernetes software layer by automating the entire stack (from BIOS and RAID settings to the OS, GPU Drivers, networking, Kubernetes Controllers and Storage) on state-of-the-art GPU Platforms such as NVIDIA GB200 NVL72
  • Software-defined dynamic, on-the-fly physical connectivity (NVIDIA NVLink) and memory (Inter-Node Memory Exchange) reconfiguration, as the customers create, update and delete their clusters to suit their AI workload needs
  • Automatic node allocation based on GPU proximity and NVIDIA NVLink domain to reduce latency and maximize GPU-to-GPU bandwidth for highly distributed jobs

Inference as a Service

  • Enables users to deploy inference services simply by selecting Large Language Models, without working with Kubernetes or the underlying infrastructure
  • OpenAI-compatible APIs, enabling drop-in integration with existing AI applications
  • Seamless scaling across multiple nodes in core and edge platforms such as NVIDIA GB200 NVL72 and other platforms

Secure Multi-tenancy and High Operability

  • Tenant isolation through encrypted cluster communications and separation
  • Automation of operational maintenance, including system monitoring and failover
  • API environment for connecting to the AI data center’s portal, customer management systems, and billing systems

These key features allow AI data center operators with customer management systems, as well as enterprises offering GPU cloud services, to add advanced capabilities that enable efficient AI model training and inference while flexibly utilizing GPU resources, to their own GPU service offerings.

Junichi Miyakawa, President & CEO of SoftBank Corp., commented:

“To further deepen the utilization of AI as it evolves toward AI agents and Physical AI, SoftBank is launching a new GPU cloud service and software business to provide the essential capabilities required for the large-scale deployment of AI in society. At the core of this initiative is our in-house developed ‘Infrinia AI Cloud OS,’ a GPU cloud platform software designed for next-generation AI infrastructure that seamlessly connects AI data centers, enterprises, service providers and developers. The advancement of AI infrastructure requires not only physical components such as GPU servers and storage, but also software that integrates these resources and enables them to be delivered flexibly and at scale. Through Infrinia, SoftBank will play a central role in building the cloud foundation for the AI era and delivering sustainable value to society.”

For more information on “Infrinia AI Cloud OS,” please visit the website below:

https://infrinia.ai/

About SoftBank Corp.

Guided by the SoftBank Group’s corporate philosophy, “Information Revolution – Happiness for everyone,” SoftBank Corp. (TOKYO: 9434) operates telecommunications and IT businesses in Japan and globally. Building on its strong business foundation, SoftBank Corp. is expanding into non-telecom fields in line with its “Beyond Carrier” growth strategy while further growing its telecom business by harnessing the power of 5G/6G, IoT, Digital Twin and Non-Terrestrial Network (NTN) solutions, including High Altitude Platform Station (HAPS)-based stratospheric telecommunications. While constructing AI data centers and developing homegrown LLMs specialized for the Japanese language, SoftBank is integrating AI with radio access networks (AI-RAN), with the aim of becoming a provider of next-generation social infrastructure. To learn more, please visit https://www.softbank.jp/en/corp/

*1: The Infrinia Team is an Infrastructure Architecture & Systems Team established within SB Telecom America, Corp., a wholly owned subsidiary of SoftBank Corp., as part of the company’s broader initiative to advance next-generation AI infrastructure. The Infrinia Team is based in Sunnyvale, California USA.

*2: A software stack is a set of software components and functions used together to build and operate systems and applications.

*3: Kubernetes is an open-source system for automating the deployment and scaling of applications and for managing containerized applications.

  • SoftBank, the SoftBank name and logo are registered trademarks or trademarks of SoftBank Group Corp. in Japan and other countries.
  • Other company, product and service names in this press release are registered trademarks or trademarks of the respective companies.

Contacts

Kyoko Shimada

SoftBank Corp.

Corporate Communications

+81-3-6889-2301

sbpr@g.softbank.co.jp

Market Opportunity
Cloud Logo
Cloud Price(CLOUD)
$0.06141
$0.06141$0.06141
-0.29%
USD
Cloud (CLOUD) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

The Surprising 2025 Decline In Online Interest Despite Market Turmoil

The Surprising 2025 Decline In Online Interest Despite Market Turmoil

The post The Surprising 2025 Decline In Online Interest Despite Market Turmoil appeared on BitcoinEthereumNews.com. Bitcoin Searches Plunge: The Surprising 2025
Share
BitcoinEthereumNews2026/01/21 14:56
Cryptos Signal Divergence Ahead of Fed Rate Decision

Cryptos Signal Divergence Ahead of Fed Rate Decision

The post Cryptos Signal Divergence Ahead of Fed Rate Decision appeared on BitcoinEthereumNews.com. Crypto assets send conflicting signals ahead of the Federal Reserve’s September rate decision. On-chain data reveals a clear decrease in Bitcoin and Ethereum flowing into centralized exchanges, but a sharp increase in altcoin inflows. The findings come from a Tuesday report by CryptoQuant, an on-chain data platform. The firm’s data shows a stark divergence in coin volume, which has been observed in movements onto centralized exchanges over the past few weeks. Bitcoin and Ethereum Inflows Drop to Multi-Month Lows Sponsored Sponsored Bitcoin has seen a dramatic drop in exchange inflows, with the 7-day moving average plummeting to 25,000 BTC, its lowest level in over a year. The average deposit per transaction has fallen to 0.57 BTC as of September. This suggests that smaller retail investors, rather than large-scale whales, are responsible for the recent cash-outs. Ethereum is showing a similar trend, with its daily exchange inflows decreasing to a two-month low. CryptoQuant reported that the 7-day moving average for ETH deposits on exchanges is around 783,000 ETH, the lowest in two months. Other Altcoins See Renewed Selling Pressure In contrast, other altcoin deposit activity on exchanges has surged. The number of altcoin deposit transactions on centralized exchanges was quite steady in May and June of this year, maintaining a 7-day moving average of about 20,000 to 30,000. Recently, however, that figure has jumped to 55,000 transactions. Altcoins: Exchange Inflow Transaction Count. Source: CryptoQuant CryptoQuant projects that altcoins, given their increased inflow activity, could face relatively higher selling pressure compared to BTC and ETH. Meanwhile, the balance of stablecoins on exchanges—a key indicator of potential buying pressure—has increased significantly. The report notes that the exchange USDT balance, around $273 million in April, grew to $379 million by August 31, marking a new yearly high. CryptoQuant interprets this surge as a reflection of…
Share
BitcoinEthereumNews2025/09/18 01:01
Strategy Makes Biggest Bitcoin Bet In Months With $2.13B Buy

Strategy Makes Biggest Bitcoin Bet In Months With $2.13B Buy

The post Strategy Makes Biggest Bitcoin Bet In Months With $2.13B Buy appeared on BitcoinEthereumNews.com. Strategy Makes Biggest Bitcoin Bet In Months
Share
BitcoinEthereumNews2026/01/21 15:07