CoreWeave Advances AI-Native Cloud Platform for the Next Phase of Production-Scale AI
New capabilities to develop, deploy, and iterate AI faster for mission critical workloads across financial services, robotics, and industrial use cases
This press release features multimedia. View the full release here: https://www.businesswire.com/news/home/20260316504524/en/
NVIDIA HGX B300, now generally available on CoreWeave.
As the industry shifts from large-scale training to iterative RL, infrastructure requirements are rapidly evolving. By combining the latest hardware with CoreWeave’s AI optimized cloud services and advanced development workflows, CoreWeave is closing the gap between training and production to support the next generation of self-improving agents and physical AI workloads.
“The next phase of AI is being defined by how efficiently AI systems can run and scale in production,” said
“The era of AI is shifting from training models to operating agents at scale,” said
Purpose-Built for Agentic AI: NVIDIA HGX B300
CoreWeave Cloud unlocks NVIDIA HGX B300 performance at scale, marking a major step forward for frontier and agentic workloads. With the general availability of NVIDIA HGX B300, part of the NVIDIA Blackwell Ultra platform, customers will have access to:
- Performance Leap: NVIDIA HGX B300 is designed for AI reasoning and inference with enhanced compute and increased memory.
- Massive Memory: Featuring 2.1 TB of HBM3e memory—a 50% increase over HGX B200 instances—enabling teams to run long-context inference with low latency and train models with 100B+ parameters on a single node.
- Unprecedented Bandwidth: Next-generation NVIDIA Quantum-X800XDR InfiniBand support doubles node-to-node bandwidth, eliminating interconnect bottlenecks.
- Liquid-Cooled Reliability: Every HGX B300 server on CoreWeave Cloud is managed by state-of-the-art liquid cooling to eliminate thermal throttling and allow sustained peak performance.
“We’ve already run production workloads with CoreWeave on NVIDIA HGX B200, and that experience built real confidence in their ability to operate at scale,” said
CoreWeave also expects to be among the first cloud providers to deploy the NVIDIA Vera Rubin NVL72 platform and NVIDIA Vera CPU rack in production in the second half of 2026. This expansion will further support large-scale inference, reasoning, and the most demanding agentic AI applications.
Direct "On-the-Job" Training for Agents with Environment-Free RL
RL is the most effective method for post-training LLMs on agentic tasks, but it traditionally requires costly, time-consuming, and difficult-to-model simulated environments.
- Efficiency: Achieve 1.4x faster training at up to 40% lower cost compared to self-managed clusters.
- Superior Results: Deliver inference with up to 5x lower cost and up to 60x lower latency without quality loss.
Research evaluations often rely on limited datasets and synthetic scenarios, leaving agents unprepared for real-world production. This gap makes it difficult to feed production insights back into the research loop, slowing time-to-market and resulting in agents that fail to meet expectations.
Production agent evaluations in W&B Weave close that gap, connecting research and production, enabling teams to learn from real user interactions. It automatically detects production failure modes, alerts researchers, and feeds insights back to improve evaluation datasets and accelerate iteration. By closing this loop, W&B Weave helps teams ship more reliable agents faster.
- Agent Self-Improvement: Incoming traces are automatically scored by LLM judges. Lower-quality outputs are routed for human review and turned into test cases, making evaluation suites smarter over time. Specialized classifier models then analyze the production traces and identify failure modes.
-
Accelerated Development : Smart alerts notify when new failure modes are detected, allowing customers to refine agents in production using data from real user interactions and continuous monitoring.
Accelerating Robotics AI Development
W&B Models now allows robotics and embodied AI customers to track, compare, and reproduce multimodal experiments faster. Teams can log training metrics, simulation outputs, and video in a single workspace, and use a new comparison panel to evaluate up to four images or videos side-by-side. In collaboration with NVIDIA,
Monitor Training Anywhere With the First Weights & Biases Mobile App
To enable AI pioneers to manage these massive production runs from anywhere, CoreWeave introduced the first iOS app designed specifically for AI model development. The
CoreWeave’s AI cloud delivers industry-leading performance and efficiency through an end-to-end technology stack optimized for AI workloads. It has achieved top MLPerf benchmark for AI workloads, Platinum rankings in both SemiAnalysis ClusterMAX™ 1.0 and 2.0, and NVIDIA Exemplar Cloud validation for training and inference on GB200 NVL72. Visit us at booth #913 at NVIDIA GTC to learn more about today’s announcements.
About CoreWeave
CoreWeave is The Essential Cloud for AI™. Built for pioneers by pioneers, CoreWeave delivers a platform of technology, tools, and teams that enables innovators to move at the pace of innovation, building and scaling AI with confidence. Established in 2017, CoreWeave completed its public listing on Nasdaq (CRWV) in
View source version on businesswire.com: https://www.businesswire.com/news/home/20260316504524/en/
Source: