Enterprise AI has learned to generate. It has learned to reason. Now companies are asking the next question: How should AI act? Early agent systems have shown what’s possible, moving beyond simple prompts to take on more complex tasks. The next step is bringing those capabilities into enterprise environments — where agents must operate with […]
NVlabs releases cuda-oxide v0.1.0, a custom rustc codegen backend that compiles #[kernel]-annotated Rust functions to PTX through a Rust → Stable MIR → Pliron IR → LLVM IR → PTX pipeline, with single-source host+device compilation from one cargo oxide build command.
The post NVIDIA AI Just Released cuda-oxide: An Experimental Rust-to-CUDA Compiler Backend that Compiles SIMT GPU Kernels Directly to PTX appeared first on MarkTechPost.
NVIDIA researchers have introduced Star Elastic, a post-training method that embeds multiple nested reasoning models — at 30B, 23B, and 12B parameter scales — inside a single checkpoint, eliminating the need for separate training runs or stored model weights per variant. Built on the Nemotron Elastic framework and applied to Nemotron Nano v3, the method trains all three variants in a single 160B-token run, achieving a 360× token reduction compared to pretraining each model from scratch. Beyond training efficiency, Star Elastic introduces elastic budget control — a novel inference scheme that uses a smaller submodel for the thinking phase and the full model for the final answer — delivering up to 16% higher accuracy and 1.9× lower latency compared to standard budget control, while nested FP8 and NVFP4 checkpoints bring the full model family within reach of RTX-class GPUs.
The post NVIDIA AI Releases Star Elastic: One Checkpoint that Contains 30B, 23B, and 12B Reasoning Models with Zero-
AI will help build the energy it needs. That’s the case U.S. Energy Secretary Chris Wright and NVIDIA Vice President of Hyperscale and High-Performance Computing Ian Buck made Thursday morning at the SCSP AI+ Expo. The 30-minute fireside chat, moderated by SCSP president Ylli Bajraktari, was called “Powering the Next American Century.” Their argument: American […]
LONDON, May 7, 2026 — The AI Summit London today unveiled its keynote speaker line-up for the event’s 10th anniversary edition, bringing together senior leaders, technologists and policymakers from across the […]
The post AI Summit London Unveils 10th Anniversary Speaker Line-Up Featuring Global Leaders in Enterprise AI appeared first on AIwire.
SANTA CLARA, Calif. and CORNING, N.Y., May 7, 2026 — NVIDIA and Corning Incorporated have announced a multiyear commercial and technology partnership to dramatically expand U.S.-based manufacturing of the advanced optical […]
The post NVIDIA and Corning Announce Long-Term Partnership to Strengthen US Manufacturing for AI Infrastructure appeared first on AIwire.
MRC (Multipath Reliable Connection) is a new open networking protocol developed by OpenAI in partnership with AMD, Broadcom, Intel, Microsoft, and NVIDIA that improves GPU networking performance and resilience in large-scale AI training clusters by spreading packets across hundreds of paths simultaneously, recovering from network failures in microseconds, and enabling supercomputers with over 100,000 GPUs to be built using only two tiers of Ethernet switches.
The post OpenAI Introduces MRC (Multipath Reliable Connection): A New Open Networking Protocol for Large-Scale AI Supercomputer Training Clusters appeared first on MarkTechPost.