At Advancing AI, AMD unveils MI355 with 35× gen-over-gen gains and doubles down on open innovation – from ROCm 7 to Helios infrastructure – to challenge NVIDIA’s AI leadership.
The deal marks a strategic move to bolster Qualcomm’s AI and custom silicon capabilities amid challenging competition and the potential start of a wave of AI silicon acquisitions.
A new partnership combines WEKA’s AI-native storage with Nebius’ GPUaaS platform to accelerate model training, inference, and innovation with microsecond latency and extreme scalability.
As the battle for AI market share continues, AMD’s recent acquisitions signal a strategic move toward optimizing both software and hardware for inference workloads and real-world AI deployment.
The HPE-owned platform combines unified observability, smart alert correlation, and automation to tackle hybrid IT complexity while also working with existing monitoring tools.
AIStor’s stateless, gateway-free design solves legacy storage issues, enabling high-performance object-native infrastructure for exabyte-scale AI and analytics workloads.
From agentic and embodied AI to trust and human capacity, Stanford’s Daniel Wu shares how infrastructure and intentionality will shape the next era of AI transformation.
From open-source models to enterprise partnerships, Together AI drives multidimensional growth — accelerating adoption, engagement, and impact across the entire AI development lifecycle.
From AI scandals to quantum breakthroughs, Allyson Klein shares what’s tracking with her 2025 tech forecast — and what caught her off guard halfway through the year.
State Street’s Anusha Nerella shares how financial institutions are laying the groundwork for responsible AI adoption, balancing innovation with compliance at scale.
Stanford’s Daniel Wu unpacks AI democratization — exploring agentic & embodied AI, multi-modal models, and trustworthy systems. Learn more at Daniel’s AI Infra Summit 2025 live presentation.
CoreWeave acquires Core Scientific in a $9B all-stock deal, unlocking 1.3 GW of power and advancing its vision of vertically integrated AI infrastructure for next-gen hyperscale workloads.
State Street’s Anusha Nerella explores AI-driven FinTech infrastructure—scalability, governance and agentic computing. Interested in finding out more about the AI Infra Summit and seeing Anusha Nerella live? Find out more here.
In this episode of In the Arena, hear how cross-border collaboration, sustainability, and tech are shaping the future of patient care and innovation.
Tune in to our latest episode of In the Arena to discover how Verge.io’s unified infrastructure platform simplifies IT management, boosts efficiency, & prepares data centers for the AI-driven future.
Join us on Data Insights as Mark Klarzynski from PEAK:AIO explores how high-performance AI storage is driving innovation in conservation, healthcare, and edge computing for a sustainable future.
Untether AI's Bob Beachler explores the future of AI inference, from energy-efficient silicon to edge computing challenges, MLPerf benchmarks, and the evolving enterprise AI landscape.
Explore how OCP’s Composable Memory Systems group tackles AI-driven challenges in memory bandwidth, latency, and scalability to optimize performance across modern data centers.
State Street’s Anusha Nerella explores AI-driven FinTech infrastructure—scalability, governance and agentic computing. Interested in finding out more about the AI Infra Summit and seeing Anusha Nerella live? Find out more here.
In this episode of In the Arena, hear how cross-border collaboration, sustainability, and tech are shaping the future of patient care and innovation.
Tune in to our latest episode of In the Arena to discover how Verge.io’s unified infrastructure platform simplifies IT management, boosts efficiency, & prepares data centers for the AI-driven future.
Join us on Data Insights as Mark Klarzynski from PEAK:AIO explores how high-performance AI storage is driving innovation in conservation, healthcare, and edge computing for a sustainable future.
Untether AI's Bob Beachler explores the future of AI inference, from energy-efficient silicon to edge computing challenges, MLPerf benchmarks, and the evolving enterprise AI landscape.
Explore how OCP’s Composable Memory Systems group tackles AI-driven challenges in memory bandwidth, latency, and scalability to optimize performance across modern data centers.