The future of AI compute is heterogenous, according to Microsoft's GM of Azure Maia Andrew Wall. The implications of this are ...
Microsoft combines accelerated computing with cloud scale engineering to bring advanced AI capabilities to our customers. For years, we’ve worked with NVIDIA to integrate hardware, software and ...
These tech stocks look particularly well positioned to benefit from this opportunity.
As AI compute costs rise, Microsoft is seeking to reduce reliance on third-party chips, extending its push from custom ...
KubeCon Europe 2026 made AI inference its central focus with major CNCF donations including llm-d, Nvidia's GPU DRA driver ...
Calling it the highest performance chip of any custom cloud accelerator, the company says Maia is optimized for AI inference on multiple models. Signaling that the future of AI may not just be how ...
The big four cloud giants are turning to Nvidia's Dynamo to boost inference performance, with the chip designer's new Kubernetes-based API helping to further ease complex orchestration. According to a ...
Gimlet Labs just raised an $80 million Series A for tech that lets AI run across NVIDIA, AMD, Intel, ARM, Cerebras and ...
Nvidia announcements show the current shortage of storage and memory could continue into the future, driving up prices and ...
Azilen launches Inference Engineering practice to optimize AI performance, reduce costs, and scale efficiently across ...