But the same qualities that make those graphics processor chips, or GPUs, so effective at creating powerful AI systems from scratch make them less efficient at putting AI products to work. That’s ...
Intel Corp. is reportedly in talks to acquire SambaNova Systems Inc., a venture-backed developer of artificial intelligence chips. Bloomberg today cited sources as saying that the discussions are in ...
Everyone is not just talking about AI inference processing; they are doing it. Analyst firm Gartner released a new report this week forecasting that global generative AI spending will hit $644 billion ...
AMD is strategically positioned to dominate the rapidly growing AI inference market, which could be 10x larger than training by 2030. The MI300X's memory advantage and ROCm's ecosystem progress make ...
The major cloud builders and their hyperscaler brethren – in many cases, one company acts like both a cloud and a hyperscaler – have made their technology choices when it comes to deploying AI ...
Nvidia is aiming to dramatically accelerate and optimize the deployment of generative AI large language models (LLMs) with a new approach to delivering models for rapid inference. At Nvidia GTC today, ...
AMD’s Q2 2025 data center revenue reached $3.24 billion, up 14% year-over-year, with EPYC CPUs powering 1,200+ cloud instances. The $800 million MI308 impairment weighed on results, but MI355 promises ...
At the GTC 2025 conference, Nvidia introduced Dynamo, a new open-source AI inference server designed to serve the latest generation of large AI models at scale. Dynamo is the successor to Nvidia’s ...
Kubernetes has become the leading platform for deploying cloud-native applications and microservices, backed by an extensive community and comprehensive feature set for managing distributed systems.
The CNCF is bullish about cloud-native computing working hand in glove with AI. AI inference is the technology that will make hundreds of billions for cloud-native companies. New kinds of AI-first ...