Global server shipments slid slightly from a quarter ago in the first quarter of 2024 and will rise around 2% both sequentially and on-yearly in the second quarter.
As air cooling grew insufficient for the heat dissipation of datacenters when running AI applications, L2A and L2L cooling methods are likely to see rising penetration to datacenters in the upcoming years.
After a double-digit on-year decline in 2023, global server shipments are expected to resume growth in 2024, but the volumes will fall back to slightly above the 2019 level.
Based on DIGITIMES Research's statistics, fourth-quarter 2023 global server shipments grew 3.1% from the prior quarter, coming somewhat short of expectation.
The three primary high performance computing (HPC) chip suppliers, Intel, AMD, and Nvidia, all experienced lackluster sales in the first half of 2023 amid weak demand for HPC applications, according to a survey of DIGITIMES Research.
In response to explosive generative AI and large language model (LLM) demand, major cloud service providers and leading server brands are stepping up efforts toward AI servers, with a focus on ramping up their procurement of high-end AI servers featuring accelerators with high bandwidth memory (HBM) integrated.
Global server shipments slipped 5.7% sequentially in the second quarter and are likely to see only single-digit on-quarter growth in the third quarter due to brand vendors and CSPs both turning conservative toward the second half of the year, even though the new mainstream server CPU platforms are ready for supply.
Demand for generative AI (Gen AI) and large language model (LLM) is rising rapidly, driven by the emergence of ChatGPT, a chatbot developed by OpenAI. For their large scale as well as the massive data sets and resources required to train LLMs, cloud service providers (CSP) are generally adopting the method of combining inference and prompt engineering for their AI solutions to support clients' customization needs.