CONNECT WITH US

Generative AI, LLM, and AI servers

Jim Hsiao, DIGITIMES Research, Taipei 0

Credit: DIGITIMES

The DIGITIMES Research special report you are trying to open is sold a-la-carte and is also available with Servers Special Report Database services. Please sign in if you wish to continue.
Demand for generative AI (Gen AI) and large language model (LLM) is rising rapidly, driven by the emergence of ChatGPT, a chatbot developed by OpenAI. For their large scale as well as the massive data sets and resources required to train LLMs, cloud service providers (CSP) are generally adopting the method of combining inference and prompt engineering for their AI solutions to support clients' customization needs.
Abstract

Demand for generative AI (Gen AI) and large language model (LLM) is rising rapidly, driven by the emergence of ChatGPT, a chatbot developed by OpenAI. For their large scale as well as the massive data sets and resources required to train LLMs, cloud service providers (CSP) are generally adopting the method of combining inference and prompt engineering for their AI solutions to support clients' customization needs.

As such, cloud inference has now become the primary running model for LLMs. However, as language applications mostly require instant responses and need to support huge simultaneous usages by multiple users, only large clusters of high-speed interconnected AI servers can perform LLM inference that satisfies most of the usage scenarios.

First-tier CSPs are aggressively deploying Gen AI cloud services. Apart from the commonly known creation of content such as texts, images, documents, and codes, CSPs have also been actively promoting Gen AI platform as a service (PaaS), providing users with pre-trained models, prompt engineering tools, and all types of APIs to allow enterprises to quickly create customized application tools.

Table of contents
Price: NT$27,000 (approx. US$900)