Fagen Wasanni Technologies August 14, 2023
Amazon Web Services (AWS) is working on two custom chips, Trainium and Inferentia, to enhance the training and inference of large language models (LLMs) on its cloud platform. LLMs are a type of artificial intelligence that can generate text, translate languages, and provide detailed answers.
Trainium, developed for LLM training, comes with numerous cores and high-bandwidth memory to support the large datasets needed for training these models. On the other hand, Inferentia is an inference chip with fewer cores but greater energy efficiency. It allows LLMs to be deployed on various devices.
While other companies like NVIDIA and Microsoft are also creating custom chips for LLM training, Amazon’s chips possess the advantage of being designed for use on AWS’s cloud...