⬤ Amazon Web Services has released its EC2 Trn3 UltraServers. Each server contains the new Trainium3 AI chip, which is built with a 3-nanometer process and built for large scale generative-AI jobs. AWS calls this its largest advance in custom silicon to date - the target users range from corporations that train huge models to research groups that run live inference.
⬤ The Trainium3 chip performs 4.4 times quicker than the prior chip and needs only one quarter of the energy. One UltraServer holds 144 Trainium3 chips plus yields 362 FP8 petaflops within a single enclosure. AWS also states that UltraClusters 3.0 can expand to millions of chips for customers that run the heaviest AI workloads.
⬤ Decart besides Ricoh have already lowered their spend by up to fifty percent while still meeting real time performance goals. Those first numbers show that custom AI hardware is turning into a requirement for firms that want to escape GPU supply limits and to keep infrastructure costs under control.
⬤ The arrival of Trainium3 marks a clear change in the way cloud giants build AI infrastructure. By investing heavily in custom silicon but also hyperscale systems, Amazon places itself as a strong entrant in the AI hardware contest and resets what customers expect for training cost and compute availability while demand for large models keeps rising.
Saad Ullah
Saad Ullah