Amazon Begins Migrating Alexa's Cloud AI From GPUs to Its Own Inferentia Chips

Amazon Begins Migrating Alexa's Cloud AI From GPUs to Its Own Inferentia Chips

Nov 15, 2020
|
70 views
Details
The Amazon Alexa team announced on November 12 that they have migrated a majority of their GPU-based machine learning inference workloads to AWS Inferentia Application Specific Integrated Circuit (ASIC). This shift from Nvidia GPU hardware to Amazon's own Inferentia chip resulted in 30-percent lower cost and 25-percent improvement in end-to-end latency on Alexa's text-to-speech workloads.

Comments
loading...