Amazon Begins Migrating Alexa's Cloud AI From GPUs to Its Own Inferentia Chips - Crossminds
CrossMind.ai logo
Amazon Begins Migrating Alexa's Cloud AI From GPUs to Its Own Inferentia Chips
Nov 15, 2020
|
70 views
Details
The Amazon Alexa team announced on November 12 that they have migrated a majority of their GPU-based machine learning inference workloads to AWS Inferentia Application Specific Integrated Circuit (ASIC). This shift from Nvidia GPU hardware to Amazon's own Inferentia chip resulted in 30-percent lower cost and 25-percent improvement in end-to-end latency on Alexa's text-to-speech workloads.
Comments
loading...
Reactions (0) | Note
    📝 No reactions yet
    Be the first one to share your thoughts!
loading...
Recommended