Google’s Custom Machine Learning Chips Are 15-30x Faster Than GPUs and CPUs

Four years ago, Google was faced with a conundrum: if all its users hit its voice recognition services for three minutes a day, the company would need to double the number of data centers just to handle all of the requests to the machine learning system powering those services, reads a PCWorld article, which talks about how Tensor Processing Unit (TPU), a chip that is designed to accelerate the inference stage of deep neural networks came into being. The article shares an update: Google published a paper on Wednesday laying out the performance gains the company saw over comparable CPUs and GPUs, both in terms of raw power and the performance per watt of power consumed. A TPU was on average 15 to 30 times faster at the machine learning inference tasks tested than a comparable server-class Intel Haswell CPU or Nvidia K80 GPU. Importantly, the performance per watt of the TPU was 25 to 80 times better than what Google found with the CPU and GPU.


Share on Google+

Read more of this story at Slashdot.

Leave a Reply

Your email address will not be published. Required fields are marked *

Home | About | Contact | Disclaimer | Terms | Privacy

Copyright © 2017 by Tom Connelly | All Rights Reserved