Matrix Multiplication Mayhem
Also in today's edition: Web3 Alternatives, Code Readability vs Performance & Can Google's Bard Earn From Ads?
The tech world has been on fire lately with Microsoft and Google's intense showdown. From GPT-4 vs LaMDA to ChatGPT vs Bard, the heat has now reached the Bing vs Google Search battle. But what's brewing under the hood is a friendly yet fierce competition between NVIDIA's CUDA-powered GPUs and Google's custom TPUs over matrix multiplication.
Solving for efficient matrix multiplication can significantly reduce the computing resources required for training and inferencing tasks. While other methods like quantisation and model shrinking have also proven to cut down on computing, they lack accuracy.
Keep reading with a 7-day free trial
Subscribe to Sector 6 | The Newsletter of AIM to keep reading this post and get 7 days of free access to the full post archives.