AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
How Google XLA Decides Which Code to Run on the CPU or GPU
I think you explained to me last time how it actually works like how it decides what to run where on the CPU or GPU but I can't recall and I probably can't even fall you into the depths of that Jose so I won't ask let's regroup on explore. It's just a single option you say I want to run on the host or on Kuda and it's a single option specify and everything figures out for you don't have to do anything else. The way it works is that we have the same called numerical definitions and so if you think about your your machine learning model it's implemented using those numerical definitions, which are used by Google XLA in their search engine