Waldman On The AI OffLoad Machine
So we found out that doing programs is great with the CPU, but then doing very tedious, high-I/O operations by the CPU was very inefficient. So we took this path to Mellanox and we do it mainly on the end points on Ethernet and InfiniBand. Then what we found out is, we can put computing inside the switch. And this is something we've done with Nvidia. … We have an AI (artificial intelligence) offload machine, floating point machines, inside the switch to increase the efficiency of artificial intelligence programs in the data center. We've seen some very interesting results. I don't know if you've shown this recently. But we are seeing more and more offloading we can take from the CPU and GPU into the network and then synergize this into the whole data center solution for AI.
Huang On The Path Forward For Nvidia Mellanox
That's our path forward. We've gone way beyond the time when CPU scaling is going to progress. And now that it's slowed down, we have to (accelerate) every workload that we can, and of course moving it into an (GPU) accelerator is one thing, but moving it into a network is completely another thing. But we should do both.