Microsoft has built one of the world’s most powerful supercomputers dedicated to machine learning, but is missing a key AI metric on power consumption.
The system has been built for OpenAI and is hosted in the Azure supercomputer with more than 285,000 CPU cores, 10,000 GPUs and 400 gigabits per second of network connectivity for each GPU server. This puts it in the top five supercomputers in the world, says Microsoft.
But unlike those discrete supercomputers which report power figures, the AI metric on power is missing from the recent announcement. This matters, as Microsoft says this is a first step toward making the next generation of very large AI models and the infrastructure needed to train them available as a platform.
A new class of massive, self-learning models has been developed by the AI research community as these tasks can be handled more efficiently in terms of computing power. These models have expanded from 1bn parameters last year to over 17bn.
“The exciting thing about these models is the breadth of things they’re going to enable,” said Microsoft Chief Technical Officer Kevin Scott, who said the potential benefits extend far beyond narrow advances in one type of AI model.
“This is about being able to do a hundred exciting things in natural language processing at once and a hundred exciting things in computer vision, and when you start to see combinations of these perceptual domains, you’re going to have new applications that are hard to even imagine right now,” he said.
This focus on the metrics for power consumption and efficiency matters. However Microsoft hasn’t detailed the cores or the interconnect being used. The company has been highly aware of the impact of power consumption on its data centres. It has a commitment to reduce its carbon footprint to zero by 2030 and compensate for its historical impact on the climate by 2050.