Serdar Yegulalp
Senior Writer

Microsoft’s Project Brainwave accelerates deep learning in Azure

news
Aug 24, 20172 mins

Using FPGAs deployed throughout the Azure cloud, Microsoft’s new system for accelerating deep learning poses a challenge to Google’s TPUs

artificial intelligence / machine learning / binary code / neural network
Credit: Thinkstock

Earlier this year, Google unveiled its Tensor Processing Unit, custom hardware for speeding up prediction-making with machine learning models.

Now Microsoft is trying something similar, with its Project Brainwave hardware, which supports many major deep learning systems in wide use. Project Brainwave covers many of the same goals as Google’s TPU: Speed up how predictions are served from machine learning models (in Brainwave case, those hosted in Azure, using custom hardware deployed in Microsoft’s cloud at scale). 

Brainwave uses reprogrammable FPGAs (field-programmable gate arrays, a type of integrated circuit) at scale in Azure to speed up many kinds of operations, like network processing and now machine learning.

Of course, developers need tools to do deep learning via Brainwave. Microsoft says Brainwave already supports Google TensorFlow and Microsoft’s own Cognitive Toolkit, and Microsoft says it plans to support “many others.” Brainwave converts existing models built with those frameworksto a format that can be used natively on Microsoft’s silicon, although it isn’t clear yet how much of a bottleneck that creates when porting existing models.

With Brainwave, Microsoft claims speed improvements above and beyond just using dedicated silicon. Its FPGAs are attached directly to the network fabric in the datacenter, letting a deep neural network be associated as closely as possible with a specific FPGA. Microsoft claims this high-throughput design makes it easier to create deep learning applications that run in real time, rather than require long offline training periods.

Microsoft also claims it’s taking a different approach using FPGAs to run predictions. Deep learning predictions can be delivered faster by swapping accuracy for speed. The Brainwave FPGAs can choose from a variety of data types needed, high or low precision, based on the needs of the specific problem.

Serdar Yegulalp

Serdar Yegulalp is a senior writer at InfoWorld. A veteran technology journalist, Serdar has been writing about computers, operating systems, databases, programming, and other information technology topics for 30 years. Before joining InfoWorld in 2013, Serdar wrote for Windows Magazine, InformationWeek, Byte, and a slew of other publications. At InfoWorld, Serdar has covered software development, devops, containerization, machine learning, and artificial intelligence, winning several B2B journalism awards including a 2024 Neal Award and a 2025 Azbee Award for best instructional content and best how-to article, respectively. He currently focuses on software development tools and technologies and major programming languages including Python, Rust, Go, Zig, and Wasm. Tune into his weekly Dev with Serdar videos for programming tips and techniques and close looks at programming libraries and tools.

More from this author