Microsoft Catapults geriatric Moore's Law from CERTAIN DEATH

starlight777

Zork Rules
Sep 12, 2013
2,668
718
1,928
Microsoft Catapults geriatric Moore's Law from CERTAIN DEATH
http://www.theregister.co.uk/2014/06/16/microsoft_catapult_fpgas/
FPGAs DOUBLE data center throughput despite puny power pump-up, we're told


Microsoft has found a way to massively increase the compute capabilities of its data centers, despite the fact that Moore's Law is wheezing towards its inevitable demise.

In a paper to be presented this week at the International Symposium on Computer Architecture (ISCA), titled A Reconfigurable Fabric for Accelerating Large-Scale Datacenter Services, a troupe of top Microsoft Research boffins explain how the company has dealt with the slowdown in single-core clock-rate improvements that has occurred over the past decade.

To get around this debilitating problem – more on this later – Microsoft has built a system it calls Catapult, which automatically offloads some of the advanced tech that powers its Bing search engine onto clusters of highly efficient, low-power FPGA chips attached to typical Intel Xeon server processors.

Think of FPGAs – field-programmable gate arrays – as chips whose circuits can be customised and tweaked as required, allowing crucial tasks to be transferred away from the Xeons and instead accelerated in FPGA hardware.

This approach may save Microsoft from a rarely acknowledged problem that lurks in the technology industry: processors are not getting much faster.
Wait. What?

For those not familiar with the chip industry, a primer. For the past 50 years, almost every aspect of our global economy has been affected by Moore's Law, which states that the number of transistors on a chip of the same size will double every 18 months – or so – resulting in faster performance and better power efficiency

One slight problem: Moore's Law is not, in fact, a law. Instead, it was an assertion by Intel founder Gordon Moore in a 1965 article that the semiconductor industry got rather carried away with. In the past ten years, the salubrious effects of Moore's Law have started to wane, because although companies are packing more and more transistors onto their chips, the performance gains that those transistors bring with them are not as great as they were during the law's halcyon days.

Intel has yoinked its entire business to the successful fulfillment of Moore's Law, and proudly announces each new boost in transistor counts. And, yes, those "new" transistors can help to increase a compute core's all-important instruction per cycle (IPC) metric – improved branch prediction, larger caches, more-efficient scheduling, beefier buffers, whatever – but the simple fact is that although chips have gone multi-core and are getting better at multi-tasking, those individual cores are not getting much faster due to any significant new discovery.

As AMD CTO Joe Macri recently told us, "There's not a whole lot of revolution left in CPUs." He did, however, note that "there's a lot of evolution left."
Microsoft's Catapult is a bit of both.

Programmable software, meet programmable hardware
Under new chief executive Satya Nadella, Microsoft is throwing billions of dollars at massive data centers in its attempt to become a cloud-first company. Part of that effort – and that investment – is to figure out a way to jump-start consistent data-center compute-performance boosts.

The solution that Microsoft Research has come up with is to pair field-programmable gate arrays with typical x86 processors, then let some data-center services such as the Bing search engine offload certain well-understood operations to the arrays.

To say that the performance improvements in this approach have been noticeable would be a gross understatement. Microsoft tells us that a test deployment on 1,632 servers was able to increase query throughput by 95 per cent, while only increasing power consumption by 10 per cent.

Though FPGA technology is well understood and used widely in the embedded technology industry, it's rare to hear of it being paired with standard off-the-shelf CPUs for accelerating web-facing software – until now, that is.
"We're moving into an era of programmable hardware supporting programmable software," Microsoft Research's Doug Burger told The Register. "We're just starting down that road now."

If Microsoft has indeed figured out how to almost double the performance of its computers while only paying a tenth more in electricity for large-scale data center tasks – and we see no reason to doubt them – that's not only a huge saving, but also one that saves the company from the slowdown in run-of-the-mill CPUs chips.
"Based on the results, Bing will roll out FPGA-enhanced servers in one data center to process customer searches starting in early 2015," Derek Chiou, the principal architect of Bing, said in a statement emailed to El Reg.

"We were looking to make a big jump forward in data center capabilities. It's an important area," Microsoft Research's Doug Burger explained to us.

"We wanted to do something that we thought could put us on a path that makes some really big leaps. Rather than banking on scaling to many, many more cores, let's take a different path – what can we do in hardware? We think specialization is going to be the next big thing."
Microsoft isn't doing this on a hunch. Burger wrote a paper [PDF] in 2011, Dark Silicon and the end of Multicore Scaling, which predicted that "left to the multicore path, we may hit a 'transistor utility economics' wall in as few as three to five years, at which point Moore's Law may end, creating massive disruptions in our industry."
 
no comments? hmm

try this part

Programmable software, meet programmable hardware
Under new chief executive Satya Nadella, Microsoft is throwing billions of dollars at massive data centers in its attempt to become a cloud-first company. Part of that effort – and that investment – is to figure out a way to jump-start consistent data-center compute-performance boosts.

The solution that Microsoft Research has come up with is to pair field-programmable gate arrays with typical x86 processors, then let some data-center services such as the Bing search engine offload certain well-understood operations to the arrays.

To say that the performance improvements in this approach have been noticeable would be a gross understatement. Microsoft tells us that a test deployment on 1,632 servers was able to increase query throughput by 95 per cent, while only increasing power consumption by 10 per cent.
 
If Microsoft has indeed figured out how to almost double the performance of its computers while only paying a tenth more in electricity for large-scale data center tasks – and we see no reason to doubt them – that's not only a huge saving, but also one that saves the company from the slowdown in run-of-the-mill CPUs chips.
 
let's take a different path – what can we do in hardware? We think specialization is going to be the next big thing."
Microsoft isn't doing this on a hunch. Burger wrote a paper [PDF] in 2011, Dark Silicon and the end of Multicore Scaling, which predicted that "left to the multicore path, we may hit a 'transistor utility economics' wall in as few as three to five years, at which point Moore's Law may end, creating massive disruptions in our industry."
 
The 41st International Symposium on Computer Architecture

I cannot find the transcript for this
can any find out what he said?


Monday, 8:45am-9:45am
Keynote I: Insight into the MICROSOFT XBOX ONE Technology
Dr. Ilan Spillinger, Corporate Vice President, Technology and Silicon,
Microsoft



Bio: Dr. Ilan Spillinger was instrumental in bringing Kinect for Xbox 360 to market, which continues to proliferate additional innovative programs and products within the Natural User Interface industry. His team's recent efforts include developing the new architecture and silicon design for Xbox One and the new Kinect, which launched in the fall of 2013. Previously, during a six-year tenure with IBM, Dr. Spillinger served as a distinguished engineer and vice president for advanced processor design. In that role he was responsible for development of all Power Architecture-based processors at IBM: server processors, embedded processors, and client-driven solutions. Prior to that, Dr. Spillinger was a principal engineer and manager of the architecture team in Intel Israel, responsible for the definition of x86-based low-cost and low-power microprocessors, specifically the first Intel mobile processor in the Intel Centrino roadmap. Spillinger holds a D.Sc. and M.Sc. in electrical engineering from the Technion Israel Institute of Technology in Haifa, Israel, and joined Microsoft in 2007.
 
Sounds familiar :) Potentially exciting stuff down the line.
 
I tried to comprehend...but it made my whole face hurt. Anyone want to explain what this means?
 
I tried to comprehend...but it made my whole face hurt. Anyone want to explain what this means?
Microsoft is doing micro hardware modifications to standard off-the-shelf chips to make certain aspects of their functionality astonishingly more effficient.
 
I tried to comprehend...but it made my whole face hurt. Anyone want to explain what this means?
Ms found a way to increase data center performance with only a minimum increase in power consumption. Should help with cloud/servers?
 
The 41st International Symposium on Computer Architecture

I cannot find the transcript for this
can any find out what he said?


Monday, 8:45am-9:45am
Keynote I: Insight into the MICROSOFT XBOX ONE Technology
Dr. Ilan Spillinger, Corporate Vice President, Technology and Silicon,
Microsoft



Bio: Dr. Ilan Spillinger was instrumental in bringing Kinect for Xbox 360 to market, which continues to proliferate additional innovative programs and products within the Natural User Interface industry. His team's recent efforts include developing the new architecture and silicon design for Xbox One and the new Kinect, which launched in the fall of 2013. Previously, during a six-year tenure with IBM, Dr. Spillinger served as a distinguished engineer and vice president for advanced processor design. In that role he was responsible for development of all Power Architecture-based processors at IBM: server processors, embedded processors, and client-driven solutions. Prior to that, Dr. Spillinger was a principal engineer and manager of the architecture team in Intel Israel, responsible for the definition of x86-based low-cost and low-power microprocessors, specifically the first Intel mobile processor in the Intel Centrino roadmap. Spillinger holds a D.Sc. and M.Sc. in electrical engineering from the Technion Israel Institute of Technology in Haifa, Israel, and joined Microsoft in 2007.
Does it have anything to do with what's under the hood of the Xbox One?
 
Interesting. It's basically hardware-programmable. Hardware drivers that can be dynamically changed, but I still don't understand really what allows an FPGA processor to exist and to exist with its boasted efficiency.
 
see the audio block / Tensilica
it is X1 FPGA/DSP
it is infact the same GPU clock
why MS showed audio block ? and GFlops ? think about it
FPGA/DSP is very hard to program

This the DPU from 2009, look at the wattage to achive 10 GMAC = 20 Gflops
http://ip.cadence.com/news/303/330/...gaMAC-sec-DSP-Performance-Tops-1-GHz-Mark.htm

"Xtensa LX3 DPU offers the industry's widest range of pre-verified DSP options ranging from a simple floating point accelerator to a 16-MAC (multiply accumulator) vector DSP powerhouse.

The base Xtensa LX3 DPU configuration can reach speeds of over 1 GHz in 45nm process technology (45GS) with an area of just 0.044 mm2 and power of 0.015 mW/MHz. When built with the new ConnX Baseband Engine DSP (ConnX BBE), the Xtensa LX3 processor delivers over 10 Giga-MACs-per-second performance, running at 625 MHz with a footprint of 0.93mm2 (post place-and-route 45GS) and consuming just 170 mW (including leakage)."
 
Think of it all as nano coding. The philosophy of this is future software paradigm which requires hardware to be designed around it. Much here in line with what MS has said technically behind its cloud infrastructure, X1 design, upcoming (new) DX12 GPUs, etc.. The future is here, just takes time to code it all.

People really fail to realize that MS spends more money on research annually than most of its competitors do put together in the hardware/software space around computing. It's why they (MS) have pretty much just given up explaining themselves and just letting things evolve and show as they come naturally while eventually everyone else will scramble to try to catch up. It is why Intel, AMD and NVidia ate all excited and part of what is coming around the bend; DX12 being a huge part of it all with new GPU architectures built around it. While existing cards will be compliant with DX12, they will not be able to touch new cards built around what DX12 is really about.

It's all related, all next gen technology and software design.
 
Last edited:
  • Like
Reactions: BIGJOHN