The best thing about manufacturing Field Programmable Gate Arrays (FPGAs) is that you can make the silicon very big. The nature of the repeatable unit design can absorb issues with a process technology, and as a result we often see FPGAs be the largest silicon dies that enter the market for a given manufacturing process. When you get to the limit of how big you can make a piece of silicon (known as the reticle limit), the only way to get bigger is to connect that silicon together. Today Intel is announcing its latest ‘large’ FPGA, and it comes with a pretty big milestone with its connectivity technology.

One of the elements driving this industry forward is packaging technology. We’ve covered in detail elements like TSMC’s 2.5D Chip-On-Wafer-On-Substrate (COWOS) packaging used in GPUs, Intel’s embedded multi-die interconnect bridge (EMIB), stacking technology like FOVEROS, and as we migrate into smaller chiplet based silicon, each will become crucial to finding the best way to produce the end chip that goes into a million systems.

Despite Intel’s best diagrams about its EMIB technology, showing many die connected together from multiple different process nodes, one major barrier has eluded the company. Until this point, all we had seen from EMIB was it connecting one high-powered die, like a GPU, to a low-powered die, like HBM. One of the criticisms to having only known Intel products connect one high-powered and one low-powered die is if the EMIB connection wasn’t thermally stable to withstand power-cycling between two die.


One of Intel's mock-ups of how future processors might look like

When connecting two die together in a substrate, especially high-powered die with vias or a BGA design, mechanical stresses have to be taken into account, especially if different metals are at play. Thermal expansion and contraction is a critical point of failure, especially when dealing with embedded and long life-cycle designs. Not only the expansion and contraction of metals, but when dealing with organic substrates holding the packaging technology, making the substrate inextricably thin also severely increases long-term feasibility concerns, especially if high-powered die are used for connectivity.

With Intel’s new FPGA, the Stratix 10 GX 10M, the concern seems to have disappeared. This new product, designed as a big FPGA for the ASIC prototyping and emulation market, combines two large 5.1M logic element FPGAs with three EMIB connections, producing an overall chip with an average TDP from 150W up to 400W with advanced cooling.


A total of 7 EMIB connections, but it's the three in the middle that are the milestone

The ASIC prototyping and emulation market, while a small market revenue wise (Intel stated ~$300-$500M/yr), is always requesting bigger and bigger FPGAs in order to be able to fit more and more of their ASIC designs onto as few FPGAs as possible in order to get the most accurate results. These chips ultimately end up running at a low frequency for accuracy, anywhere from 50 MHz to 300 MHz, but Intel states that this new Stratix 10 GX 10M design can easily replace four of its old GX 2800 FPGAs with double the connectivity and even a 40% power reduction for the same workload.

The design of the FPGA is around those two 5.1M logic element dies, connected together with three EMIB connections. These use the AIB protocol running at over 1 GHz, and form part of the 25920 connection pins across the whole chip which has another four EMIB connections to transceivers as shown in the diagram.

At 10.2 million logic elements, this FPGA eclipses the Xilinx VU19P announced in August which had 9 million elements (8172k flip-flops, 4086k LUTs). The Stratix 10 GX 10M also contains 6912 DSP elements and 48 transceiver outputs at 17.4 Gbps. Intel states that these are designed primarily as PCIe 3.0/4.0 support, and that the FPGA supports H-tiles for connectivity for customers interested in custom designs.

The launch of this new hardware coincides with Intel’s FPGA Tech Event in China, which is one of the primary markets for this product. Intel states the hardware has already been with key partners almost a year (one of the early customers is China based), but is now in production for the wider market. On the topic of Intel’s high demands for its 14nm parts, the company stated that the volume isn’t that high for this sort of product, and they won’t have any issues. The company (at least, the FPGA part of the company) did explain that the use of EMIB in this fashion means that their two-die approach assists with yield.

Personally, the fact that Intel is strapping two high powered die (~75W to ~150W each) together using multiple EMIB connections is a key step into driving the EMIB technology to the wider market. With this as a proof of concept, it paves the way for better multi-die CPU designs as well as the promise of EMIB (and Foveros) in future discrete GPU products.

Related Reading

Source: Intel

Comments Locked

31 Comments

View All Comments

  • firewrath9 - Tuesday, November 5, 2019 - link

    GLUUEEE
  • III-V - Wednesday, November 6, 2019 - link

    This is pretty fucking far from glue.
  • katsetus - Wednesday, November 6, 2019 - link

    Is not. This is the definitive reference design of glues. The youtube tutorial of gluing things together, complete with darude sandstorm and hilarious accents.
  • notashill - Wednesday, November 6, 2019 - link

    It's exactly the same kind of thing that Intel criticized AMD for by calling EPYC "4 glued-together desktop die".
  • patrickjp93 - Wednesday, November 6, 2019 - link

    @notashill. no, it's nowhere close to the same thing. AMD used what is called Glue Logic Architecture as the onboard networking and coherence protocols. This is actually much more open and flexible. You can implement any kind of communication protocol over the EMIB connection itself as long as you know how to program an FPGA (Verilog and OpenCL).
  • dullard - Wednesday, November 6, 2019 - link

    Which is a retort from people calling Intel's dual core chips "glued" together 10 years before that.

    Anand himself calling Intel's chips glued together in 2005: https://www.anandtech.com/show/1656/2

    And others:
    https://forums.tomshardware.com/threads/intels-glu...
    https://www.ifixit.com/Wiki/Computer_Processor_Cha...
    https://www.guru3d.com/articles-pages/core-i5-750-...
  • FreckledTrout - Wednesday, November 6, 2019 - link

    So was AMD's approach but Intel called it glue. So I see no issue poking fun at them.
  • dullard - Wednesday, November 6, 2019 - link

    Remember, AMD called Intel's chips glued together first:
    https://pcper.com/2006/11/intel-core-2-extreme-qx6...
    "So, as many have said, including AMD, Intel’s Kentsfield processor is two dual core processors “glued together” and seems somewhat un-elegent."
  • HStewart - Monday, November 11, 2019 - link

    Isn't this essentially the same thing that whole Zen architexture is designed is based on making multiple 8 core processor together in same cpu.package.
  • HStewart - Monday, November 11, 2019 - link

    Calling this product Gluueee is why I pretty much ignore the comments on articles here lately.
    I believe this designed where the original EMiB in XPS 15 2in1 came from and is sign of future and almost guarantee to be reproduce by competitors in the future.

Log in

Don't have an account? Sign up now