Hexagon 780: A Whole new IP for AI & DSP

Every year Qualcomm likes to talk about its new Hexagon DSPs, with the last few generations also adding to the table new Tensor Accelerators dedicated for ML inferencing. This year’s Snapdragon 888 design also hypes up the new Hexagon 780 design, with the difference being that this time around the generational improvements are truly humongous.

The new Hexagon 780 accelerator IP truly deserves a large increment in its IP numbering scheme, as it’s essentially a ground-up redesign of the company’s existing DSP with scalar and vector execution engines, and the recent Tensor Accelerators. Previously all these execution engines were acting as discrete independent blocks within the Hexagon 600 series family, but that’s now changed in the new IP design.

The new IP block fuses together all the scalar, tensor, and vector capabilities into a single monolithic IP, vastly increasing the performance and power efficiency of workloads that make a use of all the mixed capabilities of the design.

In terms of performance uplifts, scalar execution capabilities are said to be increased by 50%, while tensor execution throughput has doubled. The vector extension units seem to have remained the same this generation, but actual performance of workloads will still have been increased thanks to the new memory architecture of the new IP block.

Qualcomm states that they’ve increased the on-chip SRAM dedicated to the block 16-fold, allowing for larger machine-learning inference models to fit within the block’s memory, greatly accelerating their performance. This larger memory pool also allows for coherency between the scalar, vector and tensor units, allowing for the vastly increased workload handoff time between the different execution engines. I asked about the actual size of this new memory, but the company wouldn’t disclose any further details, just stating that it’s significant.

The company’s engineers were extremely hyped up about the new design, stating that the performance and flexibility of the new design is well beyond that of what other companies can achieve through disaggregated DSP and ML inference engines, sometimes even from different IP vendors.

The most important figure for the new design is the 3x performance per watt claim, which is just a massive generational improvement that you rarely see in the industry.

As is usual for Qualcomm, the company doesn’t actually state the per-block performance increases, but instead opts to showcase an aggregate computational throughput figure shared amongst all of the SoC’s IP blocks, including CPU, GPU, and the new Hexagon accelerator block. This new figure lands in at 26TOPs for the Snapdragon 888, which is 73% higher than the 15TOPs figure of the Snapdragon 865. Given that we’ve seen significant changes in all IP blocks this generation, I won’t attempt a breakdown estimate as it’s likely going to be wrong and off-the-mark anyhow.

The Adreno 660 - A 35% faster GPU

Amongst the improvements which lead up to that 26TOPs figure is a new vastly improved GPU in the form of the new Adreno 660.

Qualcomm still holds architectural details of their GPUs very close to their chest and thus doesn’t go disclose very much about the new GPU design and what has actually changed, but one thing they did talk about is the addition of new mixed-precision dot product as well as FP16/FP32 wave matrix-multiply instructions, which allow the new GPU to increase AI performance by up to 43%.

We’re also seeing the addition of variable rate shading (VRS) onto the Adreno GPU architecture, allowing for coarser pixel shading onto larger pixel blocks for object and screen areas which don’t require as much detail or the native resolution shading wouldn’t be noticeable. This is also a major feature that’s being introduced in the console and new-generation PC graphics cards and GPUs, which should bring greater performance uplift for new gaming titles which take advantage of the new features. It’s great to see Qualcomm bringing this to the mobile space along with the rest of the industry.

For graphics workloads, the new GPU is advertised as being able to increase performance by up to 35%, which is a very major generational performance leap.

Such a performance jump would actually signify that Qualcomm may very well regain the gaming performance crown this generation, having lost it to Apple’s SoCs over the last two generations. Apple’s latest A14 has seen rather conservative gains on the GPU side this year, so a 35% performance gain over the Snapdragon 865 should very much allow the new Snapdragon 888 to retake the leadership position.

A 35% performance increase with a 20% power efficiency increase would indicate that the new SoC would achieve the higher performance at cost of a little higher power consumption, but given the Snapdragon 865’s excellent power characteristics of below 4W, Qualcomm does have a little leeway to increase power this generation.

3200MHz LPDDR5

The new Snapdragon 888 moves from a hybrid memory controller to one that focuses on LPDDR5, and also increases the frequency support for new LPDDR5 to 3200MHz (Or LPDDR5-6400).

For the Snapdragon 865 Qualcomm was rather unenthusiastic about the LPDDR5 switch, saying that it didn’t bring all to great improvements to performance or power efficiency – something which we actually did test out and come to the same conclusion in our review of the two OnePlus 8 phones, where the LPDDR4X variant ended up being no slower and seemingly actually more efficient to us. Apple this year also kept on using LPDDR4X on their A14 and M1 SoCs – pointing out that the benefits aren’t all that great.

For the Snadpragon 888 however, Qualcomm’s engineers seemed more upbeat about LPDDR5 and the new SoC actually being able to utilise the increase memory bandwidth this generation. Without going into details, the company also stated that they’ve improved the overall design of the memory subsystem, improving aspects such as latency.

On the part of the memory subsystem, Qualcomm still employs a 3MB system-level-cache in front of the memory controllers, with the ability of all SoC IP blocks to take advantage of this cache.

The Snapdragon 888: Back to monolithic SoC on 5nm Triple ISPs: Concurrent Triple-Camera Usage
POST A COMMENT

123 Comments

View All Comments

  • MenhirMike - Wednesday, December 2, 2020 - link

    Apple bought P.A. Semi in 2008 to make their own ARM Chips, and evidently, P.A. Semi is a better semiconductor company than Samsung, Qualcomm, and anyone else in the ARM marketplace. Reply
  • Ppietra - Wednesday, December 2, 2020 - link

    P.A. Semi no longer exists and a lot of things happened since P.A. Semi was bought, like Apple buying Intrinsity which was involved in the creation of the A4 chip Reply
  • jordanl17 - Thursday, December 3, 2020 - link

    I remember the headline a long time ago.. "Apple buys Israeli based cpu developer to make their own chips" I was like, "haha, yeah, good luck what that" I. WAS. WRONG. Reply
  • jordanl17 - Thursday, December 3, 2020 - link

    maybe they weren't Israeli based... (can't edit post?) Reply
  • Luminar - Thursday, December 3, 2020 - link

    The edit functionality only exists for the first 15 seconds after posting. This is to prevent people from going back and editing their comments well after the fact to appear less wrong. Reply
  • Wilco1 - Friday, December 4, 2020 - link

    This is not true - if it was, I could edit this! Reply
  • trini00 - Saturday, December 5, 2020 - link

    https://www.zdnet.com/article/start-up-plans-new-e...

    Quite a interesting read, integration on a chip and the cache structure is some of the advantages M1 has.
    Reply
  • headeffects - Wednesday, December 2, 2020 - link

    Is this true? I knew Samsung’s 5nm was behind but behind even the TSMC 7nm sounds shocking. Reply
  • Lodix - Thursday, December 3, 2020 - link

    No Reply
  • Ppietra - Wednesday, December 2, 2020 - link

    I am no expert but I believe it happened because of very different visions/philosophies and objectives.
    ARM goes for smaller and less complex cores than Apple, believing it will consume less and that this will save space so it can add more cores in the same die, hoping for higher multithreaded performance. This would also probably be cheaper for other companies to implement.
    Apple on the other hand bet on bigger cores, maybe already envisioning that its development could in the end be more useful for computers, or at least the iPad. Apple believed that a faster core could consume less by finishing more complex tasks sooner. Costs didn’t seem to be a big concern for Apple, nor increasing the core numbers like crazy (remember when there were SoCs with 10 or more cores in phones?), nor Apple was constrained by what others might need. I imagine with these objectives Apple had to solve a lot of problems to optimize power consumption. Having to go through these challenges much sooner than ARM probably helped Apple to develop more efficient designs.
    It seems that Apple is just far more aggressive in developing its chips, and knows what it needs for its hardware and software.
    Reply

Log in

Don't have an account? Sign up now