Arm Announces The Mali-G78 GPU: Evolution to 24 Cores
by Andrei Frumusanu on May 26, 2020 9:00 AM ESTSmall Performance Improvements - Uncertain Projections
Summing up all the different microarchitectural advancements, Arm presents with us the different performance improvements we can expect of the Mali-G78:
On the part of the asynchronous top-level performance improvements the GPU can achieve by improving the geometry to shader core capabilities, Arm projects to see a roughly 8% boost in benchmarks, with a larger ~14% boost in some game titles.
These improvements are quite small, but from a SoC vendor perspective I suppose it wouldn’t be too complicated to implement this, as it would only cost an additional PLL or just a frequency divider in order to achieve the extra performance.
The generational power efficiency improvements of the G78 over the G77 in a similar configuration are 10%, likely attributed to the FMA and cache improvements of the core. It’s small, but we take what we can get.
The async feature from an energy efficiency perspective is proclaimed to be around 6-13% depending on the workload. This is actually a bit of a more complex figure in my view. The main problem in my view is that to achieve this, the SoC vendor needs to actually go ahead and employ a second voltage rail for the GPU to gain the most benefit of the asynchronous frequencies. The efficiency benefit here is small enough, that it begs the question if it’s not just cheaper to add in a few more extra cores and lock them lower, rather than incurring the cost of the extra PMIC rail, inductors and capacitors. It’s an easy efficiency gain for flagship SoCs, but I’m really wondering what vendors will be deploying in the mid-range and lower.
Mali-G68 GPU: It's the same
Alongside the Mali-G78, Arm is today also announcing the new Mali-G68 GPU:
You might be wondering why I’m including this as a footnote at the end of the article rather than covering it in more detail. The truth is, this is the exact same IP as the Mali-G78, with the only difference being that this GPU configuration only scales up to 6 cores. In essence, if the microarchitecture is implemented with up to 6 cores, it’s branded as a G68, and if uses 7 or more cores, it’s branded as a G78.
Arm actually had used this marketing with the G57, which ended up being actually the same IP as the G77, leading to some confusion with the MediaTek Dimensity 800 SoC that was announced earlier this year. We had called that GPU as a derivative of the G77 until MediaTek had reached out to us to point out that it’s actually the same GPU.
It’s pretty disappointing to see Arm do such marketing exercises, as it can be technically misleading. We asked what their rationale is, and they explained that it’s actually a customer demand for them to better differentiate their products. It’s a somewhat credible argument, but on the other hand we’ve had MediaTek outright want to point out to us this misleading branding, so it seems that not everybody is on the same page on the matter.
Arm does say that they possibly envision that future iterations in this series might actually see real microarchitectural differentiations compared to the bigger implementations. In that scenario, the branding at least would make more sense.
Mali-G78: Meagre improvements, or just bad vendor implementations?
If you didn’t already catch on until now, I’m feeling quite pessimistic about the Mali-G78. First of all, it’s just not that big of a generational upgrade compared to the Mali-G77, even by Arm’s own standards and advertised figures.
You could forgive the smaller upgrades if we had started from an excellent baseline performance. The Mali-G77 promised a whole ton of improvements in both performance and efficiency. The actual results we’ve seen out of the Exynos 990 and the MediaTek D1000 were anything but stellar. On one hand we had a SoC which seemingly had a bad implementation on a seemingly immature process node, and on the other hand we had some very mid-range performance even though it was an MP9 GPU configuration. Truth is, we still don’t know if the Mali-G77 is a good GPU or not, as we simply haven’t seen a good implementation out there. If we don’t know if the G77 is good or not, then it’s also impossible to project if the G78 will be any good.
I see Arm having the exact same problem they’ve been facing in the CPU space until the just announced Cortex-X1, as in they’re stuck with having to design a scalable GPU that fits all target markets and having to please all customer design points. Technically, that’s never the best option, as you end up with something that always has compromises.
As for potential implementers of the G78, amongst the biggest vendors it’s likely HiSilicon to be the first adopter – if they can manage to bring out the new Kirin chipsets out to market amidst the current political situation. Whether Samsung and AMD will manage to bring out an RDNA based mobile Exynos next year is also still unclear, though I’m sure that’s what they’re striving for. The biggest issue on the competitive landscape is Apple. Even if the G77 had managed to live up to its projections, the G78 certainly is showcasing too meagre improvements to be able to catch up to the Apple GPUs. We’re also supposed to be seeing the first Imagination A-series GPU SoC designs later this year which is a whole other wildcard. That’s a very tough competitive landscape for Mali – let’s hope the G78 will see more positive success in the future.
36 Comments
View All Comments
Deicidium369 - Tuesday, May 26, 2020 - link
It is a cry for help.dotjaz - Tuesday, May 26, 2020 - link
What rumor? By rumor you mean both AMD and Samsung already announced RDNA will power Exynos last year? Samsung still use exclusively Mali for at most 1 more generation. That's it. ARM can count on Samsung being a Mali customer anymore.dotjaz - Tuesday, May 26, 2020 - link
*can'tlucam - Tuesday, May 26, 2020 - link
Apple always used PowerVR, and they will again. You must have been in hibernation in the last 12months , as you don’t read news aroundSpunjji - Friday, May 29, 2020 - link
I haven't kept entirely up-to-date with this, but my understanding was that Apple went from using scaled-up versions of Imagination GPUs to using their "own design", a design which was miraculously compatible with all the low-level APIs they designed for Imagination tech. As a result of Apple's announcement, IMG's stock plummeted and they had to sell off assets and seek investment from China. Subsequently Apple went back to paying Imagination a licence fee, almost as if the whole thing was an attempt to bully a smaller company who couldn't afford extended litigation for IP theft into giving them better licencing terms.In conclusion: Apple pull this nonsense on a regular basis, and they suck.
lucam - Tuesday, May 26, 2020 - link
Apple use PowerVR, update your sourceZoolook - Tuesday, June 2, 2020 - link
Power VR is a division of Imagination, so technically you are both correct.eastcoast_pete - Tuesday, May 26, 2020 - link
My first thought was "I smell a Windows-on-ARM" laptop push here. If any of ARM's licensees (Samsung, maybe Mediatek?) wants to go down that road and join Qualcomm, they'll need a beefier graphics setup to compete with Qualcomm's 8cx designs. Maybe 24 G78s will get Samsung & Co. closer to that goal. I am actually rooting for this to happen; the low-power, always connected ultraportable space can use some new entries.eastcoast_pete - Tuesday, May 26, 2020 - link
Also, is is just me, or is ARM basically going the route used by dGPUs in the PC/Mac space (many cores/compute units)? As Andrei pointed out, that comes at the expense of some power-consuming repetition/redundancy, which is okay in a desktop or powerful laptop (big battery), but problematic if you're trying to run it all day and carry it with one hand.lucam - Tuesday, May 26, 2020 - link
New GPU, same garbage all over again..