AnandTech Year In Review 2019: Flagship Mobileby Andrei Frumusanu on December 26, 2019 8:00 AM EST
SoC Improvements: Apple vs Qualcomm vs Samsung vs HiSilicon
The system-on-chip is a device’s single most important component. The SoC powers every feature of a phone, determines the performance of the device, and also is the second biggest factor in a device’s battery life (behind the display). At the high-end, the key SoCs this year were HiSilicon’s Kirin 980, Qualcomm’s Snapdragon 855, Samsung LSI’s Exynos 9820 and Apple’s A13.
The Kirin 980 was the first released amongst the bunch, actually being announced and used late in 2018. The chip was a big step-up in execution from HiSilicon, and was an excellent product that powered most of Huawei and Honor’s devices for 2019. Whilst it was still lagging behind a bit in GPU performance, the chip presented a well-rounded product that offered top of the line CPU performance for the Android devices it powered.
HiSilicon’s follow-up Kirin 990 released with the Mate 30 just a few months ago unfortunately didn’t bring quite as large improvements as its predecessor: whilst it’s still a good and well-rounded SoC, this time around the company didn’t manage to integrate the newest IPs from Arm, as the company’s design and product cycle didn’t match up well enough with the releases of the newer Cortex-A77 CPU and Mali-G77 GPU. We’ll have to see how things pan out in 2020, but I don’t expect Huawei devices to be as competitive in this regard as in 2019.
The Galaxy S10 in March also introduced two new generation SoCs: The Snapdragon 855 and the Exynos 9820. The Snapdragon 855 was amongst Qualcomm’s best executed designs, and went on to power essentially every other flagship device in 2019.
Samsung’s Exynos 9820 – whilst not quite as performant or as efficient as the Snapdragon 855, was a big improvement over the disappointing Exynos 9810 of 2018, and served as a solid alternative for the Galaxy S10. S.LSI released a 7nm shrink of the Exynos 9820 in the form of the Exynos 9825 in the Galaxy Note10 series: functionally it’s the same chip as its predecessor, offering the same performance and only minor differences in efficiency – with the Qualcomm counterpart still being the same Snapdragon 855. Unfortunately, it’s just no better than the Qualcomm alternative, and thus S.LSI has never managed to get a design win from a vendor besides Samsung’s own mobile division.
Apple’s A13 this year brought big improvements in performance, although not exactly visible at first-glance in benchmarks. On the CPU side, the new Lightning cores were relatively straightforward upgrades in performance, but it did come at a cost in power consumption which keeps creeping up with every new A-series SoC. Power efficiency however still remains excellent, and I think that’s mostly due to the new Thunder efficiency cores which saw a very large microarchitectural upgrade in performance and efficiency. Apple’s yearly upgrades on the efficiency cores put the aging Cortex-A55s to shame in terms of performance and power, and the Android SoC vendors are very much in a dire need of an IP upgrade from Arm.
|GFXBench Aztec Normal Offscreen Power Efficiency
(System Active Power)
|Mfc. Process||FPS||Avg. Power
|iPhone 11 Pro (A13) Warm||N7P||73.27||4.07||18.00 fps/W|
|iPhone 11 Pro (A13) Cold / Peak||N7P||91.62||6.08||15.06 fps/W|
|iPhone XS (A12) Warm||N7||55.70||3.88||14.35 fps/W|
|iPhone XS (A12) Cold / Peak||N7||76.00||5.59||13.59 fps/W|
|QRD865 (Snapdragon 865)||N7P||53.65||4.65||11.53 fps/W|
|Mate 30 Pro (Kirin 990 4G)||N7||41.68||4.01||10.39 fps/W|
|Galaxy 10+ (Snapdragon 855)||N7||40.63||4.14||9.81 fps/W|
|Galaxy 10+ (Exynos 9820)||8LPP||40.18||4.62||8.69 fps/W|
On the GPU side of things, Apple has also been hitting it out of the park; the last two GPU generations have brought tremendous efficiency upgrades which also allow for larger performance gains. I really had not expected Apple to make as large strides with the A13’s GPU this year, and the efficiency improvements really surprised me. The differences to Qualcomm’s Adreno architecture are now so big that even the newest Snapdragon 865 peak performance isn’t able to match Apple’s sustained performance figures. It’s no longer that Apple just leads in CPU, they are now also massively leading in GPU.
For 2020, I’m not really expecting the competitive situation in the high-end to change much. The Snapdragon 865 and Exynos 990 will likely be excellent products, but largely won’t be able to catch up to the A13. I think the biggest surprise in 2020 will be MediaTek’s new Dimensity 1000 SoC: MediaTek’s return to the high-end comes with what seems an unusually well-rounded and solid approach. On paper, the chip should perform better than HiSilicon’s Kirin 990 and also be a very viable alternative to the Snapdragon 865, I’m predicting to see a lot more MediaTek devices and design wins than what we’ve been used to recently.
The OnePlus 7 Pro Brings 90Hz To The Masses, Starts a Trend
The OnePlus 7 Pro was one of the most important phones in 2019 for the fact that it was the first device to bring the high-refresh rate experience to the masses. The phone’s 90Hz display has been a resounding success for OnePlus to the point that the company said it’s been the key feature as to why consumers chose a OnePlus device this year. The refreshed OnePlus 7T also brought the 90Hz display to the lower-priced flagship variant.
Google’s Pixel 4 series represented the first mainstream follow-up from another vendor, however Google’s implementation here largely seemed not as well thought-out as OnePlus’, and was handicapped by lacklustre battery life, particularly on the smaller Pixel 4.
Gaming phones such as ROG Phone II even come with 120Hz displays, and the accompanying gigantic batteries of such phones means there’s no drawbacks to the user experience.
I’m expecting high refresh rate displays to be a key feature in 2020 flagships, with the usual big-name vendors also adopting them.
Battery & Display Improvements: Samsung and Apple’s Rise To Longevity
Another key improvement that I will keenly remember in 2019 will be Samsung and Apple’s battery improvements. For several years now the devices with the longest battery life were Huawei’s larger Mate-series phones with big battery capacities and efficient LCD displays, or mid-range phones which come with less power hungry SoCs and displays.
The Galaxy S10 (particularly the S10+) series I think was a game-changer in that it was the first flagship devices that combined top-of-the-line display and SoC, yet still managed to top the battery charts. Apple’s iPhone 11 Pro series follows the Galaxy S10 and Note10 series in these characteristics: the reason for the exemplary battery life of these phones is a combination of larger than usual battery capacities along with the newest OLED display technology. The S10, Note10, iPhone 11 Pro, along with the OnePlus 7T are the only devices in 2019 to use the new emitter technology to bring brighter and more efficient displays – and this advantage is it’s very much noticeable when it comes to each of these device’s battery life.
We expect the technology to trickle down to other vendors in 2020, with Samsung and Apple to continue introducing newer and even more efficient generations – a good tie-in with the expected higher refresh-rate displays next year.
Bendy Phones: The Galaxy Fold & Mate X
While certainly not relevant to the average consumer, 2019 will be remembered as the first year that foldable phones became a reality. The Galaxy Fold and the Huawei Mate X are the first serious big-name implementations of the new flexible screen technology. It’s expensive, impractical, gimmicky and has a ton of drawbacks, yet it’s also one of the most exciting things we’ve seen in years as it represents the first step in a new device form-factor category.
The new Motorola Razr seems like a much better thought-out implementation, and it seems to be the most sensible direction for foldable phones. We’ll undoubtedly hear a lot more innovation from vendors in 2020 when it comes to flexible display devices.
5G – A Reality … In Some Countries
It’s no doubt that 5G has been an incessant talking point throughout 2019, and we actually even saw the first 5G devices in the market. For anybody caring about longevity of their phones, the first generation 5G phones based on Qualcomm’s X50 modem are largely to be ignored – the modem does not support future 5G standalone networks. Second generation devices based on the X55 are certainly to be much more future-proof and much better performing.
We’ve seen the first 5G networks go live this year, but there’s a big difference in the deployment types, depending on your country and carrier. In the US, Verizon most famously launched first with the introduction of their “ultra-wideband” service in select city blocks of select cities. This is the mmWave side of 5G, which does bring the higher multi-gigabit speeds that are hyped up a lot. The reality is that such deployments are going to be extremely limited in terms of coverage and usage; it’ll take many years for such networks to be ubiquitous and even then, it’ll be mostly concentrated around high-density locations such as event centres or stadiums (Which is essentially what mmWave deployments are really intended for).
The rest of the world (And other US carriers) is focusing on bringing 5G to sub-6GHz spectrum. In a lot of countries such as South Korea we’ve already seen very wide-spread infrastructure deployments with coverage quickly catching up to 4G LTE. The speed increases here aren’t as substantial as mmWave, but the improvements here are actually more realistic, consistent and practical, and this will be what 5G will be about for the vast majority of users.
While 2019 will be the first year of 5G, that’s about it when it comes to how it’ll be remembered. I expect 5G deployment to be more of a slow-burn over the comings years as network coverage and capacity continues to improve.
Post Your CommentPlease log in or sign up to comment.
View All Comments
TheinsanegamerN - Thursday, December 26, 2019 - linkThere was a major point missed in the ROG II review, and that is the ROG II does not support VoLTE or VoWiFi in the USA. This is going to be a major headache at the end of 2020 when 3g networks are taken offline, and asus doesnt seem interested in fixing this problem.
It ruins an otherwise amazing phone.
s.yu - Friday, December 27, 2019 - linkThat's something to consider, so what's gonna happen? No more voice calls for any device that doesn't support VoLTE or VoWiFi?
name99 - Friday, December 27, 2019 - linkOn the iPhone side this is easily achieved by Apple’s power cases, if that’s what you want.
Large capacity, and feels just like a fatter iPhone.
s.yu - Friday, December 27, 2019 - linkThat power case remains the single ugliest one around.
arashi - Saturday, December 28, 2019 - linkThe same can be said for any sufficiently popular Android (Zero Lemon makes a shit ton of battery cases).
guachi - Thursday, December 26, 2019 - linkIf someone takes photography seriously they probably own a dedicated camera. It's certainly cheaper to get a low-midrange phone and a dedicated camera than it is a high-end phone.
Second, number of megapixels of a camera doesn't matter. What matters is sensor size. I wish all phone manufacturers listed the sensor size of their phone cameras.
joedish365 - Thursday, December 26, 2019 - linkDon't necessarily agree with either. I don't need more megapixels as you mention, and sensor size should result in better quality, but not necessarily depending on the software used to process the pixels. More mega pixels might let you crop more depending on sensor size. Testing "quality" is very hard and can be very personal. Most photos from camera are stored or shared in the cloud and compress somewhat. Some attempt at crowd sourcing quality...
melgross - Thursday, December 26, 2019 - linkOn smartphones, sensor size is just one number of importance. You can’t look at that and determine picture quality.
eastcoast_pete - Thursday, December 26, 2019 - linkAgree with you, however GIGO still applies. Even the best software can only put that much lipstick on a pig of a picture taken with a lousy camera unit.
s.yu - Thursday, December 26, 2019 - linkStrange that HDR+ isn't considered computational photography, because it very much is.