2018 has nearly drawn to a close, and as we’re already gearing up for the event that kicks off 2019 for the industry – the mega-show that is CES – we wanted to spend one last moment going over the highs and lows of the tech industry in 2018. So, whether you’ve been out of the loop for a while and are looking to catch up, or just after a quick summary of the year behind and a glimpse at the year to come, you’re in the right place for the AnandTech Year in Review.

GPU Market: Boom, Bust, & Broken

For better or for worse, by far the leading story in the GPU space for 2018 is the latest rise and fall of the cryptocurrency markets. While the GPU space is by no means a stranger to the market-breaking weirdness that cryptocurrency mining has on GPU prices – we’ve been through this a couple of times now – the latest cycle was the biggest boom and the biggest bust yet. And as a result it’s had immense repercussions on the industry that will play out well into 2019.

While the latest cycle was already well underway before 2018 started, it hit its peak right at the start of the year. Indeed, that’s when the Ethereum cyptocurrency itself – the predominant cryptocurrency for GPUs – hit its all-time record closing price of $1,396/token. As a result, demand for GPUs for the first eight months of the year was unprecedented, which was a boon for some, and a problem for many others.

Ethereum: From Boom To Bust (Ethereumprice.org)

For the better part of three quarters AMD and NVIDIA could sell virtually every last GPU they produced, with some eager companies going straight to the source to try to secure boards and GPUs to scale up their mining operations. As a result, video card prices spiked – of the limited number of cards to even make it to the market, they were just as quickly picked up by miners – which benefitted board vendors and the GPU vendors greatly, but put a major kink in gaming. In fact, affordable gaming cards took the brunt of this demand, as their relatively high price/performance ratios meant that for all the reasons they were good for gaming, they were good for mining as well. When there’s money to be made a Radeon RX 580 (MSRP: $229) will still find buyers, even at $370.

However with every boom comes a bust, and the latest runup on the price of Ethereum was no different. After peaking early in the year, the cryptocurrency’s price continued to trend downward over the entire year, eating away at the profitability of GPU mining. This was exacerbated by the introduction of dedicated Ethereum ASICs, which although required their own fab space to produce and components (e.g. RAM) to go with them, alleviated some of the demand for video cards for mining. Ultimately Ethereum dropping below $300/token in August put the latest boom to an end, and in the last month prices have dropped below $100/token, while the profitability of mining even with the most efficient video cards has turned negative at times.

CamelCamelCamel Price History for MSI's Radeon RX 580 Armor 8G

As a result, it’s only been in the last few months that video card prices have come back down to their reasonable, intended prices. Indeed, this year’s Black Friday to Cyber Monday period was especially robust, as RX 580s were going for as little as $169, a far cry from where they were 6 months earlier. The market may not be entirely fixed quite yet, but good video cards are finally affordable again.

With that said, as we’ve since learned from their recent earnings announcements, the break in video card prices isn’t just due to a drop in demand. Both AMD and NVIDIA ramped up their GPU production orders to try to take advantage of this extended boom period, And, judging from the inventory issues they’re now facing, both have been burnt by it. Hungover from the boom in GPU demand and the boost to revenue and profits that came from it, both companies are now in a small slump as they have built up sizable stockpiles of GPUs that came in right when demand really tapered off. As a result, both companies are trying to offload their inventory in a controlled pace, selling it in limited batches to avoid flooding the market and crashing video card prices entirely. The cryptocurrency market is all but impossible to predict, and hardware suppliers in particular are caught in the tough spot: do nothing and let cryptocurrency break the market indefinitely, or try to react and risk having excess inventory when the market needs it the least.

But for the moment, with most of the sanity restored to the video card market, there is a silver lining: cheap cards! Along with frequent sales for new cards, the second-hand market is also increasingly filled with video cards from miners and speculators who are offloading their equipment and inventory, even if it comes at a loss. So for buyers who are willing to take a bit more of a chance, right now is a very good time for snagging a late-generation video card for a good price.

GDDR6 Memory Hits The Scene

GPU pricing and market matters aside, 2018 was also an important year on the technology front. After years of planning, GDDR6, the next generation of memory for GPUs and other high-throughput processors, finally reached mass production and hit the market. This is an especially important development for mid-range cards, as these products have never had access to more exotic memory technologies like GDDR5X and HBM.

GDDR6 follows in the footsteps of both the long-lived GDDR5 and the more recent GDDR5X. First put into video cards in 2008 – a lifetime ago for the GPU industry – GDDR5 has been the backbone of most video cards over the last decade. It’s been taken to far greater frequencies than JEDEC ever originally planned, with the first cards launching with data rates of 3.6Gbps while the most recent cards have shipped at 8 and 9gbps. And while GDDR5X tried to pick up where GDDR5 left off in 2016, the lack of adoption outside of Micron meant that it never had the traction required to replace its ubiquitous predecessor.

GDDR6, on the other hand, has no such problems. The Big 3 memory vendors – SK Hynix, Samsung, and Micron – are all producing the memory. And while prices are still high as an early technology (don’t expect GDDR5 to disappear overnight), GDDR6 is primed to become the new backbone of the GPU memory industry.

Overall, GDDR6 introduces a trio of key improvements that vault the memory technology ahead of GDDR5. On the signaling front, it uses Quad Data Rate (QDR) signaling as opposed to Double Data Rate (DDR) signaling. With twice as many signal pumps per clock as before, GDDR6’s memory bus can reach the necessary data rates at lower clockspeeds, making it more efficient and easier to implement than a higher clocked DDR bus, though not without some signal integrity challenges of its own.

The second big change is to how the memory is organized and how data is prefetched: instead of a chip having a single 32-bit wide channel with an 8n prefetch, GDDR6 divides that into a pair of 16-bit wide channels, each with a 16n prefetch. This effectively doubles the amount of data that is fed to the memory bus on every memory clock cycle, matching the increased bus capacity and allowing for greater data rates out of the memory without increasing the core clock rate of the memory itself.

Finally, GDDR6 once again lowers the memory operating voltage. Whereas GDDR5 typically ran at 1.5v, the standard voltage for GDDR6 is 1.35v. The actual power savings are a bit hard to quantify here since power consumption also depends on the memory controller used, but all-told we’re expecting to see significant increases in bandwidth with minimal power consumption increases with both GPU vendors – and network controller manufacturers and other GDDR users, for that matter.

GDDR6 is currently readily available in speeds up to 14Gbps, and the standard allows for faster speeds as well. Samsung is already talking about doing 18Gbps memory, and if it ends up as long-lived as its predecessor, GDDR6 will undoubtedly go faster than that still. Meanwhile it will be interesting to see where the line between GDDR6 and HBM2 falls in the next year or two; GDDR6’s speed somewhat undermines the advantages of HBM2, but the latter memory technology recently saw its own bandwidth and capacity boost, so it may still be used as a high-end option, especially in space-constrained products such as socketed accelerators.

Turing, Polaris, and Intel's Xe
Comments Locked


View All Comments

  • PeachNCream - Friday, December 28, 2018 - link

    I can sum up 2018's GPUs more quickly. Higher TDPs and higher prices justified by pointless additional features competing with last gen rebrands and spreadsheet only iGPUs.
  • pavag - Saturday, December 29, 2018 - link

    I will not buy the stocked nvidia cards because they are obsolete. And will not buy the newer models because they don’t have the performance needed for raytracing.

    Also, this was a bad year for new games, and I buy cards when I desire to pay new games at his best.
  • StormyParis - Sunday, December 30, 2018 - link

    Interesting, Thank you.

    As I got older, my interest in the minutiae of tech, and in games, has waned. Those recap articles, and your occasional deep dive into very new stuff, keep me kinda of up to speed on the tech, so I don't feel as obsolete as my gear and games.

    So, Thanks ;-p
  • Kevin G - Monday, December 31, 2018 - link

    While I cannot discount the possibility of a mid cycle GPU refresh from nVidia to take advantage of 7 nm, I would consider it unlikely. The dies of the TU106, TU104 and TU102 are all massive. Even with a shink, the TU102 is still going to be big. I don't think any fab is ready to tackle such a large die at the volumes nVidia wants on a cutting edge process. If they could, it would have made sense to skip those Turing chips because they are so large to begin with, nVidia was maintaining the performance crown with the existing 1080Ti and they had excess inventory in the channel. While the supply of Pascal cards has finally tried up with the holiday season (and high RTX pricing), they still maintain a performance dominance that will likely remain until Navi arrives. I think for 2020 that that is a given but enough time will have passed we'll be looking at a new architecture from nVidia anyway.

    While not Xe, we will likely see Gen 11 graphics when consumer Ice Lake arrives at the end of 2019. It'll be progress there and finally move Intel to DP 1.3/1.4 and support variable rate refresh. So while the performance will be lackluster, the display technology side will pick up.

    Speaking of display technologies, I'm wondering when the first HDMI 2.1 capable GPUs arrive. I expect the first HDMI 2.1 enabled displays to be announced (though not shipping) in a few weeks at CES.
  • mapesdhs - Wednesday, January 2, 2019 - link

    Completely off-topic, but anyone noticed that, "With that said...", or, "With that being said...", have become extremely popular phrases used in YT videos & articles these days? :D

Log in

Don't have an account? Sign up now