As part of this evening’s AMD Capsaicin event (more on that later), AMD’s Chief Architect and SVP of the Radeon Technologies Group has announced a new Radeon Pro card unlike anything else. Dubbed the Radeon Pro Solid State Graphics (SSG), this card includes M.2 slots for adding NAND SSDs, with the goal of vastly increasing the amount of local storage available to the video card.

Details are a bit thin and I’ll update this later this evening, but in short the card utilizes a Polaris 10 Fiji GPU and includes 2 PCIe 3.0 M.2 slots for adding flash drives to the card. These slots are then attached to the GPU (it’s unclear if there’s a PCIe switch involved or if it’s wired directly), which the GPU can then use as an additional tier of storage. I’m told that the card can fit at least 1TB of NAND – likely limited by M.2 MLC SSD capacities – which massively increases the amount of local storage available on the card.

As AMD explains it, the purpose of going this route is to offer another solution to the workset size limitations of current professional graphics cards. Even AMD’s largest card currently tops out at 32GB, and while this is a fair amount, there are workloads that can use more. This is particular the case for workloads with massive datasets (oil & gas), or as AMD demonstrated, scrubbing through an 8K video file.

Current cards can spill over to system memory, and while the PCIe bus is fast, it’s still much slower than local memory, plus it is subject to the latency of the relatively long trip and waiting on the CPU to address requests. Local NAND storage, by comparison, offers much faster round trips, though on paper the bandwidth isn’t as good, so I’m curious to see just how it compares to the real world datasets that spill over to system memory.  Meanwhile actual memory management/usage/tiering is handled by a combination of the drivers and developer software, so developers will need to code specifically for it as things stand.

For the moment, AMD is treating the Radeon Pro SSG as a beta product, and will be selling developer kits for it directly., with full availability set for 2017. For now developers need to apply for a kit from AMD, and I’m told the first kits are available immediately. Interested developers will need to have saved up their pennies though: a dev kit will set you back $9,999.

Update:

Now that AMD’s presentation is over, we have a bit more information on the Radeon Pro SSG and how it works.

In terms of hardware, the Fiji based card is outfit with a PCIe bridge chip – the same PEX8747 bridge chip used on the Radeon Pro Duo, I’m told – with the bridge connecting the two PCIe x4 M.2 slots to the GPU, and allowing both cards to share the PCIe system connection. Architecturally the prototype card is essentially a PCIe SSD adapter and a video card on a single board, with no special connectivity in use beyond what the PCIe bridge chip provides.

The SSDs themselves are a pair of 512GB Samsung 950 Pros, which are about the fastest thing available on the market today. These SSDs are operating in RAID-0 (striped) mode to provide the maximum amount of bandwidth. Meanwhile it turns out that due to how the card is configured, the OS actually sees the SSD RAID-0 array as well, at least for the prototype design.

To use the SSDs, applications need to be programmed using AMD’s APIs to recognize the existence of the local storage and that it is “special,” being on the same board as the GPU itself. Ultimately the trick for application developers is directly streaming resources from  the SSDs treating it as a level of cache between the DRAM and system storage. The use of NAND in this manner does not fit into the traditional memory hierarchy very well, as while the SSDs are fast, on paper accessing system memory is faster still. But it should be faster than accessing system storage, even if it’s PCIe SSD storage elsewhere on the system. Similarly, don’t expect to see frame buffers spilling over to NAND any time soon. This is about getting large, mostly static resources closer to the GPU for more efficient resource streaming.

To showcase the potential benefits of this solution, AMD had an 8K video scrubbing demonstration going, comparing performance between using a source file on the SSG’s local SSDs, and using a source file on the system SSD (also a 950 Pro).

The performance differential was actually more than I expected; reading a file from the SSG SSD array was over 4GB/sec, while reading that same file from the system SSD was only averaging under 900MB/sec, which is lower than what we know 950 Pro can do in sequential reads. After putting some thought into it, I think AMD has hit upon the fact that most M.2 slots on motherboards are routed through the system chipset rather than being directly attached to the CPU. This not only adds another hop of latency, but it means crossing the relatively narrow DMI 3.0 (~PCIe 3.0 x4) link that is shared with everything else attached to the chipset.

Though by and large this is all at the proof of concept stage. The prototype, though impressive in some ways in its own right, is really just a means to get developers thinking about the idea and writing their applications to be aware of the local storage. And this includes not just what content to put on the SSG's SSDs, but also how to best exploit the non-volatile nature of its storage, and how to avoid unnecessary thrashing of the SSDs and burning valuable program/erase cycles. The SSG serves an interesting niche, albeit a limited one: scenarios where you have a large dataset and you are somewhat sensitive to latency and want to stay off of the PCIe bus, but don't need more than 4-5GB/sec of read bandwidth. So it'll be worth keeping an eye on this to see what developers can do with it.

In any case, while AMD is selling dev kits now, expect some significant changes by the time we see the retail hardware in 2017. Given the timeframe I expect we’ll be looking at much more powerful Vega cards, where the overall GPU performance will be much greater, and the difference in performance between memory/storage tiers is even more pronounced.

Source: AMD

POST A COMMENT

120 Comments

View All Comments

  • yannigr2 - Tuesday, July 26, 2016 - link

    Fanboys know better than companies. Reply
  • prisonerX - Tuesday, July 26, 2016 - link

    It doesn't cost $10K per card. That's for the developer kit. They haven't announced retail pricing. Reply
  • Demiurge - Tuesday, July 26, 2016 - link

    If two cars are stopped at a red light, which one is faster? Reply
  • pogostick - Tuesday, July 26, 2016 - link

    the red one. Reply
  • silverblue - Tuesday, July 26, 2016 - link

    Unless it's the Milky Way cars. Reply
  • Snorklax - Tuesday, July 26, 2016 - link

    Image you have a 2TB dataset you need your GPU to process.

    In all cases, this dataset will have to be read from slow storage and written to slow storage. No amount of NVLink is going to help unless you can get the entire dataset into RAM. If you have 2TB of RAM, I'm guessing that the product is not for you.

    But if the GPU has direct read/write access to the slow storage, you would get rid of all the latency in having to go through the system to fetch or write the data. So your GPU is wrangling with the 2TB of data all by itself, with 0 impact on the rest of the system

    How is this pointless?
    Reply
  • Notmyusualid - Tuesday, July 26, 2016 - link

    With a name like Communism - how do you expect to win a logical argument against him.

    Regardless, what you say makes 1000% times more sense that his comment, so rest assured many of us can see that.
    Reply
  • ddriver - Tuesday, July 26, 2016 - link

    Latency would not be an issue in this scenario, because streaming will mask it, there will only be the initial latency - so instead of processing your data set in say 60 minutes, you will be processing it in 59 minutes 59 seconds and 99 hundredths if you access data through the PCI_E bus. Reply
  • smilingcrow - Tuesday, July 26, 2016 - link

    Potentially for a sequential read but what about random? Reply
  • yannigr2 - Tuesday, July 26, 2016 - link

    This doesn't look like something that will sell in hundred of thousands of cards, more like a few thousands at best. And for a few thousand people/corporations, 1TB of SSD storage could be enough and could also offer enough acceleration and be a good alternative than going Nvidia and NVlink and then trying to build a system with half terabyte or more of system memory. Reply

Log in

Don't have an account? Sign up now