In 2018 Marvell announced the 88NR2241 Intelligent NVMe Switch: the first—and so far, only—NVMe hardware RAID controller of its kind. Now that chip has scored its first major (public) design win with Hewlett Packard Enterprise. The HPE NS204i-p is a new RAID adapter card for M.2 NVMe SSDs, intended to provide RAID-1 protection to a pair of 480GB boot drives in HPE ProLiant and Apollo systems.

The HPE NS204i-p is a half-height, half-length PCIe 3.0 x4 adapter card designed by Marvell for HPE. It features the 88NR2241 NVMe switch and two M.2 PCIe x4 slots that connect through the Marvell switch. This is not a typical PCIe switch as often seen providing fan-out of more PCIe lanes, but one that operates at a higher level and natively understands the NVMe protocol.

The NS204i-p adapter is configured specifically to provide RAID-1 (mirroring) of two SSDs, presenting them to the host system as a single NVMe device. This is the key advantage of the 88NR2241 over other NVMe RAID solutions: the host system doesn't need to know anything about the RAID array and continues to use the usual NVMe drivers. Competing NVMe RAID solutions in the market are either SAS/SATA/NVMe "tri-mode" RAID controllers that require NVMe drives to be accessed using proprietary SCSI interfaces, or are software RAID systems with the accompanying CPU overhead.

Based on the provided photos, it looks like HPE is equipping the NS204i-p with a pair of SK hynix NVMe SSDs. The spec sheet indicates these are from a read-oriented product tier, so the endurance rating should be 1 DWPD (somewhere around 876 TBW for 480GB drives).

This solution is claimed to offer several times the performance of SATA boot drive(s), and can achieve high availability of the OS and log storage without using up front hot-swap bays on a server. The HPE NS204i-p is now available for purchase from HPE, but pricing has not been publicly disclosed.


Related Reading

Source: Marvell, HPE



View All Comments

  • Dug - Tuesday, October 6, 2020 - link

    This does seem like a niche product. The host isn't generally used except for some reads and occasional updates. Reboot times are dependent on manufacturer bios, and not the drives, so I don't see a need for speed here. The space savings is questionable as most servers have OS drive space in the back already for 2.5", or even running on a pair of sd cards.

    I don't see any mention of hotswap, so if there is any downtime because you have to bring the entire server down to replace a drive, then it kind of defeats the purpose. If you had a single drive and it went down, it would take the same amount of time to replace it and restore from backup.
  • Spunjji - Wednesday, October 7, 2020 - link

    Agreed about it being niche, for sure. Downtime to replace a drive doesn't really defeat the purpose though - it allows you to control *when* that downtime happens, as opposed to having the server go down randomly whenever the drive fails. Reply
  • foobardotcom - Friday, October 9, 2020 - link

    These kind of cards are quite handy if you have for example 1U server chassis with 8x2.5" disk bays and no internal m.2 slots. These cards allow you to use pci-e slot for the OS disk and leave all 2.5" disks for data raid. This kind of non-transparent raid setup is easy when dealing with for example Debian UEFI installation because you don't have to deal with over complicated preseed configs and pull your hair out when trying to make it do another UEFI partition to provide somekind of redundancy. SD card based setups are more geared towards using it more or less like booting VMware ESX from the sd cards and not using sd cards for constant writing not even system logs.

    Usually these kind of cards have some server vendor provided management software that can communicate with controller to retrieve hardware information from those drives. Most probably MTBF of these cards is so high that total loss of other card is non-issue in big picture when compared with some more common read errors.
  • Jorgp2 - Tuesday, October 6, 2020 - link

    I don't under the point of this, hasn't high point had a quad m.2 card for a few years now?
  • MenhirMike - Tuesday, October 6, 2020 - link

    The specific thing here is that from the view of the Operating System, it's not a RAID Controller but only sees one regular NVMe drive. All the RAID Stuff is done transparently by the controller, without the need for any special driver support.

    I can't judge whether this is needed/useful, but it is a key difference between this and other NVMe RAID cards.
  • Tomatotech - Tuesday, October 6, 2020 - link

    Looks good. If I still ran ATX desktops systems, I'd definitely be interested in picking one of these card up (either 2 slot or 4 slot) and throwing a some fast nvme drives for extreme speed in general use.

    Here's a review (not by me) from

    "We needed very fast writes as well as reads. Our app generates millions of small files, and this nicely removed the write bottleneck. Running on Asus Z10PE-D16, dual Xeon E5-2699v4 with 1TB RAM. We used 4 x 2TB Samsung Evo Plus M.2 ssds on 2 x SSD7103 cards using s/w raid on WinS2019."

    Transparent RAID support is important, am fed up of being burnt by substandard RAID drivers which is why I've avoided RAID for many years.
  • Tomatotech - Tuesday, October 6, 2020 - link

    To be clear that review is for a different card that does similar things. Reply
  • Toadster - Thursday, October 22, 2020 - link

    Intel vROC does this without an add-on card... Reply

Log in

Don't have an account? Sign up now