Outrageous pricing, no thanks
I can buy a 1TB sata SSD (albeit slower) for considerably less.
If evidence from last/early this year’s large technology shows is anything to go by, then 2015 might just be the year that consumer PCIe SSD market begins stirring – something that's long overdue. Kingston HyperX Predator 480GB HHHL PCIe SSD Kingston's HyperX Predator, the company's fastest SSD to date PCIe-based drives, …
Wasn't in the review? Anyway, Googled it around £180 for anyone wondering.
Yes, you can buy a SATA SSD for less than half the price. But if you need the speed of a PCIe drive, SATA is useless however much cheaper it is! Hopefully competition will rapidly drive PCIe drive prices down to SATA SSD levels.
I want one, but don't need one!
Really, really depends on your setup. I have a 240 GB PCIe x4 and in comparison I have an IBM serveRAID x8 tied to up to 8 SSD's. Considering that I've two dozen SSD's in sets, fast is normal. I love the board for my normal lab config. It's when it's play time that things get (fun) weird. I absolutely loathe load time as I experiment. BTW, even when I use just three drives the RAID wins by a nose.
No reason I can think of why you can't run two of them in a Linux software RAID-1 set.
What I wonder about, is whether SSDs of any sort can be trusted to report data that's gone bad in the same way as hard drives can. Checksums and ECC codes are utterly fundamental to spinning rust drives, but even with these it's possible for a controller failure to allow undetected data corruption. And RAID-1 normally assumes that if a write succeeded without error, a read may be satisfied with data off either disk in the RAID array without checking that it's (still) the same on the other device.
For the paranoid, these devices might be fast enough that a new RAID class could be defined and used without crippling loss of performance. Minimum three mirrored members. On read, get the data from all three members and check for equality. If one differs from the other two, it can be assumed to be the bad one. Two members wouldn't let you know which was the good one (assuming that the bad drive wasn't detecting its own failure state).
Like navigating with one chronometer or three chronometers, but never two.
I am looking at this as part of a linux box build.
I'd put the OS and 'swap' on two SSDs (SATA) and then use these in either RAID 1 (mirroring) if I have two slots. Or RAID 10 to get yet even better performance.
Its a test bed of sorts, however, based on price / performance and density... I may opt for Intel's bigger yet more expensive kit.
Money is less of an issue because its work kit. (Business expense ;-)
A test platform for working on Spark / Hadoop / etc ... and I need a box that is quiet, I sit next to it and trust me... back in the 90's I had a rack sitting next to me... not fun.
If SSD doesn't checksum the data (and I don't know if it does), then ZFS would seem to be a good option. I have not heard of SSDs failing by returning bad data; I *have* seen lots of them fail completely though.
SSDs do seem to return the normal set of SMART stats, including reallocated sector count, so one presumes they can detect when a sector is bad.
On the Windows Server front you can set it up in Storage Spaces and ReFS to mirror at least two copies. Toss in a few spinners and it'll also do tiering and if at least one spinner force at least one copy on disks as well as the PCIe. With ReFS it'll be wandering all of that to prevent bitrot.
For Linux, ZFS up until Linux provides an equivalent or better.
You can provide reliability at a higher level. The btrfs filesystem does almost exactly as you describe: Everything it stores, it stores with checksum. If data it corrupted the check will not match and it will detect the error. If you've set it to provide protection too, it'll have another copy it can use for recovery.
It's "only" SATA so you'll be limited to 600MB/s per port—and given it's £27.58, probably rather less than 2.4GB/s across all ports—but here you go, a low-profile PCIe card that takes four mSATA sticks:
http://www.scan.co.uk/products/lycom-pe-125-ahci-6gbps-raid-4x-msata-low-profile-pcie-20-host-adapter
Try looking at the Samsung SM951. It's not quite up to the same performance as the Intel 750, as it's an ACHI device (With promises of an NVMe in the pipeline), but it's still bloody fast. 512GB capacity, and a whole lot cheaper than the HyperX or 750.
If you wait a few months you might be able to get the promised NVMe version of the SM951.
Still a bit pricey for use at home, at least for me (unless your a money bags of course).
The 480GB version works out at around 79p per GB.
Same size SATA SSDs are ~30p per GB range. (although of course limited speed to around 550MB/s).
So currently, these are 2-3 times the price per GB, to give you 2-3 times the speed of a SATA SSD.
As a comparison, SATA SSDs were around the same 79 Pence per GB in late 2011/early 2012. So hopefully these will drop to 25p per GB (or less) by 2018.
Get the price down to around 55p per GB (so mid to late 2016?), and I might be tempted.
Still a bit too expensive, but it might start replacing the SATA 3 SSD boot drive + large HDD for media combo that's currently the sweet spot. Sadly, we're still years away from dropping HDDs from that combo - once 1TB+ SSDs drop their price enough, they'll become the new "media drives" with something like this PCIe SSD for booting/apps.
Cool tech, storage speed was a huge bottleneck, I remember the days I had to wait close to a minute and a half for booting XP. now it takes like 10 seconds... from SSD. Now, what bugs me is the way we started accepting a certain degree of imperfections and failures... like HDDs used to have longer warranties, I think they were better build in the past, now they are like disposable cameras. Hopefully these devices are manufactured to a higher quality standard.
I got the 240GB version of this recently to put in a new HP MicroServer G1610T / FreeNAS build. I use it for L2ARC read cache, for which it is absolutely outrageously fast for both large sequential read and small random IOPS type workloads. It's paired with a single Crucial 80GB SSD (supercapacitor-backed) for ZIL write log, connected to the optical SATA port - SATA-II, but not much bandwidth is required for ZIL logging as by definition the writes are small. Volume storage is 4x2TB WD RED SATA. 16GB of RAM completes the picture, of which 8GB is used for running FreeNAS and for ARC read cache. FreeNAS boots off MicroSD.
This provides a bit of a beast of a SAN for my network (client access via CIFS/Plex, Proxmox VM/container storage over NFS/iSCSI - all via 2 aggregrated 1GbE NICs) in a *very* small physical package.
The expansion potential of the MicroServer is pretty limited (4xSATA 3 via SAS HBA, 1 x SATA2, 1 x PCIe) and this feels to be a reasonably balanced configuration for converged end user storage and VM NAS pool use cases. This pretty much describes my home network - file, media and print serving have to just work otherwise my wife gets upset, but I also want to be able to play with stuff ;-)
As an added bonus, I can run a few VMs directly in the remaining 8GB on the box using the FreeNAS VirtualBox plugin. These VMs then get very fast access to storage over the local PCI and SATA buses, rather than having to access over the network. There are a few temporary 'lab' style workloads (for things like Cassandra etc.) that might benefit from that, though they would have to not require much CPU given the relatively weedy Celeron processor provided as standard with the MicroServer. There are options for upgrading the CPU - e.g. to the Xeon E3-1265Lv2 4 core / 8 thread Ivy Bridge - but I'm not sure I can really be bothered with that... .
The only limitation I've so far found is that there's no real option to put 10GbE NICs in to see how fast the NAS is capable of servicing large block reads from the HyperX-cached working set - strongly suspect the limitation here will be the 1GbE NICs built into the MicroServer.
I'd like to see smaller and cheaper drives like that. I use my SSD just for my root partition, with /home on spinning rust. A 32 GB SSD would be ample, yet when I built my PC a few years ago 64 GB was already the smallest I could get. Why would I need a superfast 480GB SSD if all it does is hold documents and media? I wish I could buy a 32 GB SSD with those speeds for around €40.