You basically tell them how much usable space you need (high water mark), and they will make sure you have that amount available. It also comes with an SLA for latency and they'll even upgrade your controllers for free if it goes out of bounds.
Many people who deal with raw hardware don't realize that what the business wants is guarantees that shit is going to work, cost mostly be damned.
And so that's how you get scenarios where you pay through the nose, but you've explicitly called out what you need and you get it, even if the company ends up doing weird things like shipping the top-end hardware (with speed disabled) so that they can make sure they meet the requirements.
In this case, Pure will monitor your arrays for you and determine if the latency issues are coming from the controllers or the nvme subsystem, and they’ll either add more drives or swap out the controllers for faster hardware to make sure the SLA is met.
It’s a pretty good system especially for CFOs who prefer opex model
That’s debatable when you consider the performance you get from Pure.
Our X20 which is on the small end, with only 10 drives can easily do hundreds of thousands of IOPs, supporting over 1000 VMs and a high performance ERP solution.
The whole thing costs us about 50k a year. On an AWS/GCP that would cost a lot, lot more.
Sure, try running a thousand VMs on it though (which is an extremely random workload). A single sequential test at an optimal byte size benchmark for a consumer SSD is not representative of a real workload these arrays would see.
Doing so is a lot more complex and intensive than what a single consumer grade SSD can handle.
My point was just that if you wanted to get dedicated IOPs on AWS to match what you get with modern SANs, it’ll cost you far more.
I get 2 to 2.5 million IOPS for 4k random-reads on my personal server.
Running Linux LVM software RAID over 3 x Samsung NVMe SSDs. That's not a read-write measurement, but it's a satisfying number for a not particularly high end server.
(I use it for a side project's database engine experiments. That level of IOPS supports a very high random query rate.)
100% Read or 100% Write are as far from the real workload as it can get. The only exception would be backups read/restore.
Also 1GB test file is often fits in the SSD's RAM cache. Get iometer, 50R/50W%, blocks from 512 to 16k, at least half the size of the storage. Then you would see the real performance.
NB if you have random read way below the random write means you are measuring anything but the storage performance.
I tried CrystalMark for my desktop SSD with 64GB 4k 50R/50W and got 134K, mostly I think because it has 400K read and 130K write, so writes are bottleneck.
Sorry, lazy to learn how to use iometer, but you probably have SSD too and can report your results.
>> but you probably have SSD too and can report your results.
Uh-uh!
I intended to show you the difference between a single NVMe drive and a SSD SAN (a bit old, but still very performant to handle ~900 VMs).
Sure, I can just ramp up queue depth and see some magical numbers, but for me the real performance is in everyday tasks and running CrystalMark isn't an everyday occurrence.
> Sorry? 900 VMs equals 100% full random access. There is no sequential access there, just as I said in my first comment.
it depends on workload, if they do most of the work in RAM, and most of fs traffic is snapshoting and restoring from snapshots, then you will get 99% io seq traffic. If they do some non-trivial fs operations, then you will get q16t16 io traffic. It is very unlikely you will get q1t1 random.
> Because of your comment[0].
> This comment[1] pretty much summarized what I said in a more eloquent way.
In my view you are jumping from topic (single ssd vs nas) to another topic (your speculations about benchmark not representing real world scenarios) and then back.
I agree that if you ran a synthetic benchmark against a brand new pure x20 with 10 NVME drives you'd see astronomical numbers for iops.
That is absolutely not representative of a real workload of mixed reads and writes, different block sizes, potentially different queue depths, all coming in on hundreds or maybe thousands of different volumes.
A single consumer Samsung SSD would hilariously crumble under a real workload, it will NOT deliver hundreds of thousands of IOPs in that environment.
Your benchmark screenshot is the equivalent of showing your pickup truck can do burnouts in the parking lot, and extrapolating that to think it could keep up with a Ferrari on the Nurburgring.
> That is absolutely not representative of a real workload of mixed reads and writes, different block sizes, potentially different queue depths, all coming in on hundreds or maybe thousands of different volumes.
if you see bottleneck there, it could be that actual SSD speed is maybe irrelevant in your case since upstream software is not optimized.
We switched from EMC to PS for our on-prem stuff about 5 or 6 years ago. I sat in on some of the RFP process. PS was basically the same price for a vastly superior offering. You had to make a leap of faith, though, that PS would survive as a company.
Now there’s no doubt, so they probably have more pricing power. And EMC most likely has an offering that competes with the performance (or at least as close as they can get). Back then it was flash vs spinning disks.
We do most of our stuff in the cloud, BTW, and PS is faster for cheaper, by a wide margin. But we’re not going to bring that compute back (I’d like to).
One thing I remember was them being very open and direct, and they wanted to know a lot of heuristics about our data so they could give an accurate estimate about speed and dedup. Which was really close to the reality. Oh, and we’re using one of their devices for an Itanium OpenVMS cluster, it works like a charm. Try getting a startup to support that kind of setup these days, lol
Given their target 66% profit margin (product price = 3x cost of goods sold), their $/GB compares favorably with any flash based storage that doesn’t use wide erasure coded stripes, even ignoring dedupe and compression.
Not sure how they compare to other enterprise offerings these days, or how hard it is to roll your own RAID 6 open source thing with synchronous replication.
Personally I have years of experience with their SANs and they're a dream to work with. But I also never had to write the checks for them.