If your company is a SNIA member and you’re in the Bay Area the Storage Networking Industry Association Symposium might be the excuse you’re looking for to cut work on a lovely summer day.
I’ll be delivering a keynote address on Wednesday morning, July 23rd, at the St. Claire. Think of it as an interactive Animatronic version of StorageMojo.
Topic: Crossing the Next Storage Chasm: 5 New Technologies that will Change your Data Center. The blurb:
New technologies are changing the face of storage. Robin Harris, analyst and author of the StorageMojo blog, looks at 5 of them, including flash SSDs, 10 GigE, and Google-scale storage. Get the incisive StorageMojo take on these topics and what you really need to know.
Notice I left myself some wiggle room. What do you think the other 2 topics should be?
The StorageMojo take
The data center has more pieces in motion today than ever before. The possibilities are almost infinite, but budgets and attention spans aren’t.
As an industry we don’t do a very good job of a) listening to customers and b) responding with insightful solutions. How can the industry help itself and customers get through the maze? I have a few suggestions.
Comments welcome, of course.
How about a prediction market? Just kidding…
Sorry, can’t limit myself to two. You choose.
– ZFS and other end-to-end storage system designs. These designs
will ensure reliable, easy to manage storage as other trends
make storage less reliable and more complex. The flip side
of the coming death of traditional RAID 5, RAID 6, and possibly
SANs. These designs are enabled by …
– Massively multicore computing – hundreds of cores per computer
will transform storage system designs and requirements — as soon
as the software catches up. One dedicated core per disk channel?
Virtual machines managing storage pools, talking to other virtual
machines at memory bus speeds? How do we feed 100 simultaneously
executing threads each trying to work on part of the same file?
– Cloud storage – how high can we go before we run into WAN
bottlenecks? The (first order) economics and flexibility are
compelling; security issues can be solved; but no magic wand can
grow WAN speeds and drop costs to match the disk drive industry.
– Consumerization of storage – Enterprise storage market share will
continue shrinking as consumer storage needs explode. What will
enterprise storage look like with hundreds or thousands of cheap,
physically small, slow, unreliable disk drives? Will we need flash
drives as a buffer?
Should be interesting, wish I could be there!
— Rex
What about Clustered Storage as a topic? The “two controllers and a truckload of disks” architecture that dominates the industry today must ultimately die out in favor of N+M active clustered solutions based on commodity hardware.
Clustered block-level storage is easier, and there are several options in that space, but clustered NAS is a much harder nut to crack without creating bottlenecks for things like “metadata servers.”
Z F and S.
10 Gig Ethernet is probably inevitable – not because it will hold a candle to Infiniband but because it will be so much cheaper (both in hardware and in requirements for expertise).
For ‘Google-scale storage’ you might want to substitute ‘Amazon-style storage’, because Amazon got the basic distribution and organization mechanisms right for general-purpose use (consistent hashing and peer-oriented rather than hierarchical interaction).
Flash SSDs won’t change storage centers significantly unless their per-GB prices drop far faster than predicted (to at most 10x the price of SATA storage, which is poised to drive out ‘enterprise-class’ disks unless the price gap between them narrows dramatically). Otherwise, flash SSD is just cheaper non-volatile cache – and given how far the bottom has dropped out of conventional RAM prices there’s enough room for both volatile (read) RAM cache and battery-backed (read/write) RAM cache/conventional SSD prices to drop that their performance advantages over flash SSD (especially in cases where the cache can be situated in the host close to the point of use rather than way out on the SAN somewhere) may keep flash SSD from getting much traction.
Another trend which may help keep flash SSD from gaining traction may be ‘write anywhere’ storage (which is often functionally similar to log-structured storage but more flexible in avoiding the latter’s drawbacks) – because dramatically improving a disk’s effective ‘random-write’ bandwidth removes SSD’s (at least potential) advantages in that area. NetApp’s 15-year virtual monopoly on this technology may be coming to a close even if they succeed in their current patent litigation: there’s a lot of room to work around those patents if one makes the effort, and a lot better approaches than ZFS’s.
End-to-end data integrity checks also aren’t new, but (just as with write-anywhere approaches) may become a lot more common for block-level storage as well as for file-level storage (where I still haven’t had the chance to look at ZFS’s code to see whether its implementation is as inefficient as some of its literature suggests it may be) – and in doing so will help alleviate concerns about the wisdom of substituting SATA for SAS/SCSI/FC disks.
‘Self-healing’ storage can leverage such additional integrity checks effectively to help make storage management a negative-growth profession despite continuing exponential capacity increases – if accompanied by the kind of automated configuration management that’s possible (both in terms of handling hardware changes transparently and eliminating knob-tweaking: ZFS barely scratches the surface in this area – think Drobo at the enterprise level).
Increasingly intelligent storage software running on ever-faster, ever-cheaper (and increasingly reliable/self-checking) suitably-redundant commodity hardware is what makes all the above possible: we just have to write it.
– bill
Would be very interested to receive a copy of your SNIA presentation if possible.
Thanks!