This is the second of a two-part post about why storage arrays and appliances with SSD drives can be a good idea, here is link to the first post.


So again, why would putting drive form factors SSDs be a bad idea for  existing storage systems, arrays and appliances?


Benefits of SSD drive in storage systems, arrays and appliances:

  • Familiarity with customers who buy and use these devices
  • Reduces time to market enabling customers to innovate via  deployment
  • Establish comfort and confidence with SSD technology for customers
  • Investment protection of currently installed technology (hardware  and software)
  • Interoperability with existing interfaces, infrastructure, tools  and policies
  • Reliability, availability and serviceability (RAS) depending on  vendor implementation
  • Features and functionality (replicate, snapshot, policy, tiering,  application integration)
  • Known entity in terms of hardware, software, firmware and  microcode (good or bad)
  • Share SSD technology across more servers or accessing applications
  • Good performance assuming no controller, hardware or software  bottlenecks
  • Wear leveling and other SSD flash management if implemented
  • Can end performance bottlenecks if backend (drives) are a  problem
  • Coexist or complemented with server-based SSD caching


Note, the mere presence of SSD drives in a storage system, array  or appliance will not guarantee or enable the above items to be enabled, nor to  their full potential. Different vendors and products will implement to various  degrees of extensibility SSD drive support, so look beyond the check box of  feature, functionality. Dig in and understand how extensive and robust the SSD implementation  is to meet your specific requirements.


Caveats of SSD drives in storage systems, arrays and appliances:

  • May not use full performance potential of nand flash SLC  technology
  • Latency can be an issue for those who need extreme speed or  performance
  • May not be the most innovative newest technology on the block
  • Fun for startup vendors, marketers and their fans to poke fun at
  • Not all vendors add value or optimization for endurance of drive  SSD
  • Seen as not being technology advanced vs. legacy or mature systems


Note that different vendors will have various performance characteristics,  some good for IOPs, others for bandwidth or throughput while others for latency  or capacity. Look at different products to see how they will vary to meet your  particular needs.


Cost comparisons are tricky. SSD in HDD form  factors certainly cost more than raw flash dies, however PCIe cards and FTL  (flash translation layer) controllers also cost more than flash chips by  themselves. In other words, apples to apples comparisons are needed. In the future, ideally the baseboard or motherboard  vendors will revise the layout to support nand flash (or its replacement) with  DRAM DIMM type modules along with associated FTL and BIOS to handle the flash  program/erase cycles (P/E) and wear leveling management, something that DRAM  does not have to encounter. While that provides great  location or locality of reference (figure 1), it is also a more complex approach that  takes time and industry cooperation.


Locality of reference for memory and storage
Figure 1: Locality of reference for memory and storage


Certainly, for best performance, just like realty location matters  and thus locality of reference comes into play. That is put the data as close  to the server as possible, however when sharing is needed, then a different  approach or a companion technique is required.


Here are some general thoughts about SSD:

  • Some customers and organizations get the value and role of SSD
  • Some see where SSD can replace HDD, others see where it  compliments
  • Yet others are seeing the potential, however are moving cautiously
  • For many environments better than current performance is good  enough
  • Environments with the need for speed need every bit of  performance they can get
  • Storage systems and arrays or appliances continue to evolve  including the media they use
  • Simply looking at how some storage arrays, systems and appliances  have evolved, you can get an idea on how they might look in the future which  could include not only SAS as a backend or target, also PCIe. After all, it was  not that long ago where backend drive connections went from propriety to open parallel  SCSI or SSA to Fibre Channel loop (or switched) to SAS.
  • Engineers and marketers tend to gravitate to newer products nand technology,  which is good, as we need continued innovation on that front.
  • Customers and business people tend to gravitate towards deriving  greatest value out of what is there for as long as possible.
  • Of course, both of the latter two points are not always the case  and can be flip flopped.
  • Ultrahigh end environments and corner case applications will  continue to push the limits and are target markets for some of the newer  products and vendors.
  • Likewise, enterprise, mid market and other mainstream environments  (outside of their corner case scenarios) will continue to push known technology  to its limits as long as they can derive some business benefit value.


While not perfect, SSD in a HDD form factor with a SAS or SATA  interface properly integrated by vendors into storage systems (or arrays or  appliances) are a good fit for many environments today. Likewise, for some environments,  new from the ground up SSD based solutions that leverage flash DIMM or daughter  cards or PCIe flash cards are a fit. So to are PCIe flash cards either as a  target, or as cache to complement storage system (arrays and appliances). Certainly,  drive slots in arrays take up space for SSD, however so to does occupying PCIe  space particularly in high density servers that require every available socket  and slot for compute and DRAM memory. Thus, there are pros and cons, features  and benefits of various approaches and which is best will depend on your needs  and perhaps preferences, which may or may not be binary.


I agree that for some applications and solutions, non  drive form factor SSD make sense while in others, compatibility has its  benefits. Yet in other situations nand flash such as SLC combined with HDD and  DRAM tightly integrated such as in my Momentus XT HHDD is good for laptops,  however probably not a good fit for enterprise yet. Thus, SSD options and  placements are not binary, of course, sometimes opinions and perspectives will  be.


For some situations  PCIe, based cards in servers or appliances make sense, either as a target or as  cache. Likewise for other scenarios drive format SSD make sense in servers and  storage systems, appliances, arrays or other solutions. Thus while all of those  approaches are used for storing binary digital data, the solutions of what to  use when and where often will not be binary, that is unless your approach is to  use one tool or technique for everything.


Here are some related links to  learn more about SSD, where and when to use what:
  Why SSD based arrays and storage appliances can be a good idea (Part I)
  IT and storage economics 101,  supply and demand
  Researchers and marketers dont agree on future of nand flash  SSD
  Speaking of speeding up business  with SSD storage
  EMC VFCache respinning SSD and  intelligent caching (Part I)
  EMC VFCache respinning SSD and intelligent caching (Part II)
  SSD options for Virtual (and Physical) Environments: Part I  Spinning up to speed on SSD
  SSD options for Virtual (and Physical) Environments, Part  II: The call to duty, SSD endurance
  SSD options for Virtual (and Physical) Environments Part  III: What type of SSD is best for you?


Ok, nuff said for now.


Cheers gs