/********************************************************* END OF STYLE RULES *********************************************************/

qile网站

  • Friday, November 03, 2006

    Seeing What's Next, Fourth and Final Part

    The third group that Clayton says to watch for signs of change in an industry are the Non-consumers. In this case, people who have storage and information management problems to solve but are not buying current products because they either can't afford them, or can't adapt them to their unique needs. As I mentioned in Part I, the most interesting group of non-consumers are in High Performance Technical Computing (HPTC). They have both problems. One, their research budgets don't allow them to buy expensive supercomputers and RAID arrays, and two, today's arrays don't meet their needs anyway. To build 'cheap' supercomputers, they cluster together lots of x64 compute nodes running Linux. Then, because they spread their HPTC compute task across so many compute nodes, they need storage that enables sharing the data between all those processors. So, even if they could afford today's block or NAS arrays, they don't support this level of information sharing anyway.


    SANs have enabled IT managers to consolidate their storage HARDWARE, but not necessarily their information. Block arrays don't have the capability to effectively share the same information between multiple application servers. Even if it did, most filesystems still assume they use captive DAS disks anyway. So, SAN admins have to partition up their storage through LUN masking and switch zoning so each file system instance thinks it's talking to captive, direct attach storage.


    This won't work for highly parallel HPTC tasks. They need a filesystem with the capability to share files between multiple compute servers. Filesystems such as Sun's QFS do this with block storage but it's inefficient. Compute nodes have to spend a lot of time communicating with each other to compare which blocks are used and free, and which nodes have which blocks opened for read or write access. It's much more efficient to let the storage server do this which means, of course, it has to know how data is grouped into files and it has to manage ownership properties with each file. In other words, it needs an object storage interface. NFS is close but won't have the right ownership properties until V4++. So leading HPTC filesystems such as Lustre and Panasas' PanFS invented their own proprietary object storage protocol that runs over IP. Lustre users assemble their own object storage device using Linux on a commodity server. Panasas builds an object storage array providing one of the first commercial products for this market.


    One of the reasons this disruption is so interesting is that I've talked to several enterprise datacenter managers who want similar capability. They want to move enterprise applications from large servers to racks of scaleable x64 blades. They want the ability for multiple instances of an application running on separate compute nodes to share information (not just array hardware). They also want quick re-provisioning so an application can migrate to a new compute node without manually reconfiguring LUN masks and switch zoning as required in block SANs today. And, of course, they will need SAN security and access logging with this to comply with information laws. HPTC is evolving this technology, mostly through open source. They are evolving object protocols that allow information sharing AND provide the underlying object protocol on which security and compliance features can be added as sustaining enhancements.


    Summary
    Like the old guy in The Graduate who's advice to young Dustin Hoffman is I have just one word for you: Plastics, I have just one word for the future of storage: NAS. NAS has been a classic low-end disruption growing among overshot customers. Netapp has grown with customers who don't need the level of performance and availability of a FC SAN, and want the low cost and ease-of-use of an IP network-based filer appliance. In parallel, the sustaining enhancements to give NAS the same performance and availability as FC SANs are in process. These include NFS over RDMA and NFS V4++ enhancements to enable multipathing and client data services. IP-based NAS WILL come up and displace Fibre Channel. No question.


    Not only is NAS a low-end disruption, it is also the best candidate to become the next radical up-market sustaining innovation for undershot customers. These customers need storage servers that get data in meaningful groups with properties to let them manage compliance with information laws, manage content, put it on the right tier of storage, etc. Once NAS has the required level of performance and availability, then it just takes a series of straightforward sustaining innovations to give it those capabilities.


    Finally, object-based storage on IP is growing as a new-market disruption in HPTC. Since the FC HBA vendors appear to have no interest in growing up to high-performance computing (while they lose the low-end to IP) so these users have moved to IP. Also, they have avoided NFS because they need sharing properties not available in NFS today and they don't need the full functionality of the embedded file system in a NAS server. I predict though, that once those properties, along with RDMA are available in NAS, that someone will build a NAS server optimized for Lustre. Then, the economics of commodity NAS arrays will make those attractive to HPTC.


    Down at the spinning rust, data will always be stored in groups optimized for the hardware such as 512-byte blocks. For availability, it will also make sense to group disks together with RAID hardware optimized for those physical blocks but these block RAID arrays will become commodity devices. The questions for these vendors is to either be the winner in that commodity business, or to embed a filesystem, and meaningful data management, and compete as a NAS storage server. As for which players will pursue which strategy and who will win, I don't know. Clayton says one of the natural laws is that companies try to up-level. That says one or more existing disk drive suppliers may try to leverage their existing manufacturing capability and OEM relationships and do RAID trays. It also says existing block RAID vendors will move up to doing NAS servers. EMC has indicated it may go this way by acquiring Rainfinity. In parallel, existing NAS vendors such as Netapp will continue the sustaining innovations to give their products the availability, performance, and data integrity required to compete with Tier 1 block arrays. That will be one of the races to watch to see who gets to enterprise-class NAS first.

    Wednesday, November 01, 2006

    Seeing What's Next Part III, Overshot Customers

    In 'Seeing What's Next' Clayton describes signals indicating that a group of customers have been overshot and some of the industry changes that might result. The key signal is customers who lose interest in new features and are unwilling to pay extra for them, and who just want the product to be cheaper and easier to use. This sure sounds like what's happening with midrange and volume disk arrays. Customers used to pay a lot to get RAID arrays from a trusted supplier like EMC who could make it reliable, and they might even pay extra for features like snapshot and replication. Not anymore. Today most customers assume RAID technology has matured to the point where it 'just works' and features like snapshot are like power windows - just assumed to be there.


    Another signal of overshoot is the emergence of new standards that disaggregate parts of the value chain. This has happened at both the datapath and management interfaces. Five years ago when you bought an array, you had to get that vendor's driver stack which included their multipathing driver, and in the case of EMC included HBAs with custom firmware. Today, there is a standard for multipathing built into Solaris, available for Linux and in development for Windows, and you can use standard Qlogic or Emulex HBAs from your favorite reseller. At the management interface, SNIA has now standardized this so you can buy an array from your favorite RAID vendor and plug it into a separate third-party SAN management tool. The result, a RAID array has become an interchangeable, plug-compatible component like a disk drive.


    The next indicator of overshoot is the emergence of a new business model that emphasizes convenience and low price. Sure enough, just this morning I saw an announcement regarding Dell's arrays that they OEM from EMC. Five years ago you had to create a relationship with EMC to use their products, use their support, use their management software, and install their HBAs and stack on your servers. Today you go to a click-through order form on a website, a UPS truck delivers them, and you plug just the array component into your datacenter.


    One type of change Clayton predicts when there are overshot customers is what he calls displacing innovations. This is an innovation that introduces a new product that plugs into points of modularity (aka standard interfaces). We are seeing this as well with the emergence of virtualization devices that sit between the application servers and storage arrays. Pirus (acquired by Sun), StoreAge (just acquired by LSI), Crosswalk, and Incipient are a few of the startups in the space. Cisco and Brocade are adding this capability, and pNFS is an emerging open standard to move this higher-level virtualization out of the array. The result is RAID trays effectively become what disk drives were ten years ago - the commodity hardware that holds the data behind a higher-level virtualization device.


    I think that's the key clue to what's next. RAID arrays become what disk drives were ten years ago. Twenty years ago I developed disk firmware in Digital's disk engineering group. In the early nineties, once the disk interface standardized, it wasn't worth competing with Quantum and Seagate and we moved up the value chain to do RAID instead. Now the next cycle is happening. HP, Sun, and IBM, move up another level and OEM the whole RAID array. EMC has already stated they plan to become a software company so it won't surprise me if and when they start OEM'ing the basic RAID arrays.


    This creates OEM opportunities for the RAID vendors who want to complete in this commodity market. Because this looks so much like what disk drives went through fifteen years I think the role models are companies like Seagate the won in that disruptive cycle. They very precisely standardized form factor and functionality so they could be second-sources for each other at the big OEMs, got very good at high-yield, high volume manufacturing, reduced their product cycle times, and relentlessly drove down cost with each cycle. Who wins this battle? I don't know. Maybe one of the existing RAID suppliers like LSI or Dothill wins. On the other hand, Clayton's books explain that one of the natural laws of business is that companies try to move to the next level of value so maybe one of today's disk drive companies becomes the next big supplier of RAID trays. Like Clayton says, don't use his book to pick stocks. Use it to see higher level trends in the industry. Maybe this is one.

  • 彩客网官网

    安博电竞体育

    天和彩票注册

    博体育

    泊利彩票登录

    下载app送彩金网站

    安博电竞体育

    天禧的彩的网

    mg电子游戏官方网站