Tony Pearson is a Master Inventor and Senior IT Architect for the IBM Storage product line at the
IBM Systems Client Experience Center in Tucson Arizona, and featured contributor
to IBM's developerWorks. In 2016, Tony celebrates his 30th year anniversary with IBM Storage. He is
author of the Inside System Storage series of books. This blog is for the open exchange of ideas relating to storage and storage networking hardware, software and services.
(Short URL for this blog: ibm.co/Pearson )
My books are available on Lulu.com! Order your copies today!
Safe Harbor Statement: The information on IBM products is intended to outline IBM's general product direction and it should not be relied on in making a purchasing decision. The information on the new products is for informational purposes only and may not be incorporated into any contract. The information on IBM products is not a commitment, promise, or legal obligation to deliver any material, code, or functionality. The development, release, and timing of any features or functionality described for IBM products remains at IBM's sole discretion.
Tony Pearson is a an active participant in local, regional, and industry-specific interests, and does not receive any special payments to mention them on this blog.
Tony Pearson receives part of the revenue proceeds from sales of books he has authored listed in the side panel.
Tony Pearson is not a medical doctor, and this blog does not reference any IBM product or service that is intended for use in the diagnosis, treatment, cure, prevention or monitoring of a disease or medical condition, unless otherwise specified on individual posts.
This is a reasonable question. Since Invista 2.0 came out months ago in August, and Invista 2.1 is rumored to be out by end of this month, why put out a press release now, rather than just wait a few weeks? Thesignificant part of this announcement was that EMC finally has their first customer reference.To be fair, getting a customer to agree to be a reference is difficult for any vendor. Some non-profitsand government agencies have rules against it, and some corporations just don't want to be bothered byjournalists, or take phone calls from other prospective customers. I suspect EMC wanted to put the good folks from Purdue University in front of the cameras and microphones before they:
In Moore's terminology, Purdue University would be a "technology enthusiast", interested in exploring the technologyof the EMC Invista. Universities by their very nature often see themselves as early adopters, willing to take big risks in hopes to reap big rewards. The chasm happens later, when there are a lot of early adopters, all willing to be reference accounts. The mainstream market--shown here as pragmatists, conservatives, and skeptics-- are unwillingto accept reference claims from early adopters, searching instead for moderate gains from minimal risks. They prefer references from customers that are similar in size and industry. Whether a vendor can get a product to cross this chasm is the focus of the book.
Why "SAN" virtualization?
Technically, Invista is "storage" virtualization, not "SAN" virtualization. Virtualizationis any technology that makes one set of resources look and feel like a different setof resources, preferably with more desirable characteristics. You can virtualizeservers, SANs, and storage resources.
Virtual SAN (VSAN) technology, supported bythe Cisco MDS 9500 Series Multilayer Director Switch, partitions a single physical SAN into multipleVSANs, allowing different business functions and requirements to share a common physical infrastructure.
How does Invista advance Cisco's VSAN functionality? It doesn't, but that doesn't makethe title a falsehood, or the press release by association full of lies.If you read the entire press release, EMCcorrectly states that Invista is "storage" virtualization. Some storagevirtualization products, like EMC Invista and IBM System Storage SAN Volume Controller (SVC), require a SAN as a platform for which to perform their magic.Marketing people might use the term "SAN" torefer not just the network gear that provides the plumbing, but also to include the storage devices that are attached to the SAN. In that light, theuse of "SAN virtualization" can be understood in the title.
More importantly, it appears that EMC no longer requires that you purchase new SAN equipment from themwith Invista. When the Invista first came out, it cost over a quarter-million US dollars to cover thecost of the intelligent switches, but with the price drop to $100K, I imagine this means theyassume everyone has an appropriately-supported intelligent switch already deployed.
Why this architecture?
In his post [Storage Virtualization and Invista 2.0], EMC blogger ChuckH does a fair job explaining why EMC went in this direction for Invista, and how it is different thanother storage virtualization products.
Most storage virtualization products are cache-based. The world's first disk storagevirtualization product, the IBM 3850 Mass Storage System, introduced in 1974, and thefirst tape virtualization product, the IBM 3494 Virtual tape Server, introduced in 1997, bothused disk cache in front of tape storage. Later virtualization products, like IBM SVC and HDS USP-V, use DRAM memory cache in front of disk storage, but the concept is the same.People are comfortable with cache-based solutions, because the technology is matureand well proven in the marketplace, and excited and delighted that these can offer the following features in a mixed heterogeneous disk environment:
instantaneous point-in-time copy
None of these features are provided by Invista, as there is no cache in the switch. Instead,Invista is a "packet cracker"; it cracks open each FCP packet, inspects and modifies the contents, then passes theFCP packet along to the appropriate storage device. This process slows down each read andwrite by some amount, perhaps 20 microseconds. The disadvantage of slowing down every readand write is offset by having other benefits, like non-disruptive data migration.
To compensate for Invista's inability to provide these features,EMC offers a second solution called EMC RecoverPoint, which is an in-band cache-based appliancesimilar in design to SVC, but maps all virtual disks one-to-one to physical disks. It offersremote distance asynchronous mirroring between heterogeneous devices.EMC supports RecoverPoint in front of Invista, but if you are considering buying bothto get the combined set of features, you might as well buy an IBM SVC or HDS USP-V instead,in one system, rather than two, which is much less complicated. IBM SVC and HDS USP-Vhave both "crossed the chasm" having sold thousands of units to every type and size of customer.
Hopefully, this answers the questions you might have about EMC Invista.
Last year, I covered Chris Anderson's book [The Long Tail]. This year, Chris Anderson, editor-in-chief of Wired.com, has an upcoming book titled Free, the past and future of a radical price. Chris talked about his book here at Nokia World 2007 conference, and the [46-minute video] is worth watching.He asks the big question "What if certain resources were free?" This could be electricity, bandwidth, or storage capacity. He explores how this changes the world, and createsopportunities for new business models. However, many people are stuck in a "scarcity" modeland treat nearly-free resources as expensive, and find themselves doing traditional things thatdon't work anymore. Chris mentions [Second Life] as aneconomy where many resources are free, and seeing how people respond to that.Rather than focusing on making money, new businesses are focused on gainingattention and building their reputation. Here are some example business models:
Cross-subsidy: give away the razors, sell the razor blades; or give away cell phones and sell minutes
Ad-Supported: magazines and newspapers sell for less than production costs
Freemium: 99% use the free version, but a handful pay extra for something more
Digital economics: give away digital music to promote concert tours
Free-sample marketing: give away samples to get word-of-mouth advertising
Gift economy: give people an opportunity and platform to contribute like Wikipedia
Nick Carr writes a post [Dominating the Cloud], indicatingthat IBM, Google, Microsoft, Yahoo and Amazon are the five computing giants to watch, as they are more efficient atconverting electricity into computing than anyone else. Last month, I mentioned IBM and Google partnership on cloud computing in my post[Innovationthat matters: cell phones and cloud computing].Nick's upcoming book titled[The Big Switch] looks into "Utility Computing",comparing the change of companies generating their own electricity to using an electric grid, to the recent developments of cloud computing and software as a service (SaaS). Amazon's latest "SimpleDB" online databaseis cited as an example.
Last, but not least, Seth Godin writes in his post [Meatballs and Permeability] about the bits-vs-atoms issue, what Chris Anderson above refers to as the new digital economy. The idea here is that value carried electronically as bits (digital documents, for example) have completely different economics than value carried as atoms (physical objects), andrequires new marketing techniques. Methods from traditional marketing will not be effective in this new age.Here is a [review] of Seth's new book Meatball Sundae: Is Your Marketing Out of Sync?
All three of these books seem to be covering the same phenomenon, just from different viewpoints. I lookforward to reading them.
While Bill Gates is personally benefiting from code he wrote 30 years ago, most software engineers don't getroyalties for their creative efforts. Robin Harris on StorageMojo has a great piece on [Why are the writers striking?]The writers in this case are those who write scripts for television programs. They get 4 cents for every$19.99 DVD sold today, and want this bumped up to 8 cents. More importantly, they want the same deal forcontent shown over the internet. Currently, they get nothing when content they wrote for is shown on the internet, and they would like that fixed also.
Paying royalties to creative writers encourages them to write good stuff. The best stuff will result in moreroyalties, and we want to encourage this. What about software engineers? Don't we want them to write the beststuff also? Shouldn't they get royalties too, not just a flat salary and continued employment?
From a technology-oriented to a service-oriented approach to IT management
Companies are challenged with shifting from a technology/resource-oriented to a service-oriented approach to IT management. This involves new processes, a new reportingstructure for the IT staff, new tools and technologies, and new data to be captured.A top-down approach is recommended for large organizations, but a bottom-up approachmight be easier to implement for small and medium sized businesses.
IT service management architecture and autonomic computing
IBM has been promoting the concept of Autonomic Computing since 2001. A self-managed resource can have an autonomic manager with sensor and effector. The sensor is used to monitor status, a knowledge basecan analyze and plan for appropriate modifications, and execute these through theeffector. The Autonomic Computing Reference Architecture (ACRA) aligns with the Information Technology Service Management (ITSM) model well, with the CMDB acting asthe knowledge base for the autonomic managers. See my earlier post[Self tuning guitars and storage].
Evolving standards for IT service management
Changes to the IT infrastructure must be closely managed to avoid disruptions.IT organizations recognize that standards-based solutions enable interoperability,with less risk, to connect internal and external applications. Standards can be formally developed by standards bodies like ISO, IETF, W3C, OASIS, and DMTF; or be de facto standards that become widely used by companies, which can then laterbe adopted by standards bodies. SML and SDD are emerging standards that are incompatible with the current set of Web Services-based protocols, like WSDM, but work isunderway to try to determine a unifying standard to support all of these under ITSM.
Prospects for simplifying ITSM-based management through self-managing resources
An ideal computing system would take over a great deal of its own management.Today's IT systems are brittle, difficult to understand, and dangerous to change.The savings from automating some tasks are dwarfed by the irreducible costs of humandecision making, agreements and approvals built in formal processes. A true self-managing, scalable IT system would consist of a number of nearly-identical boxes,with a web interface to define high-level policies and provide information on utilization and performance. As the system needs to expand, it can automatically place the order. When the new boxes arrive, they are placed and connectedinto the data center, and the system configures and provisions them appropriately.
IT Autopilot: A flexible IT service management and deliver platform for smalland medium business
Using an airplane analogy, the pilot performs manual steps to get the plane safelyoff the ground, then turns it over to the autopilot for normal operations. The ITAutopilot intends to do this for IT service management in small and medium business (SMB)that may not have a large dedicated IT staff, using an SOA approach that isloosely coupled, stateless, and adhering to Web Services standards. The IT Autopilotemploys workflow-based controls, the autonomic computing MAPE model, and customizedpolicies to address SMB requirements. It could be deployed as an appliance, similarto IBM System Storage Productivity Center.
I'm continuing my coverage of IBM Systems Journal's [fifteen articles about IBM Service Management].As storage hardware cost per GB declines 25 percent per year, the cost of labor has grown to nearly 70percent of the total IT budget. This brings new focus on how we do things, rather than what things siton the raised floor. Yesterday, my post summarized[the first five articles].Here is what I got out of the next five articles:
Integrated change and configuration management
IT Infrastructure Library (ITIL) best practice covers a variety of disciplines, including incident management,problem management, release management, service help desk, change management, and configuration management.IBM has combined the last two into a single database, and this paper provides insights gained fromimplementing these in practice. A special section talks about how service providers can support multipleclients that must be kept separate from each other.
The process of building a Process Manager: Architecture and design patterns
Business processes coordinate and sequence the work done by a collection of people.Most companies define their business process from scratch, and develop their own applicationsto support their implementation. Process Managers are "out of the box" applications that help customers integrateand automate more quickly than building from scratch. These Process Managers leverage and update informationabout configuration items (CIs) in the configuration management database (CMDB). One of the first developedby IBM was the IBM Tivoli Storage Process Manager.
Integration of domain-specific IT processes and tools in IBM Service Management
ITIL tells you what needs to get done, but it doesn't tell you exactly how to do it. Completing a simplechange request to the IT environment can have a drastic impact on service level agreements (SLAs), utilization of existing storage capacity, and business operations. Sometimes it is important to use multipleProcess Manager applications together. To accomplish this, it is important to launch and land in contextat the appropriate points for smooth transition.
Using a model-driven transformational approach and service-oriented architecture for service deliver management
Companies are considering outsourcing as a way to focus on core competencies. However, the trend is towardselective outsourcing, where the customer controls the IT solution architecture and retains their legacy tools.As a result, service providers inherit the business and IT processes from their clients. IBM Research has developed the model-driven business transformation (MDBT) method that choreographs workflow tools with humanactivities. A "balanced scorecard" allows both client and outsourcer monitor progress towards strategic goals.
Catalog-based service request management
Service providers (outsourcers) are able to bring the latest IT technology, best practices, and skilledservice delivery teams. Unfortunately, unique business processes from each client limits the ability to leveragethese resources effectively. A service delivery management platform (SDMP) catalog serves as a repositoryof atomic services and the delivery teams that can perform them. This allows outsourcers to leverage resourcesacross multiple clients, while still being able to tailor business compositions of these atomic services to an individual client's requirements.