Tony Pearson is a Master Inventor and Senior IT Architect for the IBM Storage product line at the
IBM Systems Client Experience Center in Tucson Arizona, and featured contributor
to IBM's developerWorks. In 2016, Tony celebrates his 30th year anniversary with IBM Storage. He is
author of the Inside System Storage series of books. This blog is for the open exchange of ideas relating to storage and storage networking hardware, software and services.
(Short URL for this blog: ibm.co/Pearson )
My books are available on Lulu.com! Order your copies today!
Safe Harbor Statement: The information on IBM products is intended to outline IBM's general product direction and it should not be relied on in making a purchasing decision. The information on the new products is for informational purposes only and may not be incorporated into any contract. The information on IBM products is not a commitment, promise, or legal obligation to deliver any material, code, or functionality. The development, release, and timing of any features or functionality described for IBM products remains at IBM's sole discretion.
Tony Pearson is a an active participant in local, regional, and industry-specific interests, and does not receive any special payments to mention them on this blog.
Tony Pearson receives part of the revenue proceeds from sales of books he has authored listed in the side panel.
Tony Pearson is not a medical doctor, and this blog does not reference any IBM product or service that is intended for use in the diagnosis, treatment, cure, prevention or monitoring of a disease or medical condition, unless otherwise specified on individual posts.
This week, I am attending the [InterConnect Conference] in Las Vegas, Feb 21-25, 2016. This is IBM's premier Cloud & Mobile conference for the year.
Monday morning I attended the General session and a break-out session.
7030A General Session Day 1: Digital Business Transformation
The General Session was kicked off by severak clients:
Richard Holmes, Westpac Group, a 200-year-old bank with 21,000 branch locations across Australia and New Zealand. They have migrated 70 percent of their applications to the Hybrid Cloud. Provisioning server and storage resources went from 84 days to just minutes.
Matthias Rebellius, Siemens AG, Building Services. They use IBM Watson IoT to monitor the energy usage of their buildings. They have reduced energy consumption 20 to 30 percent, eliminating over 10 million cubic feet of CO2 greenhouse gas.
Robert LeBlanc, IBM Senior Vice President for Cloud, took the stage and welcomed the 23,000 attendees. Developers are turning to IBM Cloud to deliver timely, knowledgeable, and secure experiences for their customers and end-users. Business leaders are seeking new ways to enable their companies to securely implement hybrid cloud strategies that integrate mobile, IoT, and cognitive. He focused on five areas:
Choice, but with Consistency
Hybrid Cloud Integration
Powerful, Accessible data and analytics
Robert indicates that 100 percent of our strategic software products are now Hybrid-Cloud enabled. We get over 3.2 billion API calls per month, and 20,000 new IBM Bluemix users per month. More than 7,000 startups are now running on IBM SoftLayer. IBM was once again ranked #1 for Hybrid Cloud by industry analysts.
IBM predicts that 80 percent of Internet traffic will be video by year 2019. To that end, IBM offers Aspera, Ustream, and Cloudleap.
New IBM Watson APIs can analyze "tone", "emotion" and "vision".
IBM has partnered with Github to offer an Enterprise-class Github-as-a-Service offering suitable for business use.
IBM "Open for Data" has over 150 pre-populated public data sources for use with analytics. This allows applications to analyze their own data in context with public sources.
Carl Eschenbach, VMware, emphasized its partnership with IBM, announcing the ability to run VMware on IBM SoftLayer "bare metal" systems, enabling features like NSX networking and VSAN virtual storage.
Brian Cross, Apple Vice President of Product Marketing, presented the enormity of Apple's developer ecosystem:
1.5 million apps on Apple iOS application store
11 million developers making these apps
100 billion downloads of these apps
1 billion Apple devices
In the past, these developers used Xcode development environment. To take the most advantage of Apple hardware features, many developers use C or C++ programming languages to develop "Native Apps".
Apple developed a new programming language called Swift that has already made it to the top 20 development languages. He gave a demo of "Swift Playground" that allows developers see their apps running while they develop and edit the code.
Apple has made Swift open source, and extended its use across iOS, Mac OS X, Watch OS, tvOS and even Linux operating systems. This means you can write code for devices, client workstations and even servers in your datacenter or Cloud. Download it at [Swift.org].
John Ponzo, IBM Fellow, Vice President and CTO of MobileFirst, wrapped up the General session. He mentioned the "IBM Swift Sandbox" service that helps developers learn Swift programming:
Kitura -- This open source framework would allow developers to build end-to-end applications, deploy, and collaborate on web services and applications written in Swift. Kitura allows developers to build front-end and back-end code using Swift as the programming language to help simplify modern application development.
OpenWhisk -- A feature on IBM Bluemix that provides an event-driven computing service for dynamic applications. It competes against Amazon's Lambda service.
With new ways to deploy Hybrid Cloud, using new composable development tools, it is clear that "Cloud" is not merely a destination, but a new innovation platform.
1581A University of Chicago Taps into IBM Cloud Object Storage for More Effective Patient Treatments
This session was 30 minutes with Piers Nash, University of Chicago - Center for Data Intensive Science (CDIS), client testimonial, followed by Russell Kennedy, IBM, that covered an overview of Cleversafe used in the solution.
University of Chicago's Center for Data Intensive Science (CDIS) accelerates medical discoveries by democratizing access to data for scientific research. Utilizing an object storage solution, CDIS centrally stores and manages vast amounts of genomic and clinical data at web-scale, allowing researchers to collaborate via shared access to harmonized data sets, speeding discovery and enabling precision medicine.
Their initial focus is cancer research. Cancer costs over $100 Billion USD per year in healthcare costs. It is #1 killer among people under 85 years old, affecting half of all men, and a third of all women. There are 1.7 million new cancer cases in the USA every year, 15 million worldwide.
There is no "single cure" for cancer. Whereas all humans share nearly identical 3.2 billion base pairs of genetic material, there are over 15,000 different kinds of cancers, each with its own genome. Capturing RNA sequences of patients results in images 10-20 GB in size, and over the course of treatment could add up to 1 TB of image data per patient. A million patients with 1TB of data each would be an Exabyte of data (1,000 Petabytes).
To store all of this data, CDIS created the Bionimbus Protected Data Cloud, using Cleversafe as the underlying storage technology. This system goes live June 2016, and they plan to keep the data forever.
(We'll see how well that goes 10 years from now! It might be cheaper just to re-sequence a human's DNA as needed, rather than storing it forever, since an individual's DNA never changes.)
The data is "de-identified" meaning that researches using the data are unable to identify individual people associated with each case study or genomic result. They have already collected 1.66 PB of this data.
Most cancer treatments that have been effective have focused on specific genetics. The problem is targeting precise therapies to the right patients. For example, there are two very similar Lung cancers, and about 20 percent of the time, a Lung cancer is mis-identified, such that the patients has adverse reactions to the wrong treatment. By having more analytics-based medicine, the hope is to reduce this trial-and-error approach.
Russ Kennedy, IBM, wrapped up the session explaining Cleversafe, which was a Chicago-based company formed in 2004 that was acquired last year by IBM. Why did University of Chicago choose Cleversafe? Several reasons:
University of Chicago attempts to use open source projects like Gluster or Ceph failed around the 1-2 PB mark. They knew they would need much more than this!
Cleversafe was a Chicago-based company, offering local support
IDC ranked Cleversafe #1 marketshare leader of object storage in 2014 and 2015! It beats out competitors like Dell/EMC and Salty, as well as Cloud Service Providers like Amazon or Google.
Why object storage? IBM predicts a 332 percent growth in data generated from Mobile devices. As much as 90 percent of traffic on Mobile devices will be from Cloud apps rather than voice or text messages. There will be a 10-fold increase of data stored by year 2020, and at least 80 percent of this data will be unstructured content. Cleversafe estimates that managing object storage requires 15x fewer administrators than traditional storage.
Cleversafe consists of three components. The "Accessor" is software that runs bare metal, as Virtual Machine or Docker container. It offers the OpenStack Swift, HTTP/REST and Amazon S3 object-based interfaces to ingest the data. The data is encrypted, divided into pieces, then through a process called [Erasure Coding] is converted to slices. Those slices are stored on storage-rich servers called "storage nodes".
For example, five pieces of data converted to nine slices could be stored on nine machines, three machines at Site 1, three at Site 2, and three at Site 3. You only need to read back any five slices to reconstruct the data, so you could lose any four of the nine machines and still have full recoverability. If the 5/9 example above, you could lose any one site, and a machine in one of the two remaining surviving sites, and still retrieve all of your data.
There is now an "open beta" called the Transparent Cloud Storage Tiering that bridges GPFS and Spectrum Scale over to Cleversafe.
I wrapped up the morning with a lunch at Border Grill with storage clients and IBM Business Partners. This was the best steak I have had this week!
This week, I am attending the [InterConnect Conference] in Las Vegas, Feb 21-25, 2016. This is IBM's premier Cloud & Mobile conference for the year.
Sunday, I attended a series from IBM Research talking about the latest research areas.
7110A Future Directions in Enterprise Mobile Computing
Gabi Zodik (IBM) presented. Mobile and wearables are transforming all industries. Enabling technologies are required to support the new computing models that are cognitive in nature. Real-time proactive decisions can be made based on the mobile context of a user. Driven by the huge amounts of data produced by mobile devices, the next wave in computing will need to exploit data and computing at the edge of the network.
Future mobile apps will have to be cognitive to "understand" user intentions based on all the available interactions and unstructured data. A new distributed programming paradigm is emerging to meet these needs, which has to deal with massive amounts of data and devices. While the compute and storage capacity on individual devices is small, collectively they exceed all of the servers and storage in Cloud datacenters.
7107A Wearables in the Enterprise
Asaf Adi (IBM) presented. Wearable technology is booming. It is only our imagination that will limit the number of industrial, military, consumer and healthcare applications for this new emerging technology. Wearables are transforming industries and professions, enabling new business opportunities. From a show of hands, half the audience was wearing smart technology already.
In one example, he focused on construction industry. In the USA alone, there are thousands of workplace injuries, costing $190 Billion dollars. Wearable technologies can be incorporated into a hardhat to bright orange vest. In a steel mill, heat stress can be determined from ambient temperature and an employee's heart rate. Over time, we will have multiple wearables, communicating to each other.
In another example, he was able to make a hand gesture (waving his hand in front of his smartphone), and use that to generate code fragment that can be used by software developers to detect that particular hand gesture was made in any application.
Wearables cannot assume they are always connected to the Cloud. Take for example mining, where miners are deep below the ground. Technology to ensure safety needs to work regardless of connectivity.
Privacy is also a big concern. Wearables should not be used by employers to monitor every movement and activity of the employees.
7152A Cognitive IoT -- Today, Tomorrow and Beyond
Alessandro Curioni (IBM) presented. Today's sensors aren't up to the task of unlocking the complex links between people, places and things. To reach the next level, we need technologies that enable them to gather and integrate data from many sources, to reason over that data, and to learn from it. IBM calls this the Cognitive Internet of Things (IoT).
We already know IoT data can be used to predict maintenance needs, but what if it can also help designers engineer more reliable products from scratch? In addition, with advancements in nanotechnology and machine learning we can bring the power of cognitive to the edge—where the data is collected. Imagine tiny edge computers providing Watson services on every sensor?
It is estimated that we have 13 billion IoT sensors today, and that this will more than double to 29 billion by year 2020. This introduces new security threats, new levels of employee engagement, and fundamental shifts in business models.
Sadly, 88 percent of all the IoT is dark, meaning that it is not collected or processed for analysis. While the IT industry has done amazing things with the other 12 percent, we realize that programming techniques are too limited.
That is why cognitive is needed to unleash the value of the data. IBM Watson offers excellent capabilities, including Natural Language Processing (NLP), Machine Learning (ML), Image/Video analytics, and Text Analytics.
Manufacturers like Whirlpool are investigating use of IoT for home appliances, like refrigerators, washers and dryers. This is just the beginning, other industries including Healthcare, Retail, Oil, Mining and Farming will also benefit.
7108A Blockchain and the Future of Finance
Ramesh Gopinath (IBM) presented. Transferring products and funds today is inefficient, expensive, and vulnerable. Blockchain is an emerging fabric for transaction services. It has the potential to radically transform multi-party business networks, enabling significant cost and risk reduction and innovative new business models.
About 18 months ago, the "Blockchain" concept was not ready for business. Since then, Apache has accepted the "HyperLedger" project, with 17 founding companies.
Imagine a company in China or India exporting a product to a company in USA. There may be 10 or some companies or agencies involved, including multiple banks, port authorities, trucking companies, etc. The hand off the equipment, and ensure all parties are paid, some 30 different paper documents may be needed. Each company maintains their own set of records, and all the middlemen take their cut.
Blockchain represents a digitally-signed, encrypted, immutable "ledger" that records all of the steps related to a particular transaction. Since each new block has a checksum of all of the previous blocks, it prevents tampering and fraud. All parties have access to all of the ledger, eliminating discrepancies between different repositories of records.
This can be used to sell stocks, buy real estate, or transfer financial funds to your family overseas. Each party involved in a Blockchain has a node in a peer-to-peer network of nodes that can access a shared Blockchain request. A user initiates the transaction, and the nodes in the network use a Practical Byzantine Fault Tolerance [PBFT] protocol.
By providing [disintermediation], fewer middlemen in the process reduces costs, processing time, and risks. The method allows for the user's transactional privacy, but also ensures accountability and auditability.
7234A Building Cloud Infrastructure for Next-Generation Workloads
Krishna Nathan (IBM) presented. Today's cloud providers are efficient at providing today's cloud services at low costs. However, this efficiency comes with the penalty of inflexible instance types and no real guarantees on performance or quality of service.
Today's systems are organized and optimized for transactional processing, a result of evolution of the past 60 years. Relational Databases offer specific features like Atomicity, Consistency, Isolation, and Durability, known collectively as [ACID].
However, we are expanding beyond "automating our world", or "understanding our world". This means tapping into 90% unstructured workloads, multi-modal scanning, noise-tolerant with variable precision and probabilistic outcomes.
Cloud Providers have used the "best practices" of transactional datacenters. Consequently, next-generation workloads that often do not share the characteristics of traditional workloads are limited in expressing their full potential because of these infrastructure limitations. Now they need to focus on four characteristics: Locality, Composability, Heterogeneity, and Dynamic resource allocation.
New workloads need a combination of CPU, GPU, NVMe, and other resources. How do you schedule which equipment to deploy for incoming workload processing that optimizes performance? By taking these factors into account, clever Cloud providers can optimize performance results to provide best fit for each workload request.
7135A Storing and Using Data in the Cloud -- Putting Together the Puzzle Pieces
Michael Factor (IBM) presented. What do OpenStack Swift, Spark, CouchDB, Kafka and ElasticSearch have in common? They are all open source, they all are available on IBM's cloud today, and they all focus on storage and using data. The trick, though, is putting these puzzle pieces together to solve real problems. You need smart integration between data services motivated by real examples from domains such as IoT, transport and retail.
There are a plethora of of open services to manage data. A recent IDC Analyst study indicates that the worlds data will grow from 8.6 Zetabytes today to 40 Zetabytes in 2020. Michael gave some eye-opening comparisons. If the data was stored on 10-TB hard disk drives, we could make some physical comparisons:
Imagine stacking all of those disk drives one on top of each like a stack of books. the stack today would be 22,000 kilometers, more than half the way to geosynchronous orbiting satellites, but would be over 100,000 kilometers, way past those satellites in 2020.
The weight of those drives today would be comparable to the weight of 1,450 Airbus 380 airplanes. In 2020, they would weigh 6,755 Airbus 380 airplanes.
If the drives were spread across the entire Mandalay Bay convention center floor, they would be 1.7 meters deep today (about 5 feet), but would be 8 meters deep in 2020.
An example of the EMT Madrid bus company using real-time sensors to react to traffic conditions.
Here are the various pieces:
OpenStack Swift -- provides object storage
ElasticSearch, based on Apache Lucene - search engine, such as for metadata or queries
Apache Spark - combines SQL, streams and complex analytics, with filter pushdown support
Apache Parquet -- a column-based data format to replace row-based Comma-Separated-Variable (CSV) format
Apache Kafka - a message bus, works with dashDB and Secor
Beyond programming "glue", we need smart integration to get an order of magnitude boost in performance.
With over 2,000 technical sessions and 500 client testimonials, the event can be intimidating. For those of you attending this conference for the first time, I have some advice:
Review all the sessions you want to attend.
Don't limit yourself to just sessions in your area of job responsibility. Venture out. Look for adjacent spaces. Attend a session of two that you might learn something completely new and different.
Build your schedule in advance. The mobile app [IBM Events] can help you plan out your week.
Select one to five sessions per time slot, this gives you the choice to make a final selection, if there are last minute cancellations, or you discover a room is completely full.
Be mindful that some sessions are at the MGM Grand hotel, and others at the Mandalay Bay Resort, so you may need to account for travel time. This [Week-at-a-Glance] can help. I will be focusing my efforts at the Mandalay Bay.
Here's my first cut at my schedule. Maybe this will help you organize your own.
Sunday, Feb 21
7152A IBM Research Day: Cognitive IoT -- Today, Tomorrow and Beyond
7176A IBM Research Day Demo: Cognitive IoT Analytics Enable Connected Vehicle Service Innovations
7108A IBM Research Day: Blockchain and the Future of Finance
7234A IBM Research Day: Building Cloud Infrastructure for Next-Generation Workloads
7135A IBM Research Day: Storing and Using Data in the Cloud -- Putting Together the Puzzle Pieces
7085A Discover InterConnect: Discover the Value of IBM Systems for Your Business
7291A IBM Research Day Demo: Event-Driven, Serverless Cloud Compute and Storage
7112A IBM Research Day: Cloud Programming Models
7144A Open Technology Summit
Networking Reception -- Mandalay Bay South Convention Center Ballroom
Monday, Feb 22
7030A General Session Day 1: Digital Business Transformation
1581A University of Chicago Taps into IBM Cloud Object Storage for More Effective Patient Treatments
7221A Is Your Data Infrastructure Designed for Cloud, Analytics and Cognitive? Get Ready with IBM Storage!
1441A Data Resiliency: Data-Driven Analytics and Beyond
6488A Blockchain for Dummies
1267A Prudential and IBM: Integrating Application and Storage Management to Drive Cloud Service Levels
7433A Special Session: Open for Data -- An Open Invitation to Help You Solve Your Greatest Data Challenges
3050A IBM Cloud Architecture Center: A Story of 1001 Client Implementations and How You Can Leverage Them
6285A Data Management from the Cloud: Introduction to IBM Storage Insights
7215A Software Defined Storage: How Data Growth and Analytics are Driving New Innovation in Cloud Storage
3690A Meet the Experts on IBM Cloud Storage Services
7171A Solution EXPO Reception - Monday
Tuesday, Feb 23
6139A Manulife's Transformation to Agility with a Hybrid Cloud
4852A The Weather Company: How Insurers Can Differentiate with Weather Data
4955A IBM and Box: Delivering Hybrid Solutions for Enterprise Content Management
7032A General Session Day 2: IT Transformation
6524A The Role of Tape in a Cloud-Based World for Economical and Secure Data Retention
7253A Blogging 201: Plan, Maintain, and Optimize Your Blog
1013A Trends in Encryption of Data at Rest: On-Premise and in the Cloud
3757A Moving from Enterprise Premises to the Cloud in Gradual Steps
2131A Enabling Cloud Business with IBM FlashSystem
1709A Big Data—Beyond the Hype
3318A System of Systems Transformation at the Boeing Company
7408A IBM + Box: Transforming Work in the Cloud
1885A Enterprise IT "as-a-Service": A Hybrid Model for the Digital Enterprise
2692A In Wine There is Wisdom, in Beer There is Freedom, in IBM There is Digital Transformation...
5602A Successfully Architecting the Enterprise Edge: Insight from ADP
1234A Ubuntu -- Make It Your Choice for Cloud and Scale-Out Linux Workloads on POWER8
2154A Expert Panel on Hybrid Cloud Data Protection: Who Is the Service Provider?
7172A Solution EXPO Cocktail Reception
7175A InterConnect ROCKS with Elton John
Thursday, Feb 25
2419A Enhance the Agility of Your Cloud with IBM FlashSystem
7398A Cybersecurity Jeopardy! Are We Putting the Answers before the Question?
4331A Leverage zOS and Cloud Storage for Backup/Archive Efficiency and Cost Reduction
7341A IBM Storage and Catalogic: Software Defined Solutions for Hybrid Cloud and DevOps
2027A Get the Best From Your Storage: Why IBM Spectrum Virtualize Makes Sense for Your Cloud
4921A Radical Storage Simplicity for Your Cloud and How it Can Impact Your Customers
If you use Twitter, follow @IBMInterConnect, @IBMSystems and @IBMStorage for updates, and my own tweets @az990tony. If you take a photo at the event, tag it with #ileadIT to enter into the social-photo contest!
I will be there all week! Contact me if you want to get together.
Well, it's Tuesday again, and you know what that means? IBM Announcements!
This week, IBM announces the second generation of Storwize V5000 flash and disk storage systems. There are the V5000F All-flash configurations, as well as the V5000 that can support a variety of flash and spinning disk drives.
There are three models:
The V5010 has dual 2-core/2-thread processors and 16GB of cache. It supports thin provisioning, FlashCopy, Easy Tier, and remote mirroring. The base unit includes 1 GbE Ethernet ports for iSCSI host connectivity, with options to add 16GB Fibre Channel, 12Gb SAS, and 10GbE iSCSI/FCoE as well.
The 2U controllers and expansion enclosures can hold either 24 small 2.5-inch drives, or 12 larger 3.5-inch drives. A single control enclosure has two active/active IBM Spectrum Virtualize nodes, and can attach up to 10 expansion enclosures for a maximum of 264 drives.
The V5020 unit has dual 2-core/4-thread processors and up to 32GB of cache. It supports everything the V5010 does, plus encryption. The encryption is done via the Intel AES-NI instruction set to eliminate the need for special "self-encrypting drives" (SED) that other storage devices may require.
The V5030 has dual 6-core/4-thread processors and up to 64GB of cache. It supports everything the V5010 and V5020 do, plus Real-time Compression and external virtualization. The Real-time Compression can achieve up to 80 percent space savings, representing a 5:1 compression ratio.
Each control enclosure can attach to 20 expansion enclosures, which can support 504 internal drives per controller, and up to 1,008 with two controllers (four Spectrum Virtualize nodes) clustered together. This is in addition to the drives in external storage systems virtualized.
Can you believe it has been a year already since IBM announced VersaStack?
In my May 2012 blog post, [EMC Strikes Back], I poked fun at the fact that Cisco had two girlfriends "significant others": EMC and NetApp.
Cisco originally partnered with EMC to create a converged system called Vblock which combined Cisco UCS servers and switches with EMC storage. The partnership between VMware, Cisco and EMC was dubbed Virtual Computing Environment (VCE).
However, Cisco then partnered with NetApp to create Flexpod, a converged system that combined Cisco UCS servers and switches with NetApp storage. Many of my clients felt that Flexpod was an improvement over Vblock.
Before VersaStack, IBM had its own converged system, PureSystems, which combined IBM POWER and x86 servers with IBM storage. The x86 server portion of this business was sold off to Lenovo, but IBM continues to sell POWER-only and blended x86-and-POWER PureFlex systems, as well as PureApplication and PureData systems.
The [VersaStack] collaboration between IBM and Cisco offers an alternative to Vblock and Flexpod converged systems. Cisco is a leader in x86 blades and networking switches, and IBM is #1 in Flash and Software Defined Storage, including Storage Virtualization. VersaStack gives you the best of both worlds!
The VersaStack has Cisco Validated Designs for use with IBM's Spectrum Virtualize products:
Storwize V7000 Unified
This week, February 11, 2016, 12pm EDT, IBM and Cisco are hosting a webinar on VersaStack. Join us for the one year anniversary of VersaStack in a discussion with IBM, Cisco and VersaStack customers.
The speakers will be discussing VersaStack progress to date and the value VersaStack brings to client workloads. Topics of discussion will include how VersaStack can lower TCO, administrative overhead, reduce downtime and improve resource utilization, and allow for business innovation. The speakers include:
Jonathan Cox, Medicat, Director, Technology Services
Susan Martens, IBM, Director, VersaStack Sales, North America
Kent Hixson, Cisco, Sales Business Development Manager
Wow! That can seem overwhelming. While the conference spans multiple hotels on the strip, I personally will be focusing my time at the [Mandalay Bay resort]. My session will be held at the Solutions Expo on Wednesday 1:45pm. Here are the details:
YSS-1841 IBM Cloud Storage Options
This session will cover private and public cloud storage options, including flash, disk and tape, to address the different types of cloud storage requirements. It will also explain the use of Active File Management for local space management and global access to files, and support for file-and-sync.
Program: Core Curriculum Topic: Systems Hardware Sub-topic: Storage Systems & Software
To help attendees plan your week, InterConnect has a [Session Preview Tool]. I have already found over 40 sessions related to Storage that I am interested in attending!
As you can imagine, I get a lot of email from around the world. This one, from a loyal reader from overseas, was particularly interesting. Normally, I would direct them to read the fantastic manual [RTFM], but decided instead to go ahead and tackle it here in my blog.
I follow your blog for several years, it has served as a reference and training for me in my professional career and I want to thank you.
I am writing because my company has acquired a new IBM Storwize V7000 Gen2 to replace a Gen1, with 16 FC ports, 8 ports per controller node and 8-port FC FlashSystem 900. The idea is to virtualize the V7000 storage part Flash900 and other hand assign directly to the host directly. After much reading on forums and storage Redbooks I have nothing clear as it should be wiring the SAN or as zoning would be made to carry out this installation. I would appreciate if you can write on this subject as controversial as seems to be the zoning and wiring SAN and if possible be clarified by me onstage.
I will tackle this in three steps.
First, let's attach "Server 1" and the FlashSystem 900 to the SAN fabric. IBM Spectrum Virtualize can handle one, two or even four separate fabrics. Let's assume you have a dual-port Host Bus Adapter (HBA) in server 1, and two redundant fabrics. We will connect each server port to each FCP switch. Likewise, we will connect each FCP switch to the FlashSystem 900, carve up "Volume 1", and create SAN "Zone A1" and "Zone A2", which identify "Server 1" as the initiator, and "FlashSystem 900" as the target. This is all basic stuff.
"All Storwize V7000 Gen2 nodes in the Storwize V7000 Gen2 clustered system are connected
to the same SANs, and they present volumes to the hosts. These volumes are created from
storage pools that are composed of mDisks presented by the disk subsystems.
The fabric must have three distinct zones:
Storwize V7000 Gen2 cluster system zones
Create one cluster zone per fabric, and include any port per node that is designated for
intra-cluster traffic. No more than four ports per node should be allocated to intra-cluster
Create a host zone for each server host bus adapter (HBA) port accessing Storwize
Create one Storwize V7000 Gen2 storage zone for each storage system that is
virtualized by the Storwize V7000 Gen2. Some storage control systems need two
separate zones (one per controller) so that they do not 'see' each other."
Second, we connect the Storwize V7000 Gen2 to the FCP switches. You don't need to connect all of the ports, but I recommend that you have each controller node to each FCP switch, requiring four cables. Add more connections for added performance bandwidth.
Carve up "Volume 2" and this will be referred to as a "managed disk", mDisk for short, and create a "storage pool" which were formerly known as a "managed disk group" which is why you often see MDG in the naming conventions and examples. Storage pools can have one or more managed disks, and you can add more dynamically as needed.
The "storage zone" indicates the Storwize V7000 Gen2 as the initiator, and the FlashSystem 900 as target. If you want to increase the performance bandwidth, consider more cables between the FCP switches and the FlashSystem 900. We create "Zone B1" and "Zone B2". I recommend a separate "storage zones" for each additional storage system that you choose to attach to the Storwize V7000 Gen2.
The "cluster zone" that connects all of the Storwize V7000 Gen2 node ports together for node-to-node (intra-cluster) communication. Storwize V7000 Gen2 ports can serve as both initiators and targets dynamically. For example, when you write to one node, the node then copies the cache block over to the second node so there are two copies stored safely on separate nodes. Since we have two fabrics, we create "Zone C1" and "Zone C2".
Third, we connect "Server 2" to FCP switches, same as we did with "Server 1". We create "Volume 3" which is a "virtual disk, or vDisk for short, from the storage pool containing Volume 2. The "host zone"indicates Server 2 as the initiator, and Storwize V7000 Gen2 as the target. We create "Zone D1" and "Zone D2". I recommend putting each additional server in its own set of host zones.
In theory, you could have a server connected to both Volume 1 and Volume 3. For example, a Windows server would have a "C:" drive connected directly to FlashSystem 900 for high-speed performance, and have a "D:" drive on Storwize V7000 Gen2 to contain data. The Storwize V7000 Gen2 introduces 60 to 100 microseconds of added latency, but provides added value such as FlashCopy, Thin Provisioning, and Real-time compression.
Of course, there are unique situations that might require special configurations, depending on the servers, operating systems, host bus adapters, FCP switches, and storage systems involved.
In the 2004 comedy ["A Day Without a Mexican"], the director envisions how disruptive life would be in California if all the Mexicans suddenly disappeared. The point is that sometimes you take things in the background for granted.
I was reminded of this when I saw Mark Underwood's blog post [Mainframe: Still Not Crazy After All These Years]. The article reminds us how critical IBM z Systems mainframes (and related storage like the IBM DS8880 disk systems) are in our lives. Here's an excerpt:
"Warren Buffett's Berkshire Hathaway started buying up IBM stock in 2011 and bought still more of IBM later. Despite its disappointing short-term valuation, Berkshire Hathaway is standing by its IBM investment, which is one of Berkshire's top four plays. ... To make this case, some statistics may be needed:
The z13 can withstand an 8.0 earthquake.
z Systems enjoy the highest standardized security certification (FIPS 140-2, highest level 4 of 4).
23 of the world's top 25 retailers use a mainframe.
92 of the top 100 banks are mainframe users.
All 10 of the top 10 insurers have commitments in mainframe technologies.
Around 80 percent of all corporate data is managed by mainframes.
The z13 can process 2.5 billion transactions daily (that's 100 [Cyber Mondays], as IBM's Mark Anzani, VP of z Systems Strategy, Resilience and Ecosystems, observed)."
... In fact, and notwithstanding perceptions to the contrary, the mainframe's center-stage position in large corporations around the world has not budged. That's the conclusion of an industry survey sponsored by Syncsort Inc. and conducted in 2015 by Enterprise Systems Media, a publisher of magazines for IT managers and technical professionals. Seven out of 10 respondents (IT planners, architects and managers at global enterprises with $1 billion or more in annual revenues) ranked the use of the mainframe for large-scale transaction processing as very important."
What would a comparable film depicting "A Day without a Mainframe" be like? I would imagine it somewhere between a disaster movie like  and an end-of-the-world zombie horror movie like [28 Days Later]. I would gladly take a million dollars to write the screenplay!
(FCC Disclosure: I work for IBM and am a filmmaker as well. Earlier in my career, I was chief architect of IBM's Data Facility Storage Management Subsystem (DFSMS) which manages around 80 percent of the world's corporate data. This blog post can be considered a "paid celebrity endorsement" for IBM's z13 System mainframes and DS8880 Disk Systems. I have personal experience with both and highly recommend them. I am neither a Mexican nor resident of California, but work regularly with both in my job responsibilities. Like Warren Buffett, I also own stock in both IBM and Berkshire Hathaway companies. I had no involvement in the making of any of the major motion pictures mentioned in this blog post, have no financial interest in their distribution, and have not been provided any compensation for mentioning them in this blog post. They are all great movies worth watching!)
What do you think the movie would be like? Enter your comments below!
(Actually, the [XIV Model 314] was announced on Nov 10, 2015 last year, but announcements made in November and December are often overlooked between distractions like holidays and year-end processing. Today's announcement was to eliminate the "not available in some countries" restriction. The last time I mentioned on this blog that a product was not available in some countries, I had tons of questions of "why". Hopefully, waiting until a product is available in all countries eliminates that concern.)
What does the XIV model 314 offer? IBM doubled the processors, up to 180 cores, and doubled the DRAM cache, up to 1440 GB. Both of these changes were done to improve the Real-time compression capability.
To reduce test effort cycle time, IBM simplified the configuration options:
Instead of ranging from 6 to 15 modules, the model 314 is limited to 9-15 modules.
The drive sizes are reduced to just 4TB and 6TB capacities.
If you want a Solid-State drive (SSD) for cache boost, only the 800GB option is available.
Through a combination of thin provisioning and compression, you can define up to 2 PB of soft capacity per rack.
The firmware v11.6.1 reduces the minimum volume size for compression from 103GB to 51GB. Firmware perpetually licensed for Spectrum Accelerate can be used with the XIV Model 314.
From left to right: Melinda Jensen, Bill Terry, Lee Olguin, Kris Keller, Tony Pearson, and Kristy Knight.
The storage, cloud and analytics team celebrated with cake and party hats. None of us "birthday boys" eat chocolate, so this year we chose a new flavor: Strawberry Cream! It was delicious.
It was a good time to reflect on our success and accomplishments. In 2015, I helped close over $270 million USD in revenues for IBM, meaning that I helped close over a million [per day on the job].
The IT industry went through a lot of changes also. Hewlett-Packard [split into two smaller pieces]. Dell started [EMC's fade to non-existence]. Cisco and IBM joined forces to create VersaStack, a converged system that combines the most popular x86 servers with the industry's best storage. Analysts recognized IBM's leadershp in today's [Cognitive Era].
My friends over at Appcessories sent me an awesome infographic on the Internet of Things. If you happen to receive any gifts this holiday related to any of these categories, mention them in the comments below!
The State of Internet of Things in 6 Visuals – By the team at Appcessories
Last Friday, I helped students learn about Science, Technology, Engineering and Math (STEM). This was the annual [2015 Arizona STEM Adventure] event in Tucson, Arizona. This year, Pima Community College Northwest Campus provided the venue.
The event hosted more than 900 students, ranging from fourth to eighth graders. Buses collected them from 31 schools across seven cities and towns in the Tucson area. Home-schooled, private-schooled and charter-schooled children participated as well.
As I arrived, students lined up to ride this "hover chair". A lawn-blower motor floated a chair attached to a platform. A blue tarp represented water. Volunteers would pull the hover chair across the tarp, giving the kids a fun ride. I wanted to ride it myself, but it was not engineered for my body weight!
Students chose among the most interesting of 50 exhibits. IBM led two of these exhibits.
First, we had the [Bike Wheel Gyroscope]. The students would stand on a rotating swivel platform, holding a spinning bicycle wheel. When the student tipped the wheel left or right, the students body would rotate on the platform!
Second, we had Share with Storyboarding. This is the one I volunteered for. IMHO, the best part of STEM is the Arts and Design aspect needed to make products usable. Perhaps we should rename STEM to STEAM to add "A" for Arts and Design.
We held six 30-minute sessions with each group of students. Our team lead, Brenton Elmore, IBM Design Principal, explained what storyboards are, and then gave the students five topics to choose from:
Adopting homeless pets
Improving communication with teachers
A short cartoon
An idea for a mobile phone app
An idea for a new video game
Children paired up in two-person teams based on their topic interest. Why teams? Many creative collaborations involve the strengths of different teammates. For example, an author and an illustrator work together to create a comics or children's book. Broadway musicals often have a writer and composer.
Each team spent 10 minutes to draw a six-panel storyboard on [Post-it notes]. These would be stuck to a single sheet of paper. The team then would write underneath each panel the narrative of what was occurring.
Brenton taped five or six of these to the wall to share with the rest of the class. Each team would then explain to the other students what they drew, and the narrative to go with it.
When there were an odd number of students, one of us volunteers paired up with a student. Shown here is Marilynn Franco, IBM Manager, helping young Bailey in explaining their storyboard. I helped young Lili with her storyboard about a new mobile phone app idea she had.
Well it's Tuesday again, and you know what that means? IBM Announcements!
(FCC Disclosure: This official launch also includes October 6 announcements. In any case, the usual disclaimer applies: I currently work for IBM, and this blog post can be considered a "paid celebrity endorsement" of the IBM products mentioned below.)
IBM announced various updates to its Spectrum Storage product line. Here is a quick recap.
IBM Spectrum Virtualize 7.6
Spectrum Virtualize is the new name of the "storage hypervisor" code that resides in IBM SAN Volume Controller (SVC) and Storwize family products. When you buy an SVC, you will license Spectrum Virtualize software on it. It is NOT available separately as software-only that you can install on any other hardware. There are three major improvements:
Software-based Data-at-Rest Encryption
Earlier this year, IBM delivered data-at-rest encryption for the Storwize V7000 and V7000 Unified. This week, IBM extends this support to other storage hypervisors.
Since this feature is based on the Intel processor that supports the Advanced Encryption Standard New Instructions (AES-NI), it applies only to the newer hardware: SAN Volume Controller 2145-DH8, the Storwize V7000 Gen2, FlashSystem V9000, and VersaStack converged systems that contain these. You can run Spectrum Virtualize v7.6 on older hardware models, but the encryption feature will be disabled.
Basically, by taking advantage of AES-NI commands, IBM can now offer data-at-rest encryption on any virtualized flash or disk arrays, eliminating the need for special "Self-Encrypting Drives", or SED.
The encryption keys are kept on USB memory sticks, that you can either leave in the machine, or stash away in some vault or safe somewhere.
The other improvement is distributed RAID. Distributed RAID has been hugely popular on IBM XIV products, and has since found its way into the DCS3700, DCS3860 and Elastic Storage Server models.
With this new enhancement, storage admins can select "Distributed RAID-5" or "Distributed RAID-6" as alternate choices to traditional RAID ranks.
Why use it? All the drives are now active, eliminating idle spare drives that do nothing collecting dust and cobwebs waiting for an opportunity to spin up, and when they finally are used for a rebuild become a terrible bottleneck. Since all drives are reading and writing, the rebuild rate is an order of magnitude (5 to 10x) faster!
For those clients nervous about large 8TB drives and the number of days it would take to perform a traditional RAID rebuild, this should calm all of your fears.
This is one of those line-items that we have told clients that it was "just around the corner" and "coming soon, watch this space", and finally it is available. For clients using Stretched Cluster or HyperSwap across two buildings, best practices suggests keeping the quorum disk in a third building. This often met having to dedicate a single 2U disk system in a closet somewhere, with expensive Fibre Channel cables connecting to the other two buildings.
To address this, IBM now allows the quorum disk to be based on Internet Protocol (the IP portion of TCP/IP), which can be any bare-metal or virtual machine that is LAN or WAN attached. The "quorum disk" is just a little Java program. This can run on any cloud service provider as well, such as IBM SoftLayer, that both buildings have connectivity.
A minor improvement worth mentioning is that the IBM "Comprestimator" tool that estimates the capacity savings of Real-time Compression is now integrated into Spectrum Virtualize v7.6 command line interface (CLI), allowing you to run the tool on demand, as needed, on any virtual volume.
IBM Spectrum Scale v4.2
IBM plans to offer all of its solutions in any of three flavors: software-only that you can deploy on your own server hardware, pre-built system appliances, and cloud services on IBM SoftLayer, IBM Cloud Managed Services or third-party cloud providers. Spectrum Scale is the software-only flavor, and Elastic Storage Server and Storwize V7000 Unified are pre-built systems based on that software.
File and Object access
IBM published a "Redbook" on how to implement OpenStack Swift and Amazon S3 interfaces to an existing Spectrum Scale deployment. IBM supported it, but it was basically Do-it-Yourself DIY implementation. This has now been resolved, with full integration of OpenStack Swift and Amazon S3 object-protocol interfaces.
(For those unfamiliar with "Object storage", think of it like valet parking for your data. Before working for IBM, I was previously employed as a valet attendant, so I feel qualified to make this analogy.
If you park your car in a 10-story high parking structure, you have to remember where you parked to go find the car again. With valet parking, you hand over the keys to the valet attendant, the car gets parked, and you get a claim stub that you then use to get your car back. In the meantime, you don't know where your car is parked, and you don't care either!
Storing files in volume-level or file-level storage is like that 10-story high parking structure. You have to remember where you put it, which LUN or which sub-directory. With object storage, the system provides a "claim stub" in the form of an Universal Record Identifier, or URI, and simple HTTP commands like GET and POST can be used to upload and download the content.)
Policy-driven Compression and Quality of Service (QoS)
If you want to differentiate the levels of service provided by files and objects stored in your infrastructure, look no further. Simple SQL-like language is used to set up policies that are invoked when needed.
Hadoop Connector for File and Objects
The IBM Hadoop Connector allows Hadoop and Spark analytics applications to treat Spectrum Scale as a 100 percent compatible alternative to Hadoop File Systems (HDFS). Previously, this was only available for files, but now it has been extended to include objects as well.
Advanced Graphical User Interface (GUI)
Based on the award-winning GUI that has been used for IBM XIV, SVC, Storwize and various other members of the IBM System Storage family, IBM announces an HTML5-based web-browser GUI for configuring and managing Spectrum Scale and Elastic Storage Server (ESS).
Storwize V7000 Unified
The "file modules" that run IBM Spectrum Scale will get updated to R1.6 level, which supports SMB 3.0 and NFS 4.0 protocols. SMB support will now include both internal and externally-virtualized storage. You will also be able to use Active File Management to migrate to other Spectrum Scale implementations.
IBM Spectrum Control
As the former chief architect of IBM Tivoli Storage Productivity Center v1, I have been a big fan of the advancements and evolution of Spectrum Control. IBM offers three levels. The first level is "Basic Edition", entitled at no additional charge for IBM storage hardware clients. The second level is "Standard Edition" which offers configuration, provisioning and performance monitoring. The third level is "Advanced Edition", which includes advanced storage analytics, file-level reporting, storage tiering and data placement optimization.
You can imagine my skepticism when I was told that Spectrum Control was going to be enhanced to support Spectrum Scale. What could it offer? IBM Spectrum Scale already has built-in storage tiering and data placement optimization!
It turns out that having effective "management tools" was the #1 reason clients have stated were needed to implement and deploy Spectrum Scale. Since 1998, back when it was called General Parallel File System, or GPFS, the target market was High Performance Computing (HPC) familiar with Command Line Interfaces (CLI).
But IBM was to broaden the reach of IBM Spectrum Scale, to financial services, health care and life sciences, government and education, and a variety of other industries. They won't tolerate being limited to CLI interfaces.
For clients with multiple Spectrum Scale clusters, Spectrum Control can offer the following:
Visibility across the capacity utilization (file systems, pools, file sets, quotas) and cluster health across all Spectrum Scale clusters in the data center
Ability to specify alerts which are applied across all Spectrum Scale clusters, for things like relative or absolute free space in a file system, or inodes used, nodes going down, etc.
Understand the cross-cluster relationships established by remote cluster mounts, and seamlessly navigate between them
If external SAN storage is used, Spectrum Control shows the correlation between Spectrum Scale Network Shared Disks (NSD) and their corresponding SAN volumes, again with the ability to navigate between them; also it can provide performance monitoring for the volumes backing the NSD
Ability to monitor file capacity usage in the context of applications, by adding Spectrum Scale "file set containers" to application groups defined in Spectrum Control
Compare file system activity across Spectrum Scale clusters, with the ability to drill into file system and node performance charts
Support for object storage on Spectrum Scale, determine which object-enabled clusters are closest to running out of free space
While the basic built-in GUI is great for smaller deployments, if you have a dozen or more Spectrum Scale clusters, or have Spectrum Scale clusters intermixed with traditional block-level and NAS storage devices, then Spectrum Control is for you!
It used to take weeks to deploy the original versions of Tivoli Storage Productivity Center, but now, Spectrum Control is now offered in the cloud, and you can deploy it in as little as 30 minutes.
Want to check it out? You can explore Spectrum Control Storage Insights cloud service as a [Live Demo], or [Start your free trial]! The reporting capabilities of Spectrum Scale are identical between the on-premise version of Spectrum Control, and this cloud service offering.
Here's a great quote from a leading IT industry analyst:
"In multi-petabyte, multivendor installations, overall storage costs of ownership for use of IBM Spectrum Storage solutions averaged 73 percent less than EMC, and 61 percent less than Hitachi equivalents" -- Brian Jeffery, Managing Director, International Technology Group, Naples, FL
As IBM continues its transition from a hardware-oriented company founded over a century ago, manufacturing meat scales and cheese slicers, to one more focused on higher value-add software and services, the Spectrum Storage software family will play a critical role of this transformation!
Continuing my coverage of the IBM Systems Technical University in Orlando, here are the sessions that I presented or attended on Days 4 (Thursday).
Technology Trends in IBM Storage
Jack Arnold, IBM Client Technical Architect, provide an entertaining session on various technology trends in the industry. For example, What is the fastest growing storage medium for 2015? Answer: [Vinyl LP] records, which have seen a resurgence recently, growing at over 40 percent!
IBM Spectrum Scale and Elastic Storage Server offerings
Tony Pearson provided an architectural overview of both Spectrum Scale software, as well as the Elastic Storage Server pre-built system appliance.
IBM Spectrum Scale for File and Object storage
Tony Pearson explained the differences between file and object-level storage, and how IBM Spectrum Scale can provide both access methods in a single infrastructure.
IBM Storage Integration with OpenStack
IBM Spectrum Virtualize IP Replication 101
Andrea Sipka, IBM Software Developer for SVC/Storwize Copy Services from the UK Hursley lab, presented the implementation details of IP-based replication using the built-in WAN Acceleration that IBM licensed from Bridgeworks SANslide.
Storage Meet the Experts
Mo McCullough hosted the last session of Thursday with a "Meet the Experts" Q&A panel. Tony Pearson, Brian Sherman, Clod Barrera, John Wilkinson, Mike Griese and Jim Blue were among the storage experts fielding questions. Tony Pearson provided a quick overview of the LTO-7 and TS4500 tape library announcements made earlier in the week.
Most IBM conferences are 4.5 days long, which means that there are typically two or three sessions on Friday morning. Unfortunately, the two sessions I was planning to attend on Friday were both cancelled, so Day 4 was the end of my week for this conference.
Continuing my coverage of the IBM Systems Technical University in Orlando, here are the sessions that I presented or attended on Day 3 (Wednesday).
What is Big Data? Architectures and Use Cases
Tony Pearson explained what Big Data analytics are, and IBM's various products to support this, incluidng BigInsights, BigSQL and Spectrum Scale with the Hadoop Connector.
Why use IBM Spectrum Virtualize for High Availability
John Wilkinson, IBM Storage Software Engineer from the UK Hursley lab, presented the latest enhancements to Spectrum Virtualize-based products, such as SVC and Storwize V7000, related to Stretch Cluster and HyperSwap functions for High Availability.
IBM Systems Hybrid Cloud Strategy, POV and Showcase
Dave Willoughby, IBM z System Hardware Architect for Systems Cloud Emerging Technologies, provided a high-level "Point-of-View" for Hybrid Cloud, and why IBM is focused on helping clients transition from traditional IT infrastructures.
Data Footprint Reduction - Understanding IBM Storage Efficiency Options
Tony Pearson presented an overview of Thin Provisioning, Space-efficient snapshots, Data deduplication and Real-time Compression features.
IBM Spectrum Virtualize - Understnding SVC, Storwize and FlashSystem V9000
Tony Pearson provide an overview of SAN Volume Controller, the Storwize family of products and FlashSystem V9000, all of which are based on Spectrum Virtualize software.
The day ended with a trip to Universal Studios. Dinner on the City Walk offered entertainment with Dueling Pianos. This was then followed by a trip to Hogsmeade, the Harry Potter themed portion of the resort.
Continuing my coverage of the IBM Systems Technical University in Orlando, here are the sessions that I presented or attended on Day 2 (Tuesday).
Andrew Greenfield, IBM Global XIV Storage and Networking Client Technical Specialist, presented IBM's future plans for XIV and FlashSystem products. This was a special NDA session.
Eric Aquaronne, IBM Systems and Cloud Business Development lead, explained what OpenStack was, and why IBM is so heavily invested in its success. OpenStack is cloud management software that can be used to manager both on-premise and off-premise environments, including computer, storage and networking resources.
Software Defined Storage - Why? What? How?
Tony Pearson presented an overview of Software Defined Environments and how storage fits into this.
Suspiciously, there was a lot of overlap with Brian Sherman's presentation on Day 1. As Charles Caleb Colton would say, "Imitation is the sincerest form of flattery."
Making Sense of IBM Cloud Offerings
Jay Kruemcke, IBM Cloud Program Executive Client Collaboration Market Management Offering Manager, gave a high-level overview of IBM's various Cloud offerings from SoftLayer to Managed Cloud Services.
The Pendulum Swings Back - Understanding Converged and Hyperconverged environments
Tony Pearson presented IBM's involvement with Converged Systems like VersaStack and Hyperconverged systems with Spectrum Accelerate and Spectrum Scale software.
Next Generation Storage Tiering: Less Management, Lower Cost and Increased Performance
Tony Pearson presented Easy Tier, Storage Analytics Engine in Spectrum Control Advanced Edition, and Spectrum Scale tiering across flash, disk and tape media.
The second day ended with a "Networking" Reception in the Solution Center, serving food and my favorite grape-flavored beverages.
Continuing my coverage of the IBM Systems Technical University in Orlando, here are the sessions that I presented or attended on Day 1 (Monday).
Storage Keynote Session
This was a three-part kick-off keynote session. Mo McCullough, IBM Systems Lab Services and Training, coordinated the storage track of this event and provided some details on how to use the website portal and smartphone app.
Clod Barrera, IBM Distinguished Engineer and Chief Technical Strategist for Storage, presented the future of the storage industry, including trends in storage media technologies, data plane and control plane level enhancements, and broader system-wide considerations.
Tony Pearson, IBM Master Inventor and Senior Software Engineer, wrapped up the session with an overview of IBM's Smarter Storage strategy.
IBM Software Defined Storage Overview, Concepts and IBM SDS Family
Brian Sherman, IBM Distinguished Engineer and Client Technical Specialist for Advanced Technical Skills in the Americas, provided an overview of Software Defined Environments and how storage fits in that view, especially IBM's Spectrum Storage family.
IBM Cloud Storage Options
Tony Pearson presented on IBM's various Cloud Storage options.
While my original focus was on-premise storage solutions for use by Data Centers and Cloud Service providers, there was a lot of interest in IBM's storage available from SoftLayer and other Cloud providers. During this week, IBM announced its acquisition of CleverSafe, which I had not incorporated into the deck.
What's New in IBM Spectrum Protect v7.1.3
Tricia Jiang, IBM Technical Enablement Specialist for IBM Spectrum Storage, presented the latest release of IBM Spectrum Protect. That's an inside joke--this is the first release, but since it was based on IBM Tivoli Storage Manager (TSM) v7.1.2, it was easier just to continue the same numbering scheme.
The main features of v7.1.3 is the new in-line dedupe capability, the new "deduplication containers" concept, and support for backing up to object storage either on-premise or in the cloud
IBM Spectrum Scale v4.1 Overview
Glen Corneau, IBM Client Technical Specialist for Power Systems, presented the latest features of IBM Spectrum Scale, formerly known as IBM General Parallel File System (GPFS). It was interesting to hear this from a Power Systems perspective, as IBM Spectrum Scale supports both AIX and Linux on POWER.
The day ended with a Welcome Reception at the IBM Solution Center that had various z System, Power System and System Storage solutions, as well as solutions from various IBM Business Partners and other third parties.
Oh my, it is Tuesday again, and you know what that means? IBM Announcements!
This week, IBM announced its latest storage arrays in its IBM System Storage DS8000 series: the DS8880 models. Similar to the "Business Class" vs. "Enterprise Class" distinctions of the DS8870, IBM announced two new models, the DS8884 and the DS8886.
All of the new DS8880 models are based on the latest IBM POWER8 processors, and are noticeably thinner! These are now standard 19-inch wide, fitting nicely into standard IBM racks alongside most other standard 19-inch rack equipment.
The DC-UPS that used to be on the side are now at the bottom of each frame, taking up 8U of space. The High Performance Flash Enclosures (HPFE) that formerly were stored vertically above the DC-UPS will be stored horizontally with the rest of the HDD and SSD drives.
The DS8884 will have 6-core controllers, up to 256 GB Cache, 64 ports that can negotiate between 16Gbps and 8Gbps, up to 240 drives in a single-rack configuration or 768 drives in a three-frame configuration, and up to 120 flash cards in HPFEs. The performance of this one is equal or better to existing DS8870 systems.
The DS8886 will have 8-core, 16-core and 24-core controllers, offering up to three times the performance as the previous DS8870 models, with up to 2 TB of Cache, 128 ports, up to 1,536 drives across five frames, and up to 240 flash cards in HPFEs.
Field model conversion from DS8870 to DS8886 is available for existing clients with DS8870 Enterprise Configurations. This will let clients move their existing HDD, SSD, HPFE and Host Adapters over to the new DS8880 models.
In previous DS8000 models, clients would have one Hardware Management Console (HMC) inside the array, and an optional second HMC workstation somewhere else for high availability. While the second one was optional, it was always considered best practice to have it for redundancy sake. In the new DS8880 models, you can have both HMC in the array, and the Keyboard/Video/Monitor (KVM) can select between the two.
The new I/O enclosure pairs are four times faster, supporting six Device Adapters and two HPFE connections over PCIe Gen 3 network, the fastest available in the industry.
Lastly, IBM simplified the licensing of software features into three bundles, based on TB total capacity of Fixed Block (FB) LUNs and Count-Key-Data (CKD) volumes:
Base function License: Logical Configuration support for FB, Operating Environment License, Thin Provisioning, Easy Tier® automated sub-volume tiering, and I/O Priority Manager.
Copy Services License: FlashCopy®, Metro Mirror, Global Mirror, Metro/Global Mirror, z/Global Mirror (XRC), z/Global Mirror Resync, and Multi-Target PPRC.
z-Synergy Service License: Parallel Access Volumes (PAV), HyperPAV, FICON® attachment, High performance FICON (zHPF), and IBM z/OS® Distributed Data Backup (zDDB).
IBM also provided a "Product preview", announcing plans for a third member of the DS8880 family in 2016 that will be flash-optimized to provide an all-flash, higher performance storage system model.
It's Tuesday, and you know what that means? IBM Announcements! This week I am in beautiful Orlando, Florida for the [IBM Systems Technical University] conference.
This week, IBM announced its latest tape offerings for the seventh generation of Linear Tape Open (LTO-7), providing huge gains in performance and capacity.
For capacity, the new LTO-7 cartridges can hold up to 6TB native capacity, or 15TB effective capacity with 2.5x compression that for typical data. That is 2.4x larger than the 2.5TB catridges available with LTO-6. Performance is also nearly doubled, with a native throughput of 315 MB/sec, or effective 780 MB/sec effective capacity with 2.5x compression. The LTO consortium, of which IBM is a founding member, has published the roadmap for LTO generations to LTO-8, LTO-9 and LTO-10.
IBM will offer both half-height and full-height LTO-7 tape drives. All the features you love from LTO-6 like WORM, partitioning and Encryption carry forward. These drives will be supported on a variety of distributed operating systems, including Linux on z System mainframes, and the IBM i platform on POWER Systems.
The Linear Tape File System (LTFS) can be used to treat LTO-7 cartridges in much the same way as Compact Discs or USB memory sticks, allowing one person to create conent on an LTO-7 tape cartridge, and pass that cartridge to the next employee, or to another company. LTFS is also the basis for IBM Spectrum Archive that allows tape data to be part of a global namespace with IBM Spectrum Scale.
LTO-7 will be supported on the TS2900 auto-loader, as well as all of IBM's tape libraries: TS3100, TS3200, TS3310, TS3500 and TS4500. You can connect up to 15 TS3500 tape libraries together with shuttle connectors, for a maximum capacity of 2,700 drives serving 300,000 cartridges, for a maximum capacity of 1.8 Exabytes of data in a single system environment.
In addition to LTO-7 support, the IBM TS4500 tape library was also enchanced. You can now grow it up to 18 frames, and have up to 128 drives serving 23,170 cartridges, for a maximum capacity of 139 PB of data. You can now also intermix LTO and 3592 frames in the same TS4500 tape library.
For comptability, LTO-7 drives can read existing LTO-5 and LTO-6 tape cartridges, and can write to LTO-6 media, to help clients with transition.
Amy Hirst, IBM Director, z Systems, Power, & Storage Technical Training, kicked off the general session.
Dr. Seshadri "Sesha" Subbanna, IBM Corporate Innovation and Technology Evaluation, asked the audience what capability is needed to drive business growth. A recent poll indicated that the ability for businesses to innovate was the number one response.
The IT industry has had its own version of growth. Consider the Apollo 11 [Guidance Computer] used to land a man on the moon had just 4KB or RAM, and 36KB or ROM. A typical smartphone has 62,000,000 times as much.
The Appollo missions led and motivated the Integrated-Circuit technology, but soon, maybe in the next 10 years, Dr. Subbanna feels that Silicon may run its course. Today, both POWER8 and z13 servers are based on 22nm. IBM has projected possible reductions to 17nm, 13nm, 10nm, and finally 7nm. That's it, smaller than 7nm may not be possible without hitting atomic issues.
The City of Rio de Janeiro, Brazil is a good example. In 2010, heavy rains resulted in flooding and landslides that killed over 110 residents. To prevent such high death rates in the future, IBM helped the city government predictive analytics and forecasting that allows "rain simulations" to see how well the city can handle different situations.
IBM is already looking for a more holistic view of systems, and new technologies like cognitive computing. New 3D technology allows various chip technologies to be stacked as layers on a single chip. For example, you could have computer on the bottom layer, flash non-volatile storage in middle layers, and networking at top layer. Connecting the layers is merely a matter of drilling holds and filling them with metal.
The idea that compute is the center of the universe, with a mainframe server surrounded by input and output "peripheral" storage devices, is giving way to a more storage-centric model, where central storage repositories (or data lakes) are accessed by "peripheral" smartphones, tablets and variety of servers. For example, the IBM DB2 Accerlation Appliance acts as a storage-centric model that IBM z System mainframes can connect to, send data in, process complex database queries, and get the results 2000x faster.
In another client example, IBM helped a bank in China to determine optimal placement of bank branches, based on public information of average salary levels of each neighborhood.
CPU processors are also getting help from co-processor accelerators like GPU (Graphical Processing Unit) and FPGA (Field Programmable Gate Arrays). Comparing a single IBM POWER8 server that is CAPI-attached to an IBM FlashSystem to a stack of x86 servers with internal SSD, the POWER8 solution connsumes 12x less rackspace, consumes 12x less electricity, and reduces per-user costs from $24/user for x86 down to $7.50/user on POWER8.
While social media, mobile phones and the Internet of Things (IoT) generate a lot data. If you then factor the "context multiplier effect" of all the links, connections and cross-references, you quickly see that data is growing at incredible rates.
Another issue is the difficulty to identify application inter-dependencies. Forecasting disruptive anamolies can be quite difficult. In one example, adminstrators received warning messages 65 minutes before a major outage, but they did not respond in time because they were unable to understand the full implications.
Cognitive computing is different than the tabulating and programming paradigms of prior decades. It is focused on Natural Language Processing, citing evidence to base responsed, and the ability to learn and improve based on learning from experience. The IBM Watson group is working with Memorial Sloane Kettering to help oncology doctors with cancer patients.
In an interesting demo, IBM Watson computer analyzed thousands of "TED Talk" videos, and was able to respond to search queries by playing a 30-second video clip that most closely address the search topic.
Cognitive computing is also looking at "Neuro-Synaptic" chips that work very much like the neurons and synapses in the brain. I have seen some of this work already at the IBM Almaden Research Center in California.
The general session ended with a Q&A panel with Dr. Subbanna, Frank De Gilio, and Bill Starke.