This blog is for the open exchange of ideas relating to IBM Systems, storage and storage networking hardware, software and services.
(Short URL for this blog: ibm.co/Pearson )
Tony Pearson is a Master Inventor, Senior IT Architect and Event Content Manager for [IBM Systems for IBM Systems Technical University] events. With over 30 years with IBM Systems, Tony is frequent traveler, speaking to clients at events throughout the world.
Lloyd Dean is an IBM Senior Certified Executive IT Architect in Infrastructure Architecture. Lloyd has held numerous senior technical roles at IBM during his 19 plus years at IBM. Lloyd most recently has been leading efforts across the Communication/CSI Market as a senior Storage Solution Architect/CTS covering the Kansas City territory. In prior years Lloyd supported the industry accounts as a Storage Solution architect and prior to that as a Storage Software Solutions specialist during his time in the ATS organization.
Lloyd currently supports North America storage sales teams in his Storage Software Solution Architecture SME role in the Washington Systems Center team. His current focus is with IBM Cloud Private and he will be delivering and supporting sessions at Think2019, and Storage Technical University on the Value of IBM storage in this high value IBM solution a part of the IBM Cloud strategy. Lloyd maintains a Subject Matter Expert status across the IBM Spectrum Storage Software solutions. You can follow Lloyd on Twitter @ldean0558 and LinkedIn Lloyd Dean.
Tony Pearson's books are available on Lulu.com! Order your copies today!
Safe Harbor Statement: The information on IBM products is intended to outline IBM's general product direction and it should not be relied on in making a purchasing decision. The information on the new products is for informational purposes only and may not be incorporated into any contract. The information on IBM products is not a commitment, promise, or legal obligation to deliver any material, code, or functionality. The development, release, and timing of any features or functionality described for IBM products remains at IBM's sole discretion.
Tony Pearson is a an active participant in local, regional, and industry-specific interests, and does not receive any special payments to mention them on this blog.
Tony Pearson receives part of the revenue proceeds from sales of books he has authored listed in the side panel.
Tony Pearson is not a medical doctor, and this blog does not reference any IBM product or service that is intended for use in the diagnosis, treatment, cure, prevention or monitoring of a disease or medical condition, unless otherwise specified on individual posts.
The developerWorks Connections Platform is now in read-only mode and content is only available for viewing. No new wiki pages, posts, or messages may be added. Please see our FAQ for more information. The developerWorks Connections platform will officially shut down on March 31, 2020 and content will no longer be available. More details available on our FAQ. (Read in Japanese.)
Demonstrate that IBM technologies in areas like Artificial Intelligence (AI), Blockchain, Cloud, and the Internet of Things (IoT) are relevant in solving the world's biggest challenges
Encourage developers to contribute their time and talent to open source projects that benefit the greater good
Generate fresh ideas on how to tackle age-old problems that plague society
Each year will have a different focus. This year, the focus is in preventing, responding to and recovering from natural disasters, especially important with 2017 ranked as one of the worst years on record for catastrophic events, including fires, floods, earthquakes and storms.
Call for Code invites developers to create new applications to help communities and people better prepare for natural disasters. For example, developers may create an app that uses weather data and supply chain information to alert pharmacies to increase supplies of medicine, bottled water and other items based on predicted weather-related disruption. Or it could be an app that predicts when and where the disaster will be most severe, so emergency crews can be dispatched ahead of time in proper numbers to treat those in need.
Can't think of any ideas for an app? Here are some TED videos that might inspire you:
IBM's $30 million USD investment over five years will fund access to developer tools, technologies, free code and training with experts. To raise awareness and interest in Call for Code, IBM is coordinating interactive educational events, hackathons and community support for developers around the world in more than 50 cities, including Amsterdam, Bengaluru, Berlin, Delhi, Dubai, London, New York, San Francisco, Sao Paulo and Tel Aviv.
(My earliest memory of using a contest for fresh ideas was back in 1975, after the city of Tucson purchased the Tucson Rapid Transit Company. Rather than hiring an expensive marketing agency to run focus groups or surveys, the City of Tucson published in the local newspaper a "Name that Bus" contest. The winning entry was [Sun Tran], submitted by 25-year-old college student [Benjamin Rios]. He won the grand prize: $150 portable television!)
The winning Call for Cloud team will receive a financial prize and access to long-term support to help move their idea from prototype to real-world application.
Developers can register today at the [Callforcode.org] website. Projects can be submitted by individuals – or teams of up to five people – between June 18, 2018 and August 31, 2018. If you would like me on your team, as an honorary member, technical adviser or mentor, please let me know!
Thirty semi-finalists will be selected in September. A prominent jury, including some of the most iconic technologists in the world, will choose the winning solution from three finalists. The winner will be announced in October 2018 during a live-streamed concert and award event coordinated by David Clark Cause.
Additional details, a full schedule of in-person and virtual events, and training and enablement for Call for Code are available at [www.developer.ibm.com/callforcode] website.
This week, IBM clients, Business Partners and executives get together for the new IBM [Think 2018] conference. This is a combination of last year's three events: Edge, InterConnect, and World of Watson (WoW).
(The theme this week is "Putting smart to work." Some might feel that this is a grammatically-incorrect use of the adjective [smart], referring to having quick-witted intelligence or being neat and well-dressed. Many words in the English language have multiple meanings and uses. The word smart is also a noun, referring to either business acumen, technical skills, or "a sharp stinging pain")
The keynote session today was "Science Slam: Unveiling 5 Breakthrough Technologies That Will Change the World!" by Arvind Krishna, IBM Research Director. IBM has over 3,000 researchers, in 12 labs, across six continents.
This talk was based on IBM's annual five-in-five, five predictions that might change the world in the next five years. For amusement, read my 10-year-old blog post [Five in five for 2008], including predictions for smart thermostats that can be controlled remotely, and self-driving cars.
("Science Slam" is IBM Research version of [Pecha Kucha], but instead of art students having 20 minutes to show 20 PowerPoint slides, each IBM research scientist has 5-7 minutes to explain the research project they are exploring. These are done both internally, as well as to audiences outside the company.)
Jamie Garcia served as emcee, introducing each of the five experts. Each spent 5-7 minutes, Science Slam style, on what projects they were working on.
1. Crypto-anchors and blockchain technology
‘Everything you don’t understand about money
combined with everything you don’t understand
about computers’ [25-minute video]
Andreas Kind presented first. Blockchain is not just a provenance system that enables Bitcoin and other cryptocurrencies, it can be used for other goods.
(The best layman explanation of blockchain and cryptocurrencies I saw was John Oliver's humorous take on his HBO show [Last Week Tonight]!)
Counterfeit goods, from cinnamon to footwear, to medicine and automotive parts, is estimated over $1.8 trillion US dollars. IBM is working on how to use blockchain for other things, such as to restore trust into global supply chain. IBM hopes to reduce the number of counterfeit goods in half or more.
Andreas explained tamper-proof technologies called "crypto-anchors" -- from indelible ink on pharmaceuticals to computers smaller than a grain of salt -- that can be used to track products as they travel from one country to the next.
2. Lattice Cryptography and Fully Homomorphic Encryption
Cecilia Boschini from IBM Zurich presented next. As quantum computers get more powerful, the basic math involving prime numbers that most current encryption models are based on become vulnerable.
(Don't worry, she assured the audience, hackers would need a 1000-Qubit quantum computer to break today's encryption codes, which don't exist yet!)
What we need are post-quantum or quantum-resistant mathematical models. Lattice Cryptography aims to use more difficult math equations to make it more difficult for hackers to break the code, even when armed with quantum computers.
Another challenge with existing encrypted data is that we must decrypt the data to perform computations on it. Fully Homomorphic Encryption, or [FHE] for short, allows computations to be done in its encrypted state. For example, if I had a list of names with credit card or social security numbers encrypted, I could sort this list alphabetically without decrypting any of the data.
3. AI-enabled robotic microscopes to monitor ocean water
Tom Zimmerman is known as IBM Almaden's [McGyver], able to use common technologies in new and innovative ways.
By 2025, over half of the world's population will be living in water-stressed locations. IBM is working on robotic microscopes that can be deployed across the oceans, connected to the Cloud, monitoring the state of plankton.
Why plankton? Plankton produces two-thirds of all oxygen we breathe, and serves as the "baby food" for all oceanic species. Tom has re-programmed "face recognition" in smartphone cameras to recognize plankton, identifying what they are doing and eating.
Monitoring plankton provides an "early warning system", the proverbial [canaries in the coal mine] for impending water problems.
4. Eliminating Bias from Artificial Intelligence (AI)
Information overload! Overwhelmed by too much, our brains sort it out by either looking only for differences, or focusing on what we are already familiar with that confirm our beliefs.
Not enough meaning. Lacking complete information, our brains fill the gaps and connect the dots to find patterns that aren't patterns at all. Racism, prejudice, and stereotypes are examples of this.
The need to act fast! Survival in some cases demands acting fast, to avoid being eaten by an animal, for example. Unfortunately, our brains favor the quick and simple, over the more important but often delayed, distant or complicated response.
What should we remember? We decide what to remember, and what to forget. Our brains often favor generalities over specifics, as they take up less space. The details we remember when we experience it, or often edited or reinforced after the fact.
IBM is collaborating with the Massachusetts Institute of Technology [MIT] to reduce bias in Artificial Intelligence by rating different AI models on fairness.
The AI models that will win in the future are those where the biases are tamed or eliminated altogether.
5. Quantum Computing
Talia Gershon was the last speaker.
Many problems become exponentially more difficult to solve with classical computers. For example, simulating protein molecular bonding gets more difficult the larger the molecules are, because you have more electron interactions.
Quantum Computers run at a temperature of 15 millikelvin (mK), which is 460 degrees below zero. The computation unit is called a [Qubit], and a 5-Qubit quantum computer can solve problems that your laptop can solve classically. IBM now has "IBM Q" with 50-Qubit computers available.
The IT industry is still in the early stages, but IBM Quantum Information Software development kit (QISkit) allows programmers to experiment and develop algorithms for this new computational model.
Over the next five years, IBM predicts that Quantum Computing will transition from the lab, to the mainstream, to solve problems that were previously too difficult or time-consuming to solve.
(Actually, the [XIV Model 314] was announced on Nov 10, 2015 last year, but announcements made in November and December are often overlooked between distractions like holidays and year-end processing. Today's announcement was to eliminate the "not available in some countries" restriction. The last time I mentioned on this blog that a product was not available in some countries, I had tons of questions of "why". Hopefully, waiting until a product is available in all countries eliminates that concern.)
What does the XIV model 314 offer? IBM doubled the processors, up to 180 cores, and doubled the DRAM cache, up to 1440 GB. Both of these changes were done to improve the Real-time compression capability.
To reduce test effort cycle time, IBM simplified the configuration options:
Instead of ranging from 6 to 15 modules, the model 314 is limited to 9-15 modules.
The drive sizes are reduced to just 4TB and 6TB capacities.
If you want a Solid-State drive (SSD) for cache boost, only the 800GB option is available.
Through a combination of thin provisioning and compression, you can define up to 2 PB of soft capacity per rack.
The firmware v11.6.1 reduces the minimum volume size for compression from 103GB to 51GB. Firmware perpetually licensed for Spectrum Accelerate can be used with the XIV Model 314.
Well, it's Tuesday again, and you know what that means? IBM Announcements!
Here is a quick recap of the October 9, 2018 announcements this week.
IBM Elastic Storage Server V5.3.2
The new IBM Elastic Storage Server v5.3.2 offers support for new drawers, non-disruptive upgrades of older models, and an optional 100GbE switch.
When the ESS was first announced, we had GSx models and GLx models, where x represented the number of storage drawers. The "S" stood for small 2U-24 drive drawers, so for example the GS4 had two Power8 servers combined with four 2U-size flash SSD drawers. The "L" stood for large 4U-60 drive nearline HDD drawers.
The second generation models append "S" for Second, so we had GS4S and GL6S. The large models changed to larger 5U-84 drive drawers. As with the previous "L" models, two slots per system contain Solid State Drives for internal use and caching, leaving the rest for slower spinning HDD disk.
Before this week, to upgrade from one model to another meant moving the data off, installing and configuring the additional drawers, and then move the data back. With today's announcements, you can now non-disruptively upgrade GS1S to GS2S to GS4S models, and GL1S to GL2S to GL4S to GL6S.
While you can federate as many GS and GL models together, that may mean having to spend more for Power8 servers than you are comfortable with, so IBM added "GHxy" Hybrid models, with x 2U-24 drive drawers, and y 5U-84 drive drawers. Initial models included the GH14 and GH24, which had one or two flash drawers, and four large drawers. This week, IBM announced a new GH12 model. The SSD flash in the 2U drawer can be 3.84TB or 15.36TB, and the nearline drives in the 5U drawers can be 4TB, 8TB or 10TB capacities.
What did IBM call the third generation GL models? Instead of using "T" which is both the next letter in the alphabet after "S", and the initial letter of the word "third", IBM instead decided to use "C" to designate CORAL project, the Collaboration of Oakridge, Argonne, and Lawrence Livermore national labs. Since the change applied only to the GL models, not the GS models, this makes sense.
To meet the requirements to build the world's fastest supercomputer for the CORAL project, IBM created a modified Elastic Storage Server model with 4U drawers that contained 106 drives. Now, these are available to the general public! IBM announced GL1C, GL2C, GL4C and GL6C models. In these, there are 2 SSD drives, and the rest are 10TB nearline drives.
The new optional 100GbE switch has 32 ports with a total of 6.4 Tbps. These can support 10, 40, 50 and 100GbE data rates, with 300 nsec latency for 100 GbE port to port
Spectrum Scale is licensed two ways: Standard Edition based on the number of sockets, with different prices for NSD servers, FPO servers and NSD clients; and the "Data Management" edition which offered advanced features, and was based on capacity of NSD, independent of the number of servers and clients attached.
Clients liked the capacity-based license model, but did not necessarily need the advanced features. In response, IBM now offers the "Data Access" edition, which offers the same features and functions of Standard Edition, but with capacity-based licensing.
For ESS models, you can chose to license by disk as before, or by capacity in combination with Spectrum Scale capacity-based deployments.
Hortonworks Data Platform v3.0.1 has followed suit. With the merger between Hortonworks and Cloudera, Hortonworks now offers capacity-based licensing for shared storage, like the IBM Elastic Storage Server.
IBM FlashSystem A9000/A9000R software version 12.3
There are three enhancements in this release: Three-site replication, a new model of A9000R, and raising a previous pool size limit.
For three-site replication, you can now combine HyperSwap which maintains two identical copies at distance, with a third asynchronous mirroring. The first two are typically within 100 km, but the third copy can be a much greater distance, across the continent if you like.
The A9000 "Pod" had three x86-based controller and one FlashCore drawer. The A9000R "Rack" had four, six or eight x86-based controllers and two, three or four FlashCore drawers, respectively, as well as a Power Distribution Unit (PDU) and pair of InfiniBand switches to connect everything together. The new "Grid Starter" model is very much like the "Pod" with three controllers and one FlashCore drawer, but adds the PDU and IB switches. The idea is that you can start with a "Grid Starter", then later upgrade to the larger A9000R models as you grow.
Back in XIV days, the architectural limit per pool of 1PB was plenty big. But with the new capacities on the A9000 and A9000R, the 1PB limit was starting to draw complaints. This limit was lifted, so that now a single pool can be made with the entire capacity of the box.
In the mainframe world, IBM Geographically Dispersed Parallel Sysplex, now just GDPS, provide the highest BC-7 business continuity tier, providing end-to-end coordination with servers, networks and storage devices. For IBM Power Systems, similar BC-7 support is provided by IBM Geographically Dispersed Resiliency.
In this week's announcement, IBM Geographically Dispersed Resiliency (GDR) for Power Systems has been renamed and now offered in two editions: VM Recovery Manager HA and VM Recovery Manager DR. The "HA" edition provides high availability using Power Systems Live Partition Mobility for AIX, IBM i and Linux operating systems.
The "DR" edition provides both High Availability and Disaster Recovery capabilities, supporting mirrored storage systems like IBM DS8000, SAN Volume Controller, FlashSystem 9100 and V9000, and Storwize systems, as well as competitive storage from Dell EMC and Hitachi.
Next week, I will be in Hollywood, Florida for IBM Technical University (Oct 15-19), and then Rome for the IBM Technical University (Oct 22-26). I will be covering many of these announcements above, and more!
This week, May 14-18, is Business Continuity Awareness Week!
This worldwide event, sponsored by the [Business Continuity Institute], promotes education and awareness designed to increase our understanding of business continuity, teach clients on ways to understand and manage IT and business risks, and introduce new techniques and technologies designed to minimize and even to eliminate business and personal disruption.
IBM is actively involved. Monday starts off with opening statements by Andrea Sayles, IBM General Manager of Resiliency Services, and Michael Puldy, IBM Director of Global Business Continuity Management.
The event offers a variety of online webinars, as well as a wealth of educational resources.
Last week, I was in São Paulo, Brazil for IBM Systems Technical University.
Instead of separate physical rooms for each breakout session, this event had "virtual rooms". One speaker called it the "Software Defined Stage". Basically, there were five "rooms" in the main ballroom, and another eight rooms in a second ballroom.
Rather than blasting out each speaker's voice over loudspeakers, each speaker spoke softly into a headset microphone. All attendees wore headsets. Rooms 1 through 4 offered real-time translation, so attendees could chose to hear in English or Brazilian Portuguese.
In the other 13 "rooms", local speakers spoke in Brazilian Portuguese, but you still had to wear headsets to avoid speaking louder than the speaker next to you. For many of these, the charts were written in English.
My translators, Luciana and Marilia, explained to me the advantage of this approach. When speakers use English language, those who prefer must hear the real-time translation wore the "headphone of shame" which advertised to all others that an attendee's English proficiency was poor.
Sometimes, those who did not understand English well would not wear their headsets, nod or laugh with other attendees, but fail to understand the message. By forcing everyone to wear headsets, there is no stigma associated, and everyone can discreetly select the language they prefer to listen in.
Here is my recap for the breakout sessions on Day 2:
In this presentation, I gave an overview of interest in Cloud technologies, including OpenStack and RESTful APIs to manage server and storage resources. I then covered IBM Hybrid Cloud Storage configurations in five categories:
Cold storage for data infrequently accessed
Backup and Snapshot storage
Disaster Recovery storage
Daily Operations and Reporting
Special thanks to Chris Vollmar and Brian Sherman for their help in preparing this presentation.
Data Optimization: How to verify your data is being used efficiently
It is hard to believe that it was over 15 years ago that I was the chief architect for the software we now call IBM Spectrum Control. There are a variety of editions and bundles for this product, but my focus on this talk was on the advanced storage analytics found in IBM Virtual Storage Center and IBM Spectrum Control Advanced Edition.
I covered three use cases:
What storage tier to put your workload in, and how to move existing data into a faster or slower tier to meet business requirements and IT budgets.
For steady state environments, how to re-balance storage pools within a single tier to keep things even for optimal performance.
When it is time to decommission storage, how to transform volumes from one storage pool to another without downtime or outages.
Special thanks to Bryan Odom for his help in preparing this presentation.
IBM Hyperconverged Systems powered by Nutanix: Technical Overview
Ricardo Matinata, IBM Senior Technical Staff Member for Linux, KVM and Cloud on POWER, presented the latest IBM CS models for POWER systems that are pre-installed with Nutanix software running their Acropolis Hypervisor (AHV) to run Linux on POWER application virtual machines.
Managing Risks with Thin Provisioning, Compression, and Data Deduplication
This session had four parts. First, an overview of "Data Footprint Reduction" technologies, like compression, data deduplication, space-efficient snapshots and thin provisioning.
Second, a look at how these technologies can get storage administrators in trouble. Much like airlines selling more tickets than seats on the airplane, storage administrators may over-provision based on data reduction estimates, and then suddenly run out of storage capacity.
Third, an overview of IBM FlashSystem A9000 and A9000R products, often referred to as "A9000/R" to cover both as a family. These models offer data footprint reduction for all data.
Finally, I explain how the Hyper-Scale Manager GUI can help with reporting and analytics to avoid these risks. This GUI is available for the FlashSystem A9000/R, as well as XIV Gen3 and Spectrum Accelerate software clusters.
Special thanks to Rivka Matosevich for her help in preparing this presentation.
The Right Flash for the Right Workload
Fabiano Gomes, IBM Client Technical Specialist, presented IBM's portfolio of All-Flash Arrays, from FlashSystem and DS8000F to Elastic Storage Server and Storwize V7000F and V5000F models. Each of these have their own characteristics, which might favor one over the others for particular workloads and use cases.
The day was capped off with a nice evening reception at the pool bar. Bartenders were serving Caiparinhas, a Brazilian cocktail traditionally made sugar cane liquor, sugar and lime, but in this case offered in other flavors, such as pineapple or passion fruit.
Last month, I presented at the "IBM Technical University" event in beautiful Atibaia, Brazil. Here is my recap of the event.
Marcelo Porto, IBM General Manager for Brazil and Client Unit Executive for Retail
What a great way to start a conference! Marcelo asked if everyone was comfortable? Everyone cheered in the affirmative.
He then said "Well, not for long. We will take you out of your comfort zone! You will disrupt yourself, and disrupt your companies. You will learn about new technologies and solutions that will make you very uncomfortable."
He explained how everything is virtual, specifically the three companies Airbnb, Waze, Uber. All of these three have new transformational business models, and he suggested all companies should follow suit.
He then said people need to be focused on four things:
Adopting an "agile attitude"
Act like you own the company
Don't cling to the past
Have the courage to re-invent yourself and your company
Frank Koja, IBM Vice President for Sales, Enterprise Systems Hardware
(Managers and business leaders could probably raise this percentage considerably if they talked to their employees before making decisions, but that's another blog post!)
Frank showed a video of an IBM client, Plenty of Fish (POF). This is a worldwide dating site with three million POF members in Brazil. They now process over 30,000 requests and/or messages per minute. FlashSystem connected to 30 servers makes that possible.
OpenPower consortium started with just 5 companies in 2014 for technology collaboration. Today, 250 members across 26 countries in six continents collaborate to make POWER technology as ubiquitous a commodity as Intel x86.
Frank then switched to "Business models" innovation. Out of the audience of about 800 people, only 10 raised their hands that have heard of Blockchain (he asked IBMers not to raise their hands, as all IBMers have heard of Blockchain!).
Frank feels that Blockchain is the most disruptive innovation since Internet banking. Blockchain affects supply chain, finance, insurance, shipping logistics, customs inspections, and government registrations.
A video showed a woman from Everledger, which uses Blockchain for shipping diamonds. IBM offers Blockchain on LinuxOne mainframe servers.
Hybrid Cloud is point of no return, including Local, Dedicated and Public clouds. Frank feels we need to cloudify all business processes.
Mauro Angelo, IBM Enterprise Strategy & Industry Solutions Director
Mauro explained that ideas are turned into inventions, and inventions are put to good use to bring forth innovations.
If your business is not cognitive you are a full era behind. Machine learning is not knew. IBM DeepBlue beat Grandmaster in Chess tournament back in 1997.
Mauro then focused on eight specific trends:
Systems of Engagement (SoE)
This is the combination of Mobile applications and Social business. IBM invited the first smartphone, the Simon, back in 1994. Apple's iPhone came later in 2007. Pokemon Go is example of augmented reality.
Cloud offers new service and location models. IBM [SoftLayer], [Bluemix], and [Kenexa] are a few examples.
There have been a lot of enhancements in this space, including Natural Language Processing (NLP), visual recognition, even smell recognition. Cognitive solutions can also identify the appropriate context, such as GPS location. And Cognitive solutions can interact with users to ask for clarifications. It can process "Big Data", the collection of non-structured data that normal Relational Database Management Systems (RDBMS) do not touch. Finally, they can learn, something often referred to as "Machine Learning".
In 2011, IBM Watson beat two humans at the TV show game Jeopardy! Today, [Dino, a toy from CogniToys] provides Watson-like capabilities to children.
Mauro got one for his daughter. She naturally interacts with toy. "How much does an elephant weigh?" she asks. "It depends on the elephant, but a fully grown elephant weighs more than 2,000 kilos" it responds. That's cool.
Wearables like Fitbit can track blood pressure, minutes of exercise, total steps walked. IBM helped Under Armour company develop an app in this space.
Eliminates middlemen or trusted third party (TTP). The hotel chain, Hilton, is testing out a robot called Pepper, which can use Blockchain to book tennis courts.
These are technologies thinner than a strand of hair, measured in nanometers. The focus is to develop stronger, lighter materials, and macromolecules for life sciences for medicine delivery.
Mass customization meets personalization and fast design prototypes. This is not just limited to plastic, but also metal, paper, wood, biomaterials, ceramics, food, and even cement.
Cement? That's right. A Chinese company prints houses using a cement 3D printer. In a country of over one billion people, this company has figured out how to build houses without human laborers.
Internet of Things (IoT)
Olli, a 12-person self-driving bus, is the brainchild of Local Motors. They are testing it out in National Harbor, and hope to roll it out to cities like Copenhagen, Miami, and Las Vegas.
Luis Liguori, IBM Distinguished Engineer and CTO for IBM Brazil
What does IBM mean by "Digital transformation?" What separates success from failure? Developed countries from less developed countries?
Is it culture? Whether people focus on the long term, or just the short term? Does the culture encourage you to foresee the future, and adapt accordingly? Does the culture encourage you to be brave and bold? Do you hide behind Business case return on investments (ROI)? Does your culture consider conflict to be good or bad? The answer: Good!
Does your company have a purpose? When humans no longer serve purpose, they die. The same is true for companies. He said the secret to success is the four "R's" -- Relevant, Resources, Reputation and Rigor.
For example, in 1996, the Kodak was ranked the 4th largest, it filed bankruptcy in 2012 because it was no longer relevant.
Consider Samsung. Samsung has lost its reputation with the latest "Samsung Galaxy Note7" fiasco of exploding batteries!
Airbnb is an example of Digital Transformation. Who knew that there were lots of people who wanted to rent out their bedrooms and bathrooms to strangers!
Luis feels that successful companies are either born digital, or transforming to digital. Industries are merging. Lines are blurring between industries. The latest acquisition between AT&T and Time Warner is an example.
Cognitive brings intelligence to decision making. For example, Watson health has been put to task to focus on Leukemia. In one case, Watson was able to [pinpoint a rare form of Leukemia] that had misdiagnosed and being treated incorrectly with little effect.
Why cognitive? Because human beings cannot read or remember as well as computers. There are thousands of peer-reviewed articles published every day. People are afraid to act to avoid mistakes. Computers are fearless.
Did you know that Brazil celebrates "Black Friday"? There is no "Thanksgiving" in Brazil, but retailers liked the idea of having people stand outside in the middle of the night to start their Christmas shopping! A few years ago, there were [a few problems], but in most recent years, it has shown to help [boost retail sales.] Based on these initial purchases, Watson can be used to help drive the rest of the Christmas retail season.
Watson can analyze personality based on social media writings. The world will be taken over by digital natives. The last century was focused inward, or "ego-centric", but in this 21st century, we will be focused outward, towards a complete "ecosystem".
Who are your competitors? Are they the companies that make products and services similar to yours? No! They are the companies that are competing for your customer's time and attention.
While I speak English and Spanish fluently, my Brazilian Portuguese is terribly rusty. We had several rooms with a pair of real-time translators. I presented the following:
Software Defined Storage -- Why? What? How?
The Pendulum Swings Back -- Understanding Converged and Hyperconverged Environments
IBM Spectrum Scale for File and Object Storage
IBM Storage integration with OpenStack
Introduction to IBM Cloud Object Storage System and its Applications (powered by Cleversafe)
IBM's Cloud Storage Options
All of my sessions were well received, and well attended!
Photo by Dominique Salomon,
IBM Certified IT Specialist
On Wednesday night, we had a nice pool-side reception. Beers, Caiparinhas, and Caiparoskies. Caiparinhas combine a sugarcane juice-based distilled alcohol called cachaça with muddled limes and added sugar. Caiparoskies combined vodka with muddled kiwi fruit.
(Many of the IBMers from United States skipped this event to get dinner early, so they could then come back in time to watch the third and final US Presidential Debate. Because of the time zone changes, this didn't start til 11:00pm, so they could have easily attended the event and had dinner, with plenty of time to spare!)
There was also a live band! This three part band had two guitarists and one lead singer. The lead singer also did maracas and drums while singing. They covered both English and Portuguese language songs.
Rodrigo Giaffredo, IBM Engagement Catalyst
Rodrigo gave the closing session. Wearing jeans and sneakers, he reminded me of the casual storytelling style of Jeff Jonas. He organized his stories around four points:
Consider the battle between Twitter vs. Pownce in 2007. Twitter won because it offered better ways to limit what you read, or who you communicate to, through methods like Hashtags, groups, etc.
Henry Ford disrupted transportation. He realized that Time and space is money. However, as he famously said "If I asked people what they wanted, they would have said faster horses!"
Today the challenge is processing data faster. The company that is able to process faster has economic advantage.
Strong ideas focus on user needs. Weak ideas are tactical and features. Consider the [Hippo Roller]. For centuries, African women and children carried water from far away wells either on their hands on or their heads. Much of it would fall out during the long walks. The Hippo Roller holds 90 liters (about 24 gallons) and rolls easily over rough terrain.
Rodrigo showed an graph. On the y-axis was "Importance" and the x-axis "Feasibility". Solutions in the upper right corner are obvious choices. Solutions in the upper left, important but not very feasible, are considered "big bets". Solutions in the lower right, feasible but not very important, he labeled "amenities".
Most designers, architects and developers know that the later the error is found, the more expensive it is to fix. A prototype is worth a thousand meetings.
Take the company Zappos, which sells shoes online over the Internet. The founder, Nick Swinmurn, tried to get investors, getting a typical response: "What are you drinking?" (In USA, we would ask what are you smoking, but this is the way the Brazilians say it.)
With no investors, Nick built a simple website, took pictures of shoes, and fulfilled orders by purchasing the shoes from local San Francisco retailers and shipping them to the clients.
Nick started this in 1999, and finally got some $20 Million USD in funding in 2004. His simple prototype allowed him to focus on post sales support. Zappos was recognized as having the best call center, moving his operations to Las vegas, NV.
Consider the challenges of urban mobility.
Both methods eventually result in a car, but the agile prototypes allow for more effective experimental milestones.
As for Zappos, its prototype proved successful. Amazon acquired them for $1.2 Billion USD in 2009.
It is that simple: Understand, explore, prototype, and evaluate. IBM has adopted "Design Thinking" across its development organizations to better meet the needs of the marketplace.
Overall, it was a delightful event. It is nearly summer down in the Southern hemisphere, so a bit warm and humid. The attendees were all looking forward to a turn-around in the Brazilian economy, and the business opportunities that brings.
How do you define success? For some, it is based on their salary, or perhaps revenue they helped close for their company.
For others, their family life and the flexibility to handle work/life issues might be more important.
Still others look for certifications and awards from official agencies.
As a side gig, I sometimes do bartending on the weekends. Typically, these are for weddings or corporate parties.
I took weeks of bartender training and passed a three-hour exam to become state-certified to do so in Arizona. We Arizonans take our liquor seriously! If you think about it, bartending is just a notch below being a Pharmacist dispensing other drugs.
Surprisingly, some of my patrons will be condescending, "Don't you wish you can do more with your life than be a bartender?"
I am also certified "Laughter Yoga" instructor, and am called in at times to substitue for other instructors. Again, I took formal training and was certified to do so.
Again, some of my students will ask, "Don't you wish you could do more with your life than be a yoga instructor?"
In both cases, I would respond, "Dude, I earn six figures, and am happy to meet new people every week, how about you?" This usually shuts them up!
(For those interested, here are [my top 10 posts] which served as the basis of the interview!)
I am happy to be recognized externally and within IBM for my success as a blogger. Since I started blogging over 10 years ago, I have helped close over $4 Billion USD in revenue for IBM, written five books on IBM Storage, mentored dozens of other successful bloggers, and presented to thousands of clients at conferences, workshops and briefings.
Last week, I was in São Paulo, Brazil for IBM Systems Technical University.
Did the resort ask these two security guards to dress up as clowns? No, it turns out these were clowns dressed up as security guards! On other days, they were dressed in drag as housewives, or as Jamaican Rastafari in dreadlocks and tie-dyed tee shirts. Some of the attendees enjoyed their comic relief.
Here is my recap of Day 3 breakout sessions:
Demystifying Transparent Cloud Tiering for DS8000 and DFSMShsm
Ricardo Alan, IBM Client Technical Specialist, covered this recently announced synergy between DS8000 firmware and DFSMShsm, a part of the z/OS operating system for IBM Z mainframes.
(Historical note: I started my career as a software engineer for DFHSM, which was later renamed DFSMShsm, working my way up to lead architect for DFSMShsm, and later as chief architect for DFSMS overall. A good portion of my 19 patents are related to these products.)
Since the 1970s, mainframe clients were able to move less active data from expensive disk storage to lower cost tape media. DFSMShsm would be read data sets into the mainframe processor, chop them up into 16KB blocks, and then write them out to tape, often through an automated tape library.
Transparent Cloud Tiering introduces an alternative option. DFSMShsm now identifies which tracks of data need to be re-located, sends the request to IBM DS8000 storage device, and the IBM DS8000 sends the tracks as objects to the Cloud. Any application that references these data sets would automatically trigger a recall to bring the data back from the Cloud.
This feature is available for the DS8870 and DS8880 models, using the existing Ethernet ports already installed. No additional hardware is required. Enhancements to DFSMShsm will be rolled out via SPEs on z/OS releases. Initially, the system uses OpenStack Swift object protocol, but IBM has plans to support Amazon S3 protocol as well.
Data Migration Challenges and Solutions with IBM Enterprise Storage
Sidney Varoni Jr. presented this session on data migration methods. Data is migrated for three reasons. First, to re-balance across multiple storage arrays. If you bring in a new storage array, you often want to move data from older arrays to balance the workload.
The second reason is to get rid of old hardware altogether, you need to migrate the data to new hardware. With Dell acquisition of EMC, for example, many clients are using tools like TDMF to move data off of EMC and onto IBM DS8000 storage systems. IBM DS8000 storage systems are faster, easier to use and less expensive to operate from a total cost of ownership (TCO) than comparable capacity of EMC VMAX devices.
The third reason is to migrate from one data center to another. The average data center was built 10-15 years ago, and many no longer meet the needs and requirements of newer IT operations. Some clients are building new data centers, while others are moving their data to co-location facilities.
NVMe Over Fabrics: The next evolution in high performance for SSD interfaces is NVMe
Waner Dall Averde, Territory Representative from Brocade, presented this session on NVMe and NVMe Over Fabric (NVMeOF). As a joke, he showed this chart in Japanese.
(Fun Fact: The first Japanese immigrants arrived in Brazil in 1908. Brazil is home to the largest Japanese population outside Japan. Source: Wikipedia)
For the past 20 years, the Advanced Host Controller Interface (AHCI) served as the communication mechanism to send SCSI commands to SAS and SATA disk devices.
Unfortunately, AHCI is now the bottleneck between faster servers and faster Non-Volatile Memory such as Flash and Solid State Drive (SSD) storage devices. It only supports a handful of commands on a single command queue.
NVMe offers a replacement for the SCSI command set. It can support up to 64,000 commands on as many as 64,000 parallel command queues. Designed for 32 Gbps PCIe bus speeds, it is faster than traditional 6 Gbps and 12 Gbps SAS connections, reducing latency by 200 microseconds.
Unfortunately, PCIe cables are limited to just a few inches. PCIe Gen 1 supported 15 inches, PCIe Gen 2 supported 12 inches, and PCIe Gen 3 only 8 inches. To provide greater distances, NVMeOF allows the NVMe command set to be carried over long-distance networks, such as Ethernet, Infiniband or Fibre Channel.
Brocade Gen5 (16 Gbps) and Gen6 (32 and 128 Gbps) Fibre Channel switches and directors already support NVMeOF, and are designed to allow co-existence between NVMe and SCSI commands for smooth transition in mixed environments. Clients can buy their networking gear directly from IBM.
IBM Power Systems Flash Cache Acceleration
Petra Bührer, IBM Offering Manager for Power Systems software, explained recent the performance enhancement called "Flash Cache Acceleration".
This is a feature on POWER8 servers running AIX 7.1 TL4 SP2, AIX 7.2 TL0 SP0 – or higher. By using internal or direct-attach SSD, the operating system can cache most active blocks of data from external storage systems.
While this is certified for use with Oracle, it supports only single-instance databases. Oracle RAC and other active/active configurations are not supported at this time.
The Secret to IBM Disk Encryption - Deep Dive
As if Mo McCullough, one of the event coordinators for this conference, was not busy enough with keeping the conference going, he also gave technical presentations.
With the excitement over the IBM z14 end-to-end encryption announcement, there has been increased demand for everything related to encryption and security.
Unfortunately, I had to leave for the airport before the "Closing Session". The Club Med Lake Paradise resort was 60-90 minutes away from the GRU airport, and rush hour traffic in a city of 12 million people can get really bad.
IBM Master Inventor, Senior IT Architect, and Event Content Manager
Well, it's Tuesday again, and you know what that means? IBM Announcements! We have a lot today, so I will just give you the quick highlights, and then Chris and Lloyd will follow-up with more detailed posts.
New IBM Storwize V5000 models
IBM introduces several new entry-level models.
The Storwize V5010E and V5030E are the "Express" models that allow for hybrid configurations, mixing Flash and spinning HDD disk. The Storwize V5010E is a single controller, two-canister model, with basic features. The Storwize V5030E adds more memory, more CPU power, and additional features like Data Reduction Pools and data-at-rest Encryption. Hosts can attach via SAS, 16Gb FCP or iSCSI.
The Storwize V5100 is the baby model of the FlashSystem 9100, supporting both FlashCore and industry-standard NVMe Flash drives, with the option to SAS-attach expansion drawers, mixing Flash and spinning HDD disk. The Storwize V5100F is the all-flash version. Hosts can attach via 32Gbps FCP, 25GbE RoCE, 25GbE iWarp, and iSCSI.
IBM Spectrum Virtualize for Public Cloud on Amazon Web Services (AWS)
The IBM Spectrum Virtualize for Public Cloud, or what our young folks shorten unofficially to SV4PC, that has been available on IBM Cloud will now also be available on Amazon Web Services.
For those readers asking "What took so long?" Amazon was not going to put specialized equipment in their data centers, so IBM had to make Spectrum Virtualize software container-native. Yes, the SVC code now runs in its own Docker container.
Basically a 2-node cluster is represented as two AWS EC2 instances, virtualizing EBS storage. The Transparent Cloud Tiering (TCT) that let's you "FlashCopy-to-the-Cloud" can be used to go directly to Amazon's S3 object storage.
This conversion to container-native has worked so well, IBM now plans to offer container-native software-defined storage capability across the board, for object storage, block storage, and file storage.
Did you notice that the Storwize V5100/F models support 32Gbps FCP in the section above? If that raised your eyebrow, I am pleased to tell you that IBM will be supporting 32Gbps FCP on these new Storwize V5100/F, the Storwize V7000 Gen3 and the FlashSystem 9100 devices.
We have also added a new b-type SAN switch, the SAN18B-6 which is Broadcom's Gen6 technology in a sleek 1U configuration, sporting 12 FCP ports that support 32Gbps and auto-negotiate to slower speeds as needed for compatibility with 8Gbps and 16Gbps devices. The other six ports are Ethernet, and can be used for disaster recovery replication, either using native TCP/IP or FCIP protocols.
IBM has enhanced the alerting capabilities of both the on-premise IBM Spectrum Control and its "as-a-Service" sister offering IBM Storage Insights. This allows you to set up alerts for "device groups" across multiple storage devices, as well as setting up filters to make the alerts more meaningful, eliminating some of the noise.
When IBM first introduced IBM Storage Insights, it was intended as an alternative to the on-premise solution. Now, clients demand both, so now if you have one, we can offer you the other! The new [IBM Storage Insights for IBM Spectrum Control] is an IBM Cloud service that can help you predict and prevent storage problems before they impact your business.
It is complementary to IBM Spectrum Control and is available at no additional cost if you have an active license with a current subscription and support agreement for IBM Virtual Storage Center, IBM Spectrum Storage Suite, or any edition of IBM Spectrum Control.
As an on-premises application, IBM Spectrum Control doesn't send the metadata about monitored devices offsite, which is ideal for dark shops and sites that don't want to open ports to the cloud. However, if your organization allows for communication between its network and the cloud, you can use IBM Storage Insights for IBM Spectrum Control to transform your support experience for IBM block storage.
IBM Spectrum Scale has been certified to run with with Hortonworks Data Platform (HDP) 3.1 release.
(Ha, I probably could have fit all that in the title of this section, but instead I just said "IBM Spectrum Scale" and you are thinking "Oh Boy!" and then you see something that could have fit in the title and feel all disappointed. It is kind of like when the local news asks "Was the restaurant you had lunch today contaminated with Salmonella?" and then follows up with the answer "Find out at 11:00pm evening news!" And then you wait until 11:00pm for then to say, "No, there was no salmonella found in any of the restaurants.".)
So, I would not have mentioned Spectrum Scale certification of HDP 3.1 unless there was at least something else worth mentioning. There is! IBM Spectrum Scale now also enhanced its performance for SMB and NFS, and has enhanced the scalability and resiliency of its Active File Management (AFM) feature.
The IBM FlashSystem A9000 and A9000R are targeted to Cloud and Managed Service Providers (CSP/MSP). The 12.3.2 release now supports VLAN tagging for iSCSI deployments. This VLAN tagging allows multiple virtual networks and IP addresses to share iSCSI ports, making it ideal for multi-tenancy for CSP/MSP clients.
IBM manages over a hundred Blockchain networks for its clients. For those not familiar with Blockchain, it is a way to record transactions, whenever money or product changes hands, an entry is recorded into the blockchain ledger for all to see.
This has two drawbacks. Information stored in the ledger may contain information you do not want everyone to see. The other is scalability, storing photos, and other supporting documents may be nice to have, but takes up a lot of space, and slows down transaction rates.
The solution is "off-chain" data. These are supporting documents that aren't needed in the blockchain itself. To connect them, you store a checksum hash of the supporting document in the ledge, then store the supporting document as off-chain data on-premises. If you need to produce the document for an audit, its checksum hash will match what is in the ledger.
In the beginning, people thought Docker containers would just be used for microservices with no persistent storage. Then clients realized they needed persistent storage, and they needed to orchestrate that storage provisioning. The IT industry has a variety of different orchestrators like Kubernetes, Docker Swarm, and Mesos. All of these manage persistent storage differently. IBM has focused on Kubernetes, using Ubiquity open source project to manage FlexVolumes.
Container Standard Interface (CSI) is an effort to standardize the provisioning of persistent storage. Allowing containerized applications to have access to storage that persists, even after the container is shutdown or crashes. For the next few years, I suspect IBM will need to support both the old way (FlexVolumes) and the new way (CSI) until all the standards settle.
You can hear all about these exciting announcements at the upcoming IBM Systems Technical University (TechU) in Atlanta, GA (USA), April 29-May 3. Visit [ibm.biz/Atlanta2019] to learn more and register. The three of us all plan to be there! Stop by and say hello.
IBM Senior Certified Executive IT Architect
Well, it's Tuesday again, and you know what that means? IBM Announcements!
Today I want to write about a very recent enhancement for IBM storage clients. This announcement was specific to all IBM Storage clients using IBM Spectrum Control.
IBM recently announced a new solution for existing Spectrum Control clients to obtain a cloud based version and get even more value from their Spectrum Control investment. Whether you have Spectrum Control Standard, Advanced, Select or Virtual Storage Center all versions are covered under this new solution.
In this new solution for existing clients of Spectrum Control, they are entitled to this new solution using existing Spectrum Control licensed capacity. This new solution which is a cloud based Software as a Service (SaaS) offering is titled Storage Insights for Spectrum Control. For current Spectrum Control clients this new solution is provided with no additional costs.
Since the release of Storage Insights & Storage Insights Pro existing clients with Spectrum Control have asked for a similar cloud based option. Today we have that option.
Other devices: Non-IBM storage devices, VMWare, SAN Switches
On Premises / Cloud
Asset management (Type, model, serial number, firmware)
On Premises / Cloud
Support management: Ticket creation / log upload
On Premises / Cloud
Health (show status of entity) direct / call home
On Premises / Cloud
Alerting (send mesg to user) status / thresholds / eMail / SNMP / scripts
On Premises / Cloud
Storage / Fabric Performance and Error Reporting
On Premises / Cloud
Performance Interval, retention
5 min, 24 hour
5 min, 1 year
1 min, customizable
On premises / Cloud
Provisoning using Service Classes & Capacity Pools with automatic zoning
Reclamation analysis of unused volumes
On Premises / Cloud
Service Management (Chargeback & Consumer reports)
On Premises / Cloud
Custom Reporting: GUI / API
On Premises / Cloud
Tiering support across pools
Recommend and Implement
On Premises / Cloud
Balance workload across pools
User ManagementL Active DIrectory/LDAP integration
Cloud portal SLA
If your existing Spectrum Control instance is providing your requirements then consider Storage Insights for Spectrum Control for the added value of enhance IBM Storage Support, and constantly getting access to the latest features of Storage Insights for Spectrum Control without any of the maintenance or upgrade activities.
Whatever reason you may need to reach out to IBM Storage Support, with IBM Support having immediate access to your storage configuration details will reduce the time and your teams effort to get a resolution or recommendation from IBM on how to proceed.
Last week, I was in São Paulo, Brazil for IBM Systems Technical University. With over 12 million people, it is the most-populous city in the Americas. Our venue was the Club Med Lake Paradise resort on the outskirts of town. We had about 700 attendees.
We had several local speakers do the opening session. Here is my recap:
Marcelo Porto, IBM General Manager for Brazil
This year, IBM Brazil celebrates 100 year anniversary. This all happened because Valentim Boucas persuaded IBM then-President Thomas Watson, Sr. to approve the establishment of a Rio de Janeiro office for the sale of IBM machines beginning in 1917.
For 100 years now, IBM has thrived with a set of core values. In every era in the past, IBM systems have been perfect for the business needs at the time, from punch cards to personal computers. But what got us here won't get us there in the future. The biggest challenge to transformation is people and culture. We must break the chains that hold us to the past. IBM drives disruption.
To prepare for the future, Marcelo recommended the following. First, learn English, because the English language is the "API of Business". Second, keep a curious mind. Seek out new things to learn. The new world needs skills and expertise in a variety of areas. Third, watch the movie "Hidden Figures", starring the IBM mainframe computer.
IBM Watson computer now speaks and understands Brazilian Portuguese language. Groupo Fleury uses Watson for genomics research. MRV Engineering uses this for chatbots. Mae de deus Hospital uses this for Oncology, as cancer patients now dominate the percentage of patients there. Walmart uses Blockchain to focus on food safety.
IBM Watson is used at Pinacoteca de São Paulo Museum to offer "Voz de Arte", the ability to ask IBM Watson about each painting in handheld smartphone devices. An example of this was available in the Solution Center.
In addition to natural language processing (NLP), IBM Watson can also do image recognition, a task normally only humans could do.
Watson can validate signatures, perform facial recognition at different angles, and even identify shirts, pants and shoes of fashion models in photographs.
Companies and organizations that are unable to transform data into insights and business decisions will fail.
Mauro D'Angelo, IBM Strategy and Business Development for Brazil
Why are companies like Uber and Airbnb successful? Mauro felt that it was because they had a proper Cloud infrastructure combined with the right data architecture.
(In this case, "success" is based on company valuation, often billions of US dollars. However, many of these companies are not profitable, losing millions of dollars in an aggressive effort to gain customers and establish their platform. It might take 12 to 24 months before a new customer becomes profitable.)
The data explosion is driving digital transformation. Cognitive systems must understand natural language, reason, learn and interact with humans. Machine Learning is much like training a puppy. You need to reward good behavior and fix bad behavior, and be patient, as it takes a long time.
In USA, patients asking Doctors for a diagnosis get only 50 percent correct on first consultation. Often, additional doctors or additional tests are needed to finally get correct assessments. In Brazil, it is probably less than 50 percent. Hopefully, Watson will help improve this.
Watson can also detect emotional tone and personality in social media. Is a customer angry? This could help prioritize which customer issues to address first.
Schools have not changed since the days of Aristotle. Mauro showed a picture of a school taken in 1934, and a picture of the same classroom, taken recently, showing it is nearly the same. Students want to learn anytime, anywhere, and from any channel.
At Georgia Tech University, a professor told his engineering students that there were nine "Teacher Assistants" (TAs) available to help answer questions online. One of these was [Jill Watson], which was the IBM Watson computer responding to the students. The students could not tell that Jill was not human!
In traditional schools, a teacher may reach only 50 to 60 students. Compare this to [Khan Academy] that offers video instruction that have had over 1.3 million views!
Frank Koja, IBM Systems Vice President for Brazil
When you buy something over the internet, what is your decision criteria? Often, it is lowest cost. Digital transformation often requires re-invention.
Trust beats risk. The new IBM z14 mainframe focuses on trust, with end-to-end encryption, Blockchain and Machine Learning. zHyperLink drastically improves the connection between mainframe and IBM DS8880 storage. IBM is helping over 400 clients adopting Blockchain.
The FlashSystem A9000 and A9000R models are 30x faster than traditional disk systems, and more dense, able to consolidate 20 racks down to one.
The new "PowerAI" bundle combines together a complete offering for Machine Learning and Deep Learning (ML/DL) for Power systems, taking advantage of GPU and NVlink capabilities.
The "waitless" world has arrived.
This was a good start for the conference. The three speakers of the opening session were passionate of what they were talking about, and people were excited to learn more as the week progressed.
Back then, IBM allowed its employees the option to run Windows, Linux or Mac OS. Since then, dual-boot Windows/Linux configurations, like the one I had back then on my Thinkpad T410, proved too difficult for our help desk, so these are no longer allowed.
In 2015, I received my new Thinkpad T440p to replace the old T410 model. For those 20 to 25 percent of the IBM employee population that manage, support and connect directly to client networks, IBM required Linux encrypted with LUKS, using Windows as KVM guests when needed for specific applications. This is more secure than running Windows natively, preventing viruses and other malware to spread between IBM and its clients.
As I am occasionally asked to help out our colleagues in lab services or with critical situations, I decided to implement my laptop to match, just in case. RHEL is rock solid, and running Windows as KVM guests could not be easier. Not having to worry about Windows viruses while travelling on business is a huge benefit as well.
Upgrading from RHEL 6.1 all the way up to RHEL 6.9 was simply a push of a button, all the new applications and kernel get installed, followed by a quick reboot. The migration from RHEL 6.9 to RHEL 7.4, however, was a major undertaking.
In past migrations, I was moving from a working laptop to a second laptop, affording me to be fully productive on the old machine until I was ready to cut over. In this case, I am performing a fresh install on my existing machine. To avoid any problems or delays, I wrote myself an 8-page, 17 step migration plan to capture all the tasks I needed to do to minimize the impact to my productivity.
(Of cousre, IBM has a help desk. You hand over your laptop, they backup the home directory, wipe your system clean, fresh install, restore your home directory, and return the laptop to you 3-5 days later, leaving the rest of the tasks up to you. Basically, this would merely replace the first three of my 17 steps below. I did not feel like burdening our help desk, nor wait 3-5 days without a laptop!)
Here were my steps:
Backup my existing system
In addition to backing up all my individual files to the Cloud, I also used [Clonezilla] to create a full image backup of my 500GB drive to an external USB drive.
Not all data is in file form. I also exported my browser bookmarks, so that I could import them back later. I also ran an "rpm -qa" to get a list of my existing applications installed.
Initially, I thought to format the 4TB external drive in UDF format, which is readable by Windows, Linux and Mac OS and supports files that are larger than 4GB in size.
Not knowing whether I should use [ExFAT] or Universal Disk Format [UDF] format, I split the 4TB into two 1.9TB partitions, and formatted one as ExFAT, and the other as UDF. Both formats support files greater than 4GB in size, which I have, but I discovered that on the older RHEL 6.9 release, based on a 2.6 Linux kernel, you can only write 68GB of data to a UDF partition. This is fixed in later kernels, but doesn't help me with my existing RHEL 6.9 release.
Fortunately, the latest Clonezilla LiveCD chops up the cloned images into files small enough that you can write to a variety of formats, and has a newer kernel that allows writing the full capacity of UDF partition.
In a crisis, I can restore back to RHEL 6.9 within 2 hours. This was my "relief valve" if I encountered any major delays and had to go travel for business on short notice.
Fresh install of RHEL 7.4 Linux
This completely wipes clean my drive, and installs two partitions. A tiny "/boot" partition needed to boot the system, and the remaining drive capacity as a large LUKS-encrypted LVM, to be internally partitioned between "/" and "swap" logical volumes.
Copy all of my files back
The challenge is that some files might clobber some of the configurations of the new applications. For this reason, I created /home/tpearson/RHEL69 and put everything there, so that I can move them to the correct locations as appropriate.
Copying all the files back in this manner eliminated having to be tethered to the external USB drive.
Setup LAN connectivity
I have to connect to IBM and guest systems, so this configuration is important. This includes EAP, TLS and VPN configurations. I thought I could just re-use the certificates I have for RHEL 6.9, but no, I had to create and register fresh new certificates for RHEL 7.4 release.
Configure Cinnamon Desktop
RHEL 7.4 uses Gnome 3 by default, which is quite different than Gnome2 used in RHEL 6.9 release. I don't care for it, so I configured [Cinnamon desktop] instead. Many people who use Linux Mint or Ubuntu might be familiar with this, and for those switching from Windows or RHEL 6.9 Linux, Cinnamon has familiar "Start" button in lower left corner.
By default, our RHEL 7.4 image comes with Firefox and Chrome browsers, so all I needed to do was import the bookmarks that I had exported in step 1 above.
Configure KVM guests
I was able to bring over my Windows7 Kernel-Virtual Machine [KVM] from RHEL 6.9 and run without problems, but this was bloated and now consuming nearly 60GB of space. Therefore, I decided to get a fresh Windows7 and Windows10 guest images instead.
Like with Linux, I wrote down what applications I had installed on Windows, and used that to configure the Windows guests. Nearly everything I do runs natively on Linux, but I do use Microsoft Office (Powerpoint, Excel, Word) and a nice tool called [CutePDF] that allows me to print to PDF instead of an actual printer.
Windows10 comes with the "Print-to-PDF" feature built-in, so no need for CutePDF on that one.
Configure IBM Notes, Sametime and Gnote
IBM is a heavy user of [IBM Notes] (formerly called Lotus Notes), not just for email but also for its document management and database capabilities. Sametime is our "Instant Messenger" app. [Gnote] is a linux-based tool to store short notes, I use it for all of my email templates for quick copy-and-paste responses.
IBM recently made using printers super easy. Print to the common "Cloud printer", and then pick up your print-outs from any printer in the building, any IBM building, worldwide. I could print in Tucson, for example, and pick up my print-outs when I am in the IBM buildings in Austin, Texas!
I also had to configure my printer at home, for those days where I need to print a boarding pass or quick document.
Configure File Sharing
IBM has deployed IBM [Spectrum Scale] internally for employees to share files across the company called "Global Storage Architecture" (GSA). Configuration for me just meant having to find my local cell (tucgsa) for Tucson, and entering my credentials.
Install Docker and DSX Desktop
[DSX Desktop] is the local laptop version of IBM's cloud-based [Data Science Experience], allowing me to perform Hadoop and Spark analytics for the various projects I work on. It runs as a Docker container, so I had to configure Docker as well.
Install Multimedia Codecs
One of the big detractors for Linux, compared to Windows or Mac OS, is the lack of multimedia support. Linux distros, like Red Hat, don't ship with these pre-installed, leaving this as an exercise for the end user.
IBM does a lot of audio and video files, including replays of conference calls and webinars for internal training. I keep a collection of different audio and video files to ensure that I have everything configured correctly for proper playback.
Install GIMP and other software
The GNU Image Manipulation Program [GIMP] is a great tool for quick editing of graphics. Another tool, Inkscape is designed for vector graphics.
Configure file-level backup
In addition to doing full-volume image backups with Clonezilla, I back up individual files, which are sent over the IBM internal network to a central server. All I need is configure to my previous backup set, and create the appropriate include/exclude list.
Many employees might just back up their home directory, but I customize a lot of the Linux configuration, so I like to backup a few more directories. Here is what I choose to back up:
Congigure Grub2 boot configuration
RHEL 7.4 supports [Grub2], which allows you to boot iso files directly. I like to add Clonezilla and [SystemRescueCD] as boot options. These were simple enough to add, just follow instructions, copy files to the /boot directory, and create a menuentry for each.
Validate final configuration
After eight days, I have finally completed all these steps, and am able to validate that everything is working correctly. I did some sample workflows, such as:
Verify that I can launch Windows KVM guest, edit Powerpoint presentation, and print to PDF file.
Verify that I can open email, launching embedded URL links, and copy-and-paste templates from Gnote
Launch GIMP, verify that I can edit graphics, and import the results in a Powerpoint presentation.
Download and play a Webinar replay MP4 file
Fresh Clone of full volume image
Using the Clonezilla that I added to the Grub2 boot menu, I am able to backup my full 500GB drive. At this point, I will keep the RHEL 6.9 for a few weeks as emergency backup, but so far, everything seems to be working just fine.
This took longer than I expected, but am happy with the final result. Red Hat is rock-solid, and the new RHEL 7.4 allows me to run DSX Desktop, Windows 10, and some other applications that were not available on our previous RHEL 6.9 build.
Well, it's Tuesday again, and you know what that means? IBM Announcements!
(FTC Disclosure: I work for IBM, and have either written code and/or presented the DS8000 storage system and Spectrum Storage products in my professional capacity. This blog post can be considered a "paid celebrity endorsement" for the IBM DS8000 Storage System and Spectrum Storage software.)
IBM DS8880 and DS8880F Storage Systems
For those not up on the DS8000 nomenclature, here's a quick recap:
DS8880 supports a hybrid mix of Flash cards, SSD, 15K, 10K and 7200 rpm drives.
This includes the DS8884 and DS8886. The Flash cards are held in High Performance Flash Enclosures (HPFE) directly attached to the controllers, whereas the SSD and spinning disk are in shelves connected via the Device Adapters.
DS8880F is an all-flash array, with Flash cards only in HPFE. This includes the DS8884F, DS8886F and DS8888F models.
DS8880/F is convenient shorthand to refer to both the hybrid and all-flash models collectively.
Today, IBM announces new 7.68TB flash cards for the High Performance Flash Enclosures of the IBM DS8880/F. These are double the capacity of the 3.84TB cards currently available, doubling the total capacity to 368.6TB per HPFE.
Different DS8880 models support a different number of HPFE. An HPFE is a pair of 2U drawers, holding a total of 48 flash cards. You can purchase flash in groups of 16 cards, with the option to mix and match within the HPFE. For example, you can have 16 cards at the 1.6TB capacity, 16 cards with 3.68TB and 16 cards of the new 7.68TB capacity, all in a single HPFE.
The new 7.68TB support 1 Drive Write Per Day (DWPD). Some people call these "Read-Intensive" drives, but IBM refers to them as "High-Capacity Drives", to differentiate them from the "High Performance Drives" that support 10 DWPD.
In reality, the read performance is similar in both types of Flash cards offered, but the write performance is slightly slower for the High-Capacity drives due in part to additional garbage collection performed in the background. Our studies found that over 90 percent of workloads might find the High-Capacity drives good enough to handle I/O requirements.
IBM Easy Tier was updated to distinguish between High-Performance and High-Capacity flash cards, so that blocks of data that have higher or lower I/O characteristics will be relocated to the appropriate level of storage.
The newest level of IBM Spectrum Storage Suite simplifies procurement by bringing together the latest releases of the following software:
IBM Spectrum Accelerate V11
IBM Spectrum Archive Enterprise Edition V1 (Linux edition)
IBM Spectrum Control Advanced Edition V5
IBM Spectrum Protect Suite V8 (including Spectrum Protect Plus!)
IBM Spectrum Scale Data Management Edition V5
IBM Spectrum Virtualize Software for SAN Volume Controller V8 (including FlashCopy and Remote Mirror, Real-time Compression and Encryption Software)
IBM Spectrum Virtualize Software-only V8
IBM Cloud Object Storage System V3
Instead of buying software products separately, a single license enables administrators to deploy IBM Spectrum Storage Suite software when and where they need it, without having to wait. Simplified capacity pricing can significantly reduce software costs and time spent on license management.
The Spectrum Storage suite also offers a "sandbox" approach for try-and-buy. Since you have access to all the software listed, you can set up a sandbox to experiment with the functionality, without having to pay for the added capacity, until you deploy it to dev/test, quality assurance, or production.
The suite is licensed per Tebibyte [TiB]. For those not familiar with international standards, here is a comparison table:
Always decimal, 10 to the 12th power
Always binary, 2 to the 40th power
The two terms sound similar and represent nearly the same quantity within 10 percent of each other, so it is understandable when people mistakenly use the terms interchangeably.
From farm to fork, IBM Food Trust platform is a collaborative network of growers, processors, wholesalers, distributors, manufacturers, retailers and others enhancing visibility and accountability in each step of the food supply.
Powered by the IBM Blockchain Platform on IBM servers and storage systems, IBM Food Trust directly connects participants through a permissioned, permanent and shared record of food origin details, processing data, shipping details and more.
(This reminds me of a funny story: the man sitting next to me on my flight back from an IBM Systems conference in New Orleans asked me, "You look familiar. Didn't I see you at the conference this week?" I responded "Yes, were you there for the "server" or "storage" side?" He thought about it for a while, and said "I guess the server side". "Too bad", I replied, "I am on the storage side."
It took us a while, but I realized he worked in the food and restaurant industry, and that he was at a completely different conference. It happened to also have both a "server" and "storage" side!)
The IBM Food Trust platform provides new levels of transparency, quicker recalls, better standardized communication and protection of brand value. As an authorized user, you have immediate access to shared, actionable food supply data through integrated IBM Blockchain-powered modules for faster traceability and more confidence in provenance.
Today, IBM announces new services to enable clients to successfully connect to and make use of the IBM Food Trust Platform.
I was in Hollywood Florida for the IBM Systems Technical University. Here is my recap of the final two days, day 4 and 5.
The Pendulum Swings Back: Understanding Converged and Hyperconverged Systems
Once again, I presented my popular session on converged and hyperconverged systems. For converged, IBM offers IBM PureApplication systems with Power and x86 servers, as well as partnership with Cisco called VersaStack. Both support IBM Cloud Private as a platform for running applications.
For Hyperconverged, IBM offers Spectrum Accelerate and Spectrum Scale, as well as partnerships with SuperMicro that combines Spectrum Accelerate on SuperMicro x86 servers, and partnership with Nutanix for CS-models of Power servers pre-installed with Nutanix software.
Unlike other converged and hyperconverged solutions that act as isolated islands of compute and storage, IBM's solutions can be incorporated into an existing datacenter with IBM Cloud Private for orchestration, and IBM Spectrum Scale to provide common access to data.
The Seven Tiers of Business Continuity and Disaster Recovery
With all the natural disasters that happened last year in the USA, and the more recent ones all over the world, this session continues to draw a crowd.
The seven tiers range from the least expensive to most expensive. The least expensive involves restoring data from tapes stored in an offsite vault. Tape continues to be the least expensive storage medium, and can be used to bring up a company in a few days.
For faster recovery, there are options like electronic vaulting to virtual tape libraries, and now the use of Cloud storage for ubiquitous access to data from different locations.
Snapshots of entire volumes, virtual machines or databases are also quite popular. IBM offers IBM Spectrum Protect Snapshot, Spectrum Protect Plus, and Spectrum Copy Data Management for this.
Faster recovery is possible with remote mirroring. This involves sending all of the updates to a secondary location. In the event of a disaster, clients can switch processing with the data already there. IBM has over 800 clients able to do just that in less than 30 minutes.
Event Night by the Pool
Photography by Mo Reyes
While Hurricane Michael raged in upper Florida the week prior, the event coordinators were a bit nervous to offer an evening dinner event by the pool, but the weather cooperated!
Photography by Mo Reyes
I was a social butterfly, moving from table to table to talk to all of the various attendees. A light breeze and excellent food and music made for an enjoyable night!
The pool reception went on to about 10:00pm at night. IBM had lit up its logo into the pools for a great view from above. Perhaps just 30 minutes after arriving back to my hotel room, we had quite the thunderstorm! How incredibly lucky this did not happen during the event!
The following day, I presented my session on "Managing Risk with Data Footprint Reduction, a repeat of the session I did earlier that week.
This was a pleasant way to end the week! Aside from the heat and humidity being above average for October, it was a beautiful hotel in a lovely city.
Well, it's Tuesday again, and you know what that means? IBM Announcements! There were a lot of IBM Power System announcements on Tuesday, so the IBM Power team asked us to wait until Thursday to post about all of the IBM storage announcements, to avoid overwhelming excitement levels with the press and analysts.
(FTC Disclosure: I work for IBM. I have either worked on the code, developed marketing materials, and/or represented each of the products below in my professional capacity. This blog post can be considered a "paid celebrity endorsement")
A few months ago, IBM re-factored its internals. Spectrum Virtualize will continue to support its legacy storage pools, but also offered "Data Reduction Pools", or "DR pools" for short. At the time, this supported only Thin Provisioning and Compression. See fellow blogger Barry Whyte's post on [Data Reduction Pools] for more details.
Spectrum Virtualize 8.1.3 release now adds Data Duplication and RESTful API support for the Spectrum Virtualize family, including SAN Volume Controller, FlashSystem V9000 and Storwize products. These features also apply to Spectrum Virtualize as software only, and to Spectrum Virtualize for the Public Cloud.
Data Deduplication is a form of data footprint reduction. Like the deduplication in Spectrum Protect and FlashSystem A9000/R products, Spectrum Virtualize will use SHA1 hash codes to identify duplicate 8K blocks. If the hash code of the block about to be written does not match any existing hash code previously written to the cluster, it is considered unique data.
Legacy storage pools supported three kinds of volumes: fully-allocated, thin-provisioned, and compressed-thin volumes. The new DR pools support five kinds: fully-allocated, thin-provisioned, deduped-thin, compressed-thin, and deduped-compressed-thin volumes.
The new deduplication feature is included at no additional charge with the base Spectrum Virtualize license.
The RESTful API enables storage admins to easily automate common tasks with industry-standard tools. RestAPI support is available to interface with the command-line interface (CLI), create vDisk volumes and generate views normally available through the CLI, and secure authentication to the IBM Spectrum Virtualize family.
The SAN Volume Controller, FlashSystem V9000 and Storwize family now also support 12TB drives for internal storage. These are 7200 rpm 3.5 inch drives that can be in the 2U 12-bay or 5U 92-bay expansion drawers, or directly in the 12-bay Storwize controllers. Spectrum Virtualize 7.8.1 is the minimum level to support these high-capacity disks.
IBM Spectrum Virtualize for Public Cloud, available on IBM Cloud, has been enhanced to support a full eight node cluster (four node-pairs, or "I/O Groups" as they are called). This can be used as a target for remote mirror from your Spectrum Virtualize cluster on premises.
IBM offers data footprint reduction, high availability, and technical refresh guarantee programs for these products. See Ernie Pitt's blog post on [Peace of Mind with IBM Storage].
IBM Spectrum Scale 5.0 is highly scalable file and object storage system. It is available as software, pre-built appliances, and in the Cloud.
The pre-built appliances are called "Elastic Storage Server", combining Spectrum Scale software on two IBM Power servers with drawers of flash or disk drives.
IBM introduces two new "Hybrid" models to the ESS family. GH14 has one 2U drawer with 24 Solid State Drives (SSD) combined with four 5U drawers with 7200rpm spinning disk. The GH2R has two 2U drawers with four 5U drawers.
Like the GS models, the SSD are either 3.84TB or 15.3TB capacities. The 5U drawers are similar to those in the GL models, either 4TB, 8TB or 10TB capacities.
A new Enterprise Slim Rack (S42) is now available to hold these. The S42 is available for all ESS orders, including the GS, GL and new GH models.
IBM has shortened the name of "Spectrum Control Storage Insights" to just "Storage Insights" and made it available in two flavors: Storage Insights, and Storage Insights Pro.
Storage Insights is a no-cost cloud Artificial Intelligence (AI) service that provides common monitoring capabilities to all of your IBM block-level storage, including IBM FlashSystem, SAN Volume Controller (SVC), Storwize, DS8000 models and IBM XIV Storage Systems. Here are some of the capabilities offered:
View the health, performance, and capacity of all your IBM-supported devices from a single place
Filter storage device events to help you focus on the things that require your immediate attention
Act on predictive insights provided by device intelligence before anomalies have an impact on service levels
Use actionable data you get to resolve more issues on your own
Open and view IBM support tickets
Enable IBM Support to automatically collect log packages with no interaction with the client
IBM Storage Insights Pro includes everything in Storage Insights as well as these additional capabilities. This is a fee-based cloud service, licensed per TiB per month, for the added functionality:
Business impact analysis
Data placement optimization with tier planning
Capacity optimization with reclamation planning
Supports file and object storage, including IBM Spectrum Scale, Elastic Storage Server (ESS), and IBM Cloud Object Storage (IBM COS)
Both Storage Insights and Storage Insights Pro use a "data collector" that runs on premises. This can be any bare metal server or Virtual Machine running Windows, Linux or AIX operating system connected to the SAN, with access to the Internet to upload the data to the IBM Cloud.
If you have IBM block storage today, there is no reason not to try this out. You can download the "data collector" and start using Storage Insights right away. If you like it, consider upgrading to Storage Insights Pro, or the full on-premise Spectrum Control product.
Now the Thanksgiving Holidays are over, its time to catch up on some Storage Announcements! These actually came out November 20, but many readers were busy with Thanksgiving Thursday, Black Friday, Small Business Saturday, Cyber Monday and Give Back Tuesday, they might have been missed.
IBM DS8880/F Storage Systems
There were three items this time for the DS8880 family, for both Generation 1 and Generation 2 models.
The first is the expansion on the type of high capacity flash drives available for the system. With this announcement, IBM has added the high-capacity 15.3 TB flash drive set as an available option. With this, the maximum flash storage capacity supported per high performance flash enclosure (HPFE) pair is now 737 TB raw.
The second is the addition of write operations for zHyperLink. zHyperLink is a short distance, mainframe attach link that is intended to accelerate Db2 v12 for z/OS. Low I/O latencies deliver value through improved workload elapsed times and faster transactional response times, and contribute to lower scaling costs.
The third is the ability to leverage IBM Storage Insights cloud-based service for your DS8880 systems. It supports the next-generation Call Home infrastructure that enables continuous updates of the system's configuration and health. IBM Storage Insights simplifies uploading logs, speeds resolution with online configuration data, and provides an overview of open tickets, all in one place. Clients who have been using IBM Storage Insights with their their IBM Storwize and IBM FlashSystem A9000/R systems can now extend this to include their DS8880 systems.
The TS4500 library now supports IBM TS1080 tape drive model F8S. This is an LTO8 Ultrium tape drive enabled with single-mode fiber and designed for the heavy demands of backup and archive tape storage applications that require high bandwidth over long distances. TS1080 Model F8S can be installed in a TS4500 tape library and delivers maximum throughput with a native data transfer of up to 360 Mbps.
There were two other enhancements for the TS4500 were:
A new action "Export logs direct to IBM" on the web. The TS4500 will collect library logs and send via Call Home to IBM, eliminating download to the laptop or system. This improves how clients engage with IBM support teams.
A new intelligent Power Distribution Unit (PDU) with three-phase [wye] or single-phase input to provide power to the local frame control assembly. This new PDU will offer power and environmental monitoring features in the future. So stay tuned.
IBM continues its lead in tape innovation with the IBM TS1160 Enterprise Tape Drive model (60F)! Delivering a 100% increase over the TS1150 series of tape drives, the TS1160 comes in at 20TB of cartridge capacity on the new JE media type with a native data rate of 400 Mbps, or up to 900 Mbps compressed. The new TS1160 will be available for the TS4500, the TS3500 (Model 60G) as well as a stand alone design with Fibre Channel interfaces.
Of course, the new TS1160 will support the reuse of existing JD/JZ/JL media with the capability to read and write TS1150 and TS1155 formatted media, and format the same media with up to 15 TB uncompressed capacity. It will also support IBM Spectrum Archive with the LTFS format for tape. Leveraging these larger tape capacities for active archives, backups, and other long term retention requirements.
IBM Spectrum Archive can move data from Spectrum Scale flash and disk storage to less expensive tape. IBM Spectrum Archive is based on IBM Linear Tape File System (LTFS) and enables fast and easy data retention to reduce the total cost of ownership for active archive storage. Both the IBM Spectrum Archive Enterprise Edition (EE) V1.3, and the Library Edition (LE) V2.4.1, will support the new enterprise-class TS1160 tape drive.
The Data Management Edition (DME) of Spectrum Scale does not require you to separately purchase client, server, or FPO socket-based, MIPS-based, or PVU-based licenses. Instead, you can simply license the total storage capacity managed by IBM Spectrum Scale. DME is now available through IBM Shopz. Shopz is a web service to order z Systems software, manage software licenses, and view software inventory. It provides an easy way to plan and order your product ServerPac or CBPDO, running on z.
This month, IBM Tucson Development Lab is celebrating 40 year anniversary! IBM has been operating in Arizona for the past 70 years, and of course IBM has been in the storage business for the past 90 years if you consider "punched cards" as storage on paper.
This year also marks the 40 year anniversary of DFHSM, the first product I worked on when I started here back in 1986. DFHSM stands for the Data Facility Hierarchical Storage Manager, which effectively moves data between disk and tape storage.
IBM put up two banners to celebrate! The first was for IBM Enterprise Tape storage. My first question was "What are punched cards doing on a banner for magnetic tape?"
A bit of history will explain that the first tape storage was non-magnetic. Back in 1725, Basile Bouchon developed the control of a loom by punched holes in paper tape. These were used to create intricate patterns in woven cloth.
In the late 1880s, Herman Hollerith, a young technical whiz at the US Census Bureau, had an idea for a machine that could count and sort census results far faster than human clerks. The bureau funded Hollerith’s work, and the [first tabulating machines] helped count the 1890 census, saving the bureau several years’ work and more than US$5 million.
Hollerith left the bureau to form the Tabulating Machine Company, selling his system to other countries’ census offices and then to businesses such as railroads and retailers. Hollerith had little competition, and his machines and punched cards became the standard for the industry.
In 1911, financier Charles Flint bought the Tabulating Machine Company and merged it with the International Time Recording Company and the Computing Scale Company of America to form the Computing-Tabulating-Recording Company, or C-T-R, later renamed IBM in 1924.
In 1928, IBM introduced a new version of the punched card with rectangular holes and 80 columns. The 80-character standard was used from everything from the first computer screens, to the first file layouts
It wasn't until 1952 that the first magnetic tape system hit the scene: the IBM model 726. Tape reels were the size of pizzas, and were prominently shown spinning around in various Hollywood movies to represent computers "working" on a problem.
In my now infamous 2007 post [Hu Yoshida should know better], I explain the 3850 Mass Storage System (MSS). In 1974, The IBM 3850 MSS was one of the first hybrid disk-and-tape storage systems. It was an automated tape library pretending to be disk, with tape cartridges stored in hexagonal honeycomb shelves. The tape cartridges were cylindrical, about the size of a can of soda. The spool of 770 feet of tape media held just 5MB of data.
A full IBM 3850 MSS configuration with thousands of tape cartridges was used for the 1980 US Census, holding 102 GB database, representing the data collected about 226.5 million U.S. residents. That's about 450 bytes per resident, enough to fill six punched cards.
The second banner was for IBM Enterprise Disk storage.
IBM introduced the IT industry's first commercial disk system in 1956. While the banner says "RAMAC 305", that is the name of the server. The storage system was called the [350 Disk Storage Unit]. It was the size of two refrigerators and held 5 MB of data.
In the early 1990s, I visited a client in Germany that had a 3990 controller with two 3390 disk systems attached, holding 90 GB of data in the size of three refrigerators. They had five storage administrators to manage this configuration.
A few years later at another client, they had roughly 7000 GB (7 TB) of data on their mainframe, and an equal amount across all of their Windows and UNIX servers. I met with their storage administrators, there were two for the mainframe, and about three dozen for the distributed servers.
I had two questions for them. First, why were there two storage admins for the mainframe? The mature policy-based automation on the platform would mean only one person required. Their response: when one of us is on a two-week vacation, the other can handle the workload.
My second question was for the remaining storage admins: When was the last time any of you took a two-week vacation? None had, of course, since the storage administration tools back then meant they were all working overtime on various tedious and manual tasks!
In February 2006, the folks in IBM Germany asked the IBM Storage Marketing team what events or celebration were planned for September 13, 2016, the 50 year anniversary of disk. My marketing colleagues responded, "that is only seven months away, you didn't give us enough lead time notice to plan!"