On his blog post on preparation
, Seth Godin mentioned an appropriate Swedish saying:
There is no bad weather, just bad clothing.
Appropriate because it snowed here in Tucson, Arizona on Sunday evening, leaving many of us here figuring out how to drive through the stuff on Monday. In my entire lifetime, I have only witness snow down in the Tucson valley a handful of times. It got me thinking about coats, and the wonderful schemes for coat check rooms, as an analogy for data access. A lot of people ask me to compare and contrast one technology from another, say block-level virtualization from content-addressable storage, and so on, and I always try to find a good analogy to help explain things.
Let's start with the setting. It is snowing outside and people are wearing coats. When they come inside, they check their coats at a coat check room, a large room with rows and rows of racks with hangers. A coat check attendant takes your coat and puts it on a hanger, and gives you a ticket or other identifier that will allow you to retrieve your coat later. The ticket must have sufficient information to retrieve the coat quickly, rather than searching rows and rows of hangers for it.
- Block-based disk storage
You walk to the coat-check desk, tell the attendant to hang your coat on a specific hanger, say hanger number 387. When you come back, you ask for the coat on hanger 387. The coat-check attendant knows exactly where hanger 387 is, and is able to retrieve it quickly. Most disk systems use this approach, including IBM SAN Volume Controller and DS family of disk systems.
- Name-based disk storage
You walk to the coat-check desk, tell the person the name that you want to call your coat. An empty hanger is located, and a list of coat names, with their associated hanger number, is then kept. Upon return, you ask for your coat by name, and the coat-check attendant looks up the hanger number to match, and retrieves your coat. This is the scheme used by the IBM System Storage DR550, N series for NAS storage, and the IBM Healthcare and Life Sciences Grid Medical Archive Solution (GMAS).
- Content-addressable storage (CAS)
You walk to the coat-check desk and hand them your coat. The attendant weighs your coat, checks the brand, the size, the number of buttons and zippers, types it all in, and the computer spits out a "hash code" from 1 to 99999. An empty hanger is found, and the hash code is associated to the hanger number. Upon return, you provide the hash code you were given, and the coat-check attendant looks up the hanger number to match, and retrieves your coat.This is the scheme used for some non-erasable, non-rewriteable storage, such as the EMC Centera.
IBM invented hash codes in 1953 as a way to speed up searches. For example, if you want to look up a word in the dictionary, knowing the first letter of the word makes it much quicker, because you can thumb directly to that section. A hash code was intended to give a more even distribution, so that if a million words are stored in a "hash code dictionary" then you would calculate the hash code, then look up only that section of words associated with that specific hash code number.
A problem arises when you generate "hash codes" for storage. It is possible for two different pieces of data to resolve to the same hash code. When an application tries to write a piece of data, and it resolves to a hash code that already exists, that is called a collision. One response is to either compare the incoming data to the data that is already stored, confirm they are identical, but that can be time consuming. The other response is to just assume they are identical, and reject the secondary copy, a process often referred to as "de-duplication".
What's the chance of getting a collision for data that is really different? Let's take for example the famousBirthday paradox. Suppose the coat check room assigned the hanger based on your birthday (month and day). How may coats before you run the risk of having two people turn in coats with the same birthday? After only 23 people, the likelihood is 50%. At 60 people, it goes up to 99%.
For this reason, IBM does not offer content-addressable storage. For non-erasable, non-rewriteable storage, the IBM System Storage DR550 requires the application to give each object a name, and that name is then used to storage the data, eliminating the possibility that data might accidently be thrown away.
It's safer that way.
technorati tags: Seth Godin, Swedish, saying, bad, weather, clothing, snow, Tucson, coat, check, room, IBM, block-based, disk, storage, DR550, N series, NAS, healthcare, life sciences, grid, medical, archive, solution, GMAS, content-addressable, CAS, EMC, Centera, hash code, collision, de-duplication, birthday, paradox
Today, January 16, IBM launches its latest disk system, the DS3000 series.
There are actually three products in the DS3000 series:
The DS3200 is a 2U high, 12 drive system that attaches to servers via 3Gbps Serial Attach (SAS) interface.You can expand this to 48 drives by added EXP3000 expansion units. Here are theDS3200 specifications.
The DS3400 is a 2U high, 12 drive system that attaches to servers via 4Gbps Fibre Channel (FC) interface.You can expand this to 48 drives by added EXP3000 expansion units. Here are the DS3400 specifications.
The EXP3000 is a 2U high, 12 drive expansion drawer. It was announced back in August 2006, but is part of theoverall DS3000 series. It can be used directly with servers, but is also designed to be attached to the back of the DS3200 or DS3400 to increase capacity.Here are the EXP3000 specifications.
With this announcement, IBM provides entry-level storage at the "less-than-$5000" price point, withsupport for intermix of 10K and 15K RPM drives, and scalable up to 14.4 TB capacity.This would be ideal storage for HP, Dell, IBM System x and BladeCenter servers.
technorati tags: IBM, disk, DS3000, DS3200, DS3400, EXP3000, HP, Dell, SAS, SATA, FC
On his "Data Storage - Dullness becomes Mainstream" blog, Chris Evans is
amazed athow low they can go!
.He compares the latest 100GB Toshiba 1.8" drive designed for portable music players, to the size andweight of older technology, like the IBM 3380 Direct Access Storage Device (DASD).
Chris couldn't find the dimensions of the 3380, so I thought I would provide the missing detail.The IBM 3380 History Archivesprovides a nice summary:
- The CJ2 model that Chris mentions was announced September 1, 1987 and shipped in 1988. Earlier models of the 3380 were announced 1980-1986.
- Capacity and performance were measured in 7-bit "characters", since we were not yet storing full 8-bit bytes.
- By today's standards, having such a large box to hold a few GB might seem amusing, but at the time, this unit was four times the capacity as its predecessor, the IBM 3350 DASD. Compare that with our first disk system, the IBM 350 Disk Storage Unit, introduced in 1956, that stored only 5 million characters (5MB) and was the size of two refrigerators.
- The term "DASD", pronounced daz-dee, was used as some earlier devices were based on magnetic drums or strips of magnetic tape. Today, DASD is still a common term for disk systems among mainframe administrators.
- The 3380 was also twice as fast as the IBM 3350, at 3 million characters per second (3 MB/sec). The irony was thatthe mainframe servers could not keep up, so a Speed Matching Buffer feature was invented to slow it down to half-speed, when used with certain models of mainframe.
As for the dimensions, I too had a hard time finding a publicly available resource that listed 3380 dimensions,so I searched internal IBM resources, and finally, asked someone over in the next building just to measure one ofthe 3380K models we still have in the Tucson test lab floor. The dimensions are ... (drumroll please)
- 70 inches (1778mm) tall
- 44 inches (1117mm) wide
- 32 inches (812mm) deep
The result is that the box could actually hold a much more impressive 52,500 of the new Toshiba drives, twicethe original, albeit conservative, estimate. Before anyone"tries this at home", however, keep in mind that around each Toshiba drive,as with any ATA drive, you need to have all the electronics to communicate to the outside world, and provide cooling. Running tens of thousands of these little guys in the spaceof 60 square feet would probably melt the floor or set off your smoke alarm system.
At least take a backup first.
technorati tags: Chris Evans, Toshiba, IBM, 3380, DASD, CJ2, 3350, ATA
Chris Anderson, of Wired magazine, wrote a great article called The Long Tail
This article became a book by the same name published earlier this year, and I just discovered it on a recent visit to Second Life. A lot of IBMers are now alsoSecond Lifers, and I suspect it is just a matter of time before we are conductingour customer briefings there, and getting our year-end bonuses paid directly in Linden bucks.(Those of you not familiar with Second Life can watch this 3-minute video fromthe folks at Text100)
Anyways, the Long Tail describes the new economy of entertainment thanks to digitalstorage. Here are some of the key insights.
- In the past, entertainment was all about hits: hit songs, hit movies,hit novels, and this was primarily because of the economic realities restricted byphysical space. Chris writes: "An average movie theater will not show a film unless it can attract at least 1,500 people over a two-week run; that's essentially the rent for a screen. An average record store needs to sell at least two copies of a CD per year to make it worth carrying; that's the rent for a half inch of shelf space."
- Things have changed. To drive the point home, Robbie Vann-Adibe (CEO of eCast), poses the trick question"What percentage of the top 10,000 titles in any online media store (Netflix, iTunes, Amazon, or any other) will rent or sell at least once a month?" The answer will surprise you. Write down your guess first, then go read here. His digital jukeboxes are able to play from a list of150,000 songs, not the few hundred you'd find at the Tap Room which is rated as having the best jukebox in Tucson.
- The phenomenon is not just limited to music. "Take books," Chris writes, "The average Barnes & Noble carries 130,000 titles. Yet more than half of Amazon's book sales come from outside its top 130,000 titles. Consider the implication: If the Amazon statistics are any guide, the market for books that are not even sold in the average bookstore is larger than the market for those that are..."
This has incredible implications for the storage industry. For one, content providers are going to dig deep into their archives to digitize and deliver "long tail" offerings. If they don't have a deep archive, many will start to build one. Second, the need to search through that large volume of content will become more critical. Classifying and indexing with the appropriate tags and metadata will be an important task.
technorati tags: Chris Anderson, Wired, magazine, IBM, Secondlife, Linden bucks, Text100, Long Tail, Robbie+Vann-Adibe, eCast, NetFlix, iTunes, Amazon, Tap Room, Barnes Noble, deep, archive, metadata, tags
Last week, Steve Jobs demonstrated the latest evidence that theinmates are running the asylum
over at Apple
I wasn't at the event, but thought it would be good to explain some basic concepts ofInformation Lifecycle Management (ILM),using the files on my iPod as an example. (Disclosure: IBM makes the technology inside many of Apple's computers, and so IBMers get to buy Appleproducts at employee prices. I own a Mac Mini based on IBM's POWER4 processor, and an iPod Photo 60GB model).
I have 20,000 MP3 music files, representing 106GB of data. This fits nicely on my 250GB external disk system attached to my Mac Mini, but won't all fit on my little 60GB iPod. I needed a way to decide what music I keep on bothmy iPod and Mac Mini, and which I keep only on my Mac Mini. When I am traveling, I am able to listen only to the musicin the first group, but when I am at home, I am able to listen to all my music in both groups.(Another disclosure: I use my Tivo connected to my LAN to play all my MP3 music through my home stereo system.I had my entire house wired with Cat5 to make this possible.)
Apple's iTunes software lets me decide which MP3 files are copied to my iPod using "playlists". A playlist is a list of songs. Fixed playlists are created manually, each song copied to its list in a specific order. Smart playlists are createdautomatically, via policy. I give it the criteria, and it finds the songs for me. If I import a new music CD,none of the songs will be added to any fixed playlists, but could be added to my smart playlists if I set the policiescorrectly. Apple iTunes supports both "include" and "exclude" methodologies.
I use primarily smart playlists, based on genre and rating. I have tried to keep the number of genre down to a small manageable list:
Of course, what I have for genre may not match what's in theGracenote database, so I sometimes have to makeupdates to match my convention. I've picked these based on my different "applications" for my music. For example, I listen to Ambient music to help me fall asleep on airplanes, but Rock when I exercise at the gym.
- Rhythm & Blues
- Hip Hop
Next, I use the ratings from one to five stars. The advantage to the rating is that I can change them on-the-fly directly on my iPod. All other "metadata" has to be entered only from the keyboard of my Mac Mini.
|*||Files for Mac Mini only, not copied to my iPod|
|**||Non-mix, copied to my iPod, but typically spoken words, such as language lessons|
|***||Mix, music to include in my music mixes|
|****||Keep on my iPod, but re-evaluate|
So, I have five smart playlists, "One Star", "Two Stars", etc. for each rating, and have decidedto keep only the 2, 3, 4 and 5 star songs on my iPod, by simply putting check marks on those playlists to copythem over. I have about 50 songs with 5 stars, and 8000 with 3 stars, and the rest in the other categories,leaving me a few GB to spare.
I also have playlists for each genre, "Rock mix", "Pop Mix", "Ambient Mix", etc. where I have selected thosethat match the genre, AND have 3, 4 or 5 stars. In this manner, I can listen to a mix. If I find a song mis-classified for that genre, I change it to four stars, which serves as myreminder to re-evaluate when I am back at home on my Mac Mini. If I don't want a song in my mix, I just lowerit to 2 stars. I want it off my iPod altogether, I lower it to one star.
This method is simple enough, and allows me to enjoy my music right away, and more effectively, without having to wait for completely finishing my classification process.
Next week, I'm traveling to Africa (purely vacation, not related to my job, my senator, or myinvolvement in anycharitable organizations). My Canon camera has only a 1GB IBM Microdrive, but I am able to offloadmy pictures to my iPod, connected via USB cable, and review the pictures on the little 2-inch screen. By simply "unchecking" my 2-star and 3-starplaylists, and checking only those mixes I plan to take with me, I was able to clear 17GB of space, plenty ofroom for all my photos of elephants and giraffes, but still plenty of music to listen to. Thanks to my simple methodology, I was able to do this with minimal effort, and willhave no problem putting all my music back when I return.
When evaluating an ILM process, many people are overwhelmed by their fear of the classification process, when in reality it doesn't have to be so complicated.
Is there an "iTunes" for the storage in your datacenter? Yes! It's called IBM TotalStorage Productivity Center. It can help you list and classify all the files in your IT environment,including files in your internal disks inside the servers, your NAS and SAN external disk systems, across both IBM and non-IBM hardware.It's a good thing to consider as part of your overall ILM strategy.
technorati tags: Apple, Steve Jobs, inmates, running, asylum, IBM, information, lifecycle, management, iPod, music, genre, star, rating, iTunes, datacenter, TotalStorage, Productivity Center, NAS, SAN, ILM[Read More]
Welcome to my new blog on IBM Developerworks!
I am Tony Pearson, IBM brand marketing strategy, located in Tucson, Arizona. I have degrees in Computer Engineering and Electrical Engineering from the University of Arizona. Over the past 20 years, I have worked in a variety of storage roles, including development projects, product and portfolio management, testing, field support, and now bring that technical experience to marketing.
There are a lot of things to discuss related to storage, and I am never short of opinions. As such, the standard IBM disclaimer applies: “The postings on this site solely reflect the personal views of the author and do not necessarily represent the views, positions, strategies or opinions of IBM or IBM management.”
I have invited other IBMers to post their opinions, and when they do, their opinions may not necessarily match mine either.
This is an open two-way conversation between IBM, Business Partners, Independent Software Vendors, prospects and existing clients. I encourage everyone to post comments about our products, services, and marketing efforts.
IBM Senior Certified Executive IT Architect
| || |
On April 2nd IBM announced several key enhancements across the Storwize V5000 portfolio.
The models for the V5000 now include V5010E, 5030E and the V5100.
The V5010E includes two single socket Broadwell DE 2.2GHz, 2 core processor canisters. Each canister supports a maximum of 32GB of RAM.
The V5030E includes the Broadwell DE 1.9GHz, 6 core processor in its two canisters. Each canister supports a maximum of 32GB of RAM.
The V5100 boasts a single Skylake 1.7Ghz processor with 8 cores in each canisters. RAM is increased to a total of 576GB for the entire controller, or 288GB maximum per canister.
From a scale out perspective the V5010E model supports a single controller configuration, while the V5030E and V5100 both support up to two controller clusters. This provides for a maximum of 392 drives in the V5010E and a massive 1520 drives in either the V5030E or V5100 dual controller clusters.
While the new lineup for the V5000 is impressive; regarding the quantity of drives, and the storage available per model will blow your mind.
The V5000E & V5100 versions support the following drive types.
|Storwize V5000E & V5100 - SAS Drive Support|
|Tier 1 Flash||-||1.9TB||3.84TB||7.68TB||15.36TB|
|High Performance, Enterprise Class Disk Drives||10000 rpm||900GB||1.2TB||1.8TB||2.4TB|
|High-Capacity, Archival-Class Nearline Disk Drives||7200 rpm||2TB||-||-||-|
|High-Capacity, Archival-Class Nearline Disk Drives||7200 rpm||4TB||6TB||8TB||10TB||12TB|
Along with the compute & cache layer enhancements across the V5000 platform, the V5100 receieved a large boost in its storage support. IBM has delivered a solid Flash / Hybrid storage controller player to the markey. In the V5100 model we now support both the Flash Core Modules (FCM) and or NVMe Industry Standard SSD drives. In addition in the expansion drawers you can add even more SAS SSD, 10K and NL-SAS capacity.
The combination of Flash Core Modules and NVMe Industry Standard drives makes the V5100 the perfect low to medium class storage controller.
|2.5" (LFF)||Hardware Compressing||Capacities|| || || || |
|IBM Flash Core Modules||Yes||4.8TB||9.6TB||19.2TB|| |
|Industry Standard NVMe Flash Drives||No||1.92TB||3.84TB||7.68TB||15.36TB|
Both the V5010/30E and the V5100 models offer the following on board interface ports per canister:
|1GbE||1x1GbE techport + iSCSI||1x1GbE dedicated tech port||1x1GbE dedicated tech port|
|1GbE||1x1GbE iSCSI only rpm|| || || || |
|10GbE||n/a||2x10GbE (iSCSI only)||4x10GbE (iSCSI only)|
|12G SAS||1xSAS expansion||2xSAS expansion|| |
The Storwize V5000E models support 1 additional interface card, or 1 SAS host attachment adapter per canister. The optional interface cards available are:
|Storwize V5000 Supported Adapter Cards|
|4 port 16Gb Fibre Channel only|
|2 port 25GbE iSCSI only|
|2 port 12Gb SAS host attach|
The Storwize V5100 model also supports 1 host interface card, and also optionally up to two SAS Storage Expansion adapters. The adapter interface cards available for the V5100 are:
|Storwize V5000 Supported Adapter Cards|
|4 port 16Gb Fibre Channel, FC NVMeOF|
|2 port 25GbE ROCE ISER, iSCSI|
|2 port 25GbE iWARP ISER, iSCSI|
|2 port 12Gb SAS to allow SAS Expansions|
Regarding the software features for the V5000E and V5100 models, all of the models discussed still include the following features as shown. Specifics on the actual software features is documented below:
|Storwize V5000 V5100 Software Features|
|DRP Compression||No||Yes||License Per Enclosure||*Required minimum 32GB RAM|
|Hardware Encryption||No||DFSA License Key||DFSA License Key|| |
|Software Encryption||No||No||DFSA License Key (for use on virtual storage)|| |
|DRAID - Distributed RAID||DRAID 6 recommended / DRAID 5 supported||DRAID 6 recommended / DRAID 5 supported||DRAID 6 recommended / DRAID 5 supported|| |
|TRAID - Traditional RAID||R0, 1 & 10||R0, 1 & 10||R0, 1 & 10|| |
|External Virtualization||Data Migration Only||Data Migration Only||License Per Enclosure|| |
|VVols - Supported with Spectrum Connect||Yes||Yes||Yes|| |
|Flash Copy||DFSA License Key||DFSA License Key for each IO Group||License Per Enclosure|| |
|Remote Copy||DFSA License Key||DFSA License Key for each IO Group||License Per Enclosure|| |
|Spectrum Virtualize 8.3||Yes||Yes||Yes|| |
|EasyTier||DFSA License Key||DFSA License Key for each IO Group||License Per Enclosure|| |
|Thin Provisioning||Yes||Yes||Yes|| |
|Volume Mirroring||Yes||Yes||Yes|| |
|Secure Disable USB Ports||Yes||Yes||Yes|| |
|T10 Diff||Yes||Yes||Yes|| |
|DRP (Data Reduction Pools)||Yes||Yes||Yes|| |
|Data Deduplication||Yes||Yes||Yes|| |
Reviewing this information has been eye awaking and though valuable I have only touched on pieces of this latest update for IBM storage.
For those seeking an even deeper depth of knowledge I refer you to my learned colleague Barry Whites blog,
Visit Barry's blog [Storage Virtualization] for more details.
Leaving a discussion like this and not providing a review across the Storwize Block Family would not be welcomed by you my readers.
Here is a table reflecting the features by model for your quick reference.
IBM Storage Built with IBM Spectrum Virtualize
|Storwize V5010E||Storwize V5030E||Storwize V5100/F||Storwize V7000 Gen3||Storwize V9100 Family|
|3-Site Data Copies (In Combination with IBM Spectrum Copy Data Management)|
|Local and remote replication (snapshots, DR, copy/migrate to cloud)|
|Easy Tier AI-driven automated tiering|
|Transparent Data Migration|
| ||Data Reduction Pools|
| ||Scale-out Clustering|
| ||HyperSwap High Availability|
| || ||NVMe Flash and FC-NVMe Host Connections|
| || ||Penalty-Free Compression and Encryption with IBM FlashCore Modules|
| || ||External Storage Virtualization of more than 450 storage systems|
| || ||Storage Class Memory Capable|
Below is a chart reflecting details on the IBM FlashWatch Program by Storwize Model
|Product||100% Availability Guarantee||Data Reduction Guarantee||Controller Upgrade Program||
7-year 24x7 Support
|Flash Endurance||Data Migration|
|Storwize V7000 Gen3||Yes||Yes||Yes||Yes||Yes||Yes|
To learn more, see [Meet the new IBM Storwize V5000] YouTube video, and the [IBM Storwize V5000E models], [IBM Storwize V5100], [IBM SAN Volume Controller, IBM FlashSystem 9100, and IBM Storwize families offer new drive options] announcement letters.
Watch for my next blog write up soon.
Well, it's Tuesday again, and you know what that means? IBM Announcements!
(This week I am in Pennsylvania and New York speaking to clients. The weather this week has not been cooperative!)
- Spectrum Protect Plus 10.1.2
Just in time for the upcoming VMworld conference, IBM announces the following features added to Spectrum Protect Plus, a snapshot-based backup software for VMware, Hyper-V and databases.
- Data-at-Rest Encryption for local backups stored in the vSnap repository
- IBM Db2 support with point-in-time recovery
- VMware vSphere 6.7 support
- Dashboard enhancements
- Alerting for backup and restore jobs and storage thresholds limits
- Drill-down capabilities for dashboard widgets
- Spectrum Protect 8.1.6
IBM also continues to enhance its traditional file-based backup product. Here are some of the features:
- Tier data by backup state for container pools. When you have multiple backup versions, the most recent version is called the "active", the older versions are called "inactive" versions. Rarely do you recover inactive versions, so this feature allows them to be migrated off to object or cloud storage.
- Ransomware detection for Virtual Environment workloads. This is an enhancement of the "Ransomware detection" introduced earlier this year, but for VMware and Hyper-V images.
- IBM DS8882F All-Flash Array
When IBM announced the DS8880, it shocked folks that it changed them from the previous 33-inch wide, to a standard 19-inch width. The IBM Z team followed up with 19-inch wide models of its mainframe servers.
Now, IBM can bring these together. There are two flavors of the new DS8882F:
- The "Rackless" model is 17U in height with the optional keyboard/monitor, and can be put into existing 19-inch racks. These can be used with VMware, Linux, Windows, AIX and z/OS.
- The "Flex Frame" model, which is 16U, allowing it to fit nicely inside a single-rack IBM Z Z14 ZR1 model, or LinuxOne RockHopper II model. It is 16U instead of 17U because it shares the existing 1U-high keyboard/monitor unit.
Like the DS8888F, DS8886F, and DS8884F models, the new DS8882F uses the High Performance Flash Enclosure (HPFE) gen2 drawers, supporting either high-performance/high-endurance drives (400GB to 3.2TB each), or high-capacity/standard-endurance drives (3.8TB to 15.3 TB each).
The R8.5 release of firmware that accompanies this announcement also supports data-in-flight encryption for Transparent Cloud Tiering. It also supports a new feature called "Safeguarded Copies", up to 500 copies to protect against hackers and ransomware.
IBM Spectrum Access blueprints have been extended to support IBM Z and LinuxOne. These blueprints show how to run IBM Cloud Private with Spectrum Connect with IBM block storage, including IBM DS8880/F, SVC, Storwize and FlashSystem models.
- IBM Storage Solutions for Virtual Desktop Infrastructures (VDI)
IBM offers a new blueprint to configure Virtual Desktops with its newly announced IBM FlashSystem 9100 device. The low latency/high IOPS capability of the FlashSystem 9100 is perfect for the type of "boot storms" that are often encountered with VDI deployments.
- IBM Spectrum Scale 5.0.2 and Elastic Storage Server
At recent IBM Technical University, I joked that the IBM Elastic Storage Server is only "part of a complete breakfast" because it only supported the NSD POSIX interface. To make it useful in most situations, you needed to buy additional servers outside of the ESS to run Spectrum Scale protocol nodes to provide industry-standard file and object protocols.
Today, IBM announced that you can order a new "IBM Elastic Storage Server Data Server" (5148-22L) which is a POWER server with the Spectrum Scale software pre-installed for protocol node support. It has [similar specifications] to the IBM Elastic Storage Server Management Server (5148-21L).
If you prefer to run Spectrum Scale in the cloud, you can "Bring your own license" (BYOL) to Amazon Web Services.
To learn more, see [IBM Spectrum Scale Protocol Node Server] press release.
- IBM Cloud Object Storage
IBM Cloud Object Storage is software that can be deployed on a variety of x86 servers. Today, IBM announces additional servers qualified for use with this software, including 1U-high Quanta servers.
I will not be attending VMworld this year. If you attend, send me pictures!
This week, I was part of an all-day event called "Healthcare and Research Trends & Directions in a Cognitive World" at the IBM Executive Briefing Center (EBC) in Rochester, MN. I was one of many presenters covering Information Technology to improve healthcare outcomes. Todd Stacy, IBM Director Server Sales for US Public Market, served as our emcee.
This was a great day. Special thanks to Kathy Lehr, Trish Froeschle, and Scott Gass for organizing this event! We had clients from a variety of Health Care and Life Science industry backgrounds. I certainly learned a few things myself.
technorati tags: IBM, Michael Weiner, Todd Stacy, Watson Health, Obesity Epidemic, Aging Population, HITECH, Greg Tevis, Raj Tandon, Spectrum Storage, Spectrum Virtualize, Spectrum Control, Spectrum Protect, Spectrum Scale, Spectrum Copy Data Management, IBM Cloud Object Storage, EPIC, Cache Database, Jason Crites, Wayland Vacek, IBM Merge, HIMSS, Jane Yu, Frank Lee, Sidra Medical Research Center, Qatar, Spectrum Compute, VersaStack, FlashSystem, Kathy Lehr, Trish Froeschle, Scott Gass
- The Cognitive Healthcare Organization
Dr. Michael Weiner, IBM Chief Medical Information Officer, Watson Health, covered some of the real challenges not just facing the United States, but also other countries. On average, healthcare in USA [costs over $10,000 USD per American citizen]! Compare that to only $3,700 USD for the folks in the United Kingdom! In fact, nearly all industrial nations spend between $2,000 and $5,000 per person. Where does all the U.S. money go?
- Aging Population
A big challenge is our ever-aging population. Every day, there are 10,000 [Baby Boomers] reaching their 65th birthday, with fewer people in the 25-44 age group to work as nurses to take care of them. About 15 percent of the US population are elderly (over age 65) and this is expected to grow to 20 percent in year 2040. The situation is even worse in Japan, where 25 percent of the population today is elderly, and this is expected to be 40 percent by year 2060.
- New Care Models
In some countries, like Australia and Japan, post office workers who spent their time delivering mail, now can stop in to check in on elderly people. As people ship less mail, using social media or email instead, this keeps the postal workers employed, in a manner that provides society value.
The USA enjoys one of the lowest costs for food, but then suffers from an epidemic of obesity, with over 34 percent of Americans are obese. When New York City eliminated Trans Fats, heart attacks dropped considerably.
In 2009, the Health Information Technology for Economic and Clinical Health [HITECH] Act required the digitization of medical information, known as "Meaningful Use", which has greatly influenced healthcare facilities. This was implemented by a combination of incentives and penalties. Now, more than than 92 percent of hospitals in the USA have digitized medical information! The rest are still using paper and Xray film images. Some places were initially exempted, such as Assisted Living Homes for example, so there is still more work to be done.
- New Technology
An advantage of using computer-based solutions like Artificial Intelligence is that it eliminates bias. When a woman walks into an Emergency Room complaining about chest pains, few health staff would consider this a sign of heart attack. When a man does same, health staff considers heart attack as the first diagnosis, at the risk of missing out on other possibilities.
Every year, over a million articles related to healthcare research are published. Who can read all this in a timely manner? IBM Watson! After [winning in Jeopardy], IBM Watson was "sent to medical school" to learn how to assist doctors in diagnosing patients.
- Transforming Health Care Data Management with IBM Spectrum Storage
Greg Tevis, IBM Software Defined Storage Architect, and Raj Tandon, IBM Senior Strategist, co-presented this introduction to IBM Spectrum Storage family of products. They covered examples with IBM Spectrum Virtualize, IBM Spectrum Control, IBM Spectrum Protect, IBM Spectrum Scale, IBM Cloud Object Storage, and IBM Copy Data Management. The latter having support directly for EPIC and Cache databases.
- Cognitive Imaging Solutions for Healthcare Providers
Jason Crites, IBM Healthcare and Life Sciences Data Solutions Leader, and Wayland Vacek, Enterprise Sales Manager for Merge, presented IBM Watson Imaging Clinical Review, from IBM's acquisition of the Merge company. The solution is based on IBM Spectrum Scale as the back-end storage repository.
Merge has been around for more than 20 years, with clinical workflow offerings in Cardiology, Radiology, Orthopedics and Eye care. Often, IBM Watson is able to identify things in medical images that escape the review or radiologists or other medical specialists.
At HIMSS conference earlier this year, The human radiologists were shown a collection of images used to train IBM Watson. The human radiologists only identified 20 percent of the images correctly, while IBM Watson got all of them, every time. In many cases, human radiologists have only a few seconds to look at an Xray image. Computers like IBM Watson are now fast enough to compete directly with human radiologists in the same number of seconds.
- Building a Foundation for the Cognitive Era in Healthcare and Life Sciences
Dr. Jane Yu, IBM Systems Architect, Healthcare & Life Sciences, and Dr. Frank Lee, IBM Global Sales Leader, IBM Software Defined Infrastructure & Life Sciences, co-presented this topic. They present five challenges:
- Growing data volumes are making it more difficult to manage, process and store this data.
- Scientists find themselves spending more than 80 percent of their time manually integrating data from silos, and less than 20 percent of their time doing actual research and deriving insights from their analyses.
- Compute- and data-intensive workflows may take days to complete on existing server and storage systems.
- IT organizations must keep up with rapidly evolving applications, development frameworks, and databases for preferred. Health care Life Science (HCLS) applications. This includes SAS, Matlab, Hadoop, Spark, NoSQL databases, as well as Deep Learning and Machine Learning workloads.
- Scientific integrity and government mandates increasingly require collaboration across organizational boundaries.
In one example, Sidra Medical and Research Center plans to map the genomes of all 250,000 citizens in the Middle Eastern country of Qatar. Imagine that processing each Qatari citizen will generate 200 GB of data for this project, resulting in 50 Petabytes (PB) of data!
Combining IBM Spectrum Compute products with IBM Spectrum Scale storage, can help address these challenges.
- Modernize & Transform Helathcare with IBM Storage Solutions
Finally, I presented a 90-minute breakout session that covered three solution areas:
- Flash storage to speed up medical records and research. Those who have already implemented Electronic Health Records (EHR) for "Meaningful Use" compliance recognize the value this provides to improving healthcare. Adding All-Flash Arrays such as IBM FlashSystem, Storwize V7000F or DS8000F can drastically improve application performance.
- Spectrum Scale and IBM Cloud Object Storage for Vendor Neutral Archive. It seems silly that each PACS vendor has its own little island of storage. A better approach is to send all PACS data from various vendors into a "Vendor-Neutral" storage repository. Both IBM Spectrum Scale and IBM Cloud Object Storage System, either linked together or used separately, can be part of a VNA solution.
- VersaStack to simplify deployments. VersaStack is a Converged System that combines best-of-breed Cisco servers and switches with best-of-breed IBM storage, pre-cabled, pre-configured, and pre-loaded with all the necessary software to manage the environment as a single entity. This can reduce the time it takes to deploy new medical applications from weeks to just hours.
Last November, fellow blogger Chris Mellor from The Register wrote an interesting article [EMC crying two SAN breakup tears].
(Back in 2010, I poked fun at EMC with my post [VPLEX: EMC's Latest Wheel is Round]. I pointed out that EMC's announcement of "new features" that already existed in IBM's SAN Volume Controller. Oops! They did it again!)
Basically, Dell EMC is working on a new "2 Tiers" approach that combines high-performance flash tier with high-capacity object storage. Guess what? IBM already offers this! Why wait?
IBM Spectrum Scale, formerly known as the General Parallel File System (GPFS), supports POSIX, HDFS, OpenStack Swift, Amazon S3, NFS, SMB and iSCSI protocols.
Spectrum Scale can provide this front-end abstraction layer between flash and object storage, including IBM Cloud Object Storage system and IBM Bluemix (formerly SoftLayer) cloud services.
But why limit yourself to just two tiers? IBM Spectrum Scale can also support 15K, 10K and 7200 RPM spinning disk drive tiers, as well as virtual or physical tape tier, the ultimate low-cost high-capacity tier!
Several years ago, IBM coined the phrase "FLAPE" to discuss the two-tier approach of combining Flash with Tape using Spectrum Scale as the front-end abstraction layer.
Perhaps we should call combinations of Flash and Object "FLobject" storage? If the name catches on, you read it here first!
technorati tags: IBM, Chris Mellor, The Register, Dell EMC, 2 Tiers, Spectrum Scale, GPFS, POSIX, HDFS, OpenStack Swift, Amazon S3, NFS, SMB, iSCSI, IBM Cloud Object Storage, IBM Bluemix, IBM SoftLayer, tape, FLAPE, FLobject
Last month, I had the pleasure to help train Watson in its latest mission, to help answer questions from sellers, this are not just for the IBM feet on the street, but also for IBM distributors and IBM Business Partners as well.
In their post [Workers Spend Too Much Time Searching for Information], Cottrill Research explains the problem all too well. Here is an excerpt:
"... [survey by SearchYourCloud] revealed 'workers took up to 8 searches to find the right document and information.' Here are a few other statistics that help tell the tale of information overload and wasted time spent searching for correct information -- either external or internal:
- 'According to a McKinsey report, employees spend 1.8 hours every day -- 9.3 hours per week, on average -- searching and gathering information. Put another way, businesses hire 5 employees but only 4 show up to work; the fifth is off searching for answers, but not contributing any value.' Source: [Time Searching for Information]
- '19.8 percent of business time -- the equivalent of one day per working week -- is wasted by employees searching for information to do their job effectively,' according to Interact. Source: [A Fifth of Business Time is Wasted]
- IDC data shows that 'the knowledge worker spends about 2.5 hours per day, or roughly 30 percent of the workday, searching for information ... 60 percent [of company executives] felt that time constraints and lack of understanding of how to find information were preventing their employees from finding the information they needed.' Source: [Information: The Lifeblood of the Enterprise]."
In the early days of the Internet, before search engines like Google or Bing, I competed in [Internet Scavenger Hunts]. A dozen or more contestants would be in a room, and would be given a list of 20 questions to find answers for. Each of us would then hunt down answers on the Internet. The person to find the most documented answers before time runs out wins. It was quite the challenge!
Over the years, I have honed my skills as a [Search Ninja]. With over 30 years of experience in IBM Storage, many sellers come to me for answers. Sometimes sellers are just too lazy to look for the answers themselves, too busy trying to meet client deadlines, or too green to know where to look.
A good portion of my 60-hour week is spent helping sellers find the answers they are looking for. Sometimes I dig into the [SSIC], product data sheets, or various IBM Redbooks.
Other times, I would confer with experts, engineers and architects in particular development teams. Often, I learn something new myself. In a few cases, I have turned some questions into ideas for blog posts!
It was no surprise when I was asked to help train Watson for the new "Systems SmartSeller" tool. This will be a tool that runs on smartphones or desktops to help answer questions that sellers might need to respond to RFP or other client queries.
The premise was simple. Treat Watson as a student at "Cognitive University" taking classes from dozens of IBM professors, in a series of semesters, or "phases".
Phase I involved building the "Corpus", the set of documents related to z Systems, POWER systems, Storage and SDI solutions; and a "Grading Tool" that would be used as the Graphical User Interface. I was not involved in phase I.
Phase II was where I came in. Hundreds of questions are categorized by product area. I worked on 500 questions for storage. For each question, Watson had up to eleven different responses, typically a paragraph from the Corpus. My job as a professor was to grade the responses to some 500 storage questions:
|★ (one star)
||Irrelevant, answer not even storage-related
|★★ (two stars)
||Relevant, at least it is storage-related, but does not answer the question, or answers it poorly
|★★★ (three stars)
||Relevant, adequately answers the question
|★★★★ (four stars)
||Relevant, answers the question well
Most of the answers were either 1-star (not storage related) or 2-star (mentioned storage, but poor response). I would search through the existing Corpus looking for a better answer, and at best found only 3-star responses, which I would add to the list and grade as a 3-star response.
I then searched the Internet for better answers. Once I found a good match, I would type up a 4-star response, add it to the list, and point it to the appropriate resources on the Web.
Other professors, who were also looking at these questions, would then get to grade my suggested responses as well. Watson would learn based on the consensus of how appropriate and accurate each response was graded.
I don't know where the Cognitive University team got some of the questions, but they were quite representative of the ones I get every week. In some cases, the seller didn't understand the question he heard from the client, making it difficult for me to figure out what they were actually asking for.
It reminds me of that parlor game ["Telephone" or "Chinese Whispers"], in which one person whispers a message to the ear of the next person through a line of people until the last player announces the message to the entire group. I have actually played this at an IBM event in China!
Watson needs to parse the question into nouns and verbs, and use that Natural Linguistic Programming (NLP) to then search the Corpus for appropriate answer. I determined three challenges for Watson in this case:
- The questions are not always fully formed sentences. For example, "Object storage?" Is this asking what is object storage in general, or rather what does IBM offer in this area?
- The questions often do not spell the names of products correctly, or use informal abbreviations. "Can Store-wise V7 do RtC?" is a typical example, short for "Can the IBM Storwize V7000 storage controller perform Real-time Compression?"
- The questions ask what is planned in the future. "When will IBM offer feature x in product y?" I am sorry, but Watson is not [Zoltar, the fortune teller]!
I managed to grade the responses in the two weeks we were given. Part of my frustration was the grading tool itself was a bit buggy, and I spent some time trying to track down some of its flaws.
The next phase is in late January and February. This will give the Cognitive University team a chance to update the Corpus, improve the grading interface, and find more professors and different set of questions. I volunteered the most recent four years' worth of my blog posts to be added to the Corpus.
Maybe this tool will help me turn my 60-hour week back to the 40-hour week it should be!
technorati tags: IBM, Watson, Cottrill Research, SearchYourCloud, McKinsey, IDC, Google, Bing, Search Ninja, Internet Scavenger Hunts, SSIC, Telephone Game, Chinese Whispers, NLP, RFP, Storwize, RtC, Zoltar, Cognitive University
Well, it's Tuesday again, and you know what that means? IBM Announcements!
(OK, yes, today is Friday, but I was busy getting married on Tuesday, so IBM pushed the announcements out one day to Wednesday, and technically I am writing this blog post during my honeymoon vacation, so the IBM marketing team and my new wife both cut me some slack. Work/Life balance is all about compromises, right?)
- IBM DS8880 Storage System
The IBM DS8880 comes in three models, the DS8884 entry level, the DS8886 enterprise level, and the DS8888 all-flash array. IBM offers 1, 2, 3 and 4 year warranties.
The new High Performance Flash Enclosure (HPFE) Gen2 delivers more capacity than Gen1. The 2U flash enclosures are configured in pairs with each enclosure supporting up to twenty-four 2.5-inch flash cards in capacities 400 GB, 800 GB, 1.6 TB and 3.2 TB.
The HPFE Gen2 are currently available for both the DS8884 and DS8886 models. The maximum flash capacity for the DS8886 increases from 96 TB to 614.4 TB, delivering reduced storage costs through lesser cost per IOPS with this new flash enclosure. IBM has made a statement of direction to offer these HPFE Gen2 on the DS8888 as well.
To improve security, IBM DS8880 now supports customer-defined digital certificates for authentication, and configurable Hardware Management Console (HMC) firewall support.
For IBM's mainframe clients, IBM now offers "Extents-level" space release support for z/OS®, DSCLI (Command Line Interface) support for z/OS environment, and FICON® Information Unit (IU) pacing improvements.
My blog post [Re-Evaluating RAID-5 and RAID-6 for slower larger drives] helped to convince upper management to make RAID-6 the default protection level in R8.2 release.
To learn more, see [ IBM DS8880 storage family delivers a series of flash-enclosure models] press release.
- IBM Spectrum Virtualize™ V7.8
IBM Spectrum Virtualize™ V7.8 delivers support for the latest SAN Volume Controller, FlashSystem V9000 and Storwize® product family, and adds new software functionality and improvements
In conjunction with [IBM Spectrum Copy Data Management], Spectrum Virtualize v7.8 offers flexible data protection with transparent cloud tiering to leverage the cloud as FlashCopy targets and restore these snapshots from the cloud on select platforms.
In my September blog post [IBM Edge 2016 Day 3 Wednesday Breakout Sessions] I gave a quick recap of how IBM Spectrum Virtualize offers data-at-rest encryption for both internal and external drives.
However, the encryption keys are kept on USB thumb drives, which are either left in the USB ports on the back of the hardware, or locked away in a safe, only to be retrieved as needed when rebooting the systems or upgrading the firmware.
Now, IBM Spectrum Virtualize v7.8 supports the IBM Security Key Lifecycle Manager (SKLM) to manage encryption keys. IBM continues to support USB thumb drives if you prefer, but SKLM is used to manage keys for most of the rest of IBM products, and provides centralized management.
To learn more, see [ IBM Spectrum Virtualize Software V7.8] press release.
- IBM SAN Volume Controller and Storwize
The SVC and Storwize models can directly attach via 12Gb SAS to expansion drawers. At the time, we supported 2U-high 12-bay that support Large Form Factor (LFF) 3.5-inch Nearline (7200 rpm) drives, and 2U-high 24-bay that support the Small Form Factor (SFF) 2.5-inch drives (SSD, 15K, 10K and 7200 rpm).
With Spectrum Virtualize v7.8, IBM now offers a third option, the 5U-high 92-bay that supports both LFF and SFF drives. This new expansion can be attached to Storwize V5000 Gen2, Storwize V7000 (models 524/Gen2 and 624/Gen2+), and SVC (models DH8 and SV1).
For the 12-bay and 92-bay, IBM now supports 10TB capacity 3.5-inch Nearline drives. For the 24-bay and 92-bay, IBM now supports 7.68 TB and 15.36 TB capacity Solid State Drives (SSD).
For those concerned about the phrase "lower endurance" in the press release, let me explain. SSD have a bit of extra capacity included. If you write the full capacity of the drive every day for a year, you will "burn up" about one percent of the capacity.
To handle ten "Full Drive Writes per Day" (10 FDWP) over the course of five years, IBM adds 50 percent extra spare capacity above the 400 GB, 800 GB, 1.6 TB and 3.2 TB capacities. So, a 400GB full-endurance drive is really 600 GB inside. These were sometimes referred to as "Enterprise" SSD.
For the larger device sizes, the IT industry has determined that 1 FDWP is sufficient, so instead of 50 percent spare capacity, IBM adds only 5 percent extra. The 7.68 TB is really 8.06 TB inside. These were earlier referred to as "Read-Intensive" SSD. These come in 1.92 TB, 3.84 TB, 7.68 TB and 15.36 TB capacities.
IBM is also offering non-disruptive model conversions. Storwize V5010 can now be converted to V5020, and V5020 can be converted to V5030. The Storwize V7000 Model 524 (Gen2) can be converted to model 624 (Gen2+).
To learn more, see [ IBM SAN Volume Controller and Storwize family high-density expansion] press release.
- IBM FlashSystem V9000
The IBM FlashSystem V9000 will also support its own version of 5U-high, 92-bay, but to simplify ordering, will only support the following drive types:
- High-capacity, archival-class Nearline disk drives in 8 TB and 10 TB 7,200 rpm
- Flash drives in 1.92 TB, 3.84 TB, 7.68 TB, and 15.36 TB
To learn more, see [ IBM FlashSystem V9000 HD Expansion Enclosure Model 92F] press release.
- IBM DeepFlash Elastic Storage Server (ESS)
The DeepFlash 150 is the perfect JBOF addition to the ESS family. The current ESS models had either 2U-high 24-drive bays, or 4U-high 60-drive bays. This new model is 3U-high with 64 high-capacity (8 TB) Board Solid State Drives (BSSD).
The ESS includes all the features of IBM Spectrum Scale, including both 8+2 and 8+3 Erasure Coding data protection. This provides file and object access to data, including POSIX compliance for Windows, Linux and AIX operating systems, as well as HDFS-compliant access for big data analytics.
To learn more, see [IBM DeepFlash Elastic Storage Server] landing page.
By now, there are multitude news articles on these announcements, so I recommend you go look for them.
technorati tags: IBM, DS8880, DS8884, DS8886, High Performance Flash Enclosure, HPFE, DSCLI, FICON, Spectrum Virtualize, Data-at-Rest Encryption, Spectrum Copy Data Management, Transparent Cloud Tiering, SAS, NL-SAS, LFF, SFF, FDWP, FlashSystem V9000, DeepFlash 150, DeepFlash ESS, Elastic Storage Server, Erasure Coding, Widows, Linux, AIX, BSSD
SAP HANA is an in-memory, relational database management system supported on Linux for x86 and POWER servers. The "HANA" acronym is short for "High-Performance Analytic Appliance" software. By keeping the data in memory, analytics and queries can be performed much faster than from traditional disk repositories.
Server memory, however, is volatile storage, so the data needs to be stored on persistent storage such as flash or disk drives. SAP has certified several configurations, some involve IBM Spectrum Scale solutions. I will use the following graphic to explain the three configurations.
- Linux on x86-64 with Spectrum Scale FPO
With SAP HANA on Lenovo x86-64 servers, SAP has certified internal flash or disk drives running IBM Spectrum Scale in "File Placement Optimization" (FPO) mode. FPO provides a shared-nothing architecture that matches the SAP HANA architecture. IBM Spectrum Protect can backup this configuration, providing data protection and disaster recovery support.
- Linux on POWER with Elastic Storage Server
With SAP HANA on POWER servers, SAP has certified external Elastic Storage Server (ESS). Not only is POWER the better platform to run SAP HANA than x86-64, but Elastic Storage Server offers excellent erasure coding to provide excellent rebuild times and storage efficiency.
The ESS is a pre-built system that combines IBM Spectrum Scale software with server and storage hardware. IBM Spectrum Protect can also backup this configuration, providing data protection and disaster recovery support.
- Block-level Storage over Storage Area Network (SAN)
Various IBM block-level devices are support for SAP HANA on both Linux on x86-64 and Linux on POWER. Unfortunately, SAP only has certified (to date) the use of the XFS file system. The problem many clients mention about this configuration is the lack of end-to-end backup and disaster recovery. This is solved by the Spectrum Scale configurations in the previous two examples.
Other combinations, such as SAP HANA on POWER with Spectrum Scale FPO, or on x86-64 servers with Elastic Storage Serer, are either not SAP-certified, or not directly supported by SAP without their approval.
IBM and SAP have worked closely together for many years, and I am glad to see SAP HANA and IBM Spectrum Scale based solutions continue this tradition.
technorati tags: IBM, SAP, SAP HANA, x86, x86-64, Linux, POWER, Spectrum Scale, FPO, Elastic Storage Server, ESS, Spectrum Protect, Disaster Recovery
Well it's Tuesday again, and you know what that means? IBM announcements!
(For those wondering where I went in July, then perhaps the better question should be "where didn't I go?". I started in Boston, MA, then Iceland, England, Hungary, Romania, Qatar, Kenya, Dubai UAE, and finally Seattle, WA. Whew! This week, I am visiting clients in Tennessee.)
Today, IBM launches a whole set of updated offerings based on the IBM Spectrum Virtualize software code base.
- IBM Spectrum Virtualize v7.7.1 software-only offering
Like the rest of the IBM Spectrum Storage family of products, IBM Spectrum Virtualize can now be purchased as software only, allowing you to install it on your own x86 servers, rather than purchasing pre-built systems from IBM.
The software license comes in two flavors. The traditional "perpetual license" allows you to move the software from one x86 server to another. Say after 4 years, you have depreciated the server, or the hardware components fail, and you want to get a newer server. This is the same perpetual license that clients with IBM SAN Volume Controller and Storwize family have enjoyed since 2003.
The other is a "monthly license", which allows you to stand up your own "SVC" using your own x86 servers, for a period of months needed for a development/test project, disaster recovery, or some other purpose. After the project is over, you can discontinue the license, and re-purpose the x86 servers for something else. This is especially handy for Managed Service Providers (MSP) and Cloud Service Providers (CSP), but certainly can prove useful in traditional datacenters as well. The "monthly licensing" option is also available for IBM SAN Volume Controller (SVC) as well.
The software license is based on Tebibyte [TiB]. For those not familiar with international standards, here is a comparison table:
||Always decimal, 10 to the 12th power
||Always binary, 2 to the 40th power
For more information, read the [IBM Spectrum Virtualize Software-only Announcement Letter].
- IBM SAN Volume Controller Model SV1
The new SV1 model is based on two 8-core [Intel Broadwell] processors, which IBM has clocked at up to 30 percent performance improvement over the DH8 model. It also offers up to 256GB of cache memory per node, which sadly only the first 64GB are usable at the current software level. Someday, a future release of software will address all 256GB of memory.
The IBM SAN Volume Controller now offers "Enterprise Class Support" as an option. In the past, the SVC was a "customer setup" box, similar to midrange and entry-level products. Now, you can upgrade your support to match that of IBM DS8000 and XIV enterprise class offerings. This means that IBM experts will maintain your microcode levels for you.
For more information, read the [IBM SAN Volume Controller SV1 Announcement Letter], and the [Enterprise Class Service Announcement Letter]..
- New software features
The v7.7.1 also provides new features for existing SVC, Storwize and FlashSystem V9000 products. Here are a few:
- Manageability with CLI support for host groups
- Scalability with support for up to 10,000 virtual disks, depending on the model; and up to 20 Expansion Enclosures on SVC 2145 DH8 and SV1 models, and the FlashSystem V9000
- RAS and performance enhancements for distributed RAID (DRAID)
- Flexibility with iSCSI virtualization support for XIV Gen3, IBM Spectrum Accelerate, FlashSystem A9000 and FlashSystem A9000R arrays
For more information, read the [IBM Spectrum Virtualize v7.7.1 Features Announcement Letter].
- IBM Storwize V7000 Gen2+ Model 624
The new 624 model is based on a single 10-core Intel Broadwell processor, which IBM has clocked at up to 45 percent performance improvement over the previous model. It also offers up to 128GB of cache memory per system, 64GB per node, double what came standard on the 524 model!
Why "Gen2+"? Moving from an 8-core Haswell to a 10-core Broadwell CPU, and doubling the cache memory didn't seem to be enough "architectural change" to justify calling it a "Gen3", so marketing decided on Gen2+ instead.
For more information, read the [Storwize V7000 Announcement Letter].
- IBM FlashSystem V9000
I refer to the IBM FlashSystem V9000 as my "Superman" product. When Superman dons on his glasses he becomes "Clark Kent", mild-mannered newspaper reporter. But behind the glasses, he is always Superman! Likewise, the FlashSystem V9000 is an all-flash array with an impressive set of features, but take off the fancy bezel, and you find that it is a pair of fully-loaded SAN Volume Controllers (which we call "Control Enclosures AC3") and a FlashSystem 900 drawer of the world's fastest flash storage.
The new FlashSystem V9000 is based on the new SV1 models of SVC. Each V9000 can attach up to 20 expansion enclosures over 12Gb SAS connections. The expansion enclosure can hold either 24 of the smaller 2.5-inch drives, or 12 of the larger 3.5-inch drives. Of course, the FlashSystem V9000 can also virtualize any of almost 400 different kinds of storage arrays, from all the major vendors, similar to SAN Volume Controller. This provides tiering options that match well with the FlashSystem 900 inside.
For more information, read the [FlashSystem V9000 Announcement Letter].
- IBM Storwize V7000F and V5030F all-flash array models
The FlashSystem V9000 was originally going to be called the Storwize V9000, but the FlashSystem folks wanted to keep all of the "FlashCore" technology under one name. In perhaps a bit retaliation, or maybe sibling rivalry, the Storwize team added the letter "F" to refer to the All-flash models of the Storwize V7000F and V5030F.
The "flash" in the V7000F and V5030F are just Solid-state drives, not nearly as fast as the cards in the FlashSystem models. The drives come in 1.92TB and 3.84TB capacities. You might see these rounded up to 2TB and 4TB on some presentations, but IBM officially never likes to exaggerate.
For more information, read the [Storwize All-Flash Announcement Letter], and the [New Flash Drive Options Announcement Letter].
technorati tags: IBM, SAN Volume Controller, SVC, Storwize, FlashSystem V9000, All-flash Array
Can you believe it has been a year already since IBM announced VersaStack?
In my May 2012 blog post, [EMC Strikes Back], I poked fun at the fact that Cisco had two
girlfriends "significant others": EMC and NetApp.
Cisco originally partnered with EMC to create a converged system called Vblock which combined Cisco UCS servers and switches with EMC storage. The partnership between VMware, Cisco and EMC was dubbed Virtual Computing Environment (VCE).
However, Cisco then partnered with NetApp to create Flexpod, a converged system that combined Cisco UCS servers and switches with NetApp storage. Many of my clients felt that Flexpod was an improvement over Vblock.
A lot has happened since then. In 2014, [drastically reduced its investment in VCE]. Last year, Dell then spent $67 Billion dollars to effectively take EMC out of the storage business. While this was a huge birthday present for IBM, not everyone is happy to see EMC fade away. Whitney Garcia has a great article titled [Crying at the Dell-EMC wedding: Why VCE customers should consider alternatives].
Before VersaStack, IBM had its own converged system, PureSystems, which combined IBM POWER and x86 servers with IBM storage. The x86 server portion of this business was sold off to Lenovo, but IBM continues to sell POWER-only and blended x86-and-POWER PureFlex systems, as well as PureApplication and PureData systems.
The [VersaStack] collaboration between IBM and Cisco offers an alternative to Vblock and Flexpod converged systems. Cisco is a leader in x86 blades and networking switches, and IBM is #1 in Flash and Software Defined Storage, including Storage Virtualization. VersaStack gives you the best of both worlds!
The VersaStack has Cisco Validated Designs for use with IBM's Spectrum Virtualize products:
- FlashSystem V9000
- Storwize V7000
- Storwize V7000 Unified
- Storwize V5000
This week, February 11, 2016, 12pm EDT, IBM and Cisco are hosting a webinar on VersaStack. Join us for the one year anniversary of VersaStack in a discussion with IBM, Cisco and VersaStack customers.
The speakers will be discussing VersaStack progress to date and the value VersaStack brings to client workloads. Topics of discussion will include how VersaStack can lower TCO, administrative overhead, reduce downtime and improve resource utilization, and allow for business innovation. The speakers include:
- Jonathan Cox, Medicat, Director, Technology Services
- Susan Martens, IBM, Director, VersaStack Sales, North America
- Kent Hixson, Cisco, Sales Business Development Manager
Here is the [Registration Link] to participate. Hope you can make it!
technorati tags: IBM, Cisco, EMC, VCE, VMware, Vblock, NetApp, Flexpod, VersaStack, #VersaStack, POWER, x86, Lenovo, PureSystems, PureFlex, PureApplication, PureData, Whitney Garcia, Jonathan Cox, Susan Martens, Kent Hixson, FlashSystem V9000, Storwize V7000, Storwize V7000 Unified, Storwize V5000, Medicat
Well it's Tuesday again, and you know what that means? IBM Announcements!
(FCC Disclosure: This official launch also includes October 6 announcements. In any case, the usual disclaimer applies: I currently work for IBM, and this blog post can be considered a "paid celebrity endorsement" of the IBM products mentioned below.)
IBM announced various updates to its Spectrum Storage product line. Here is a quick recap.
- IBM Spectrum Virtualize 7.6
Spectrum Virtualize is the new name of the "storage hypervisor" code that resides in IBM SAN Volume Controller (SVC) and Storwize family products. When you buy an SVC, you will license Spectrum Virtualize software on it. It is NOT available separately as software-only that you can install on any other hardware. There are three major improvements:
- Software-based Data-at-Rest Encryption
Earlier this year, IBM delivered data-at-rest encryption for the Storwize V7000 and V7000 Unified. This week, IBM extends this support to other storage hypervisors.
Since this feature is based on the Intel processor that supports the Advanced Encryption Standard New Instructions (AES-NI), it applies only to the newer hardware: SAN Volume Controller 2145-DH8, the Storwize V7000 Gen2, FlashSystem V9000, and VersaStack converged systems that contain these. You can run Spectrum Virtualize v7.6 on older hardware models, but the encryption feature will be disabled.
Basically, by taking advantage of AES-NI commands, IBM can now offer data-at-rest encryption on any virtualized flash or disk arrays, eliminating the need for special "Self-Encrypting Drives", or SED.
The encryption keys are kept on USB memory sticks, that you can either leave in the machine, or stash away in some vault or safe somewhere.
- Distributed RAID
The other improvement is distributed RAID. Distributed RAID has been hugely popular on IBM XIV products, and has since found its way into the DCS3700, DCS3860 and Elastic Storage Server models.
With this new enhancement, storage admins can select "Distributed RAID-5" or "Distributed RAID-6" as alternate choices to traditional RAID ranks.
Why use it? All the drives are now active, eliminating idle spare drives that do nothing collecting dust and cobwebs waiting for an opportunity to spin up, and when they finally are used for a rebuild become a terrible bottleneck. Since all drives are reading and writing, the rebuild rate is an order of magnitude (5 to 10x) faster!
For those clients nervous about large 8TB drives and the number of days it would take to perform a traditional RAID rebuild, this should calm all of your fears.
- IP-based Quorum
This is one of those line-items that we have told clients that it was "just around the corner" and "coming soon, watch this space", and finally it is available. For clients using Stretched Cluster or HyperSwap across two buildings, best practices suggests keeping the quorum disk in a third building. This often met having to dedicate a single 2U disk system in a closet somewhere, with expensive Fibre Channel cables connecting to the other two buildings.
To address this, IBM now allows the quorum disk to be based on Internet Protocol (the IP portion of TCP/IP), which can be any bare-metal or virtual machine that is LAN or WAN attached. The "quorum disk" is just a little Java program. This can run on any cloud service provider as well, such as IBM SoftLayer, that both buildings have connectivity.
A minor improvement worth mentioning is that the IBM "Comprestimator" tool that estimates the capacity savings of Real-time Compression is now integrated into Spectrum Virtualize v7.6 command line interface (CLI), allowing you to run the tool on demand, as needed, on any virtual volume.
- IBM Spectrum Scale v4.2
IBM plans to offer all of its solutions in any of three flavors: software-only that you can deploy on your own server hardware, pre-built system appliances, and cloud services on IBM SoftLayer, IBM Cloud Managed Services or third-party cloud providers. Spectrum Scale is the software-only flavor, and Elastic Storage Server and Storwize V7000 Unified are pre-built systems based on that software.
- File and Object access
IBM published a "Redbook" on how to implement OpenStack Swift and Amazon S3 interfaces to an existing Spectrum Scale deployment. IBM supported it, but it was basically Do-it-Yourself DIY implementation. This has now been resolved, with full integration of OpenStack Swift and Amazon S3 object-protocol interfaces.
(For those unfamiliar with "Object storage", think of it like valet parking for your data. Before working for IBM, I was previously employed as a valet attendant, so I feel qualified to make this analogy.
If you park your car in a 10-story high parking structure, you have to remember where you parked to go find the car again. With valet parking, you hand over the keys to the valet attendant, the car gets parked, and you get a claim stub that you then use to get your car back. In the meantime, you don't know where your car is parked, and you don't care either!
Storing files in volume-level or file-level storage is like that 10-story high parking structure. You have to remember where you put it, which LUN or which sub-directory. With object storage, the system provides a "claim stub" in the form of an Universal Record Identifier, or URI, and simple HTTP commands like GET and POST can be used to upload and download the content.)
- Policy-driven Compression and Quality of Service (QoS)
If you want to differentiate the levels of service provided by files and objects stored in your infrastructure, look no further. Simple SQL-like language is used to set up policies that are invoked when needed.
- Hadoop Connector for File and Objects
The IBM Hadoop Connector allows Hadoop and Spark analytics applications to treat Spectrum Scale as a 100 percent compatible alternative to Hadoop File Systems (HDFS). Previously, this was only available for files, but now it has been extended to include objects as well.
- Advanced Graphical User Interface (GUI)
Based on the award-winning GUI that has been used for IBM XIV, SVC, Storwize and various other members of the IBM System Storage family, IBM announces an HTML5-based web-browser GUI for configuring and managing Spectrum Scale and Elastic Storage Server (ESS).
- Storwize V7000 Unified
The "file modules" that run IBM Spectrum Scale will get updated to R1.6 level, which supports SMB 3.0 and NFS 4.0 protocols. SMB support will now include both internal and externally-virtualized storage. You will also be able to use Active File Management to migrate to other Spectrum Scale implementations.
- IBM Spectrum Control
As the former chief architect of IBM Tivoli Storage Productivity Center v1, I have been a big fan of the advancements and evolution of Spectrum Control. IBM offers three levels. The first level is "Basic Edition", entitled at no additional charge for IBM storage hardware clients. The second level is "Standard Edition" which offers configuration, provisioning and performance monitoring. The third level is "Advanced Edition", which includes advanced storage analytics, file-level reporting, storage tiering and data placement optimization.
You can imagine my skepticism when I was told that Spectrum Control was going to be enhanced to support Spectrum Scale. What could it offer? IBM Spectrum Scale already has built-in storage tiering and data placement optimization!
It turns out that having effective "management tools" was the #1 reason clients have stated were needed to implement and deploy Spectrum Scale. Since 1998, back when it was called General Parallel File System, or GPFS, the target market was High Performance Computing (HPC) familiar with Command Line Interfaces (CLI).
But IBM was to broaden the reach of IBM Spectrum Scale, to financial services, health care and life sciences, government and education, and a variety of other industries. They won't tolerate being limited to CLI interfaces.
For clients with multiple Spectrum Scale clusters, Spectrum Control can offer the following:
- Visibility across the capacity utilization (file systems, pools, file sets, quotas) and cluster health across all Spectrum Scale clusters in the data center
- Ability to specify alerts which are applied across all Spectrum Scale clusters, for things like relative or absolute free space in a file system, or inodes used, nodes going down, etc.
- Understand the cross-cluster relationships established by remote cluster mounts, and seamlessly navigate between them
- If external SAN storage is used, Spectrum Control shows the correlation between Spectrum Scale Network Shared Disks (NSD) and their corresponding SAN volumes, again with the ability to navigate between them; also it can provide performance monitoring for the volumes backing the NSD
- Ability to monitor file capacity usage in the context of applications, by adding Spectrum Scale "file set containers" to application groups defined in Spectrum Control
- Compare file system activity across Spectrum Scale clusters, with the ability to drill into file system and node performance charts
- Support for object storage on Spectrum Scale, determine which object-enabled clusters are closest to running out of free space
While the basic built-in GUI is great for smaller deployments, if you have a dozen or more Spectrum Scale clusters, or have Spectrum Scale clusters intermixed with traditional block-level and NAS storage devices, then Spectrum Control is for you!
It used to take weeks to deploy the original versions of Tivoli Storage Productivity Center, but now, Spectrum Control is now offered in the cloud, and you can deploy it in as little as 30 minutes.
Want to check it out? You can explore Spectrum Control Storage Insights cloud service as a [Live Demo], or [Start your free trial]! The reporting capabilities of Spectrum Scale are identical between the on-premise version of Spectrum Control, and this cloud service offering.
Here's a great quote from a leading IT industry analyst:
"In multi-petabyte, multivendor installations, overall storage costs of ownership for use of IBM Spectrum Storage solutions averaged 73 percent less than EMC, and 61 percent less than Hitachi equivalents" -- Brian Jeffery, Managing Director, International Technology Group, Naples, FL
As IBM continues its transition from a hardware-oriented company founded over a century ago, manufacturing meat scales and cheese slicers, to one more focused on higher value-add software and services, the Spectrum Storage software family will play a critical role of this transformation!
technorati tags: IBM, Spectrum Virtualize, data-at-rest, encryption, SVC, Storwize, Storwize V7000, FlashSystem V9000, VersaStack, storage hypervisor, distributed RAID, RAID-5, RAID-6, Spectrum Scale, Elastic Storage Server, OpenStack, OpenStack Swift, Amazon S3, HTTP, Compression, Quality of Service, QoS, Hadoop, Spark, Hadoop Connector, HDFS, GUI, XIV, DCS3700, DCS3860, Spectrum Control, Tivoli Storage, Productivity Center, TPC, CLI, NAS, Storage Insights, SoftLayer, IBM Cloud Managed Services,
Oh my, it is Tuesday again, and you know what that means? IBM Announcements!
This week, IBM announced its latest storage arrays in its IBM System Storage DS8000 series: the DS8880 models. Similar to the "Business Class" vs. "Enterprise Class" distinctions of the DS8870, IBM announced two new models, the DS8884 and the DS8886.
All of the new DS8880 models are based on the latest IBM POWER8 processors, and are noticeably thinner! These are now standard 19-inch wide, fitting nicely into standard IBM racks alongside most other standard 19-inch rack equipment.
The DC-UPS that used to be on the side are now at the bottom of each frame, taking up 8U of space. The High Performance Flash Enclosures (HPFE) that formerly were stored vertically above the DC-UPS will be stored horizontally with the rest of the HDD and SSD drives.
- DS8884 model
- The DS8884 will have 6-core controllers, up to 256 GB Cache, 64 ports that can negotiate between 16Gbps and 8Gbps, up to 240 drives in a single-rack configuration or 768 drives in a three-frame configuration, and up to 120 flash cards in HPFEs. The performance of this one is equal or better to existing DS8870 systems.
- DS8886 model
- The DS8886 will have 8-core, 16-core and 24-core controllers, offering up to three times the performance as the previous DS8870 models, with up to 2 TB of Cache, 128 ports, up to 1,536 drives across five frames, and up to 240 flash cards in HPFEs.
Field model conversion from DS8870 to DS8886 is available for existing clients with DS8870 Enterprise Configurations. This will let clients move their existing HDD, SSD, HPFE and Host Adapters over to the new DS8880 models.
In previous DS8000 models, clients would have one Hardware Management Console (HMC) inside the array, and an optional second HMC workstation somewhere else for high availability. While the second one was optional, it was always considered best practice to have it for redundancy sake. In the new DS8880 models, you can have both HMC in the array, and the Keyboard/Video/Monitor (KVM) can select between the two.
The new I/O enclosure pairs are four times faster, supporting six Device Adapters and two HPFE connections over PCIe Gen 3 network, the fastest available in the industry.
Lastly, IBM simplified the licensing of software features into three bundles, based on TB total capacity of Fixed Block (FB) LUNs and Count-Key-Data (CKD) volumes:
- Base function License: Logical Configuration support for FB, Operating Environment License, Thin Provisioning, Easy Tier® automated sub-volume tiering, and I/O Priority Manager.
- Copy Services License: FlashCopy®, Metro Mirror, Global Mirror, Metro/Global Mirror, z/Global Mirror (XRC), z/Global Mirror Resync, and Multi-Target PPRC.
- z-Synergy Service License: Parallel Access Volumes (PAV), HyperPAV, FICON® attachment, High performance FICON (zHPF), and IBM z/OS® Distributed Data Backup (zDDB).
IBM also provided a "Product preview", announcing plans for a third member of the DS8880 family in 2016 that will be flash-optimized to provide an all-flash, higher performance storage system model.
To learn more, read the [IBM Press Release] and [Function authorizations].
technorati tags: IBM, DS8000, DS8870, DS8880, DS8884, DS8886, HPFE, HDD, SSD, HMC, KVM, FB, CKD, Easy Tier, FlashCopy, FICON, zHPF, zDDB, all-flash
Next week, May 11-15, I will be in Las Vegas, Nevada for the [IBM Edge 2015 conference], covering IBM System Storage, z Systems and POWER Systems. Is this your first time going?
We have sold out, with over 6,000 registrants, that is way more than the 4,200 we had last year. If this is your first time to Edge, then I thought I would get some of my colleagues to help set some expectations.
So, I asked several of my colleagues a few questions, to help provide a variety of viewpoints. Here were my questions:
- What should new attendees expect at Edge?
- What are the trends that IBM is focused on? How is IBM differentiating themselves in the marketplace?
- Why does it matter to our attendees? What's in it for them?
- What is the call to action for my readers?
- What should new attendees expect at Edge?
"IBM Real-time Compression technology is field-proven and allows seamless scaling as demand for data storage capacity expands. The wave of new applications and data flowing from systems of engagement create a data management challenge. With Real-time Compression, which compresses data in-line and can store up to five times more data in the same physical space, IBM can help to address these challenges. Compression performance is optimized for enterprise-level and cloud workloads. Expect some exciting announcements in regards to this technology!"
-- Ori Bauer,
Director, WW Development and IBM Systems Israel Development Center
"They should expect Demonstrations and Exhibits! One featuring the capabilities of the FlashSystem V9000 through its slick and easy to use graphical user interface. The other featuring breakthrough technology from IBM Almaden Research.
A 'Digital Rack' introducing five configurations of the FlashSystem V9000 and 900 models that address five pain points customers face. A FlashSystem V9000 module showcasing its rich Tier-1 storage features and virtualization capabilities. A full rack / full configuration FlashSystem V9000 deployment illustrating over two Petabytes (2 PB) effective capacity."
-- Bill Bostic
Manager, FlashSystem Development and Exploitation
"The attendees can expect to see how the IBM DS8870 leadership allows clients to achieve the most value from the latest generation of IBM Mainframe, the z13!
Together, the DS8870 and z13 deliver end-to-end FICON speeds using 16Gbps Fibre Channel. These faster links have industry leading reliability as they are protected by new IBM-led standards to make 16 GB FC 300x more reliable. Using z Systems High Performance FICON (zHPF) protocols, DB2 utilities and Log writes can be reduced by up to 68 percent at 100 KM distances and 7 percent at machine room distances. This capability complements the GDPS and TPC-R HyperSwap function for continuous availability by mitigating the distance penalty when the secondary disks are in a send site up to 100 KM away.
Along with better performance and reliability the DS8870 extends the autonomic I/O management function of the z/OS work load manager into the SAN fabric with Fabric I/O Priority Management. For write operations the fabric priority is applied to the Metro Mirror traffic to provide a consistent policy across FCP and FICON sharing the same ISLs."
-- Harry Yudenfriend,
IBM Fellow, Storage Development
"There will be a VersaStack demo at the Cisco booth in the solutions center. There will be a client and IBM Business Partner giving testimony to the value of VersaStack."
-- Eric Stouffer
Director, IBM Storwize Business Line Executive
"Attendees can expect technology previews of exciting new features, like a highly flexible active-active solution for Storwize, next generation of cloud management integration technology for VMware, providing policy-based storage controls to cloud infrastructures that can take full advantage of advanced storage system features, and for integrating cloud storage into your traditional storage infrastructure."
-- Tommy Rickard,
Director, IBM Systems Development, United Kingdom
- What are the trends that IBM is focusing on?
"Accelerating applications. The IBM FlashSystem 900 is designed for uncompromising performance, macro efficiency and enterprise reliability.
Radically improving data center efficiency, economics, and performance by exploiting the rich set of FlashSystem V9000 features including virtualization, dynamic tiering, real time compression, thin provisioning, snap shots, cloning, replication, high availability, with the performance of FlashSystem technology."
-- Bill Bostic
Manager, FlashSystem Development and Exploitation
"VersaStack Integrated Infrastructure Solution combines Cisco Network MDS & Nexus, Cisco UCS compute, and Storage with IBM Storwize V7000K, with Cisco UCS Director for single point of management, to address one of the most rapidly growing markets. Storwize improves integrated infrastructure in three key ways - 1) by virtualizing storage to transform utilization, including older or other less capable storage; 2) optimizing speed and cost of devices automatically using EasyTier; 3) squeezing the most data in to the least space with no performance penalty using real time compression. No other integrated infrastructure from the likes of Dell, HP or Oracle has the value of VersaStack."
-- Ian Shave,
Global Business Unit Executive for VersaStack & BD&A, IBM Storage
"Increasing performance of ethernet networks, especially in new cloud deployments, that increases the importance of iSCSI as an enterprise, high-performance storage network. IBM provides rich support for iSCSI and is continuing to invest in this area. Increasing dominance of cloud infrastructures, including private, hybrid, and public forms, that is driving the need for delegated policy-based management to take advantage of storage system features for the software defined environment. IBM provides the most mature, capable, and flexible storage infrastructure, with leading integration into Openstack. VMware and other key cloud infrastructures.
Increasing opportunity for Cloud storage to satisfy traditional storage needs."
-- Tommy Rickard,
Director, IBM Systems Development, United Kingdom
- Why does it matter to our attendees? What's in it for them?
"Leading the way in helping the industry and clients exploit flash technology. IBM FlashSystem 900 is the acknowledged technology for application acceleration. That foundation is extended in the FlashSystem V9000 with a rich set of features to enable replacement of disk in the data center. The IBM Data Engine for NoSQL, released last year, uses FlashSystem technology to leverage flash as memory. Now, IBM will be demonstrating the next step to expand to exploitation of flash beyond flash as memory. This technology will dramatically expand flash applicability for NoSQL use cases. It will simplify storage management and accelerate node recovery, while dramatically improving availability and reliability.
Flash technology promises the advantages of better performance, better economics, and better environmentals. More important is the opportunity to radically reshape IT and the ability to deliver new customer value. IBM R&D is focused on delivering leading edge extensible technology as well as blazing the trail to enable the industry and customers exploit it. "
-- Bill Bostic
Manager, FlashSystem Development and Exploitation
"The end-to-end value that only IBM can provide improves resilience, lowers costs and is better able to handle the demands of the emerging Cloud, Analytics, Mobile and Social (CAMS) workloads."
-- Harry Yudenfriend,
IBM Fellow, Storage Development
"Integrated infrastructure speeds time to value with much of the integration and customization laid out in a recipe called a Cisco Validated Design (CVD). This saves time and money for clients and partners. It also provide confidence for these clients in the implementation and use in their environments. Cisco provides a single solution support offering for VersaStack, too, simplifying the technical support for clients and partners."
-- Eric Stouffer
Director, IBM Storwize Business Line Executive
"Cloud architectures are driving the technical agenda for infrastructure. Many clients are making use of cloud technologies, either to satisfy storage needs using public cloud, or to streamline their processes using private cloud. The largest clients, such as some financial clients, are building new data centers modelled on cloud infrastructure, and the growth of cloud investment is defining the technical agenda for these installations."
-- Tommy Rickard,
Director, IBM Systems Development, United Kingdom
- What is the call to action for my readers?
"Come to the demonstrations and exhibits at the Solution Center, which runs Monday lunch to Wednesday lunch, including Monday and Tuesday night receptions! This will be your opportunity to see these solutions live, and get to talk to the experts about them."
-- Bill Bostic
Manager, FlashSystem Development and Exploitation
"Check out my session cSY1467: DS8870 Exploitation of the z13 and z/OS I/O Enhancements, which I will present twice on Tuesday.
In the mean time, read the latest IBM Redpapers: [Enhancing Value to Existing and Future Workloads with IBM z13 (REDP-5135)] and
[Get More Out of Your IT Infrastructure With IBM z13 I/O Enhancements (REDP-5134)], or watch our latest videos on YouTube: [DS8870 and z13 FEC Demo] and [IBM DS8870 and z Systems]."
-- Harry Yudenfriend,
IBM Fellow, Storage Development
"See the VersaStack demo, go to the various sessions on VersaStack, some of which from IBM and the others from Cisco, visit the [IBM on VersaStack" and/or [Cisco on VersaStack] web sites for more information, or watch the YouTube video [New VersaStack Solution by Cisco and IBM]. If you are seller or IBM Business Partner, find a VersaStack Academy near you for more detailed education in person with experts from IBM and Cisco."
-- Ian Shave,
Global Business Unit Executive for VersaStack & BD&A, IBM Storage
If you didn't get into Edge, we are going to have a LiveStream of some of the keynote sessions on our [Digital Event Center] that you can watch from the comfort of your home or office. Go to the [Registration page] for details.
I arrive this Sunday afternoon. If you see me, stop and say "Hi!"
technorati tags: IBM, #edge2015, #ibmedge, #ibmstorage, Ori Bauer, William Bostic, Bill Bostic, Harry Yudenfriend, Real-time Compression, FlashSystem, FlashSystem 900, FlashSystem V9000, DS8000, DS8870, z13 System, IBM Mainframe, FICON, zHPF, Fabric I/O Priority Management, NoSQL, IBM Almaden Research, Solution Center, VersaStack, Storwize, Tommy Rickard, Eric Stouffer, Ian Shave
Happy New Year everyone!
Well it's Tuesday again, and you know what that means? IBM Announcements!
(Update: I thought it was quite clever to announce the new z13 mainframe on January the 13th. A few [triskaidekaphobic] employees pointed out that certain [Greek and Spanish-speaking cultures] consider Tuesday the 13th to be an unlucky day. However, superstitious people should probably not work in IT, as it would be difficult for a worldwide company like IBM to avoid all the numbers that different cultures consider unlucky.)
IBM will have a live streaming event on Jan 14, Redefining Digital Business: The new generation of IBM z Systems], for those who want to hear the announcement in more detail. Here is what the invitation page has to offer:
You are cordially invited to join IBM on January 14 from 2:00pm to 4:30pm Eastern Standard Time (US) when IBM will share a whole new generation of IBM z Systems™ built to meet the needs of your digital business. Join us and learn how IBM z Systems are designed to:
Support the transaction growth and needs of the mobile generation and the Internet of things
Integrate data, transactions and analytics, for in-transaction insights and right-time actions
Provide secure, trusted and efficient cloud services with new economic models
Exploit new modern, open development environments, tooling and skills for greater returns
At this live streaming event, you will hear from a remarkable group of business and technology leaders who will share success stories, best practices and the exciting technology innovations and capabilities of the new generation of IBM z Systems. Go to the Registration page] to participate.
But what does this really mean? Are you thinking BFD?
(Update: For those not familiar with IT acronyms, BFD refers to "Bigger, Faster, Denser" -- the trend in IT to announce new generations that are merely bigger, faster, and/or denser versions of the previous generations. Fortunately, the z13 takes up the same amount of data center floor space -- 2 floor tiles = 2 square meters = 20 square feet -- and weighs approximately the same as the z196 and zEC12, so raised floor struts do not have to be strengthened or reinforced to take in this new system.)
You may have noticed that we are now talking about "z System" instead of "System z". This change was made to line up with IBM's change to "POWER Systems" from "System p". Leadership felt that dropping the stodgy old zEnterprise and giving the mainframe a "hip" new name would attract new emerging digital workloads like Cloud, Analytics, Mobile and Social.
This is not the first time IBM has renamed products in a series. While the IBM mainframe just celebrated its 50th anniversary last year, the "13" refers to the 13th generation of CMOS-based mainframe technology introduced in 1994. Here is a quick table to show you the names that have evolved over the years:
1 to 6
S/390 G1 to G6
zSeries z800 and z900
zSeries z890 and z990
9 to 10
System z9 and z10
System z196 and z114
zEnterprise zEC12 and zBC12
(Note: This change also corresponds to a completely restructuring of IBM into business units, eliminating its former hardware and software groups. The design and development of all mainframe-related hardware, software and middleware will be consolidated under the IBM Systems business unit. I will wait for IBM's 4Q financial results announcement on or after January 20 before I cover this in any more detail.)
The z13 machine itself has some unique differences from previous generations. Instead of a "Multi-chip Module" (MCM) that contained multiple processor and storage controllers on a single slab, the z13 uses Single-Chip Modules (SCM) that are either a single 8-core processor, or a single system controller, allowing them to be field replaceable units (FRU).
Previous generations organized the processors in 1 to 4 vertical "books". The problem was that if you had a single book system, you bought a lot of hardware infrastructure designed to support a full four books. In the new design, processors are organized into horizontal Central Processor Complex (CPC) drawers, with additional hardware infrastructure provided per drawer. This makes the lower-end models more affordable. Each drawer has six processor SCMs and two system controller SCMs, providing 39 to 42 usable cores per drawer. Models ranges from 30 to 141 usable cores, with the option to upgrade from one model to another as your needs grow.
The z13 provides N-2 generation compatibility. This means you can have the z196, zEC12 and z13 all participate in the same Parallel Sysplex. You will also be able to upgrade your z196 or zEC12 to the new z13 system.
The new z13 can have up to 10TB of memory, and this can be assigned entirely to a single Logical Partition, or LPAR. The system can be subdivided up to 85 LPARs, versus 60 on the previous generation. Currently, z/OS v1 can only have up to 1TB per LPAR, and z/OS v2 can only go up to 4TB, so I suspect this 10TB is planning for future OS releases.
The new z13 now offers Simultaneous Multithreading [SMT]. Initially, this will double the number of threads for IFL engines (supporting Linux and z/VM), and zIIP engines supporting DB2, Java, XML and IPsec workloads. IBM is eliminating the zAAP engines, since zIIP engines can do all of that. The new [Preview z/OS v2.2] will take advantage of this SMT capability.
To assist with database, analytics and multimedia workloads, the z13 offers Single Instruction, Multiple Data [SIMD] capability. This allows a single instruction to perform the same update or action across many data fields.
For those clients with zBX models 2 and 3, allowing you to run POWER-based AIX and x86-based Microsoft Windows and Linux operating systems on your mainframe, they will be able to upgrade to [zBX model 4 for the z13 System]. Now that IBM has sold off its x86 server business to Lenovo, I suspect it will also phase out the zBX offerings as well.
To handle emerging workloads of Cloud, Mobile and other Web applications, IBM will offer a new stronger and faster Crypto Express5S cryptographic adapter. The z13 will enhance public key support for constrained digital environments using Elliptic Curve Cryptography (ECC) for users such as Chrome, Firefox, and Apple's iMessage. The z13 will also minimize reformatting of databases with new exploitation of VISA format preserving encryption (FPE) for credit card numbers.
The z13 also made some enhancements for Linux clients. The zAware analytics that analyzes internal traces and logs for z/OS has been extended to support Linux on System z. For those who want to use GDPS Business Continuity and Disaster Recovery services, but don't want to develop z/OS skills for the "K" system, there will now be a Virtual GDPS appliance that will run self-contained z/OS. Lastly, IBM has made a statement of direction that it will support open source Linux KVM as a Linux-only alternative to z/VM hypervisor. OpenStack will support both this new Linux KVM as well as z/VM 6.3 release.
The PCIe bus has been upgraded to Gen3 at 16Gbps, from Gen2 used in the zEC12. These can be used for Coupling Facility Links, which are faster than the legacy 6 Gbps InfiniBand, which are also supported for legacy migration. People with z196 and zEC12 can either carry forward their I/O drawers they have previously purchased, or move the PCiE Gen2 cards into the new Gen3 drawers.
The new z13 will also support 16Gb FICON, using the new FICON Express5S cards. Here is my segue into storage, as you are probably now wondering when I was going to get to the storage part of the announcement!
IBM is also announcing corresponding changes to the DS8870 firmware and accessories to go with the z13 System. This includes:
FICON Dynamic Routing
Reduce cost with improved and persistent performance for supporting I/O devices. This feature will allow SAN directors to have both FICON and FCP share the same Inter-switch Links (ISL). This is especially useful for clients who use FCP with their Linux, z/VM, AIX or Windows workloads.
16Gb Host Adapters
Improve network performance with twice faster FC and FICON adapters, and minimize latency for database log writes with zHyperWrite and Metro Mirror. These 16Gb adapters can auto-negotiate down to 4Gb and 8Gb, so that the DS8870 can connect to both the new z13 mainframe, as well as older models.
Forward Error Correction
Preserve data integrity with more redundancy on the information transmitted via 16Gb adapters.
zHPF Extended Distances improvements
Increase remote data speed with 50 percent better I/O performance when accessing remote disk, typically after a HyperSwap. The "zHPF" acronym is short for z System High Performance FICON.
Improved resiliency capabilities while enhancing the value of FICON Dynamic Routing mentioned above. IBM is extending Workload Manager (WLM) Quality of Service (QoS) optimization that exists now for compute and storage to the SAN Fabric, allowing WLM policies to influence FICON traffic.
IBM zHyperWrite™ capability
Helps you achieve better DB2 log write performance when using Metro Mirror (PPRC) in a HyperSwap-managed environment. Log writes are sent directly to both DASD, freeing up Metro Mirror resources.
But don't just take my word for it, here are reviews of the new system from various journalists:
"They seem a computing odd couple: the mainframe, the old workhorse, and the smartphone, the cool-kid computer of today. But IBM has designed the latest version of the mainframe, which is being introduced on Wednesday, with the smartphone in mind. The new mainframe, the z13, has been engineered to cope with the huge volume of data and transactions generated by people using smartphones and tablets."
-- New York Times
"One customer enthusiastic about such features is Citigroup Inc., a longtime IBM user that favors mainframes for both reliability and security. 'Security is in the DNA of the mainframe,' said Martin Kennedy, Citi's managing director for platforms and storage. Another factor shaping the bank's needs, Mr. Kennedy said, is the rising volume of transactions carried out using smartphones and other mobile devices. Mainframes are particularly good at combining data from a variety of systems and presenting them to a user's mobile app, he said."
-- The Wall Street Journal
"IBM is introducing a new mainframe in a bet that clients will need its souped-up speed and security to handle a surge in consumers using smartphones for everything from banking to checking health-care records. The z13 system can encrypt and analyze data in real time and process 30,000 transactions a second, International Business Machines Corp. (IBM) announced today. That means faster and safer transactions for consumers on mobile phones."
"With the unveiling of the z13, IBM has taken its MobileFirst Platform to deliver even better performance and security than before, as it incorporates the fastest microprocessor in the world, server processors that are twice as fast as existing products, 300 percent additional memory and 100 percent more bandwidth analytics speed. Last year, IBM formed a partnership with Apple to help bring Apple's iDevices to business customers to boost sales, with IBM providing cloud and mobile analytics support."
-- The Street
"'We're driving toward a world where more and more people are using mobile devices, or embedded devices, to interact with systems,' John Birtles, director of IBM z Systems, tells WIRED. 'We need to make sure that those devices are secure, that the transaction's secure, and that our clients get the level of analytics that gives them opportunities to improve their businesses.'"
IBM mainframes are used to process the majority of financial transactions around the world, is well positioned to handle Cloud, Analytics, Mobile and Social workloads. The IBM DS8000 series disk is the #1 market leader for disk storage on mainframe environments.
To learn more, read the [IBM z13 announcement letter] or [Press Release].
technorati tags: IBM, z13, mainframe, z+System, DS8000, DS8870, HyperWrite, FICON, PCiE, #z13
Last week, in my post [IT Support for the Holidays], I mentioned that I was scrubbing computers in preparation to give them to charity. A local reader asked if I would be willing to donate one of the computers to her kindergarten class. She teaches a class of 20 kids, at the very same elementary school that I went to when I was that age.
So here is the beefiest machine of the set.
Make/Model: Sony PCV-RC850
Processor: 2.4GHz Intel 32-bit
Hard disk: 40GB
Removable media: CD/DVD-ROM and CD/DVD-RW
Keyboard/mouse: standard PS/2
Sound: headphone jack
Ethernet port: 100Mbps
USB ports: two
IBM likes grand challenges, like [Deep Blue computer] to play chess against Grandmaster Garry Kasparov, and the [Watson computer] to play against two experts on the game show Jeopardy! My "Kidergarten Desktop" challenge is certainly on a smaller scale--to install software on this machine that will neet the following requirements
Have age-appropriate educational software and games for the students to learn reading, writing and math. This will also help them be more technology-savvy, learn the [QWERTY keyboard], and be more computer literate.
Have software for the teacher to use for her own job, after the kids have gone for the day, including submitting grades, sending email to parents, typing up lesson plans, data collection, researching the latest trends in education, for example.
Require minimal maintenance, be easy to rescue, repair and recover if necessary.
The 512MB is not enough to run Microsoft Windows 7, but certainly enough to run some flavors of Linux. Inspired by this review of [Top 6 Linux Distributions for Children], I thought I would give a few a spin.
Many of these have LiveCD/LiveDVD/LiveUSB versions that can be booted directly to try them out, and install directly to hard disk if you like it. Unfortunately, this often requires 1GB of memory or more, so I will need a different approach.
I had already scrubbed the [Windows XP] and replaced with [Linux Mint 12 LXDE]. Can I just install the Edubuntu-desktop on Linux Mint? While Linux Mint is Ubuntu-based, it is not binary compatible, so I will need to install fresh.
The [Edubuntu] LiveDVD requires 1GB of memory to try out, so to get this installed, I used the "Alternate Ubuntu 12.04" installer DVD.
Why 12.04 release of Ubuntu? The current release is 13.10 will only be supported for 9 months, and in keeping with "Requirement #3 Minimal Maintenance", the [Edubuntu team recommends installing a Long Term Support (LTS) release], and 12.04.3 is the most recent LTS that will be supported through 2017.
Edubuntu recommends 20GB of disk space to run, so I have partitioned the 40GB drive as follows:
For this machine, I will have three users configured:
admin - Administrator (that would be me for now) assigned to the "wheel" group to allow special priveleges
teacher - Teacher will have her own userid/password, so that she can do her own work
student - One userid/password shared by all students. This should eliminate kindergarten students from having to remember a userid and password that is unique to them. They are only five and six years old, after all!
Ubuntu's [Alternate Installer] uses basic graphic mode that can run in 512MB, and once installed, I was then able to install the Edubuntu Desktop and both preschool and primary-level educational software, to account for all learning ability levels of the children.
admin-$ sudo bash
admin-# apt-get install edubuntu-desktop
admin-# apt-get install ubuntu-edu-preschool
admin-# apt-get install ubuntu-edu-primary
I am not a big fan of Ubuntu's "Unity" panel on the left, and was hoping that Edubuntu-desktop would remove it, but no luck. so I removed it manually.
On the second partition, sda2 I put a few system utilities, including [Clonezilla] and [SystemRescue CD].
This system does not boot USB files natively, and getting Grub2 boot loader to boot ISO files was more difficult than I imagined. I was able to extract the necessary files over to sda2 hard disk to get them to work. I took "Clonezilla" full system backups to a separate SSH server over my local subnet.
Well, that's my start. Any suggestions? Has anyone done this before? Please enter comments below.
Wrapping up my week on All-Flash arrays, I thought I would cover some of the Enterprise Reliability features of the IBM FlashSystem.
On Monday, [IBM FlashSystem versus EMC XtremeIO all-Flash Arrays], I discussed some of the features of the IBM FlashSystem that differentiate it from EMC's ExtremeIO and other all-Flash arrays. On Tuesday, [IBM 2013 Storage Announcements for November 19] included discussion of the all-Flash model of the IBM System Storage DS8870 disk system.
Just as light bulbs burn out eventually after repeatedly being turned on and off, Flash does not last forever either.
A set of transistors can represent a single bit of informaiton (Single-level cell, or SLC for short), or multiple bits (Multi-level Cell, MLC). MLC typically refers to two bits, with a new "Triple-level cell" or TLC technology, able to store three bits per set of transistors.
SLC is faster and can endure more "Program-erase" write cycles, but MLC is less expensive to manufacture and therefore used in most consumer products, like digital cameras, smart phones, music players and USB memory sticks. To learn more on this, see this 6-page IBM whitepaper on [Comparison of NAND Flash Technologies Used in Solid-State Storage].
In between, "Enterprise MLC" (or eMLC for short) refers specifically to a different grade of chips IBM gets from the flash manufacturer. eMLC chips use a similar MLC bit arrangement, but are typically selected from higher bins, and most importantly have much longer program-erase cycle times which yield greater chip endurance, at the expense of long data retention when power is off (but seriously, when is anything off for very long in a data center?)
As a result, eMLC has 10x the endurance of regalar MLC, approaching parity with SLC at half the cost!
In the IBM FlashSystem, DRAM cache is used to buffer the writes first, then written out to the Flash. This helps to further improve the endurance.
For enterprise reliability, each Flash chip on the IBM FlashSystem has Error Correcting Codes (ECC), and then each set of 10 chips is placed in a 9+P RAID-5 configuration.
The chips are sub-divided into 16 planes. In the event a cell fails, the data for that plane can be reconstructed from parity, and written to spare space on the other planes of that same chip set. That plane is then reformated as an 8+P RAID-5, bypassing the failed plane.
In this manner, a cell failure only results in losing a small portion of one chip. If the same plane fails another failure on another chip, it will drop down to 7+P, 6+P, 5+P, and finally 4+P. This is known as "Variable Stripe RAID" or VSR for short.
IBM FlashSystem can survive over 1,000 such cell failures without an outage. By comparison, a single cell failure on an SSD often marks the entire drive as a failure.
But wait, there's more. Why stop at just RAID-5 across 10 chips. The chips are organized into modules, and IBM FlashSystem can perform RAID-5 across modules, in a 10+P+S RAID-5 configuration. This is referred to as "Two dimensional RAID" or 2D-RAID for short.
Even if you lost an entire module, the system will automatically rebuild on the spare module, and you can replace the bad one non-disruptively.
Many use cases for all-Flash arrays do not require such high levels of Enterprise reliability. Several of the all-Flash competitors have adopted a "design-for-failure" approach common among Cloud Service Providers like Amazon Web Services.
The idea is to assume that the data stored on them is just a copy from some other storage media. In the event of a Flash failure, it can easily be restored from a mirrored copy or backup.
For the IBM FlashSystem, The newer 800 series are based on eMLC, ideal for the majority of business applications, databases and virtual machine images placed on all-Flash arrays. The older 700 series are based on more expensive SLC, designed specifically for sustained write-intensive workloads.
Within each series, the "tens" models (710, 810) offer RAID-0 striping across ECC and VSR protected modules. For higher levels of availability, the "twenties" models (720, 820) offer ECC, VSR and 2D-RAID protection.
technorati tags: IBM, FlashSystem, SSD, Flash, endurance, high availability, enterprise reliability, SLC, MLC, eMLC, ECC, VSR, 2D-RAID, EMC, XtremeIO, DS8870
Wrapping up last week's theme on [New Year's Resolutions] to Eat Less and Exercise More. Yesterday, I talked about [tracking your diet], in this post, I will discuss tools to track your exercise and results.
(Note: I am neither a medical doctor nor registered dietician. I can share with you ideas that have worked for me, that might help you achieve your goals. I strongly suggest you read books and consult with medical experts as necessary.)
Back in 2009, studies found that [exercise alone does not lead to weight loss]. I had been working out at a gym, three times a week with a personal trainer, for six years, but was the same weight as when I started.
Much to the chagrin of my personal trainer, the article convinced me to quit the gym, discontinue her services, and focus on my diet instead. She warned I would gain 10 to 20 pounds within the year. Guess what? I didn't! I actually lost two pounds.
Here are my suggestions:
- Be Patient about Weight Loss
When I grew up, we all learned that 3,500 calories equals a pound of body fat, so to lose just one pound per week, you need to eat 500 calories less than you burn every day. Many dieters get impatient, even frustrated, that they are not losing weight fast enough.
A recent study led by Diana Thomas, a professor at Montclair State University in New Jersey, found that [as you lose weight, your metabolism drops].
Doing more exercise might help you build muscle, reduce stress and ligthen your mood, but it won't raise your metabolism as was once thought, nor even keep it at the levels you were at your prior weight.
But Dr. Thomas has taken the new findings and created a new [Weight Loss Estimator] that takes into account the drop in metabolism.
The example shows a 45-year old male, 200 pounds, eating 500 calories less than his normal 2,791-calorie diet. Over the course of 12 months, the tool estimates losing only 15.8 pounds, much less than one pound per week!
- Moderate Exercise can be Healthful
Does this mean you should just give up on exercise altogether? No.
The trick is to do moderate exercise, which you can do in and around your home, or at a gym. The goal is to [keep your heartbeat between 105 and 134 beats per minute]. This has been found to maximize the use of your own fat stores for energy.
Using my Android smartphone, I like the apps from [VirtuaGym]:
- [Fitness Home & Gym]. This app has a variety of workout circuits such as calisthentics and weight-lifting that you can easily follow. An avatar demonstrates how to perform each exercise, and does them with you to keep the pace. Here is an example [2-minute YouTube video] to show the app in action.
- [Cardio GPS]. This app is for cardio activities, such as walking, cycling, hiking, jogging and running. The GPS keeps track of your location, determines your speed, and the distance you travelled.
Both apps allow you to upload your activites to their website. This allows you to track which activities you did when, and share your progress with your friends on Facebook.
My favorite low-impact cardio exercise is simply walking. I start up my Cardio GPS app, put on my noise-cancelling headphones, and listen to music, podcasts or audiobooks on my iPod music player. I like the [Freakonomics Podcast] series.
To help me keep my iPod charged and loaded with the latest podcasts, my friends over at [Startech.com] sent me two iPod cables last month for my birthday. Sweet!
- The white one on the left for home use is two feet long, and has a 90-degree neck on the connector side so that my iPod can be propped up against a stack of books while I sync up my music and podcasts.
- The black one on the right for travel has both iPod and micro-USB connections, so that I can use it with both my iPod and my Samsung Galaxy smartphone!
- Determine the Right Metrics to Measure
"If you can't measure it, you can't manage it."
-- Peter Drucker
Tracking the right metric is important. Here are some metrics, and why they are useful or not.
- Body weight. The problem with measuring pounds is that this combines muscle weight that you want more of, and body fat that you want less of. Dieters who exercise often lose very little weight, some gain. This can be a misleading measure of progress.
- Body Mass Index. BMI is [calculated from your weight and height]. What do fellow actors George Clooney, Brad Pitt, Matt Damon and Tony Pearson have in common? We all have BMI indexes over 25, and therefore deemed unhealthy. Ha!
Richard Alleyne from the Telegraph has a great article on this. Here is an excerpt:
"A BMI of 25 to 29.9 is considered overweight and one of 30 or above is considered obese.
People with BMIs between 19 and 22 live longest. Death rates are noticeably higher for people with indexes 25 and above.
BMI does not identify how fat is distributed on the body. Storing more fat on the waist is a higher risk for cardiovascular disease, than storing it mostly elsewhere. "
For me, my doctor is happy with any index less than 27.
- Waist-to-Hip Ratio. This is simply the [the ratio of the circumference of your waist to that of your hips]. This is the measurement I track, as it focuses on the unhealthy body fat around my belly rather than in other areas. Men should aim for 0.9, while women should aim for 0.7 instead.
(FTC Disclosure: I work for IBM. I do not have any financial interest in, nor have been paid to mention any other companies, products or services on this blog post. Per FTC guidelines, this post can be considered a celebrity endorsement of cables from Startech.com, who provided me the cables at no charge.)
If you have resolved to lose weight, get fit, manage stress, or sleep better, and this series of posts has given you ideas or helped you in any way, I would love to hear about it. Please post a comment below!
Well, it's Tuesday again, and you know what that means! IBM announcements!
Today, I am in New York visiting clients. The weather is a lot nicer than I expected. Here is a picture of the Hudson River through some trees with leaves turning color. Something we don't see in Tucson! Our cactus and pine trees stay green year-round!
The announcements today center around the IBM PureSystems family of expert integrated systems. The PureFlex is based on Flex System components. The Flex System chassis is 10U high that hold 14 bays, consisting of 7 rows by 2 columns. Computer and Storage nodes fit in the front, and switches, fans and power supplies in the back. Here is a quick recap:
- IBM Flex System Compute Nodes
The x220 Compute Node is a single-bay low-power 2-socket x86 server. The x440 Compute Node is a powerful double-bay (1 row, 2 columns). The p260 Compute Node is a single-bay server based on the latest POWER7+ CPU processor.
- IBM Flex System Expansion Nodes
Do you remember those old movies where a motorcycle would have a sidecar that could hold another passenger, or extra cargo? IBM introduces "Expansion Nodes" for the x200 series single-bay Compute nodes. The idea here is that in a single column, you have one bay for the Compute node, and then on the side in the next bay (same column) you have an Expanions node. There are two choices:
- Storage Expansion Node allows you to have eight additional drives
- PCIe Expansion Node allows to to have four PCIe cards, which could include the SSD-based PCIe cards from IBM's recent acquisition, Texas Memory Systems.
There are times where one or two internal drives are just not enough storage for a single server, and these expanion nodes could just be the perfect solution for some use cases.
- IBM Flex System V7000 Storage Node
I saved the best for last! The Flex System V7000 Storage Node is basically the IBM Storwize V7000 repackaged to fit into the Flex System chassis. This means that in the front of the chassis, the Flex System V7000 takes up four bays (2 rows by 2 columns). In the back of the chassis are the power supplies, fans and switches.
The new Flex System V7000 supports everything the Storwize V7000 does except the upgrade to "Unified" through file modules. For those who want to have Storwize V7000 Unified in their PureFlex systems, IBM will continue to offer the outside-the-chassis original Storwize V7000 that can have two file modules added for NFS, CIFS, HTTPS, FTP and SCP protocol support.
- IBM Flex System Converged Network Switch
The Converged Network Switch provide Fibre Channel over Ethernet (FCoE) directly from the chassis. This eliminates the need for a separate "Top-of-Rack" switch, and allows the new Flex System V7000 Storage Node to externally virtualize FCoE-based disk arrays.
- Patterns of Expertise for Infrastructure
The original patterns of expertise focused on the PureApplication Systems. Now IBM has added some for the Infrastructure on PureFlex systems.
IBM has sold over 1,000 Flex System and PureFlex systems, across 40 different countries around the world, since their introduction a few months ago in April! These latest enhancements will help solidify IBM's industry leadership,
technorati tags: IBM, PureFlex, Storwize V7000, Flex System V7000, PureSystems, PureApplication, Patters of Expertise
Every year, I teach hundreds of sellers how to sell IBM storage products. I have been doing this since the late 1990s, and it is one task that has carried forward from one job to another as I transitioned through various roles from development, to marketing, to consulting.
This week, I am in the city of Taipei [Taipei] to teach Top Gun sales class, part of IBM's [Sales Training] curriculum. This is only my second time here on the island of Taiwan.
As you can see from this photo, Taipei is a large city with just row after row of buildings. The metropolitan area has about seven million people, and I saw lots of construction for more on my ride in from the airport.
The student body consists of IBM Business Partners and field sales reps eager to learn how to become better sellers. Typically, some of the students might have just been hired on, just finished IBM Sales School, a few might have transferred from selling other product lines, while others are established storage sellers looking for a refresher on the latest solutions and technologies.
I am part of the teach team comprised of seven instructors from different countries. Here is what the week entails for me:
- Monday - I will present "Selling Scale-Out NAS Solutions" that covers the IBM SONAS appliance and gateway configurations, and be part of a panel discussion on Disk with several other experts.
- Tuesday - I have two topics, "Selling Disk Virtualization Solutions" and "Selling Unified Storage Solutions", which cover the IBM SAN Volume Controller (SVC), Storwize V7000 and Storwize V7000 Unified products.
- Wednesday - I will explain how to position and sell IBM products against the competition.
- Thursday - I will present "Selling Infrastructure Management Solutions" and "Selling Unified Recovery Management Solutions", which focus on the IBM Tivoli Storage portfolio, including Tivoli Storage Productivity Center, Tivoli Storage Manager (TSM), and Tivoli Storage FlashCopy Manager (FCM). The day ends with the dreaded "Final Exam".
- Friday - The students will present their "Team Value Workshop" presentations, and the class concludes with a formal graduation ceremony for the subset of students who pass. A few outstanding students will be honored with "Top Gun" status.
These are the solution areas I present most often as a consultant at the IBM Executive Briefing Center in Tucson, so I can provide real-life stories of different client situations to help illustrate my examples.
To learn more about a Top Gun class in your area, see the [Top Gun class schedule].
The weather here in Taipei calls for rain every day! I was able to take this photo on Sunday morning while it was still nice and clear, but later in the afternoon, we had quite the downpour. I am glad I brought my raincoat!
technorati tags: IBM, Sales Training, Top Gun, Taipei, Taiwan, NAS, SONAS, disk, virtualization, unified+storage, SAN Volume Controller, SVC, Storwize V7000, Storwize V7000 Unified, Infrastructure Management, Tivoli Storage, Productivity Center, TPC, Unified Recovery Management, TSM, FlashCopy, FCM
Well it's Tuesday again, and you know what that means.. IBM announcements! Today, IBM announces that next Monday marks the 60th anniversary of first commercial digital tape storage system! I am on the East coast this week visiting clients, but plan to be back in Tucson in time for the cake and fireworks next Monday.
Note that I said first commercial tape system as tape itself, in various forms, [has been around since 4000 B.C.]. A little historical context might help:
- 1845 - surgical tape
- 1921 - the Band-Aid, self-adhesive bandage
- 1925 - masking tape (which 3M sold under its newly announced Scotch® brand)
- 1930 - clear cellulose-based tape (today, when people say Scotch tape, they usually are referring to the cellulose version)
- 1935 - Allgemeine Elektrizitatsgesellschaft (AEG) presents Magnetophon K1, audio recording on analog tape
- 1942 - Duct tape
- 1947 - Bing Crosby adopts audio recording for his radio program. This eliminated him doing the same program live twice per day, perhaps the first example of using technology for "deduplication".
According to the IBM Archives the [IBM 726 tape drive was formally announced May 21, 1952]. It was the size of a refrigerator, and the tape reel was the size of a large pizza. The next time you pull a frozen pizza from your fridge, you can remember this month's celebration!
When I first joined IBM in 1986, there were three kinds of IBM tape. The round reel called 3420, and the square cartridge called 3480, and the tubes that contained a wide swath of tape stored in honeycomb shelves called the [IBM 3850 Mass Storage System].
My first job at IBM was to work on DFHSM, which was specifically started in 1977 to manage the IBM 3850, and later renamed to the DFSMShsm component of the DFSMS element of the z/OS operating system. This software was instrumental in keeping disk and tape at high 80-95 percent utilization rates on mainframe servers.
While visiting a client in Detroit, the client loved their StorageTek tape automation silo, but didn't care for the StorageTek drives inside were incompatible with IBM formats. They wanted to put IBM drives into the StorageTek silos. I agreed it was a good idea, and brought this back to the attention of development. In a contentious meeting with management and engineers, I presented this feedback from the client.
Everyone in the room said IBM couldn't do that. I asked "Why not?" The software engineers I spoke to already said they could support it. With StorageTek at the brink of Chapter 11 bankruptcy, I argued that IBM drives in their tape automation would ease the transition of our mainframe customers to an all-IBM environment.
Was the reason related to business/legal concerns, or was their a hardware issue? It turned out to be a little of both. On the business side, IBM had to agree to work with StorageTek on service and support to its mutual clients in mixed environments. On the technical side, the drive had to be tilted 12 degrees to line up with the robotic hand. A few years later, the IBM silo-compatible 3592 drive was commercially available.
Rather than put StorageTek completely out of business, it had the opposite effect. Now that IBM drives can be put in StorageTek libraries, everyone wanted one, basically bringing StorageTek back to life. This forced IBM to offer its own tape automation libraries.
In 1993, I filed my first patent. It was for the RECYCLE function in DFHSM to consolidate valid data from partial tapes to fresh new tapes. Before my patent, the RECYCLE function selected tapes alphabetically, by volume serial (VOLSER). My patent evaluated all tapes based on how full they were, and sorted them least-full to most-full, to maximize the return of cartridges.
Different tape cartridges can hold different amounts of data, especially with different formats on the same media type, with or without compression, so calculating the percentage full turned out to be a tricky algorithm that continues to be used in mainframe environments today.
The patent was popular for cross-licensing, and IBM has since filed additional patents for this invention in other countries to further increase its license revenue for intellectual property.
In 1997, IBM launched the IBM 3494 Virtual Tape Server (VTS), the first virtual tape storage device, blending disk and tape to optimal effect. This was based off the IBM 3850 Mass Storage Systems, which was the first virtual disk system, that used 3380 disk and tape to emulate the older 3350 disk systems.
In the VTS, tape volume images would be emulated as files on a disk system, then later moved to physical tape. We would call the disk the "Tape Volume Cache", and use caching algorithms to decide how long to keep data in cache, versus destage to tape. However, there were only a few tape drives, and sometimes when the VTS was busy, there were no tape drives available to destage the older images, and the cache would fill up.
I had already solved this problem in DFHSM, with a function called pre-migration. The idea was to pre-emptively copy data to tape, but leave it also on disk, so that when it needed to be destaged, all we had to do was delete the disk copy and activate the tape copy. We patented using this idea for the VTS, and it is still used in the successor models of IBM Sysem Storage TS7740 virtual tape libraries today.
Today, tape continues to be the least expensive storage medium, about 15 to 25 times less expensive, dollar-per-GB, than disk technologies. A dollar of today's LTO-5 tape can hold 22 days worth of MP3 music at 192 Kbps recording. A full TS1140 tape cartridge can hold 2 million copies of the book "War and Peace".
(If you have not read the book, Woody Allen took a speed reading course and read the entire novel in just 20 minutes. He summed up the novel in three words: "It involves Russia." By comparison, in the same 20 minutes, at 650MB/sec, the TS1140 drive can read this novel over and over 390,000 times.)
If you have your own "war stories" about tape, I would love to hear them, please consider posting a comment below.
technorati tags: IBM, AEG, Bing Crosby, Duct+Tape, Band Aid, DFHSM, RECYCLE, DFSMShsm, z/OS, StorageTek, VTS, VTL, LTO-5, TS1140, LTFS, Woody Allen
This week, I am in beautiful Sao Paulo, Brazil, teaching Top Gun class to IBM Business Partners and sales reps. Traditionally, we have "Tape Thursday" where we focus on our tape systems, from tape drives, to physical and virtual tape libraries. IBM is the number #1 tape vendor, and has been for the past eight years.
(The alliteration doesn't translate well here in Brazil. The Portuguese word for tape is "fita", and Thursday here is "quinta-feira", but "fita-quinta-feira" just doesn't have the same ring to it.)
In the class, we discussed how to handle common misperceptions and myths about tape. Here are a few examples:
- Myth 1: Tape processing is manually intensive
In my July 2007 blog post [Times a Million], I coined the phrase "Laptop Mentality" to describe the problem most people have dealing with data center decisions. Many folks extend linearly their experiences using their PCs, workstations or laptops to apply to the data center, unable to comprehend large numbers or solutions that take advantage of the economies of scale.
For many, the only experience dealing with tape was manual. In the 1980s, we made "mix tapes" on little cassettes, and in the 1990s we recorded our favorite television shows on VHS tapes in the VCR. Today, we have playlists on flash or disk-based music players, and record TV shows on disk-based video recorders like Tivo. The conclusion is that tapes are manual, and disk are not.
Manual processing of tapes ended in 1987, with the introduction of a silo-like tape library from StorageTek. IBM quickly responded with its own IBM 3495 Tape Library Data Server in 1992. Today, clients have many tape automation choices, from the smallest IBM TS2900 Tape Autoloader that has one drive and nine cartridges, all the way to the largest IBM TS3500 multiple-library shuttle complex that can hold exabytes of data. These tape automation systems eliminate most of the manual handling of cartridges in day-to-day operations.
- Myth 2: Tape media is less reliable than disk media
For any storage media to be unreliable is to return the wrong information that is different than what was originally stored. There are only two ways for this to happen: if you write a "zero" but read back a "one", or write a "one" and read a "zero". This is called a bit error. Every storage media has a "bit error rate" that is the average likelihood for some large amount of data written.
According to the latest [LTO Bit Error rates, 2012 March], today's tape expects only 1 bit error per 10E17 bits written (about 100 Petabytes). This is 10 times more reliable than Enterprise SAS disk (1 bit per 10E16), and 100 times more reliable than Enterprise-class SATA disk (1 bit per 10E15).
Tape is the media used in "black boxes" for airplanes. When an airplane crashes, the black box is retrieved and used to investigate the causes of the crash. In 1986, the Space Shuttle Challenger exploded 73 seconds after take-off. The tapes in the black box sat on the ocean floor for six weeks before being recovered. Amazingly, IBM was able to successfully restore [90 percent of the block data, and 100 percent of voice data].
- Myth 3: Most tape restores fail
Why do people still believe that most tape restores fail? Curtis Preston, on his Backup Central blog, has a great post [Gartner Never Said 71 percent of Tape Restores Fail].
Analysts are quite upset when they are quoted out of context, but in this case, Gartner never said anything closely similar to this. Nor did the other analysts that Curtis investigated for similar claims. What Garnter did say was that disk provides an attractive alternative storage media for backup which can increase the performance of the recovery process.
Back in the 1990s, Savur Rao and I developed a patent to help backup DB2 for z/OS by using the FlashCopy feature of IBM's high-end disk system. The software method to coordinate the FlashCopy snapshots with the database application and maintain multiple versions was implemented in the DFSMShsm component of DFSMS. A few years later, this was part of a set of patents IBM cross-licensed to Microsoft for them to implement a similar software for Windows called Data Protection Manager (DPM). IBM has since introduced its own version for distributed systems called IBM Tivoli FlashCopy Manager that runs not just on Windows, but also AIX, Linux, HP-UX and Solaris operating systems.
Curtis suspects the "71 percent" citation may have been propogated by an ambitious product manager of Microsoft's Data Protection Manager, back in 2006, perhaps to help drive up business to their new disk-based backup product. Certainly, Microsoft was not the only vendor to disparage tape in this manner.
A few years ago, an [EMC failure brought down the State of Virginia] due to not just a component failure it its production disk system, but then made it worse by failing to recover from the disk-based remote mirror copy. Fortunately, the data was able to be restored from tape over the next four days. If you wonder why nobody at EMC says "Tape is Dead" anymore, perhaps it is because tape saved their butts that week.
(FTC Disclosure: I work for IBM and this post can be considered a paid, celebrity endorsement for all of the IBM tape and software products mentioned on this post. I own shares of stock in both IBM and Google, and use Google's Gmail for my personal email, as well as many other Google services. While IBM, Google and Microsoft can be considered competitors to each other in some areas, IBM has working relationships with both companies on various projects. References in this post to other companies like EMC are merely to provide illustrative examples only, based on publicly available information. IBM is part of the Linear Tape Open (LTO) consortium.)
Last year, Google lost the email data for half a million Gmail accounts due to a software error. Once again, tape came to the rescue, with [Google restoring lost Gmail data from tape backups].
- Myth 4: Vendors and Manufacturers are no longer investing in tape technology
IBM and others are still investing Research and Development (R&D) dollars to improve tape technology. What people don't realize is that much of the R&D spent on magnetic media can be applied across both disk and tape, such as IBM's development of the Giant Magnetoresistance read/write head, or [GMR] for short.
Most recently, IBM made another major advancement with tape with the introduction of the Linear Tape File Systems (LTFS). This allows greater portability to share data between users, and between companies, but treating tape cartridges much like USB memory sticks or pen drives. You can read more in my post [IBM and Fox win an Emmy for LTFS technology]!
Next month, IBM celebrates the 60th anniversary for tape. It is good to see that tape continues to be a vibrant part of the IT industry, and to IBM's storage business!
technorati tags: IBM, Google, Microsoft, EMC, Brazil, LTO, TS2900, TS3500, Space Shuttle, Challenger
On my last blog post [Is this what HDS tells our mainframe clients?], I poked fun at Hu Yoshida's blog post that contained a graphic with questionable results. Suddenly, the blog post disappeared altogether. Poof! Gone!
Just so that I am not accused of taking a graph out of context, here is Hu's original post, in its entirety:
"Since my last post on Storage Performance Efficiency, Claus wrote on the use of HDP, Hitachi Dynamic Provisioning and HDT, Hitachi Dynamic Tiering for mainframes on Virtual Storage Platform (VSP). Naturally, this prompted me to think of the specific performance efficiency implications for mainframes.
HDP brings the performance benefits of automated wide striping and HDT automatically keeps the hot pages of data on the highest performance tier of storage for mainframes, just as it does for open systems. There are differences between open systems and mainframe implementation due to mainframe CKD and CCHHR formats for instance, the page size is optimized for mainframe storage formats and storage reclamation must be host initiate. For more information check out our website: http://www.hds.com/assets/pdf/how-to-apply-latest-advances-in-hitachi-mainframe-storage.pdf
There are also additional performance efficiencies specific for mainframes.
Mainframe HDP is the foundation for Extended Addressable Volumes, which increases the size of 3390 volumes from 65,520 cylinders to 262,668 cylinders. This, along with HyperPAV--which facilitates multiple accesses to a volume, addressing the problem of queuing on a very large volume with a single UCB--enhances throughput with many more concurrent I/O operations.
The thin provisioning of HDP also increases the performance of mainframe functions that move, copy, or replicate these thin volumes like Concurrent Copy, FlashCopy V02, and HUR, since the actual volumes are smaller.
If you have mainframes, check out the capacity and performance efficiency of VSP with HDP and HDT.
For other posts on maximizing storage and capacity efficiencies, check these out:
At this point, you might be wondering: "If Hu Yoshida deleted his blog post, how did Tony get a copy of it? Did Tony save a copy of the HTML source before Hu deleted it?" No. I should have, in retrospect, in case lawyers got involved. It turns out that deleting a blog post does not clear the various copies in various RSS Feed Reader caches. I was able to dig out the previous version from the vast Google repository. (Many thanks to my friends at Google!!!).
The graph itself was hosted separately has been deleted, but it was just taken from slide 10 of the HDS presentation [How to Apply the Latest Advances in Hitachi Mainframe Storage], so it was easy to recreate.
(Lesson to all bloggers: If you write a blog post, and later decide to remove it for whatever legal, ethical, moral reasons, it is better to edit the post to remove offending content, and add a comment that the post was edited, and why. Shrinking a 700-word article down to 'Sorry Folks - I decided to remove this blog post because...' would do the trick. This new edited version will then slowly propagate across to all of the RSS Feed Reader caches, eliminating most traces to the original. Of course, the original may have been saved by any number of your readers, but at least if you have an edited version, it can serve as the official or canonical version.)
Perhaps there was a reason why HDS did not want to make public the FUD its sales team use in private meetings with IBM mainframe clients. Whatever it was, this appears to be another case where the cover-up is worse than the original crime!
technorati tags: HDS, Hu Yoshida, VSP, EAV
Five years ago, I sprayed coffee all over my screen from something I read on a blog post from fellow blogger Hu Yoshida from HDS. You can read what cased my reaction in my now infamous post [Hu Yoshida should know better]. Subsequently, over the years, I have disagreed with Hu on a variety of of topics, as documented in my 2010 blog post [Hu Yoshida Does It Again].
(Apparently, I am not alone, as the process of spraying one's coffee onto one's computer screen while reading other blog posts has been referred to as "Pulling a Tony" or "Doing a Tony" by other bloggers!)
Fortunately, my IBM colleague David Sacks doesn't drink coffee. Last month, David noticed that Hu had posted a graph in a recent blog entry titled [Additional Storage Performance Efficiencies for Mainframes], comparing the performance of HDS's Virtual Storage Platform (VSP) to IBM's DS8000.
For those not familiar with disk performance graphs, flatter is better, lower response time and larger IOPS are always desired. This graph implies that the HDS disk system is astonishingly faster than IBM's DS8000 series disk system. Certainly, the HDS VSP qualifies as a member of the elite [Super High-End club] with impressive SPC benchmark numbers, and is generally recognized as a device that works in IBM mainframe environments. But this new comparison graph is just ridiculous!
(Note: While SPC benchmarks are useful for making purchase decisions, different disk systems respond differently to different workloads. As the former lead architect of DFSMS for z/OS, I am often brought in to consult on mainframe performance issues in complex situations. Several times, we have fixed performance problems for our mainframe clients by replacing their HDS systems with IBM DS8000 series!)
Since Hu's blog entry contained very little information about the performance test used to generate the graph, David submitted a comment directly to Hu's blog asking a few simple questions to help IBM and Hu's readers determine whether the test was fair. Here is David's comment as submitted:
(Disclosure: I work for IBM. This comment is my own.)
I was quite surprised by the performance shown for the IBM DS8000 in the graph in your blog. Unfortunately, you provided very little detail about the benchmark. That makes it rather difficult (to say the least) to identify factors behind the results shown and to determine whether the comparison was a fair one.
Of the little information provided, an attribute that somewhat stands out is that the test appears to be limited to a single volume at least, that's my interpretation of "LDEV: 1*3390-3"? IBM's internal tests for this kind of case show far better response time and I/Os per second than the graph you published.
Here are a few examples of details you could provide to help readers determine whether the benchmark was fair and whether the results have any relevance to their environment.
- What DS8000 model was the test run on? (the DS8000 is a family of systems with generations going back 8 years. The latest and fastest model is the DS8800.)
- What were the hardware and software configurations of the DS8000 and VSP systems, including the number and speed of performance-related components?
- What were the I/O workload characteristics (e.g., read:write ratio and block size(s))?
- What was the data capacity of each volume? (Allocated and used capacity.)
- What were the cache sizes and cache hit ratios for each system? (The average I/O response times under 1.5 milliseconds for each system imply the cache hit ratios were relatively high.)
- How many physical drives were volumes striped across in each system?"
Unlike my blog on IBM, HDS bloggers like Hu are allowed to reject or deny comments before they appear on his blog post. We were disappointed that HDS never posted David's comment nor responded to it. That certainly raises questions about the quality of the comparison.
So, perhaps this is yet another case of [Hitachi Math], a phrase coined by fellow blogger Barry Burke from EMC back in 2007 in reference to outlandish HDS claims. My earliest mention was in my blog post [Not letting the Wookie Win].
By the way, since the test was about z/OS Extended Address Volumes (EAV), it is worth mentioning that IBM's DS8700 and DS8800 support 3390 volume capacities up to 1 TB each, while the HDS VSP is limited to only 223 GB per volume. Larger volume capacities help support ease-of-growth and help reduce the number of volumes storage administrators need to manage; that's just one example of how the DS8000 series continues to provide the best storage system support for z/OS environments.
Personally, I am all for running both IBM and HDS boxes side-by-side and publishing the methodology, the workload characteristics, the configuration details, and the results. Sunshine is always the best disinfectant!
technorati tags: IBM, DS8000, DS8800, HDS, Hu Yoshida, USP, VSP, mainframe, EAV
Well, it's Tuesday, and you know what that means... IBM announcements!
In today's environment, clients expect more from their storage, and from their storage provider. The announcements span the gamut, from helping to use Business Analytics to analyze Big Data for trends, insights and patterns, to managing private, public and hybrid cloud environments, all with systems that are optimized for their particular workloads.
There are over a dozen different announcements, so I will split these up into separate posts. Here is part 1.
- IBM Scale Out Network Attach Storage (SONAS) R1.3
I have covered [IBM SONAS] for quite some time now. Based on IBM's General Parallel File System (GPFS), this integrated system combines servers, storage and software into a fully functional scale-out NAS solution that support NFS, CIFS, FTP/SFTP, HTTP/HTTPS, and SCP protocols. IBM continues its technical leadership in the scale-out NAS marketplace with new hardware and software features.
The hardware adds new disk options, with 900GB SAS 15K RPM drives, and 3TB NL-SAS 7200 RPM drives. These come in 4U drawers of 60 drives each, six ranks of ten drives each. So, with the high-performance SAS drives that would be about 43TB usable capacity per drawer, and with the high-capacity NL-SAS drives about 144TB usable. You can have any mix of high-performance drawers and high-capacity drawers, up to 7200 drives, for a maximum usable capacity of 17PB usable (21PB for those who prefer it raw). This makes it the largest commercial scale-out NAS in the industry. This capacity can be made into one big file system, or divided up to 256 smaller file systems.
In addition to snapshots of each file system, you can divide the file system up into smaller tree branches and snapshot these independently as well. The tree branches are called fileset containers. Furthermore, you can now make writeable clones of individual files, which provides a space-efficient way to create copies for testing, training or whatever.
Performance is improved in many areas. The interface nodes now can support a second dual-port 10GbE, and replication performance is improved by 10x.
SONAS supports access-based enumeration, which means that if there are 100 different subdirectories, but you only have authority to access five of them, then that's all you see, those five directories. You don't even know the other 95 directories exist.
I saved the coolest feature for last, it is called Active Cloud Engine™ that offers both local and global file management. Locally, Active Cloud Engine placement rules to decide what type of disk a new file should be placed on. Management rules that will move the files from one disk type to another, or even migrates the data to tape or other externally-managed storage! A high-speed scan engine can rip through 10 million files per node, to identify files that need to be moved, backed up or expired.
Globally, Active Cloud Engine makes the global namespace truly global, allowing the file system to span multiple geographic locations. Built-in intelligence moves individual files to where they are closest to the users that use them most. This includes an intelligent push-over-WAN write cache, on-demand pull-from-WAN cache for reads, and will even pre-fetch subsets of files.
No other scale-out NAS solution from any other storage vendor offers this amazing and awesome capability!
- IBM® Storwize® V7000
Last year, we introduced the [IBM Storwize V7000], a midrange disk system with block-level access via FCP and iSCSI protocols. The 2U-high control enclosure held two cannister nodes, a 12-drive or 24-drive bay, and a pair of power-supply/battery UPS modules. The controller could attach up to nine expansion enclosures for more capacity, as well as virtualize other storage systems. This has been one of our most successful products ever, selling over 100PB in the past 12 months to over 2,500 delighted customers.
The 12-drive enclosure now supports both 2TB and 3TB NL-SAS drives. The 24-drive enclosures support 200/300/400GB Solid-State Drives (SSD), 146 and 300GB 15K RPM drives, 300/450/600GB 10K RPM drives, and a new 1TB NL-SAS drive option. For those who want to set up "Flash-and-Stash" in a single 2U drawer, now you can combine SSD and NL-SAS in the 24-drive enclosure! This is the perfect platform for IBM's Easy Tier sub-LUN automated tiering. IBM's Easy Tier is substantially more powerful and easier to use than EMC's FAST-VP or HDS's Dynamic Tiering.
Last week, at Oracle OpenWorld, there were various vendors hawking their DRAM/SSD-only disk systems, including my friends at Texas Memory Systems, Pure Storage, and Violin Memory Systems. When people came to the IBM booth to ask what IBM offers, I explained that both the IBM DS8000 and the Storwize V7000 can be outfitted in this manner. With the Storwize V7000, you can buy as much or little SSD as you like. You do not have to buy these drives in groups of 8 or 16 at a time.
The Storwize V7000 is the sister product of the IBM SAN Volume Controller, so you can replicate between one and the other. I see two use cases for this. First, you might have a SVC at a primary location, and decide to replicate just the subset of mission-critical production data to a remote location, and use the Storwize V7000 as the target device. Secondly, you could have three remote or branch offices (ROBO) that replicate to a centralized data center SAN Volume Controller.
Lastly, like the SVC, the Storwize V7000 now supports clustering so that you can now combine multiple control enclosures together to make a single system.
- IBM® Storwize® V7000 Unified
Do you remember how IBM combined the best of SAN Volume Controller, XIV and DS8000 RAID into the Storwize V7000? Well, IBM did it again, combining the best of the Storwize V7000 with the common NAS software base developed for SONAS into the new "Storwize V7000 Unified".
You can upgrade your block-only Storwize V7000 into a file-and-block "Storwize V7000 Unified" storage system. This is a 6U-high system, consisting of a pair of 2U-high file modules connected to a standard 2U-high control enclosure. Like the block-only version, the control enclosure can attach up to nine expansion enclosures, as well as all the same support to virtualize external disk systems. The file modules combine the management node, interface node and storage node functionality that SONAS R1.3 offers.
What exactly does that mean for you? In addition to FCP and iSCSI for block-level LUNs, you can carve out file systems that support NFS, CIFS, FTP/SFTP, HTTP/HTTPS, and SCP protocols. All the same support as SONAS for anti-virus checking, access-based enumeration, integrated TSM backup and HSM functionality to migrate data to tape, NDMP backup support for other backup software, and Active Cloud Engine's local file management are all included!
- IBM SAN Volume Controller V6.3
The SAN Volume Controller [SVC] increases its stretched cluster to distances up to 300km. This is 3x further than EMC's VPLEX offering. This allows identical copies of data to be kept identical in both locations, and allows for Live Partition Mobility or VMware vMotion to move workloads seamlessly from one data center to another. Combining two data centers with an SVC stretch cluster is often referred to as "Data Center Federation".
The SVC also introduces a low-bandwidth option for Global Mirror. We actually borrowed this concept from our XIV disk system. Normally, SVC's Global Mirror will consume all the bandwidth it can to keep the destination copy of the data within a few seconds of currency behind the source copy. But do you always need to be that current? Can you afford the bandwidth requirements needed to keep up with that? If you answered "No!" to either of these, then the low-bandwidth option is you. Basically, a FlashCopy is done on the source copy, this copy is then sent over to the destination, and a FlashCopy is made of that. The process is then repeated on a scheduled basis, like every four hours. This greatly reduces the amount of bandwidth required, and for many workloads, having currency in hours, rather than seconds, is good enough.
I am very excited about all these announcements! It is a good time to be working for IBM, and look forward to sharing these exciting enhancements with clients at the Tucson EBC.
technorati tags: IBM, SONAS, GPFS, SAS, NL-SAS, Active Cloud Engine, Global+Namespace, Storwize+V7000, V7000U, V7000 Unified, block-only, block-and-file, SVC, SSD, Easy Tier, Flash-and-Stash, Texas Memory Systems, Pure Storage, Violin Memory
I gotten several emails expressing worry that I have fallen off the face of th earth. The last two weeks have been educational and eye-opening for me. I can't provide details in my blog, so I will just say that it involved government agencies that IBM refers to as "dark accounts", and that I am now back safely in the USA. Between adjusting to time zone differences, ridiculously long hours, and restricted access to the internet, I was unable to blog lately.
Instead, I will resume my coverage of the [IBM System Storage Technical University 2011]. The "Solutions Expo" runs Monday evening through Wednesday lunch. This is a chance for people to explore all the solutions that are part of IBM's large "eco-system" for IBM System storage and System x products. There were several sponsors for this event.
As is often the case at these conferences, the various booths hand out fun items. The hot items this year were tie-dyed tee-shirts from Qlogic, and propeller beanies from the IBM rack and power systems team. Here is Amanda, one of the bartenders showing off the latter.
After the expo on Tuesday night, my friends at [Texas Memory Systems] held an after-party. Unlike the pens, tee-shirts and keychains at the Expo, these guys had a raffle for real storage products. Here is Erik Eyberg handing out a RamSan PCIe card, valued at $14,000 or so. IBM recently certified the TMS RamSan as External SSD storage for the IBM SAN Volume Controller (SVC). The SVC can optimize performance using this for automated sub-LUN tiering with the IBM System Storage Easy Tier feature.
technorati tags: IBM, Storage, University, Qlogic, TMS, RamSan, SVC, Easy Tier