I’m the performance architect for, among other things, the pureQuery platform. My team has the responsibility for not just ensuring that our products perform well, but also to help produce verifiable performance numbers that we can share with confidence.
I’m happy to say that we are ready to share our performance numbers for pureQuery access to DB2 for Linux, UNIX and Windows. (We already have some great numbers published for z/OS for both Java and .NET).
The goal of this particular performance test was to measure throughput improvement using static SQL execution, which is possible to do even for existing JDBC applications with no change to the application source code. The increased throughput comes mainly as a result of saving the cost of preparing the SQL when using static vs dynamic SQL. We typically don’t see the same level of interest in static execution from DB2 for LUW customers as we do from DB2 for z/OS customers because the LUW platform does not have the same memory constraints as z/OS – and therefore LUW customers might be more likely to throw hardware at the problem to achieve greater dynamic cache hit ratios and hence improve throughput.
However, static SQL also provides predictable performance because the access plan is pre-determined and I often find users are happier with predictable response times rather than ultra-fast response which can deteriorate over time.
Static SQL execution also provides much more than predictable performance. By using it, you can significantly improve problem determination and traceability. You can also reduce the risk of static SQL injection from dynamically executing applications. You can read about some of those benefits in this article. And there are additional benefits to pureQuery usage such as literal consolidation or the ability to make emergency fixes to application SQL without changing the application, which you can read about in Sonali’s article on 2.2 features.
OK, now that I’ve hopefully convinced you that there are many, many reasons to consider pureQuery and static SQL execution for DB2 LUW environments, I would like to go ahead and share our performance results.
The measurement environment
Our measurements were done with a typical 3-tier environment of a client, application server, and database server as shown here.
A word about the “ERWW” application we use. ERWW is an OLTP application based on an order entry and tracking system that is designed to exercise the database tier much more than the application tier (that is to say, there is not a lot of business logic in the application). The ERWW workload models a wholesale supplier managing orders, and consists of seven transaction types. The frequency of transactions is set to simulate a realistic scenario; the mix used in the benchmark environment was 47 percent update transactions, 53 percent read-only transactions. The workload is triggered by a Java client program which generates HTTP requests for the required transaction mix.
Before I go into our results, I have to offer up the standard disclaimer that any of you who are familiar with performance work are used to hearing. The tests that we ran were done in a controlled environment where we were able to carefully control extenuating factors that can influence the results. In particular, the type of application you run can significantly affect the results in terms of the mix of database-intensive work versus application-intensive work. The ERWW workload is a very database intensive workload and most of the work is done by the database server processing SQL requests. Therefore, by using pureQuery to optimize the database server side processing, we are in fact optimizing a large chunk of the workload. Consequently the performance gains for this workload are significant. We chose ERWW because it was readily available to us, and not because we thought it would give us the best results. I guess what I am trying to say is that your results will vary.
OK, now that that’s out of the way. We measured static execution using both client optimization of an existing JDBC application and also as a ‘new’ version of the application written in pureQuery annotated method style. The performance is reported in Normalized Throughput Rate - Transactions Per Second (ITR). The ITR is the notional throughput rate assuming that the CPUs are 100 percent busy. For example, consider an application with a transaction rate of 200 transactions per second at 75 percent CPU consumption. The ITR for this application would be 200 * 100/75 = 267 tps. This is the notional transaction rate that could be achieved if the CPUs were 100 percent busy, and no other bottleneck is hit first.
We measured the JDBC workload with both 90% and 95% package cache hit ratios. To achieve a 90% package cache hit ratio with the ERWW workload, the DB2 Package Cache (PCKCACHESZ) was sized to 180 x 4k pages, and for a 95% hit ratio it was sized to 210 x 4k pages.
Here are the results with a 90% cache hit ratio. The results are shown on the vertical axis as the database ITR improvements over the baseline of JDBC.
As you can see, client optimization almost doubled throughput over the existing JDBC application. The new application that uses pureQuery method style API more than doubled the database transaction throughput.
The results with a 95% cache hit ratio are shown here.
Note that we achieved significant throughput improvements even with a high package cache hit ratio.
In summary, pureQuery and static execution can offer many benefits, one of which may be improving the performance of your data servers with your applications. By changing the dynamic SQL to static SQL, pureQuery should help you either achieve better throughput on your existing hardware, or reduce CPU consumption of your existing hardware, allowing you to load more tasks onto it. I highly recommend that you also check out the bigger picture around Java acceleration (including the other benefits I mention) as shown in this video.
We have been busy working on an update to Data Studio just in time for DB2 10.1 for Linux, Unix and Windows
. This release includes enhancements throughout the product as well as the added support for DB2® V10.1 for Linux, UNIX, and Windows databases which include the following features:
- Adaptive compression for table rows
- Special registers for temporal tables in server profiles
- Time-based data management with temporal tables
- Data management using multi-temperature storage
- Data security with row and column access control (RCAC)
These DB2® V10.1 features are fully supported by Data Studio making it easier for you to take advantage of them.
Other product enhancements include updated syntax checking, additional support for pureScale, improvements to SQL routine support, improved performance information gathering, as well as enhancements to query tuning, shell sharing, and the Data Stuido web console, here are the details:
- We updated the syntax checking in the SQL and XQuery editor for all DB2 for Linux, UNIX, and Windows Command Line Processor (CLP) commands to provide validation of your commands and utilities as well as SQL.
- For pureScale we added support to help verify DB2 Cluster Services Status and Manage DB2 Cluster Services Configuration to further assist in the use of pureScale in your enterprise.
- We have made the following improvements around SQL routine support:
- You can now use the routine editor to create, edit, deploy, run and debug routines directly from the Data Source Explorer without having to create projects, simplifying the task of routine development when there is no other need for a development project.
- Likewise, you can re-compile a routine directly from the Data Source Explorer to enable/disable debugging.
- You can now create unit routine test cases with saved parameters and pre/post SQLs and then compare the current results with a saved base line result. This helps catch regressions when editing routines.
- To make it easier to tune your SQL before deployment, there is now an improved user interface for profiling DB2 LUW routines and to simplify integration, you can invoke InfoSphere Optim Query Workload Tuner from the profiling view.
- You can now use the Configuration Checker to verify that all essential server components are in place for routine development on DB2 Z.
- We made it easier to gather database performance information for stored procedures in DB2 for Linux, UNIX, and Windows databases.
The query tuning features in Data Studio have been enhanced to include the following:
- We added a Start Tuning wizard to the Task Launcher for a better up and running experience
- Our access plan graphs can now display three new nodes:
- Show the rebalancing of rows between SMP subagents with REBAL
- Show multi-distinct processing with MGSTRM
- Show a zigzag join with ZZJOIN
In addition to the Data Studio query tuning enhancements, the client for InfoSphere Optim Query Workload Tuner is now part of the Data Studio v3.1.1 product. After you install Data Studio v3.1.1, all that you need is a license to enable InfoSphere Optim Query Workload Tuner, simplifying the install process for those of you who use both Data Studio and InfoSphere Optim Query Workload Tuner.
On top of all the new functionality, the Data Studio v3.1.1 clients are now built off of Eclipse 3.6, which means the Data Studio full client is now compatible (for shell-sharing) with newer Eclipse based offerings such as the latest IBM Rational Application Developer and Rational Developer for System z.
For the Data Studio web console, we have included the following enhancements:
- New alerts so that you can configure for the following:
- High availability disaster recovery (HADR) databases
- DB2 pureScale instance status
- When the DB2 pureScale cluster facility is not found
- For DB2 V10.1 for Linux, UNIX, and Windows databases, we also added the following enhancements:
- Your storage group information is provided in alerts both under the 'Storage' category as well as on the 'Current Table Spaces' dashboard.
- You can create and manage user-defined alert types using custom scripts so you are no longer limited to the out-of-box alerts.
- You can import existing tasks from the DB2 Task Center into the Job Manager to ease the move from the DB2 Task Center to the Job Manager in Data Studio.
- You have more control over your jobs as we have added the ability to cancel a job that has started and specify a timeout period for your jobs.
Visit the Data Studio product page
and download from developerWorks
. Data Studio is bundled with a number of IBM data servers such as DB2 for Linux, Unix and Windows, DB2 Connect and IBM Informix. For a good overview, check out the "Getting started with IBM Data Studio for DB2
Big news last week on the Optim tools front.
- DB2 introduced a new server, Advanced Enterprise Server Edition, that bundles up tons of tools for a phenomenal value, including Optim Performance Manager! Anyone looking at or owning Enterprise Server Edition needs to take a look at this high value offering. See the announcement here.
- All DB2 for Linux, Unix, and Windows customers with current subscriptions now get access to Optim Database Administrator and Optim Development Studio in addition to Data Studio. Use Optim Database Administrator to save time and reduce errors in administering DB2 databases. Use Optim Development Studio to accelerate application development against DB2 and Oracle. Let the downloading begin! (Links further down)
- We released mod releases and fixpacks today for several tools with great added function with more coming later in the month.
Let me take you through the highlights.
Administration and Availability
- Data Studio provides an easy to use development and administration tool. Data Studio 2.2.1, still available in either a standalone or integrated development environment (IDE) configuration, is enhanced for first time user experience, customizable catalog navigation, customizable, template-based routine development (check out the new article from Michael on this), enhanced data browsing, a new query tuning UI, a new health monitor, and updates for use with DB2 for z/OS V10. Early user feedback has given rave reviews to the new task launcher for improving the experience for first-time users, particularly those not familiar with the Eclipse environment. The new health monitor is a separately installed component with Web UI, but is seamlessly launchable from the Eclipse client to monitor DB2 for Linux, UNIX, and Windows health and view alerts, application, utilities, and storage. Download it here.
- Optim Database Administrator saves time and reduces errors associated with complex database changes for DB2 for Linux, UNIX, and Windows databases. Installable together with the Data Studio IDE, it provides copy and paste, compare and synch, and data migration features that identify and manage dependent objects automatically. It provides seamless integration with InfoSphere Data Architect for governing and accelerating the database design through deployment process. Optim Database Administrator 2.2.3 inherits most of the Data Studio enhancements and enhances compare filtering. It also improves performance for building change scripts when working with a small number of objects in very large databases. Use with Optim Test Data Management to streamline test data creation, reduce storage consumption, and manage data privacy. Download Optim Database Administrator 2.2.3 here.
- Optim High Performance Unload reduces demand on system resources, shrinks batch window requirements, accelerates database migrations, and minimizes impact on business operations by speeding up unload processes by 4X or more. Optim High Performance Unload 4.2 is enhanced with DB2 pureScale support and a new option for unloading from a schema.
Design and Development
- InfoSphere Data Architect enhances design productivity, data quality, enterprise consistency and data governance. It supports business and IT collaboration to keep IT aligned with business objectives and helps to define and enforce compliance to enterprise standards. New in InfoSphere Data Architect 7.5.3 is enhanced multi-dimensional modeling including automatic discovery and notation of facts, measures, dimensions, and outriggers, automatic de-normalization of operational data for designing warehouses, and lossless integration with InfoSphere Warehouse and Cognos Framework Manager. Watch for a demo of this new capability later this month. Download InfoSphere Data Architect 7.5.3 here.
- Optim Development Studio increases development efficiency up to 50% for Java data access and facilitates cross-system development and migration. A quick review: it supports development for DB2, Oracle, and Informix. Its SQL outline feature facilitates developer and DBA collaboration quickly isolating all the SQL for review and enabling impact analysis by correlating SQL with source code, database objects, and ALTER requests. Its application metadata can be transferred to Optim Performance Manager to pinpoint problem SQL to the Java method and line of code. Plus it positions the organization to take advantage of Optim pureQuery Runtime to enhance and lockdown performance and mitigate SQL injection risk. New in Optim Development Studio 2.2.1 is the ability to deploy SQL scripts, procedures, and functions to multiple development and test servers with a single gesture. The customer feedback has been fabulous. Clifford will have an article coming out in a couple weeks. Included with Optim Development Studio are development-only copies of Optim pureQuery Runtime and the DB2 Connect JDBC driver, so download all of these together from here.
- Optim Performance Manager provides out-of-the-box performance monitoring and management to improve quality of service and prevent impacts to business operations. Its intuitive web UI provides use-anywhere monitoring, alerting, and diagnosis of potential performance bottlenecks. Integrated tooling enables easy setup exploitation of DB2 workload manager to make sure that system database resources are applied according to business priorities. And its performance warehouse provides needed context for proactive performance management, trend analysis, and capacity planning. The new Optim Performance Manager fix pack (timed for delivery with DB2 Advanced Enterprise Server Edition) does double-duty providing health monitoring for all your other development and test systems at no additional charge providing a single integrated console for monitoring all your DB2 servers. Plus it has added more flexibility in alert management, DB2 Workload Manager configuration, more performance metrics, more reports, and more consumability improvements. Want more? Upgrade to Optim Performance Manager Enterprise Edition. It enables monitoring of end-to-end database transaction times to assure that critical workloads are meeting their response time targets. New in this delivery, tell us the response time goal for the workload and let us figure out how to make it happen. Plus there’s now support for .NET and Type 2 connection workloads, enhanced Tivoli integration, and more granular performance detail.
- Optim Query Tuner gives DBAs and developers expert advice on query structure, access paths, statistics, and indexes to improve query performance. Optim Query Tuner for Linux, UNIX, and Windows 2.2.1 sports a new user interface to streamline query tuning workflow.
- Optim pureQuery Runtime improves performance and predictability while reducing risk. Optim pureQuery Runtime 2.2.1 adds support for ODBC and CLI applications. Using pureQuery, you can configure a DB2 application to run the application's dynamic SQL statements statically. You can also convert SQL statement string literals to parameter markers to improve performance, replace poor performing SQL statements with optimized statements without changing the application, and reduce the risk of SQL injection attacks by restricting what SQL statement are allowed to run against the database.
OK, let me take a breath! As you can see, we’ve been busy. Hope you’re planning to stop in and see us in Las Vegas at the Information On Demand conference. Kimberly will post a summary of our key sessions and events soon.
Now that IBM Data Studio Version 3.1.1 is available for download, we thought it would be helpful to share this list of documentation links to help you with everything from finding what's new in this release to getting you up and running in IBM Data Studio:
IBM Data Studio Version 3.1.1 release information:
What's New in IBM Data Studio Version 3.1.x
Data Studio V3.1.x features: by data server and component
Determine which features are available in each component of Data Studio as well as what databases are supported for each feature.
Fixed APAR list for IBM Data Studio Version 3.1.1
Roadmap, overview, and help information:
IBM Data Studio 3.1.x Information Roadmap
Links to information and resources for each phase of the data lifecycle in IBM Data Studio.
IBM Data Studio Version 3.1.1 Overview
IBM InfoSphere Optim Query Workload Tuner Version 3.1.1 Overview
IBM InfoSphere Optim pureQuery Runtime Version 3.1.1 Overview
IBM Data Studio Version 3.1.1 system requirements
Download IBM Data Studio Version 3.1.1
Quick Start Guide for IBM Data Studio Version 3.1.1
Shell-share IBM Data Studio:
A guide to help you with a standard installation of the IBM Data Studio components.
IBM Software products installed together that share a common environment
Review the list of compatible products to determine if you can shell share IBM Data Studio.
Limitations for shared Eclipse 3.6.x-based IBM product installations
Still can't find what you're looking for? Send me an email @ email@example.com
IBM Data Studio
I’m hoping that you
saw the big DB2 10 for Linux, UNIX, and Windows announcement or will be attending the virtual conference The Future of Database and Data Warehousing Software today through Thursday. Concurrent with
the upcoming delivery of DB2 10, scheduled for April 30th, our
InfoSphere Optim tools team is planning a set of releases designed to provide
support for the new DB2 10 features. We hope these enhancements will help
you get going with new features including time travel query, row and column
access control, multi-temperature storage, adaptive compression, and query
performance improvements. The new releases are:
Data Architect V8.1
Optim Query Workload Tuner V3.1.1
Optim pureQuery Runtime V3.1.1
Optim Performance Manager V5.1.1
Optim Configuration Manager V2.1.1
Optim High Performance Unload V4.2.1
Merge Backup V1.1.1
Recovery Expert V3.1.1
The new capabilities help developers, DBAs, or security
develop, and deploy temporal tables, develop and validate time travel
queries, or access and view row history.
recommendations for adaptive compression use, deploy compression, and
validate compression savings.
to safeguard data enabling security administrators to define, view, and
deploy row and column access control and DBAs to analyze the impact of new
security controls and update routines accordingly.
the value of multi-temperature storage by monitoring storage group access,
analyzing storage group growth and contents, enabling WLM configurations
to prioritize queries against high performing storage, and automating
period roll-off for range partitioned tables to lower cost storage.
access plan improvements and regressions based on new optimization
strategies and get actionable recommendations for improvements.
multiple standby availability and alert availability risks.
Furthermore, InfoSphere Optim tools continue to extend support
for production deployments of DB2 pureScale environments with added
configuration, administration, and monitoring capabilities. Developers will
benefit from new tool support that facilitates application migration and
co-existence, with added support for PL/SQL procedure profiling and Oracle
database federation definition.
And those are just the
enhancements focused on DB2 10. Over the next few weeks, we’ll be blogging on
the releases in detail including enhancements not mentioned here.
And stay tuned!
Holly Hayes, Optim Tools Product Manager
I've been around the block enough times to see that the bricks look the same. I have seen the same performance issues repeated at untold companies with the biggest issue being identification and performance tuning for Java applications.
To be specific, it is the Java communications with DB2 for z/OS that has profoundly perplexed me and my z/OS colleagues. Java, from my perspective, has sometimes been a performance nightmare on z/OS systems. By using the Distributed Data Facility (DDF) for communication, we use a one-size-fits-all approach by using a single Workload Manager service policy. This isn’t because we want to, but because we have to. To make matters worse, these connections come in sufficiently generic that there is NO WAY to figure out what Java application created this thread or multiple threads.
I can't tell you the number of times I would hear a Java programmer wander into my cube and tell me that they have a problem but they don't know where it is. "..Jeff, can you look at...” in which case I'd say, "STOP! Look at what? A thread? I have hundreds of them. Which one?". And you wonder why DBAs get a reputation as being difficult to work with! But that is another story...
Recently, I joined the IBM Data Studio Enablement team. One of our charters is to articulate the value of Data Studio, pureQuery, DB2 Optimization Expert, and a few other tools. OK, I am an old "green screen" guy. A teammate accused me of not embracing our products to which I answered that I am reluctant to jump on any bandwagon unless I see a true value - Not just as an IBMer, but as a z/OS DB2 System Programmer. How can I "sell" a product if I am not "sold" on the value myself?
Guess what? I found the product combination that is going to change the way Java and DB2 for z/OS work in performance. It is Data Studio Developer with pureQuery. Why? Because I can now uniquely identify the Java thread correlation ID with a unique name which means I can now see it on OMEGAMON for DB2 and I can redirect this work to a service class in Workload Manager specifically used for Java work. (See this demo for an example of how you can see the unique names in OMEGAMON.)
The tooling in Data Studio Developer allows the programmer to quickly develop the Java structures necessary to access DB2 for z/OS. In addition, Java application developers can now easily bind a package, which accesses DB2 for z/OS, used by the Java application. Finally!! We’ve had this DB2/application access technique forever with COBOL using CICS. Now we have it with Java, too.
As a result of having a statically-bound package for Java code with a type 4 driver, I can now set up a Workload Manager service policy for Java DB2 calls as it passes through subsystem type DDF using a PK rule. To go one step further, WLM could be set up with a JAVAHIGH and a JAVALOW service classes. Then, these classes can be prioritized, have a time goal applied for period 1 and a velocity goal applied for period 2. Then, using the naming rule PK for package name, these service classes can be referenced to specific service classes.
Not too many products really flip my switch, but this combination of Data Studio Developer and Data Studio pureQuery Runtime is one of them. Dynamic SQL tuning with Java and DB2 for z/OS has been my nemesis for a very long time. pureQuery gets Java code as close to "well-tuned" as I have ever seen. I would recommend Data Studio Developer and pureQuery to anyone struggling with an out-of-control distributed environment going against DB2 for z/OS as the DB-tier.
-- Jeff Sullivan
After IBM bought Telelogic
I was (and still am) frequently asked “What is the difference between InfoSphere Data Architect and Telelogic (now Rational System Architect). Here is the information I have about the product roadmaps and how these two products complement each other.Before the acquisition:
First, some background to set the stage. As you know, even before the acquisition IBM marketed a data modeling tool which at that time was called Rational Data Architect, which had originated from the old Rational Rose and Rational Software Architect product line. It has since evolved into a robust data modeling solution that is today called InfoSphere Data Architect
Meanwhile, Telelogic was in the business of creating Enterprise Architecture
solutions. The best way to explain Enterprise Architecture is to look at a common application: an IT consolidation project. The goal of the business is to leverage the architecture blueprint of strategy, process, applications, systems, and data to best identify consolidation candidates, then understand the impact of change to the organization in order to successfully plan and implement the transformation. The business goal is to leverage architectural information to maintain, even increase, operational efficiency while reducing costs. Now how do you accomplish this? This is precisely what Enterprise Architecture enables you to do. It lets you come up with the plans and processes needed to move from the goals of what you are trying to achieve from a business perspective. To successfully build a current and accurate architecture for planning, information needs to be imported from various sources within the organization. This information can already exist within Visio (process), Excel or asset management tools (IT content), or data modeling tools (information). Telelogic System Architect was (and is) dedicated to building this higher level Enterprise Architecture, but other tools are more suitable for the detailed work in an individual domain (breadth versus depth). However, synchronization among these point solutions and enterprise architecture solutions is critical for consolidation of data and elimination of duplicate work. The situation today after the acquisition:
Because of the acquisition, IBM and Rational’s robust Solution and Data Architecture line, including InfoSphere Data Architect, provide the capabilities required by focused solution and data architecture teams. System Architect and InfoSphere Data Architect provide customers the solutions they need to address both their high level EA initiatives, and focused data architecture initiatives, while providing a path of alignment between the two efforts. Today, Telelogic Solution Architect has been renamed to Rational System Architect
. Sure, there is some overlap, but this overlap becomes a value-add to IBM customers. Going back to the IT consolidation example, more accurate decisions can be derived from the Enterprise Architecture if data in the architecture comes from current and accurate sources. System Architect’s import from InfoSphere Data Architect extracts information from an accurate and current source. This is one reason that Rational System Architect has a data modeling component.
So to say all of that in one sentence, InfoSphere Data Architect is meant to be the data modeling tool focused on data architects and database modeling, whereas the data modeling components in Rational System Architect are meant to be the tools that support the data aspects of the broader enterprise architecture strategy.Going forward
IBM’s plans for the Rational System Architect product are to expand the concept of Enterprise Architecture Management, making Enterprise Architecture more relevant for more stakeholders within the organization. This is being accomplished by expanding the harvesting, analysis and reporting capabilities in core System Architect, while expanding supported integrations between System Architect and other focused domain solutions. This means that you will see more and more requirements for the lower level tooling being fulfilled by legacy Rational products. As for InfoSphere Data Architect, we are continuing to compete with our peers for the hearts and minds of data architects, enhancing data modeling capabilities, while extending the value proposition of what data designers do by enabling policy-based, management-by-intention capabilities supporting data governance initiatives.
Hope this helps.
Well, I'm glad I got your attention... One of the things that I've been pondering lately is what percentage of DBAs manage more than one database platform. Since this is an IBM blog, how many DBAs manage both DB2 for Linux, UNIX, and Windows and DB2 for z/OS? I've heard that they exist, but maybe I haven't been asking, and y'all haven't been telling. Does your shop have any of these elusive creatures? If so, what platforms do they support? If not, why not? Does this model work well for you? Why or why not? Is this something that you'd like to see happen?
One of our objectives in creating heterogeneous database tooling is to help reduce the learning curve when using another database platform. While this is still a work in progress for parts of our portfolio, I'd sure like to know how important this is to DBAs.
So, please tell... Post your answer here, or on the IBM Optim Facebook fan page (click on the Fans tab if you don’t see it), or, if you’re shy, you can just send me an email at bfsmith at us.ibm.com.
Technorati Tags: smith
Technorati Tags: ibm
Hello, long time reader, first time blogger here. I work as a tech lead managing the advancement of heterogeneous database support for Optim Development Studio and Optim Database Administrator product offerings.
Pleasantries out the way, I am here to tell you about the new ways of packaging our no-charge capabilities that we hope you’ll like. We’re calling this is no-charge capability Data Studio and are using the Optim name for the is value-added capability. The goal is that this naming convention should be less confusing and simpler. You can get your basic admin and development toolingwith Data Studio, and then add additional functionality if needed byacquiring other Optim products (or Rational or InfoSphere etc...). You can try some of these additional capabilities by downloading and using the trials
Data Studio comes in two flavors:
- The first being the standard Eclipse IDE (Integrated Development Environment) which is the way most of the Optim offerings are released.
- The second is what we call stand-alone. This is built as an Eclipse RCP (Rich Client Platform) package.
The stand-alone package is the one I will focus on in this blog. Eclipse RCP in simplistic terms refers to the absolute minimum set of plug-ins required to create an Eclipse-based Rich Client Application. One of the primary goals for the Data Studio (RCP) stand-alone package was to provide a lightweight executable that would help DB2 and Informix DBAs perform day-day simple admin tasks. So, all references to JDT (Java Development Tools) do not exist which also helps in keeping the stand-alone image lightweight.
The stand-alone package is only 196 MB (network download size) and available on 32/64 bit platforms for Windows XP/Vista and Linux RedHat/SUSE which you can download here
. The installation process itself is trivial. It’s a self-extracting binary that lays out the files in appropriate directories. You will notice that the installer creates a default workspace for you in $HOME/IBM/Data Studio 2.2 stand-alone directory. You can change that at a later time. Since the stand-alone version is completely self-contained, a JRE v1.6 (Java runtime) is also bundled and installed with the product. The built-in help and welcome experience provide appropriate context-sensitive help and tutorials.
As depicted in the table below, the stand-alone version is rich with features that enable DBAs perform their day-day tasks effectively. The main difference between the two packages is that the IDE package also has support for Java stored procedures, Web Services, SQLJ development and XML because it targets developers as well.
Default perspectives are also different. For the stand-alone, you will be presented with Database Administration; for the IDE, the default is the Data Perspective. Both packages support Data Development Project creation, with the IDE flavor able to create/debug Java stored procedures (in addition to SQL stored procedures, supported by both.)
A key highlight of both no-charge offerings is to let you know when another offering can help you perform a task. The IDE version is installed via IBM Installation Manager (IM) and by definition can shell share with other Eclipse-based products such as Optim, Rational and InfoSphere.With the stand-alone package, if you want to shell-share with another product, you will need to switch to the Data Studio IDE package. Not to worry though if you started with the stand-alone package but then want to shell share with other products. Any/all work done with the stand-alone can be reused after moving to the IDE package.
Please remember to read the system requirements
before you download. It references important information like Java Runtime versions, Linux download tips, etc. Also, you can check out the discussion forum
if you have questions.
It is our sincere hope that you give this a spin and drop us a line (or add a comment below) about what you think about Data Studio.
-- Srini Bhagavan
I want to take a moment of your time to share the new support page that highlights the features and capabilities between the stand-alone and IDE packages of IBM Data Studio:
Feature in IBM Data Studio (Stand-alone versus IDE)
Previously, this feature comparison table was available only from the Features and Benefits page but now it has its own dedicated page and is available through the IBM Support site. Use this page to compare features and help you determine which package is appropriate for your needs.
Using software can be like trying to get to a freeway onramp that you know well by driving through an unfamiliar part of a city. You know what the outcome should look like (cruising on the open highway), but to get there you have to learn new landmarks ("Ah, so I turn left at the donut shop!"), new shortcuts ("So, I saved 5 minutes by taking Jefferson. I'll have to remember that."), the streets that go only one way (*pant, pant* "No, I guess I can't drive that way down Sycamore after all."), and more.
The information roadmap for InfoSphere Optim Query Tuner and InfoSphere Optim Query Workload Tuner will help you get to know the streets and landmarks of those two products, so that you can tune single SQL statements or workloads and get to your open highways faster. You'll find it at http://www.ibm.com/developerworks/data/roadmaps/roadmap_ioqt_ioqwt.html
If you have any suggestions for links to include in it or have any other comments about it, please let me know.
Information Development for InfoSphere Optim Query Tuner and InfoSphere Optim Query Workload Tuner
As I have now written a few blogs, some folks have asked, "who are you and why are you blogging", so here it goes :-)
My name is Eric Naiburg and I am a program director at IBM responsible for product marketing and strategy of the Integrated Data Management products. I started my career working with databases as a product manager at a small company Logic Works where I was responsible for a product called ERwin. I am sure most of you have heard of it. Before getting into product management though, I learned about architecture and modeling in a very different way than computer science, I built about 50 houses in central New Jersey with a few partners. After spending about 5 years with Logic Works which later became Platinum Technologies and is now CA, I left to join Rational Software and build out Rational Rose Data Modeler which was the precursor to InfoSphere Data Architect. Over 7 years with Rational and IBM Rational, I held several positions including product management, solutions marketing and product marketing. I also was privileged enough to work with a good friend and colleague Robert Maksimchuk where we wrote 2 books:
About 3 years after IBM had acquired Rational, I decided to take some time and try my hand working for a much smaller company where I joined Dr. Ivar Jacobson and his consulting company as Vice President of Sales and Marketing. I spent more than 2 years helping to grow Ivar’s business, while learning great things from him, but in the end, I decided that IBM was the place for me and last September, I decided to come back. I had a great opportunity to join the Optim (Princeton Softech) team about a year after they had been acquired and here I live happy and enjoying working with a new group of people, new products and a team with a ton of energy.
Here is my promised update on InfoSphere Optim Performance Manager (OPM) 5.1.1 that will be available April 30th. As I mentioned in my earlier posting Optim Releases Support DB2 10, the majority of the enhancements are in support of DB2 10 features, so let me go over those first:
- Multi-temperature support – DB2 10 multi-temperature data management allows you to store different portions of your data on different classes (speed and cost) of storage so that you can match access requirements to underlying storage and non-disruptively change them over time as data ages and as access requirements change. “Storage groups” are used to define the different classes of storage and tables spaces are now defined within a storage group. OPM adds monitoring support for storage groups showing the table space to storage group relationship on alerts and dashboards and provides a new storage group report to understand utilization and growth. In addition, DB2 Workload Manager (WLM) was enhanced to enable prioritization based on the storage group of the data accessed. These features are enabled through the OPM user interface as well
- Multiple HADR – DB2 10 supports configuring multiple standbys now. OPM has improved alerting for high availability and disaster recovery (HADR) environments including additional alerts regarding standby connection and readiness, alerts regarding logging issues between the primary and standby, and distinguishing a single failed standby from last remaining failed standby.
- pureScale – DB2 10 continues to enhance the pureScale feature and OPM continues to enhance its monitoring of pureScale deployments. New alerts recognize cluster alerts failure, additional state changes in the cluster facility, and threshold alerts on group buffer pool hit ratio and cluster cache facility virtual memory, and page reclaim rate. DB2 10 enables WLM for use with pureScale and OPM is enhanced to support that configuration as well.
- WLM Dispatcher – DB2 10 provides a new WLM dispatcher feature that allows you to manage what percentages of CPU resources are assigned to service classes. OPM surfaces the options for dispatcher shares and limits in the OPM WLM configuration panels.
In addition, there are a few enhancements apart from DB2 10 support I want to mention that I think you’ll be excited about.
- User-defined Alerts – We get requests for additional alerts that are beyond our ability to fit into plan in the short term. Thus we are enabling user-defined alerts which will give you the ability to write your own SQL or scripts to detect issues and link them into the OPM standard alerting, notification, and blackout infrastructure. Now you can generate alerts for errors in db2diag.log, canary queries, or other alerts important to your environment. We will have some sample scripts in the Information Center. .
- More granular collection intervals – OPM configuration is based on user-specified collections intervals. Now you can specify different collection intervals for different kinds of data. For example you might want to collect buffer pool information every minute but table space information every hour.
- At-a-Glance performance overview and baseline – Finally, and for me the most exciting, is our introduction of baseline support. This is actually just the first step in a series of planned enhancements in this area. The At-a-Glance view appears in the Overview dashboard as a new tab and provides a higher level view of system activity: CPU distribution overall as well as within DB2, workload, and specific tabs for examining locking, I/O, and SQL processing. From here you can set baselines using the time slider to select the baseline timeframe. Once a baseline is set, the At-a-Glance values are automatically compared to the baseline and significant deviations indicated. Here’s a simple example of the At-a-Glance view
More information, including the Information Center updates, will be available with the general availability.
Holly Hayes, InfoSphere Optim Tools Product Manager
In today's economy doing more with less is imperative. You want to get the most out of your DB2 investment because your enterprise depends on it. IBM Data Server Manager is a simple, scalable and smart solution to administer, monitor, manage and optimize performance for DB2 for Linux, UNIX and Windows databases across the enterprise. Let's take a look at the different facets of where IBM Data Server Manager can empower you to do more.
Let's start with performance. Think about all those many times where you were responsible for diagnosing and troubleshooting a performance problem. There are many factors that you need to consider in order to find out the root cause of performance problems like CPU time, constrained network, data skew, poor statistics and many more.. Solving database performance problems in a multi-tiered environment can be very daunting to say the least. It's good to know that IBM Data Server Manager is there to help by providing access to real time and historical performance data, capability to identify potential data access bottlenecks, comparing SQL performance over time, and even get alerted of emergent issues using proactive alerts.
Let's now look at query tuning. Use the query tuning capabilities and let all the built in advisors and statistics make recommendations on how to fix poorly performing queries. Don't know where you can take advantage of BLU Acceleraiton yet? With IBM Data Server Manager you can get recommendations on what tables will get the most benefit if you convert to BLU...This means you can spend less time wondering where BLU would be of value and more time implementing the next generation of in memory computing for DB2.
Having troubles identifying all the different clients that are connecting to your database server? Let IBM Data Server Manager help you control what those clients are doing and even take an inventory of different connection properties from one central location so that you can control how they should be set. You can also control memory management and let the tool advise you on where to take advantage of adaptive compression in DB2 10.5.
From a database administration capability IBM Data Server Manager is there to help the novice user run various utilities against DB2 like setting up BLU tables, but it also helps the non-DBA user to do things like edit and run SQL statements with help using content assist. Other areas of interest for the DBA are catalog navigation, object management, privilege management, maintenance and availability management.
Think about all of these capabilities contained in one product. IBM Data Server Manager provides a simple quick and easy install and simplified user experience with an integrated web console. Look at real time and historical metrics that can scale to 100's of databases in your enterprise. Use the smart alerts in IBM Data Server Manager to optimize query performance and give guided problem diagnosis and workflow analysis from investigation to action.
In order to be the most efficient out there you want to stay ahead of your competition by being more lean and fit. DB2 database software already leads the way in terms of performance, scale and reliability. Let IBM Data Server Manager help you make the most out of your database investment allowing you to increase SLA attainment, reduce IT costs and align with business priorities.
updated 6/16/ to fix minor typo.
Following up on Curt's blog
about the new releases in June, let's take a deeper look at what is new in InfoSphere Data Architect
Building on top of the privacy specifications for generating test data that was already built in the product from December 2008 you will now be able to pick from a predefined list of categories for specific data privacy information. It's probably best to explain this with an example. Let's say you have a credit card column that you want to mark as private by generating a random number where you maintain the first 4 digits of the card. Within InfoSphere Data Architect you can specify that you want to use the credit card masking policy, and IDA will be able to connect to Optim Test Data Management and Data Privacy
solutions to get the appropriate masking method that should be used. Not only can you generate this in the design phase of your model you can now share this with Optim Development Studio
so now when developing applications you have the ability to view what data is private and even look at the SQL that accesses the sensitive data.
Also new in IDA 7.5.2 is the capability to size storage requirements and estimate for data capacity and growth. This is often called volumetrics
support and as per customer requests we have implemented this in the new release.
Finally, building on the fact that InfoSphere Data Architect is more than just a data modeling tool, we have leveraged all the different use cases that customers have implemented to improve on the different integration scenarios that we provide with IDA. We already know that Data Architect is built on top of the Rational Software Delivery platform (reminder, this product used to be called Rational Data Architect) and we continue to improve in those areas, but we have also enhanced integration scenarios related to Information Management as well. Since most of the Optim Solutions for Integrated Data Management are built on Eclipse you can utilize the sharing of connection information feature that was introduced in the June releases. Also new in IDA 7.5.2 is improved integration with IBM Industry Models
and glossary information. All Industry Models and the newly added glossary information can now be managed in InfoSphere Data Architect.
The trial of this release will be available in a few weeks at the current trial download
location. The announcement letter is here
. Oh, by the way, the announcement letter also contains information about the updated Learning Services course for IDA
that has been enhanced to cover more product capabilities. I always strongly recommend that new users get education, and this new and improved course can help you get what you need to get started.
-- Anson Kokkat