DB2 Version 9.7 for Linux, UNIX, and Windows

Best practices for data warehousing environments: planning and design

The best practice papers described in this topic help you plan and implement solutions to efficiently manage your DB2® data warehousing environments.

Building a data migration strategy with IBM InfoSphere Optim High Performance Unload

IBM Optim High Performance Unload for DB2 for Linux, UNIX, and Windows V4.02 is a high-speed tool for unloading, extracting, and migrating data in DB2 for Linux, UNIX, and Windows databases. High Performance Unload (HPU) is designed to extract and migrate data from DB2 table space containers. A data migration strategy that uses HPU minimizes storage needs and automates many manual tasks.

HPU uses named-pipes and parallel LOAD operations to stream data from the source to the target database without the need to ever stage the data on disk. You can direct HPU to determine where different partition maps, software levels, and key constraints are used and automatically handle these events during data migration.

HPU can also unload subsets of data from target database without the need to access the DB2 software layer. With this functionality, you can migrate data from larger production systems to smaller pre-production or development environments.

You can download the Building a data migration strategy with IBM InfoSphere Optim High Performance Unload best practice paper from the IBM® developerWorks® best practices community: Building a data migration strategy with IBM InfoSphere Optim High Performance Unload.

Transforming IBM Industry Models into a production data warehouse

Implementing an industry model can help accelerate projects in a wide variety of industry sectors by reducing the effort required to create a database design optimized for data warehousing and business intelligence.

IBM Industry Models cover a range of industries which include banking, healthcare, retail, and telecommunications. The example that is chosen in this document is from a subset of the IBM Insurance Information Warehouse model pertaining to Solvency II (SII) regulations.

Many of the most important partitioned database design decisions are dependent on the queries that are generated by reporting and analytical applications. This paper explains how to translate reporting needs into database design decisions. This paper guides you through the following recommended process for transforming a logical data model for dimensional warehousing into a physical database design for production use in your environment.

You can download the Transforming IBM Industry Models into a production data warehouse best practice paper from the IBM developerWorks best practices community: Transforming IBM Industry Models into a production data warehouse.

Physical database design for data warehouse environments

This paper provides best practice recommendations that you can apply when designing a physical data model to support the competing workloads that exist in a typical 24x7 data warehouse environment. It also provides a sample scenario with completed logical and physical data models. You can download a script file that contains the DDL statements to create the physical database model for the sample scenario.

This paper targets experienced users who are involved in the design and development of the physical data model of a data warehouse in DB2 for Linux, UNIX, and Windows or IBM InfoSphere Warehouse Version 9.7 environments.

You can download the Physical database design for data warehouse environments best practice paper from the IBM developerWorks best practices community: Physical database design for data warehouse environments.

Implementing DB2 workload management in a data warehouse

Using a staged approach, the Implementing DB2 workload management in a data warehouse best practice paper guides you through the steps to implement the best practices workload management configuration on IBM DB2 for Linux, UNIX, and Windows software. The steps create sufficient controls to help ensure a stable, predictable system for most data warehouse environments. This initial configuration is intended to be a good base on which you can perform additional tuning and configuration changes to achieve your specific workload management objectives.

This best practice paper presents a set of definitions representing the different stages of maturity for a workload management configuration in a DB2 for Linux, UNIX, and Windows database. These stages range from the basic stage 0 configuration to the advanced stage 3 configuration. A specific configuration template and process are provided so that you can progress from a stage 0 configuration to a stage 2 configuration. General descriptions and advice about common stage 3 scenarios are also given.

You can download this best practice paper from the IBM developerWorks website: Implementing DB2 workload management in a data warehouse.

Ingesting data into an IBM Smart Analytics System

The key design goals of ingest application design are to balance the required rate of ingest with the availability of processing resources, to ingest data without affecting data availability, and to maintain flexibility in the volumes of data ingested. Identifying service level objectives for the speed and volume of data that is ingested into the data warehouse helps determine the architecture, design, and development of the data ingest application.

The Ingesting data into an IBM Smart Analytics System best practice paper is targeted at people who are involved in the design and development of data ingest applications that are based on DB2 for Linux, UNIX, and Windows database software. In particular, the focus is on ingesting data into the IBM Smart Analytics System environment with configurations that are based on System x® and Power Systems™ servers. You can download this best practice paper from the IBM developerWorks website: Ingesting data into an IBM Smart Analytics System.

Building a recovery strategy for an IBM Smart Analytics System data warehouse

Identifying and planning for recovery scenarios that are most likely to occur are the key factors in determining the speed with which you can recover from data loss or corruption. The recommendations in the Building a recovery strategy for an IBM Smart Analytics System data warehouse best practice paper help you develop a strategy that meets your data warehouse backup and recovery needs. The paper focuses on IBM DB2 for Linux, UNIX, and Windows software in the IBM Smart Analytics System environment with configurations that are based on System x® and Power Systems servers.

This article is targeted at those involved in planning, configuring, designing, implementing, or administering a data warehouse that is based on DB2 for Linux, UNIX, and Windows software. In particular, this paper focuses on the IBM Smart Analytics System environment with configurations based on System x® and Power Systems servers.

You can download this best practice paper from the IBM developerWorks website: Building a recovery strategy for an IBM Smart Analytics System data warehouse.

Using IBM InfoSphere Optim High Performance Unload as part of a recovery strategy in an IBM Smart Analytics System

IBM InfoSphere® Optim™ High Performance Unload (HPU) for DB2 for Linux, UNIX, and Windows V4.2 software is a high-speed tool for unloading, extracting, and repartitioning data in DB2 for Linux, UNIX, and Windows databases. HPU unloads large volumes of data quickly by reading directly from full, incremental, and delta backup images or table space containers rather than through the DB2 software layer.

The Using IBM InfoSphere Optim High Performance Unload as part of a recovery strategy in an IBM Smart Analytics System paper describes how to incorporate HPU into a recovery strategy and when to use HPU to meet your recovery objectives. In addition, the paper contrasts the use of output files and named pipes, covers introducing parallelism, describes the HPU command and control file, and provides best practices for creating control files. Finally, the paper details how to install and configure HPU in an IBM Smart Analytics System environment. You can download this best practice paper from the IBM developerWorks website: Using IBM InfoSphere Optim High Performance Unload as part of a recovery strategy in an IBM Smart Analytics System.

Deploying Optim Performance Manager in large-scale environments

Large-scale environments often have special considerations when it comes to managing performance. In such environments, it is very important to have a comprehensive, proactive performance management approach. IBM InfoSphere Optim Performance Manager software offers such an approach.

Use the Deploying Optim Performance Manager in large-scale environments paper to learn about best practices for deploying InfoSphere Optim Performance Manager Extended Edition software in large-scale environments. You can download this best practice paper from the IBM developerWorks website: Deploying Optim Performance Manager in large-scale environments.