Tips for operating Data Integrator
You can use the following tips to enhance how your company operates the IBM® TRIRIGA® Data Integrator tool.
Tip | Description |
---|---|
Associate existing records | After all records are created, reuse the Header File to create an Associative upload Header File. Establishing associations after all records are created eliminates issues that surround nonexistent records. |
Checking for data upload errors | Every time the Data Integrator runs and has errors, it creates a Data Upload Error record. To find a Data Upload Error record, select | . Scroll through the records in the result page. Click the file name hyperlink.
Confirm through BO result page | After the data produces a successful notification, review the data through the BO result page. For hierarchical objects, ensure that the child records appear under the proper parent in the hierarchy window. |
Data Integrator cautions | Data Integrator does less checking
of the data than a form into which you manually enter data:
|
Data Upload records | Every time that you use the Data Integrator to initiate the reading of data, it creates a Data Upload record. If you initiate a schedule of batch uploads, there is still just one Data Upload record. After an upload completes, you can learn how the upload went by looking at the Data Upload record that corresponds to the upload. The Data Upload record has a smart section that contains Data Upload Error records. If any errors occurred, you see them in the smart section. To find a Data Upload record, select | . Scroll through the records in the result page. Click the file name hyperlink.
Folder cleanup | System Administration processes for Data Integrator batch upload must include periodic cleanup of the input and log folders. |
Managing files for upload | When files are uploaded in batch mode, the upload happens without the direct participation of a person. The platform looks for the files on the computer it runs on rather than on a computer that a person is using. To identify the computer and the directory that is used for batch data uploads, you might need to consult with the person who administers the IBM TRIRIGA Application Platform environment for your organization. To ensure that files are properly processed,
observe the following rules that regard when and how to put a file
in the directory for processing:
|
Processing an entire batch | It is possible to have a workflow that processes all the records in the batch after the individual records are created or updated. When an upload of a tab-delimited file is finished, the platform runs an Apply action on the Data Upload record that is associated with the upload. As records are processed, an association is created from the Data Upload record to the affected records. The association name is Data Load Status Of. You can use the Association Manager to define an association with this name from the Data Load Status Of business object to the other relevant business object. If you do so, a workflow that is started by an Apply action on a Data Upload record can process all of the uploaded records. You can use a Retrieve Workflow task to retrieve records that are sorted in the order of their value in a particular field. Therefore, you can use any field in the records to determine the order in which the workflow processes them. |
Retire formula, rollup, and calculation workflows | In Production environments, all formula, rollup, and calculation workflows must be retired against the business object before the upload is run. This procedure ensures that the records are being created without forcing the platform to run unnecessary workflows against partial data. It also helps eliminate looping that might occur against hierarchical objects. After the upload is successful, republish and run the workflows against your new data set. |
Run all files during off peak hours | Depending on the amount of information that is uploaded, server resources can become taxed and performance can degrade. Users might notice a performance hit, which might result in false positives and calls to your help desk. |
Timing of record processing | The IBM TRIRIGA Application Platform does not necessarily process uploaded records in the same order they appear in the tab-delimited file. It reads the data from each tab-delimited record in sequence and then uses a different thread to process the data to create or update a record. In most cases, it does not matter in what order records are processed. If you must process the batch records in the same order that they are read, you must force the Data Integrator to process all the records with the same thread. To accomplish this task, set the number of threads that are used by the Data Import Agent to 1. |
TRIRIGAWEB.properties for batch upload | The TRIRIGAWEB.properties file contains file settings for batch upload, specifically for input, processing, output, errors, and log. The files that are defined in TRIRIGAWEB.properties must exist and the file names must match. |
Upload records first | Do not upload a Data Integrator file with all fields and all associations in one file. Create all records for all business objects first. After you import all records, begin the task of associating the records to one another. |
Upload sets of records | Do not push massive amounts of data in a single Data Integrator file. Use sets of 5000 records per Data Integrator file. This practice reduces errors that are related to memory and system capacity. |
Uploading images | If you must assign images to newly uploaded records, you can use the following format. Assume that the image field is named triImageIM. The file name must start with the word "file" such as filemyimagename.jpg or file12345.gif. Based on the example, the file must be placed in the [TRIRIGA install directory]\userfiles\Company-1 directory before you upload the Data Integrator file. |