IBM Content Analytics with Enterprise Search, Version 3.0.0                  

Crawler setup requirements

To crawl some types of data sources, or to ensure that security can be enforced when users search or mine collections, you might need to run setup scripts on the crawler server.

You might also need to configure the servers to be crawled so that the system servers can communicate with the data source servers, access the data to be crawled, and obtain security data. The following table summarizes the post-installation tasks that are required by each crawler type. You must run the scripts, and configure the target servers, before you create the crawler.

Table 1. Crawler setup requirements
Crawler type Setup requirements
Case Manager You must install the IBM® FileNet® Content Engine Client on the crawler server. You must then run the escrfilenet.sh or escrfilenet.vbs script on the crawler server.
Content Integrator You must run the escrvbr.sh or escrvbr.vbs script on the crawler server.
DB2 You must run the escrdb2.sh or escrdb2.vbs script on the crawler server. You must also ensure that the DB2 Administration Client is installed on the crawler server.
Content Manager You must run the escrcm.sh or escrcm.vbs script on the crawler server.
Domino Document Manager, Notes, and Quickr for Domino To crawl Lotus® Domino® servers that use the Notes® remote procedure call (NRPC) protocol:
  • On an AIX® system, ensure that the I/O Completion Port module is installed and available on the crawler server.
  • Run the escrnote.sh or escrnote.vbs script on the crawler server.
  • Install a Domino server on the crawler server. The Domino server must be a member of the Domino domain to be crawled and, for maximum security, configure the Domino server as a Lotus Domino Trusted Server.
To crawl Lotus Domino servers that use the Domino Internet Inter-ORB Protocol (DIIOP):
  • On an AIX system, ensure that the I/O Completion Port module is installed and available on the crawler server.
  • Configure the crawler server to use the DIIOP protocol.
Exchange Server To use Exchange Server key management and the Secure Sockets Layer (SSL) protocol, the keystore file must be on the crawler server (you specify the password to access the file when you configure the crawler).
FileNet P8 You must install the IBM FileNet Content Engine Client on the crawler server. You must then run the escrfilenet.sh or escrfilenet.vbs script on the crawler server.
JDBC database You must install the JDBC drivers for accessing the target databases on the crawler server.
NNTP No setup requirements.
SharePoint No setup requirements.
UNIX file system No setup requirements.
Web The Web crawler abides by the robots exclusion protocol and rules in robot.txt files.

Depending on your security requirements, ensure that the Web server can access sites that are protected by HTTP basic authentication or form-based authentication.

Seed list, Web Content Management, and WebSphere Portal No setup requirements.
Windows file system To validate current user credentials when a user submits a search request, ensure that domain accounts are correctly configured. Requirements for setting up domain accounts for files that are crawled on the local computer are different from requirements for files that are crawled on a remote Windows server.
Agent for Windows file systems You must install an agent server on the target Windows file server or on a Windows server that hosts the file servers that you want to crawl. The agent server and the file servers to be crawled must all belong to the same Windows domain or workgroup.

Feedback

Last updated: May 2012

© Copyright IBM Corporation 2004, 2012.
This information center is powered by Eclipse technology. (http://www.eclipse.org)