The next step in provisioning a Db2® Big SQL instance involves setting up a connection to
one or more remote data sources.
About this task
You can connect to a Hadoop cluster, an object store, or to both a Hadoop cluster and an object
store. In the third scenario, you must configure the remote cluster to connect to the object store
because Db2 Big SQL uses the Hive metastore as its catalog. For more information
about each scenario, see Db2 Big SQL architecture.
Note: If updates are made to the remote Hadoop cluster's core-site.xml
configuration file to set up the cluster access to an object store service, its Hive service must be restarted, even if the cluster
manager doesn't indicate that a restart is needed.
Procedure
- To connect to a Hadoop cluster, select the Configure Hadoop
cluster checkbox, and specify the following information:
- The Hadoop cluster manager (Ambari server or Cloudera cluster manager server) host URL
Note: If
you are connecting to a Cloudera cluster manager server over SSL, you must configure SSL. For more
information, see the post-provisioning task
Configuring SSL access.
- The Hadoop cluster manager username
- The Hadoop cluster manager user password
- To allow Db2 Big SQL to automate the
creation of principals and keytabs when MIT Kerberos security is enabled on the Hadoop cluster, select the Configure Kerberos
checkbox, and specify the following information for the user that can create principals for Db2 Big SQL in the Kerberos Key Distribution Center (KDC):
- The Kerberos admin principal
- The Kerberos admin password
Note: If Active Directory is used for Kerberos,
then you can specify a dummy value (for example, BLANK) in the Kerberos admin principal and password fields.
- To connect to an object store, select the Configure object store
checkbox, and specify the following information:
- The object store service endpoint
- The object store service access ID key
- The object store service secret key
Remember: If the Db2 Big SQL
instance is also set up to connect to a Hadoop cluster, the cluster must be configured to access the
same object store service.
- If you are accessing a public object store, select the With SSL
checkbox.
Note: For most public cloud object stores, it is generally not necessary to provide a Secure Sockets
Layer (SSL) certificate in the SSL Certificate box. But select the
With SSL checkbox so that the connection to the object store is
secure.
- If you are accessing an on-premises, self-hosted object store that is accessed by using
SSL, do the following steps.
- Select the With SSL checkbox.
- In the SSL Certificate box, copy the SSL certificate in base64
encoded Privacy Enhanced Mail (PEM) format, including the PEM header and footer.
For
example,
-----BEGIN CERTIFICATE-----
MIIFVzCCAz+gAwIBAgIJAM+JlcdkA2RaMA0GCSqGSIb3DQEBCwUAMEIxCzAJBgNV
BAYTAlhYMRUwEwYDVQQHDAxEZWZhdWx0IENpdHkxHDAaBgNVBAoME0RlZmF1bHQg
...
ZzV0pevZpJWFCt2QYEprZppj0KyiGHKQcEXAn/953YPTOmmdzGVOu5eLoTncICte
oBzKDOdxT6CTenizfaiP5LlWH1LfPvw/+0Nz
-----END CERTIFICATE-----
- If you are connecting to an object store or to both a Hadoop cluster and an object store,
and you want to limit access to a single object store bucket, select the Specify the
object store bucket name checkbox, and enter the bucket name.
- Click Next.
- In the Summary page, review the information that you
specified.
- If you want to make changes, click Previous and go to the page
where you want to edit information.
- When all configuration information is correct, click
Create.
Results
A Db2 Big SQL instance is created.
What to do next
Depending on your environment, you might have to do some post-provisioning tasks. For more
information, see Post-provisioning tasks.
After you provision an instance, you must add one or more users to the instance. You (instance
owner) are not automatically added as a user. For more information, see the section on specifying
which users can access Db2 Big SQL instances in
Configuring, monitoring, and managing access to Db2 Big SQL instances.