Managing Analytics Engine powered by Apache Spark instances

You can manage the Analytics Engine powered by Apache Spark instances by using the Cloud Pak for Data web client. It supports the following functions:

  • View and edit Spark instance details
  • Manage user access to the instance or delete the instance
  • View submitted Spark applications
  • View Spark history details

A user with Administrator or Developer role can view the instance details.

To manage a service instance for Analytics Engine powered by Apache Spark:

  1. Log in to Cloud Pak for Data.
  • From the navigation menu Cloud Pak for Data navigation menu, select Services > Instances, and select the instance name to view the service instance details. You can find the following information:

    • The Configuration tab includes:

      • Instance summary details
      • Default runtime details
      • Storage claim name
      • Instance resource quota information
      • The endpoint to start and stop history server
      • The url to the Spark History server user interface
      • Spark kernel endpoint
      • Spark job v4 endpoint
      • Spark job v3 endpoint (deprecated)
      • Default Spark configuration

      • Click the edit icon (Edit icon) to edit the default runtime details, the instance resource quota details, and default Spark configuration details.
    • The Applications tab lists submitted applications. For more information, see Spark job status.

    • The Spark history tab provides details about the Spark history server (whether server is stopped or running, amount of resources being consumed by it, a link to open the Spark history UI). You can start or stop the Spark history server and open the Spark history server UI page to view the application run status in detail.

  • If spec.serviceConfig.sparkAdvEnabled is enabled in the Analytics Engine custom resource (CR), you see:

    • The name of the deployment space
    • The deployment space ID

From the options menu on the right side of the window, you can:

- **Manage access**: Only a user with **Administrator** role can manage user access to the Analytics Engine powered by Apache Spark instances. From here, an administrator can grant users **Developer** role to the instance so that they can submit Spark jobs. See [Managing user access](manage-access.html).
- **Delete**: Only a user with **Administrator** role can delete an Analytics Engine powered by Apache Spark instance.
Important: If the advanced features are enabled for this service instance, you must delete the deployment space that is associated with the service instance if you want to create an service instance again with the same name. When you delete the deployment space, you will also delete all assets and jobs in that space.

To delete a deployment space:

  1. From the navigation menu Cloud Pak for Data navigation menu in Cloud Pak for Data, click Deployments.
  2. On the Spaces tab, search for the space named <InstanceName>_space. From the Actions menu on the right, select Delete.

If you can't delete the deployment space, check to see whether any jobs are stuck in Starting state. See Troubleshooting for Analytics Engine powered by Apache Spark for how to remove jobs that are stuck in Starting state.

Note: The data files in the instance user's `home` directory, which is created at the time the Analytics Engine Powered by Apache Spark instance is provisioned, are not deleted when the instance is deleted. Delete this data yourself.

Generating an access token

All users must generate their own access token to use the Spark jobs API. For instructions on how to generate an access token, see Generating an API authorization token.

What to do next

Parent topic: Administering Analytics Engine powered by Apache Spark