Kubernetes

This guide will run through how to set up the RAPIDS Accelerator for Apache Spark in a Kubernetes cluster. At the end of this guide, the reader will be able to run a sample Apache Spark application that runs on NVIDIA GPUs in a Kubernetes cluster.

This is a quick start guide which uses default settings which may be different from your cluster.

Kubernetes requires a Docker image to run Spark. Generally everything needed is in the Docker image - Spark, the RAPIDS Accelerator for Spark jars, and the discovery script. See this Dockerfile.cuda example.

You can find other supported base CUDA images for from CUDA dockerhub. Its source Dockerfile is inside GitLab repository which can be used to build the docker images from OS base image from scratch.

  • Kubernetes cluster is up and running with NVIDIA GPU support

  • Docker is installed on a client machine

  • A Docker repository which is accessible by the Kubernetes cluster

These instructions do not cover how to setup a Kubernetes cluster.

Please refer to Install Kubernetes on how to install a Kubernetes cluster with NVIDIA GPU support.

On a client machine which has access to the Kubernetes cluster:

  1. Download Apache Spark. Supported versions of Spark are listed on the RAPIDS Accelerator download page. Please note that only Scala version 2.12 is currently supported by the accelerator.

    Note that you can download these into a local directory and untar the Spark .tar.gz as a directory named spark.

  2. Download the RAPIDS Accelerator for Spark jars and the GPU discovery script.

    Put rapids-4-spark_<version>.jar and getGpusResources.sh in the same directory as spark.

    Note

    If here you decide to put above jar in the spark/jars directory which will be copied into /opt/spark/jars directory in Docker image, then in the future you do not need to specify spark.driver.extraClassPath or spark.executor.extraClassPath using cluster mode. This example just shows you a way to put customized jars or 3rd party jars.

  3. Download the sample Dockerfile.cuda in the same directory as spark.

    The sample Dockerfile.cuda will copy the spark directory’s several sub-directories into /opt/spark/ along with the RAPIDS Accelerator jars and getGpusResources.sh into /opt/sparkRapidsPlugin inside the Docker image.

    You can modify the Dockerfile to copy your application into the docker image, i.e. test.py.

    Examine the Dockerfile.cuda file to ensure the file names are correct and modify if needed.

    Currently the directory in the local machine should look as below:

    Copy
    Copied!
                

    $ ls Dockerfile.cuda getGpusResources.sh rapids-4-spark_<version>.jar spark

  4. Build the Docker image with a proper repository name and tag and push it to the repository

    Copy
    Copied!
                

    export IMAGE_NAME=xxx/yyy:tag docker build . -f Dockerfile.cuda -t $IMAGE_NAME docker push $IMAGE_NAME

Submitting a Simple Test Job

This simple job will test if the RAPIDS Accelerator can be found. ClassNotFoundException is a common error if the Spark driver can not find the RAPIDS Accelerator jar, resulting in an exception like this:

Copy
Copied!
            

Exception in thread "main" java.lang.ClassNotFoundException: com.nvidia.spark.SQLPlugin

Here is an example job:

Copy
Copied!
            

export SPARK_HOME=~/spark export IMAGE_NAME=xxx/yyy:tag export K8SMASTER=k8s://https://<k8s-apiserver-host>:<k8s-apiserver-port> export SPARK_NAMESPACE=default export SPARK_DRIVER_NAME=exampledriver $SPARK_HOME/bin/spark-submit \ --master $K8SMASTER \ --deploy-mode cluster \ --name examplejob \ --class org.apache.spark.examples.SparkPi \ --conf spark.executor.instances=1 \ --conf spark.executor.resource.gpu.amount=1 \ --conf spark.executor.memory=4G \ --conf spark.executor.cores=1 \ --conf spark.task.cpus=1 \ --conf spark.task.resource.gpu.amount=1 \ --conf spark.rapids.memory.pinnedPool.size=2G \ --conf spark.executor.memoryOverhead=3G \ --conf spark.sql.files.maxPartitionBytes=512m \ --conf spark.sql.shuffle.partitions=10 \ --conf spark.plugins=com.nvidia.spark.SQLPlugin \ --conf spark.kubernetes.namespace=$SPARK_NAMESPACE \ --conf spark.kubernetes.driver.pod.name=$SPARK_DRIVER_NAME \ --conf spark.executor.resource.gpu.discoveryScript=/opt/sparkRapidsPlugin/getGpusResources.sh \ --conf spark.executor.resource.gpu.vendor=nvidia.com \ --conf spark.kubernetes.container.image=$IMAGE_NAME \ --conf spark.executor.extraClassPath=/opt/sparkRapidsPlugin/rapids-4-spark_<version>.jar \ --conf spark.driver.extraClassPath=/opt/sparkRapidsPlugin/rapids-4-spark_<version>.jar \ --driver-memory 2G \ local:///opt/spark/examples/jars/spark-examples_2.12-3.0.2.jar

Note

local:// means the jar file location is inside the Docker image. Since this is cluster mode, the Spark driver is running inside a pod in Kubernetes. The driver and executor pods can be seen when the job is running:

Copy
Copied!
            

$ kubectl get pods NAME READY STATUS RESTARTS AGE spark-pi-d11075782f399fd7-exec-1 1/1 Running 0 9s exampledriver 1/1 Running 0 15s

To view the Spark driver log, use below command:

Copy
Copied!
            

kubectl logs $SPARK_DRIVER_NAME

To view the Spark driver UI when the job is running first expose the driver UI port:

Copy
Copied!
            

kubectl port-forward $SPARK_DRIVER_NAME 4040:4040

Then open a web browser to the Spark driver UI page on the exposed port:

Copy
Copied!
            

http://localhost:4040

To kill the Spark job:

Copy
Copied!
            

$SPARK_HOME/bin/spark-submit --kill spark:$SPARK_DRIVER_NAME

To delete the driver POD:

Copy
Copied!
            

kubectl delete pod $SPARK_DRIVER_NAME

Running an Interactive Spark Shell

If you need an interactive Spark shell with executor pods running inside the Kubernetes cluster:

Copy
Copied!
            

$SPARK_HOME/bin/spark-shell \ --master $K8SMASTER \ --name mysparkshell \ --deploy-mode client \ --conf spark.executor.instances=1 \ --conf spark.executor.resource.gpu.amount=1 \ --conf spark.executor.memory=4G \ --conf spark.executor.cores=1 \ --conf spark.task.cpus=1 \ --conf spark.task.resource.gpu.amount=1 \ --conf spark.rapids.memory.pinnedPool.size=2G \ --conf spark.executor.memoryOverhead=3G \ --conf spark.sql.files.maxPartitionBytes=512m \ --conf spark.sql.shuffle.partitions=10 \ --conf spark.plugins=com.nvidia.spark.SQLPlugin \ --conf spark.kubernetes.namespace=$SPARK_NAMESPACE \ --conf spark.executor.resource.gpu.discoveryScript=/opt/sparkRapidsPlugin/getGpusResources.sh \ --conf spark.executor.resource.gpu.vendor=nvidia.com \ --conf spark.kubernetes.container.image=$IMAGE_NAME \ --conf spark.executor.extraClassPath=/opt/sparkRapidsPlugin/rapids-4-spark_<version>.jar \ --driver-class-path=./rapids-4-spark_<version>.jar \ --driver-memory 2G

Only the client deploy mode should be used. If you specify the cluster deploy mode, you would see the following error:

Copy
Copied!
            

Cluster deploy mode is not applicable to Spark shells.

Also notice that --conf spark.driver.extraClassPath was removed but --driver-class-path was added. This is because now the driver is running on the client machine, so the jar paths should be local filesystem paths.

When running the shell you can see only the executor pods are running inside Kubernetes:

Copy
Copied!
            

$ kubectl get pods NAME READY STATUS RESTARTS AGE mysparkshell-bfe52e782f44841c-exec-1 1/1 Running 0 11s

The following Scala code can be run in the Spark shell to test if the RAPIDS Accelerator is enabled.

Copy
Copied!
            

val df = spark.sparkContext.parallelize(Seq(1)).toDF() df.createOrReplaceTempView("df") spark.sql("SELECT value FROM df WHERE value <>1").show spark.sql("SELECT value FROM df WHERE value <>1").explain :quit

The expected explain plan should contain the GPU related operators:

Copy
Copied!
            

scala> spark.sql("SELECT value FROM df WHERE value <>1").explain == Physical Plan == GpuColumnarToRow false +- GpuFilter NOT (value#2 = 1) +- GpuRowToColumnar TargetSize(2147483647) +- *(1) SerializeFromObject [input[0, int, false] AS value#2] +- Scan[obj#1]

Running PySpark in Client Mode

Of course, you can COPY the Python code in the Docker image when building it and submit it using the cluster deploy mode as showin in the previous example pi job.

However if you do not want to re-build the Docker image each time and just want to submit the Python code from the client machine, you can use the client deploy mode.

Copy
Copied!
            

$SPARK_HOME/bin/spark-submit \ --master $K8SMASTER \ --deploy-mode client \ --name mypythonjob \ --conf spark.executor.instances=1 \ --conf spark.executor.resource.gpu.amount=1 \ --conf spark.executor.memory=4G \ --conf spark.executor.cores=1 \ --conf spark.task.cpus=1 \ --conf spark.task.resource.gpu.amount=1 \ --conf spark.rapids.memory.pinnedPool.size=2G \ --conf spark.executor.memoryOverhead=3G \ --conf spark.sql.files.maxPartitionBytes=512m \ --conf spark.sql.shuffle.partitions=10 \ --conf spark.plugins=com.nvidia.spark.SQLPlugin \ --conf spark.kubernetes.namespace=$SPARK_NAMESPACE \ --conf spark.executor.resource.gpu.discoveryScript=/opt/sparkRapidsPlugin/getGpusResources.sh \ --conf spark.executor.resource.gpu.vendor=nvidia.com \ --conf spark.kubernetes.container.image=$IMAGE_NAME \ --conf spark.executor.extraClassPath=/opt/sparkRapidsPlugin/rapids-4-spark_<version>.jar \ --driver-memory 2G \ --driver-class-path=./rapids-4-spark_<version>.jar \ test.py

A sample test.py is as below:

Copy
Copied!
            

from pyspark.sql import SQLContext from pyspark import SparkConf from pyspark import SparkContext conf = SparkConf() sc = SparkContext.getOrCreate() sqlContext = SQLContext(sc) df=sqlContext.createDataFrame([1,2,3], "int").toDF("value") df.createOrReplaceTempView("df") sqlContext.sql("SELECT * FROM df WHERE value<>1").explain() sqlContext.sql("SELECT * FROM df WHERE value<>1").show() sc.stop()

Using Spark Operator is another way to submit Spark Applications into a Kubernetes Cluster.

  1. Locate the Spark Application jars/files in the docker image when preparing docker image.

    For example, assume /opt/sparkRapidsPlugin/test.py is inside the docker image.

    This is because currently only cluster deployment mode is supported by Spark Operator.

  2. Create spark operator using helm.

    Follow Spark Operator quick start guide

  1. Create a Spark Application YAML file

    For example, create a file named testpython-rapids.yaml with the following contents:

    Copy
    Copied!
                

    apiVersion: "sparkoperator.k8s.io/v1beta2" kind: SparkApplication metadata: name: testpython-rapids namespace: default spec: sparkConf: "spark.ui.port": "4045" "spark.rapids.sql.concurrentGpuTasks": "1" "spark.executor.resource.gpu.amount": "1" "spark.task.resource.gpu.amount": "1" "spark.executor.memory": "1g" "spark.rapids.memory.pinnedPool.size": "2g" "spark.executor.memoryOverhead": "3g" "spark.sql.files.maxPartitionBytes": "512m" "spark.sql.shuffle.partitions": "10" "spark.plugins": "com.nvidia.spark.SQLPlugin" "spark.executor.resource.gpu.discoveryScript": "/opt/sparkRapidsPlugin/getGpusResources.sh" "spark.executor.resource.gpu.vendor": "nvidia.com" "spark.executor.extraClassPath": "/opt/sparkRapidsPlugin/rapids-4-spark.jar" "spark.driver.extraClassPath": "/opt/sparkRapidsPlugin/rapids-4-spark.jar" type: Python pythonVersion: 3 mode: cluster image: "<IMAGE_NAME>" imagePullPolicy: Always mainApplicationFile: "local:///opt/sparkRapidsPlugin/test.py" sparkVersion: "3.1.1" restartPolicy: type: Never volumes: - name: "test-volume" hostPath: path: "/tmp" type: Directory driver: cores: 1 coreLimit: "1200m" memory: "1024m" labels: version: 3.1.1 serviceAccount: spark volumeMounts: - name: "test-volume" mountPath: "/tmp" executor: cores: 1 instances: 1 memory: "5000m" gpu: name: "nvidia.com/gpu" quantity: 1 labels: version: 3.1.1 volumeMounts: - name: "test-volume" mountPath: "/tmp"

  2. Submit the Spark Application

    Copy
    Copied!
                

    sparkctl create testpython-rapids.yaml

    Note

    sparkctl can be built from the Spark Operator repo after installing golang:

    Copy
    Copied!
                

    cd sparkctl go build -o sparkctl

  3. Check the driver log

    Copy
    Copied!
                

    sparkctl log testpython-rapids

  4. Check the status of this Spark Application

    Copy
    Copied!
                

    sparkctl status testpython-rapids

  5. Port forwarding when Spark driver is running

    Copy
    Copied!
                

    sparkctl forward testpython-rapids --local-port 1234 --remote-port 4045

    Then open browser with http://localhost:1234/ to check Spark UI.

  6. Delete the Spark Application

    Copy
    Copied!
                

    sparkctl delete testpython-rapids

Please refer to Running Spark on Kubernetes for more information.

Previous Azure Synapse Analytics
Next Spark Workload Qualification
© Copyright 2024, NVIDIA. Last updated on Apr 23, 2024.