Cloud Example Jupyter: Difference between revisions

From Computer Science Wiki
Jump to navigation Jump to search
Carnold (talk | contribs)
No edit summary
Carnold (talk | contribs)
 
(23 intermediate revisions by the same user not shown)
Line 1: Line 1:
''Work in progress''
=== Introduction ===
=== Introduction ===
The goal of this project is to support a class teaching basic programming using [https://jupyter.org Jupyter Notebook]  Jupyter notebook is a single process that supports only one person.  To support a whole class, a jupyter notebook process will need to be run for each student.  Jupyter offers a [https://jupyter.org/hub Jupyter Hub] that automatically spawns these singe user processes.  However, a single machine can only support around 50-70 students before suffering performance issues.  Kubernetes allows this process to scale out horizontally by spreading the single-user instances across physical nodes.  I will give a break down of different pieces needed to make this work and go into more detail on certain aspects.
The goal of this project is to support a class teaching basic programming using [https://jupyter.org Jupyter Notebook]  Jupyter notebook is a single process that supports only one person.  To support a whole class, a [https://jupyter.org jupyter notebook] process will need to be run for each student.  Jupyter offers [https://jupyter.org/hub Jupyter Hub] that automatically spawns these singe user processes.  However, a single machine can only support around 50-70 students before suffering performance issues.  [https://kubernetes.io/docs/home/ Kubernetes] allows this process to scale out horizontally by spreading the single-user instances across physical nodes.  I will give a break down of the different pieces needed to make this work on https://cloud.cs.vt.edu and go into more detail on certain aspects.


=== Basics ===
=== Workloads ===
Here is what the '''Workloads''' tab looks like:
Here is what the '''[https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/ Workloads]''' tab looks like:


[[File:Jupyter1.JPG]]
[[File:Jupyter1.JPG]]
'''Details of each ''[https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]:'''''
==== continuous-image-puller ====
This ''[https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]'':
* Starts a [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] on all ''[https://kubernetes.io/docs/concepts/architecture/nodes/ physical nodes]'' of the cluster.  This is called a ''[https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ DaemonSet]'' by Kubernetes.
* The purpose of this [https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ DaemonSet] is to pull the [https://docs.docker.com/engine/reference/commandline/image/ image(s)] need to run the single-user [https://jupyter.org Jupyter Notebook] and automatically download on any new nodes to the cluster -- basically pre-caching the [https://docs.docker.com/engine/reference/commandline/image/ images].  The process of downloading the [https://docs.docker.com/engine/reference/commandline/image/ images] the first time it is run on a node can take some time and detract from the user experience.
* The ''[https://kubernetes.io/docs/concepts/containers/ containers]'' in each [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] are:
** The primary [https://docs.docker.com/engine/reference/commandline/image/ image] that you can see is '''gcr.io/google_containers/pause:3.0'''  This is just a simple lightweight [https://kubernetes.io/docs/concepts/containers/ container] that doesn't do any processing, just to keep the ''[https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]'' alive
** The other two [https://docs.docker.com/engine/reference/commandline/image/ images] are the single-user [https://jupyter.org Jupyter Notebook] and a networking tools image (updates iptables).  Both of these [https://docs.docker.com/engine/reference/commandline/image/ images] are run with a simple echo log command, so again they don't do any processing but the [https://docs.docker.com/engine/reference/commandline/image/ image] is downloaded and cached by kubernetes
==== hub ====
This [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]:
* [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet], currently running a single [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod]
* The purpose of this [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet] is to run the JupyterHub process that authenticates the user, configures the proxy, and calls the spawner to create a single-user instance of Jupyter for the user.
* The only [https://kubernetes.io/docs/concepts/containers/ container] in the [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] is docker.cs.vt.edu/carnold/jupyterhub/jupyterhub:v1.00  This is a customized version of the jupyterhub [https://docs.docker.com/engine/reference/commandline/image/ docker image] to add the CAS authentication plug-in.
==== jupyter-carnold ====
This [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]:
* Runs a single [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod]
* This is an example of a [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] automatically spawned for the user ''carnold'' and hosts their single-user instance of [https://jupyter.org Jupyter Notebook]
* The [https://kubernetes.io/docs/concepts/containers/ containers] in this [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] are:
** The primary [https://docs.docker.com/engine/reference/commandline/image/ image] is docker.cs.vt.edu/carnold/jupyterhub/singleuser:v1.06  This is a customized verison of the [https://jupyter.org jupyter notebook] single user [https://docs.docker.com/engine/reference/commandline/image/ docker image] to add a custom script that runs on start up.
==== proxy ====
This [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]:
* [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet], currently running a single [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod]
* The purpose of this [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet] is to run Nginx as a reverse proxy redirecting users to their single-user instances of [https://jupyter.org Jupyter Notebook].  The proxies get configured by the '''hub''' [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment].
* The only [https://kubernetes.io/docs/concepts/containers/ container] in this [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] is jupyterhub/configurable-http-proxy:4.1.0  This is based on Nginx.
* There is an external IP mapped to this service with ports 443 and 80 forwarding to this [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod].  This allows connections from outside the cluster.
==== user-placeholder ====
This [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]:
* [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet], currently running a single [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod]
* The purpose of this [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet] is to hold a spot (or multiple spots) on the cluster for single-user instances.  This effectively reserves the resources (memory and cpu requests) necessary to run a single-user instance of [https://jupyter.org Jupyter Notebook].  In theory, you could scale this up to reserve as many places as you need.
* The only [https://kubernetes.io/docs/concepts/containers/ container] in this [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pod] is gcr.io/google_containers/pause:3.0  Again, this [https://docs.docker.com/engine/reference/commandline/image/ image] doesn't do any processing, but keeps the instance alive.
==== user-scheduler ====
This [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment]:
* [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet], currently running two [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pods] for redundancy
* The purpose of this [https://kubernetes.io/docs/concepts/workloads/controllers/replicaset/ ReplicaSet] is to schedule and spawn the single-user instances of [https://jupyter.org Jupyter Notebook].
* The only [https://kubernetes.io/docs/concepts/containers/ container] in these [https://kubernetes.io/docs/concepts/workloads/pods/pod/ pods] is gcr.io/google_containers/kube-scheduler-amd64:v1.11.2  This is a complex process, and the [https://docs.docker.com/engine/reference/commandline/image/ image] is created by Kubernetes itself.  Find out more at: https://kubernetes.io/docs/reference/command-line-tools-reference/kube-scheduler
=== Services ===
Here is what the '''[https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/service-discovery/ Service Discovery]''' tab looks like:
[[File:Jupyter2.JPG]]
'''Details of each ''[https://kubernetes.io/docs/concepts/services-networking/service/ service]:'''''
==== hub ====
* Points to the '''hub''' [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment] API port
* Allows other instances to connect to the Hub's API
==== proxy-api ====
* Points to the '''proxy''' [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment] API port
* Used by '''hub''' to configure reverse proxy entries as needed
==== proxy-public ====
* Points to the '''proxy''' http ports
* Maps the proxy ports to an external IP
=== Volumes ===
Here is what the '''[https://rancher.com/docs/rancher/v2.x/en/cluster-admin/volumes-and-storage/ Volumes]''' tab looks like:
[[File:Jupyter3.JPG]]
'''Details of each ''[https://kubernetes.io/docs/concepts/storage/volumes/ volume]:'''''
==== claim-carnold ====
* This is a [https://kubernetes.io/docs/concepts/storage/persistent-volumes/ persistent volume claim]
* This volume is dedicated to a single-user [https://jupyter.org Jupyter Notebook], specifically, the user ''carnold''
* The purpose of this volume is to persist user's changes to their notebook across nodes and restarts
==== hub-db-dir ====
* This is a [[https://kubernetes.io/docs/concepts/storage/persistent-volumes/ persistent volume claim]
* This volume is used by the '''hub''' [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment] to support a sqlite database for configuration.
* The purpose of this volume is to persist DB changes across nodes and restarts
=== Config Maps ===
Here is what the '''[https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/configmaps/ Config Maps]''' tab looks like:
[[File:Jupyter4.JPG]]
'''Details of each ''[https://kubernetes.io/docs/tasks/configure-pod-container/configure-pod-configmap/ ConfigMap]:'''''
==== hub-config ====
Used by a python script inside the '''hub''' [https://rancher.com/docs/rancher/v2.x/en/k8s-in-rancher/workloads/deploy-workloads/ deployment] to configure the JupyterHub process.  This is what allows custom configuration with a base [https://docs.docker.com/engine/reference/commandline/image/ image].
==== jupyterhub.v19 ====
==== user-scheduler ====
=== Deployment ===
==== Helm Chart ====
This example was deployed using a [https://helm.sh/docs/developing_charts/ Helm Chart]  The chart used is based on https://zero-to-jupyterhub.readthedocs.io/en/latest/  I only made minor changes to the chart to work with https://cloud.cs.vt.edu  You can see the full chart code at: https://version.cs.vt.edu/carnold/jupyterhub/tree/master/charts/jupyterhub  Charts make deploying, modifying, and updating complex kubernetes deployments more simple.  The CS cloud web UI allows you to use helm charts, and even custom helm charts like the once I created, it calls them [https://rancher.com/docs/rancher/v2.x/en/catalog/apps Apps] and collections of Apps called [https://rancher.com/docs/rancher/v2.x/en/catalog/ Catalogs].  Charts use templates to create the kubernetes resources, and pulls custom values for the template from a YAML file.  Here is the yaml file used for this deployment:
<pre>
--
  auth:
    type: custom
    custom:
      className: jhub_cas_authenticator.cas_auth.CASAuthenticator
      config:
        cas_login_url: "https://login.vt.edu/profile/cas/login"
        cas_service_url: "https://testing131.cs.vt.edu/hub/login"
        cas_client_ca_certs: "/etc/vtca.crt"
        cas_service_validate_url: "https://login.vt.edu/profile/cas/serviceValidate"
  hub:
    image:
      name: docker.cs.vt.edu/carnold/jupyterhub/jupyterhub
      tag: v1.00
  proxy:
    secretToken: "********"
  singleuser:
    image:
      name: docker.cs.vt.edu/carnold/jupyterhub/singleuser
      tag: v1.06
    lifecycleHooks:
      postStart:
        exec:
          command: ["/tmp/entrypoint.sh"]
    storage:
      capacity: "500Mi"
    memory:
      limit: 128M
      guarantee: 128M
    cpu:
      limit: .5
  cull:
    timeout: 300
</pre>
==== Custom [https://docs.docker.com/engine/reference/commandline/image/ Images] ====
I customized a couple of the [https://docs.docker.com/engine/reference/commandline/image/ docker images] used for this project.  I use [https://docs.docker.com/engine/reference/commandline/tag/ tagging] to control versioning and allow upgrades, this is very important when it comes to the single user [https://docs.docker.com/engine/reference/commandline/image/ image] since it is pre-cached.  Here are details about the customization.
===== Jupyter Hub =====
The default Jupyter Hub [https://docs.docker.com/engine/reference/commandline/image/ docker image] does not include an authentication plug-in for CAS which is used by VT for login.  I needed to add the plug-in and add the CA certificate to the [https://docs.docker.com/engine/reference/commandline/image/ docker image].  I used our gitlab instance to store my [https://docs.docker.com/engine/reference/builder/ Dockerfile] and store the customized [https://docs.docker.com/engine/reference/commandline/image/ docker image] in [https://docs.docker.com/registry/ registry].
* See my [https://docs.docker.com/engine/reference/builder/ Dockerfile] in the repository: https://version.cs.vt.edu/carnold/jupyterhub/tree/master/jupyterhub
* I then [https://docs.docker.com/engine/reference/commandline/build/ build], [https://docs.docker.com/engine/reference/commandline/tag/ tag], and upload the [https://docs.docker.com/engine/reference/commandline/image/ docker image]:
** Build: <code>docker build -t docker.cs.vt.edu/carnold/jupyterhub/jupyterhub:v1.00 .</code>
** Upload: <code>docker push docker.cs.vt.edu/carnold/jupyterhub/jupyterhub:v1.00</code>
===== Jupyter Single User =====
Similarly, I needed to customize the [https://jupyter.org Jupyter Notebook] Single User [https://docs.docker.com/engine/reference/commandline/image/ docker image].  This time, it needs a script that is to be run when the [https://jupyter.org Jupyter Notebook] starts.  The script will merge in changes from a git repository.
* See my [https://docs.docker.com/engine/reference/builder/ Dockerfile] in the repository: https://version.cs.vt.edu/carnold/jupyterhub/tree/master/singleuser
* I then [https://docs.docker.com/engine/reference/commandline/build/ build], [https://docs.docker.com/engine/reference/commandline/tag/ tag], and upload the [https://docs.docker.com/engine/reference/commandline/image/ docker image]:
** Build: <code>docker build -t docker.cs.vt.edu/carnold/jupyterhub/singleuser:v1.06 .</code>
** Upload: <code>docker push docker.cs.vt.edu/carnold/jupyterhub/singleuser:v1.06</code>

Latest revision as of 12:43, 21 August 2019

Introduction

The goal of this project is to support a class teaching basic programming using Jupyter Notebook Jupyter notebook is a single process that supports only one person. To support a whole class, a jupyter notebook process will need to be run for each student. Jupyter offers Jupyter Hub that automatically spawns these singe user processes. However, a single machine can only support around 50-70 students before suffering performance issues. Kubernetes allows this process to scale out horizontally by spreading the single-user instances across physical nodes. I will give a break down of the different pieces needed to make this work on https://cloud.cs.vt.edu and go into more detail on certain aspects.

Workloads

Here is what the Workloads tab looks like:

Details of each deployment:

continuous-image-puller

This deployment:

  • Starts a pod on all physical nodes of the cluster. This is called a DaemonSet by Kubernetes.
  • The purpose of this DaemonSet is to pull the image(s) need to run the single-user Jupyter Notebook and automatically download on any new nodes to the cluster -- basically pre-caching the images. The process of downloading the images the first time it is run on a node can take some time and detract from the user experience.
  • The containers in each pod are:
    • The primary image that you can see is gcr.io/google_containers/pause:3.0 This is just a simple lightweight container that doesn't do any processing, just to keep the deployment alive
    • The other two images are the single-user Jupyter Notebook and a networking tools image (updates iptables). Both of these images are run with a simple echo log command, so again they don't do any processing but the image is downloaded and cached by kubernetes

hub

This deployment:

  • ReplicaSet, currently running a single pod
  • The purpose of this ReplicaSet is to run the JupyterHub process that authenticates the user, configures the proxy, and calls the spawner to create a single-user instance of Jupyter for the user.
  • The only container in the pod is docker.cs.vt.edu/carnold/jupyterhub/jupyterhub:v1.00 This is a customized version of the jupyterhub docker image to add the CAS authentication plug-in.

jupyter-carnold

This deployment:

  • Runs a single pod
  • This is an example of a pod automatically spawned for the user carnold and hosts their single-user instance of Jupyter Notebook
  • The containers in this pod are:
    • The primary image is docker.cs.vt.edu/carnold/jupyterhub/singleuser:v1.06 This is a customized verison of the jupyter notebook single user docker image to add a custom script that runs on start up.

proxy

This deployment:

  • ReplicaSet, currently running a single pod
  • The purpose of this ReplicaSet is to run Nginx as a reverse proxy redirecting users to their single-user instances of Jupyter Notebook. The proxies get configured by the hub deployment.
  • The only container in this pod is jupyterhub/configurable-http-proxy:4.1.0 This is based on Nginx.
  • There is an external IP mapped to this service with ports 443 and 80 forwarding to this pod. This allows connections from outside the cluster.

user-placeholder

This deployment:

  • ReplicaSet, currently running a single pod
  • The purpose of this ReplicaSet is to hold a spot (or multiple spots) on the cluster for single-user instances. This effectively reserves the resources (memory and cpu requests) necessary to run a single-user instance of Jupyter Notebook. In theory, you could scale this up to reserve as many places as you need.
  • The only container in this pod is gcr.io/google_containers/pause:3.0 Again, this image doesn't do any processing, but keeps the instance alive.

user-scheduler

This deployment:

Services

Here is what the Service Discovery tab looks like:

Details of each service:

hub

  • Points to the hub deployment API port
  • Allows other instances to connect to the Hub's API

proxy-api

  • Points to the proxy deployment API port
  • Used by hub to configure reverse proxy entries as needed

proxy-public

  • Points to the proxy http ports
  • Maps the proxy ports to an external IP

Volumes

Here is what the Volumes tab looks like:

Details of each volume:

claim-carnold

  • This is a persistent volume claim
  • This volume is dedicated to a single-user Jupyter Notebook, specifically, the user carnold
  • The purpose of this volume is to persist user's changes to their notebook across nodes and restarts

hub-db-dir

  • This is a [persistent volume claim
  • This volume is used by the hub deployment to support a sqlite database for configuration.
  • The purpose of this volume is to persist DB changes across nodes and restarts

Config Maps

Here is what the Config Maps tab looks like:

Details of each ConfigMap:

hub-config

Used by a python script inside the hub deployment to configure the JupyterHub process. This is what allows custom configuration with a base image.

jupyterhub.v19

user-scheduler

Deployment

Helm Chart

This example was deployed using a Helm Chart The chart used is based on https://zero-to-jupyterhub.readthedocs.io/en/latest/ I only made minor changes to the chart to work with https://cloud.cs.vt.edu You can see the full chart code at: https://version.cs.vt.edu/carnold/jupyterhub/tree/master/charts/jupyterhub Charts make deploying, modifying, and updating complex kubernetes deployments more simple. The CS cloud web UI allows you to use helm charts, and even custom helm charts like the once I created, it calls them Apps and collections of Apps called Catalogs. Charts use templates to create the kubernetes resources, and pulls custom values for the template from a YAML file. Here is the yaml file used for this deployment:

--
  auth:
    type: custom
    custom:
      className: jhub_cas_authenticator.cas_auth.CASAuthenticator
      config:
        cas_login_url: "https://login.vt.edu/profile/cas/login"
        cas_service_url: "https://testing131.cs.vt.edu/hub/login"
        cas_client_ca_certs: "/etc/vtca.crt"
        cas_service_validate_url: "https://login.vt.edu/profile/cas/serviceValidate"
  hub:
    image:
      name: docker.cs.vt.edu/carnold/jupyterhub/jupyterhub
      tag: v1.00
  proxy: 
    secretToken: "********"
  singleuser:
    image: 
      name: docker.cs.vt.edu/carnold/jupyterhub/singleuser
      tag: v1.06
    lifecycleHooks:
      postStart:
        exec:
          command: ["/tmp/entrypoint.sh"]
    storage: 
      capacity: "500Mi"
    memory:
      limit: 128M
      guarantee: 128M
    cpu:
      limit: .5
  cull:
    timeout: 300

Custom Images

I customized a couple of the docker images used for this project. I use tagging to control versioning and allow upgrades, this is very important when it comes to the single user image since it is pre-cached. Here are details about the customization.

Jupyter Hub

The default Jupyter Hub docker image does not include an authentication plug-in for CAS which is used by VT for login. I needed to add the plug-in and add the CA certificate to the docker image. I used our gitlab instance to store my Dockerfile and store the customized docker image in registry.

Jupyter Single User

Similarly, I needed to customize the Jupyter Notebook Single User docker image. This time, it needs a script that is to be run when the Jupyter Notebook starts. The script will merge in changes from a git repository.