diff --git a/docs/index.md b/docs/index.md index 5f4af4ec..82148904 100644 --- a/docs/index.md +++ b/docs/index.md @@ -16,7 +16,6 @@ Explorer The Hub Use VS Code Use GitHub Codespaces -Batch Jobs Using QGIS Changelog ``` diff --git a/docs/overview/batch.md b/docs/overview/batch.md deleted file mode 100644 index f47b9204..00000000 --- a/docs/overview/batch.md +++ /dev/null @@ -1,58 +0,0 @@ -# Running batch jobs - -The [Planetary Computer Hub][hub] is a great option for *interactive* data analysis. But JupyterHub isn't primarily designed for -asynchronous (or batch) workflows, where you submit some kind of job and let it run to completion. - -The Hub includes a [kbatch](https://kbatch.readthedocs.io/en/latest/) service, which lets you submit jobs to run on the same -computing infrastructure as your interactive workflows. See the [installation instructions](https://kbatch.readthedocs.io/en/latest/#install) -for kbatch to get started. - -## Configuration - -Once you have `kbatch` installed, you can configure the Hub URL and token to use. - -First, visit the [token generation page][token] to generate a token - -![JupyterHub Admin page kkto generate a token.](../concepts/images/hub-token.png) - -Next, use `kbatch configure` - -```{code-block} console -$ kbatch configure --kbatch-url=https://pccompute.westeurope.cloudapp.azure.com/compute/services/kbatch --token='' -``` - -## Submit a Job - -Use `kbatch submit` to submit a "job", which is just some commands to run. At a minimum, your job needs to include - -1. A name to identify your job -2. A container image, which defines the software environment the job will run in -3. A command to run - -```{code-block} console -❯ kbatch job submit ... -``` - -List your jobs with `kbatch job list`: - -![kbatch job list output showing a few jobs](images/kbatch-job-list.png) - -See the [kbatch examples gallery][gallery] for more. - -## Job runtime - -These batch jobs run in the same compute environment as the JuptyerHub. In particular, they include access to Dask Gateway, so you're able to start -and stop Dask clusters just as if you were running on the hub. - -```{note} -Your jobs will *not* have access to your JupyterHub home directory. You'll need to submit any [code files][code] along with your job. -``` - -## Limitations - -`kbatch` jobs are simply commands to be executed. It doesn't offer any fancier workflow orchestration features like alerting, automatically parallelization, artifact management, etc. - -[hub]: environment.md -[token]: https://pccompute.westeurope.cloudapp.azure.com/compute/hub/token -[gallery]: https://kbatch.readthedocs.io/en/latest/examples/index.html -[code]: https://kbatch.readthedocs.io/en/latest/user-guide.html#submitting-code-files \ No newline at end of file diff --git a/docs/overview/environment.md b/docs/overview/environment.md index a25018ba..795dcc51 100644 --- a/docs/overview/environment.md +++ b/docs/overview/environment.md @@ -26,8 +26,9 @@ Select *Stop My Server* to stop your server and release all of the resources you ![JupyterHub menu to stop the server](images/hub-home.png) -Note that we will automatically stop notebook servers that appear idle or are older that 24 hours. If you expect a job to take longer -than 24 hours, then see [running batch jobs](./batch) for a way to submit long-running jobs. +Note that we will automatically stop notebook servers that appear idle or are older that 24 hours. +The Planetary Computer Hub is primarily intended for interactive computation on datasets +from our catalog. ## Using JupyterLab