Skip to content

Commit

Permalink
doc-remove-include-attribute-file (#7)
Browse files Browse the repository at this point in the history
  • Loading branch information
MelissaFlinn authored Nov 30, 2023
1 parent 80d9240 commit 4edf63d
Show file tree
Hide file tree
Showing 15 changed files with 0 additions and 31 deletions.
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='automating-workflows-with-pipelines']
= Automating workflows with data science pipelines

include::_attributes.adoc[]

In previous sections of this {deliverable}, you used a notebook to train and save your data model. Optionally, you can automate these tasks by using {productname-long} pipelines. Pipelines offer a way to automate the execution of multiple notebooks and Python code. By using pipelines, you can execute long training jobs or retrain your models on a schedule without having to manually run them in a notebook.

In this section, you create a simple pipeline by using the GUI pipeline editor. The pipeline uses the notebook that you used in previous sections, to train a model and then save it to S3 storage.
Expand Down
2 changes: 0 additions & 2 deletions workshop/docs/modules/ROOT/pages/creating-a-workbench.adoc
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='creating-a-workbench']
= Creating a workbench and selecting a Notebook image

include::_attributes.adoc[]

A workbench is an instance of your development and experimentation environment. Within the workbench you can select a notebook image for your data science work.

.Prerequisite
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='creating-data-connections-to-storage']
== Creating data connections to your own S3-compatible object storage

include::_attributes.adoc[]

NOTE: If you do not have your own s3-compatible storage, or if you want to use a disposable local Minio instance instead, skip this section and follow the steps in xref:running-a-script-to-install-storage.adoc[Running a script to install local object storage buckets and create data connections].

.Prerequisite
Expand Down
2 changes: 0 additions & 2 deletions workshop/docs/modules/ROOT/pages/deploying-a-data-model.adoc
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='deploying-a-data-model']
= Deploying a data model

include::_attributes.adoc[]

Now that the data model is accessible in storage and saved in the portable ONNX format, you can use an {productname-short} model server to deploy it as an API.

{productname-short} multi-model servers can host several models at once. You create a new model server and deploy your model to it.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='enabling-data-science-pipelines']
= Enabling data science pipelines

include::_attributes.adoc[]

NOTE: If you do not intend to complete the pipelines section of the workshop you can skip this step and move on to the next section, xref:creating-a-workbench.adoc[Create a Workbench]


Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='importing-files-into-jupyter']
= Importing the {deliverable} files into the Jupyter environment

include::_attributes.adoc[]

The Jupyter environment is a web-based environment, but everything you do inside it happens on *{productname-long}* and powered by the *OpenShift* cluster. This means that, without having to install and maintain anything on your own computer, and without disposing of lots of local resources like CPU, GPU and RAM, you can conduct your Data Science work in this powerful and stable managed environment.

.Prerequisite
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='navigating-to-the-dashboard']
= Navigating to the {productname-short} dashboard

include::_attributes.adoc[]

.Procedure

. After you log in to the OpenShift console, access the {productname-short} dashboard by clicking the application launcher icon on the header.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
[id='preparing-a-data-model-for-deployment']
= Preparing a data model for deployment
include::_attributes.adoc[]

After you train a model, you can deploy it by using the {productname-short} model serving capabilities.

Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='running-a-pipeline-generated-from-python-code']
= Running a data science pipeline generated from Python code

include::_attributes.adoc[]

In the previous scetion, you created a simple pipeline by using the GUI pipeline editor, it's often desirable to create pipelines by using code that can be version-controlled and shared with others. The https://github.com/kubeflow/kfp-tekton[kfp-tekton] SDK provides a Python API for creating pipelines. The SDK is available as a Python package that you can install by using the `pip install kfp-tekton` command. With this package, you can use Python code to create a pipeline and then compile it to Tekton YAML. Then you can import the YAML code into {productname-short}.

This {deliverable} does not delve into the details of how to use the SDK. Instead, it provides the files for you to view and upload.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='running-a-script-to-install-storage']
= Running a script to install local object storage buckets and create data connections

include::_attributes.adoc[]

For convenience, the provided script installs two data connections (and associated secrets) and two Minio buckets as s3-compatible storage The script creates a random user and password for security. This script is based on the instructions for installing Minio in this https://ai-on-openshift.io/tools-and-applications/minio/minio/[guide].

IMPORTANT: The storage buckets that this script creates are *not* meant for production usage.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='running-code-in-a-notebook']
= Running code in a notebook

include::_attributes.adoc[]

NOTE: If you're already at ease with Jupyter, you can xref:training-a-data-model.adoc[skip to the next section].

A notebook is an environment where you have _cells_ that can display formatted text or code.
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='setting-up-your-data-science-project']
= Setting up your data science project

include::_attributes.adoc[]

Before you begin, make sure that you are logged in to *{productname-long}* and that you can see the dashboard:

image::projects/dashboard-enabled.png[Dashboard Enabled]
Expand Down
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='storing-data-with-data-connections']
= Storing data with data connections

include::_attributes.adoc[]

For this {deliverable}, you need two S3-compatible object storage buckets, such as Ceph, Minio, or AWS S3:

* *My Storage* - Use this bucket for storing your models and data. You can reuse this bucket and its connection for your notebooks and model servers.
Expand Down
2 changes: 0 additions & 2 deletions workshop/docs/modules/ROOT/pages/testing-the-model-api.adoc
Original file line number Diff line number Diff line change
@@ -1,8 +1,6 @@
[id='testing-the-model-api']
= Testing the model API

include::_attributes.adoc[]

Now that you've deployed the model, you can test its API endpoints.

When you created the model server, you did *not* create a route for external access to the API and you did not protect it with an authentication token. By default, if you do not specify external access, the model server provides an internal endpoint with no authentication.
Expand Down
4 changes: 0 additions & 4 deletions workshop/docs/modules/ROOT/pages/training-a-data-model.adoc
Original file line number Diff line number Diff line change
@@ -1,12 +1,8 @@
[id='training-a-data-model']
= Training a data model

include::_attributes.adoc[]

Now that you know how the Jupyter notebook environment works, the real work can begin!

//need to describe the provided data model? train a model for inference, fraud detection?

In your notebook environment, open the file `1_experiment_train.ipynb`, and follow the instructions directly in the notebook. The instructions guide you through some simple data exploration, experimentation, and data model training tasks.

image::workbenches/jupyter-notebook-1.png[Jupyter Notebook 1]
Expand Down

0 comments on commit 4edf63d

Please sign in to comment.