Skip to content

Latest commit

 

History

History

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Chapter 9 - Scalable Inference Serving on Cloud with TensorFlow Serving and KubeFlow

Note: All images in this directory, unless specified otherwise, are licensed under CC BY-NC 4.0.

Figure List

Figure number Description Notes
9-1 A high-level overview and comparison of different inference serving options
9-2 Navigate to http://localhost:5000/hello within a web browser to view the “Hello World!” web page
9-3 Listing page for machine learning models on the Google Cloud ML Engine dashboard
9-4 Model creation page on Google Cloud ML Engine
9-5 Model listings page on Google Cloud ML Engine
9-6 Details page of the just-created Dog/Cat classifier
9-7 Creating a new version for a machine learning model
9-8 Creating a new Google Cloud Storage bucket within the ML model version creation page
9-9 Google Cloud Storage Browser page showing the uploaded Dog/Cat classifier model in TensorFlow format
9-10 Add the URI for the model you uploaded to Google Cloud Storage
9-11 An end-to-end pipeline illustrated in KubeFlow
9-12 Creating a new Jupyter Notebook server on KubeFlow
9-13 Creating a KubeFlow deployment on GCP using the browser
9-14 Google Cloud ML Engine showing incoming queries and latency of serving the calls, with end-to-end latency at user’s end of about 3.5 seconds
9-15 Cost comparison of infrastructure as a service (Google Cloud ML Engine) versus building your own stack over virtual machines (Azure VM) (costs as of August 2019)