Although in many cases it's not necessary to understand all of the many ways
to configure a SavedModel, this method has a few practical implications:
It will be treated as a graph for inference / serving (i.e. uses the tag
saved_model.SERVING)
The SavedModel will load in TensorFlow Serving and supports the
Predict
API.
To use the Classify, Regress, or MultiInference APIs, please
use either
tf.Estimator
or the lower level
SavedModel
APIs.
Some TensorFlow ops depend on information on disk or other information
called "assets". These are generally handled automatically by adding the
assets to the GraphKeys.ASSET_FILEPATHS collection. Only assets in that
collection are exported; if you need more custom behavior, you'll need to
use the
SavedModelBuilder.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2022-11-04 UTC."],[],[]]