TensorFlow Serving using TensorFlow 1.13.0-rc1
Major Features and Improvements
- Support for TensorRT 5.0 (GPU docker image built against CUDA 10 and TensorRT 5.0)
- Support for listening gRPC over UNIX socket (commit: a25b0da)
- New GPU version of TensorFlow Serving API PIP package. This depends on the
tensorflow-gpu
instead oftensorflow
PIP package, but is otherwise identical. (commit: 525c1af) - TF Serving end-to-end colab! Training with Keras, serving with TF Serving and REST API (commit: 1ff8aad)
Breaking Changes
- No breaking changes.
Bug Fixes and Other Changes
- Make error message for input size mismatch in
Predict
call even more actionable. (commit: 7237fb5) - Document how to use the version policy to pin a specific version, or serve multiple versions, of a model. (commit: 2724bfe)
- Document config reloading and model version labels. (commit: f4890af)
- Fix the compile error on ARM-32 in net_http/server. (commit: 5446fd9)
- Adds ModelSpec to SessionRunResponse. (commit: 58a2263)
- Add MKL support (commit: 8f79253)
- Fix default path of Prometheus metrics endpoint (commit: 9d05b0c)
- Add monitoring metrics for saved model (export_dir) warm up latency. (commit: de0935b)
- Add more details/clarification to model version labels documentation. (commit: f9e6ac4)
- Split
--tensorflow_session_parallelism
flag into two new flags:--tensorflow_intra_op_parallelism
and--tensorflow_inter_op_parallelism
(commit: 71092e4) - Update CPU Docker images to Ubuntu 18.04 (commit: 8023fba)
- Upgrade to Bazel 0.20.0 (commit: fc0b75f)
- Update Python 2 scripts to be compatible with both Python 2 and 3 (commit: 846d443)
Thanks to our Contributors
This release contains contributions from many people at Google, as well as:
Daniel Shi, Karthik Vadla, lapolonio, robert, Shintaro Murakami, Siju, Tom Forbes, Ville TöRhöNen