github tensorflow/serving 2.4.0

latest releases: 2.17.0, 2.16.1, 2.15.1...
3 years ago

Major Features and Improvements

  • Update TF Text to v2.3.0.
  • Upgrade to CUDA Version 11.0.
  • Update CUDNN_VERSION to 8.0.4.30.
  • Adds user guide for Remote Predict Op.
  • Add support for serving regress/classify (native keras) TF2 models.

Breaking Changes

Bug Fixes and Other Changes

  • Adding /usr/local/cuda/extras/CUPTI/lib64 to LD_LIBRARY_PATH in order to unblock profiling (commit: 1270b8c)
  • Improve error message when version directory is not found (commit: d687d3e)
  • Migrate the remaining references of tf.app to compat.v1. (commit: 06fbf87)
  • Cleanup TraceMe idioms (commit: f22f802)
  • Adds LICENSE file to tensorflow-serving-api python package. (commit: 41188d4)
  • Enable a way to 'forget' unloaded models in the ServableStateMonitor. (commit: 53c5a65)
  • Added abstract layer for remote predict op over different RPC protocols with template. (commit: c54ca7e)
  • Add an example which call the Remote Predict Op directly. (commit: d5b980f)
  • For batching session in TF serving model server, introduce options to enable large batch splitting. (commit: f84187e)
  • Add multi-inference support for TF2 models that use (commit: abb8d3b)
  • Use absl::optional instead of tensorflow::serving::optional. (commit: c809305)
  • Use absl::optional instead of tensorflow::serving::optional. (commit: cf1cf93)
  • Remove tensorflow::serving::MakeCleanup and use tensorflow::gtl::MakeCleanup. (commit: 6ccb003)
  • Use absl::optional and remove tensorflow::serving::optional. (commit: e8e5222)
  • Deprecate tensorflow::CreateProfilerService() and update serving client. (commit: 98a5503)
  • Change the SASS & PTX we ship with TF (commit: 0869292)
  • Adding custom op support. (commit: 892ea42)
  • Upgrade to PY3 for tests. (commit: 02624a8)
  • Makes clear how to make a default config file for serving multiple models. (commit: 084eaeb)
  • Use TraceMeEncode in BatchingSession's TraceMe. (commit: 78ff058)
  • Export metrics for runtime latency for predict/classify/regress. (commit: c317582)
  • Refactor net_http/client to expose request/response functionality as a public API (not yet finalized) for usage testing ServerRequestInterface and HttpServerInterface instances. (commit: 0b951c8)
  • In model warm-up path, re-write error code out-of-range (intended when reading EOF in a file) to ok. (commit: d9bde73)
  • fix Client Rest API endpoint (commit: b847bac)
  • Support multiple SignatureDefs by key in TFLite models (commit: 2e14cd9)
  • Add dedicated aliases field to ModelServerConfig. (commit: 718152d)
  • Remove deprecated flag fail_if_no_model_versions_found from tensorflow serving binary (commit: 4b62462)
  • Fix TraceMe instrumentation for the padding size. (commit: 0cb94cd)
  • Add vlog to dump updated model label map (for debugging) each time the map is updated. (commit: ac10e74)
  • Add python wrapper for remote predict op and clean the build and include files. (commit: d0daa10)
  • Add portpicker module required to run modelserver e2e tests. (commit: 82f8cc0)
  • changing "infintiy" to "really high value" (commit: c96474c)
  • Minimal commandline client to trigger profiler on the modelserver. (commit: c0a5619)
  • Add signature name to RPOp. (commit: 84dfc8b)
  • When RPC error occurs, the output tensors should still get allocated. (commit: 9113de2)
  • Fix BM_MobileNet benchmark (commit: af66562)
  • Add JSPB BUILD targets for inference and example proto files. (commit: f1009eb)
  • Fall back to legacy TFLite tensor naming when parsing signature defs in TFLiteSession. (commit: 3884187)

Thanks to our Contributors

This release contains contributions from many people at Google, as well as:

Adarshreddy Adelli, Lescurel

Don't miss a new serving release

NewReleases is sending notifications on new releases.