Google develops TensorFlow Serving library

Google TensorFlow

Google TensorFlow Serving library

Google has released a stable version of its TensorFlow Serving. The new open source library is designed to serve machine-learned models in a production environment by offering an out of the box integration with TensorFlow models.

First released in beta back in February, TensorFlow Serving is aimed to facilitate the deployment of algorithms and experiments while maintaining the same server architecture and APIs. The library can help developers push multiple versions of machine learning models and even roll them back.

Developers can use TensorFlow Serving to integrate with other model types along with TensorFlow learning models. You need to use a Docker container to install the server binary on non-Linux systems. Notably, the complete TensorFlow package comes bundled with a pre-built binary of TensorFlow Serving.

TensorFlow Serving 1.0 comes with servables, loaders, sources and managers. Servables are basically underlying objects used for central abstraction and computation in TensorFlow Serving. Loaders, on the other hand, are used for managing a servable life cycle. Sources include plugin modules that work with servables while managers are designed to handle the life cycle of servable.

Open source existence

The major benefit of TensorFlow Serving is the set of C++ libraries that offer standards for support for learning and serving TensorFlow models. The generic core platform is not linked with TensorFlow. However, you can even use the library as a hosted service with Google Cloud ML platform.


Please enter your comment!
Please enter your name here