Skip to content

Latest commit

 

History

History
 
 

README.md

HeAR serving

This folder contains the source code and configuration necessary to serve the model on Vertex AI. The implementation follows this container architecture.

The serving container can be used in both online and batch prediction workflows:

Description of select files and folders

  • data_processing/: A library for data retrieval and processing.

  • serving_framework/: A library for implementing Vertex-compatible HTTP servers.

  • vertex_schemata/: Folder containing YAML files that define the PredictSchemata for Vertex AI endpoints.

  • Dockerfile: Defines the Docker image for serving the model.

  • entrypoint.sh: A bash script that is used as the Docker entrypoint. It sets up the necessary environment variables, copies the TensorFlow SavedModel(s) locally and launches the TensorFlow server and the frontend HTTP server.

  • predictor.py: Prepares model input, calls the model, and post-processes the output into the final response.

  • requirements.txt: Lists the required Python packages.

  • server_gunicorn.py: Creates the HTTP server that launches the prediction executor.