-
Notifications
You must be signed in to change notification settings - Fork 202
Open
Description
It seems that mlflow runtime only handle predict methods of its model in a synchronous way when called from
model_output = self._model.predict(decoded_payload, params=params) |
raw_predictions = self._model.predict(data, params=inference_params) |
I know mlflow embedded serve() uses Flask and is not designed to be async, but there is no hard interface contract expecting predict() to be synchronous right ?
Let's say i develop my model and set the predict() method to be async. would it make sense to work on the possibility to await it in mlserver mlflow runtime to make all the chain async ? ( maybe with some option)
jharb-mirakl, leroyvi, Jeremynadalmirakl, mroodnejad and VincentFarcette
Metadata
Metadata
Assignees
Labels
No labels