How to Create a Batch Inference API for a Model?

Solved!
Suhail
Level 3
How to Create a Batch Inference API for a Model?

Hello Dataiku Community,

I'm looking for guidance on how to set up a batch inference API for a machine learning model. Specifically, I want to create an API endpoint that can take a batch of data and return predictions from my model.

Here are a few details about my setup:

- I have a trained model.

- I want to provide it with a batch of input data.

- I need the API to return predictions for each data point in the batch.

Could someone please provide step-by-step instructions, best practices, or point me to relevant documentation or tutorials on achieving this in Dataiku?

Any help or insights would be greatly appreciated.

Thank you in advance!

0 Kudos
1 Solution
AdrienL
Dataiker

Hi,

This is natively handled by the API node, in which you can deploy a model prediction endpoint. You can call this endpoint with a batch of records (of reasonable size), and it will return the prediction for all these records.

See the โ€œFirst APIโ€ sections for guide in the API node documentation.

View solution in original post

2 Replies
AdrienL
Dataiker

Hi,

This is natively handled by the API node, in which you can deploy a model prediction endpoint. You can call this endpoint with a batch of records (of reasonable size), and it will return the prediction for all these records.

See the โ€œFirst APIโ€ sections for guide in the API node documentation.

Suhail
Level 3
Author

Hi @AdrienL ,

thanks for the reply.

i already had followed the guide you shared.

However i missed the part where it mentions for batch inference the endpoint would change from /predict to /predict-multi.

I am able to run batch inferences now.

Thanks

0 Kudos