1712n / challenge

Challenge Program
65 stars 27 forks source link

Service to deploy NLP models for inference #98

Closed alinapark closed 1 year ago

alinapark commented 1 year ago

Overview

Create a service that deploys five NLP models for inference, then receives messages through an exposed POST API endpoint, and finally returns inference results (of all five models) in a single response body. Expected deliverable is a service packed in the Docker image.

Your service could be a well-configured framework or a self-made API server; use any ML model deployment tool you see fit. There's no language limitation. The most important here is the reusability of a final project.

All submissions will be deployed on a g4dn.2xlarge instance, and the assessment will be performance (mainly, throughput)-based

More details in the challenge repository


Additional Resources

khushpatel2002 commented 1 year ago

Hello @alinapark , For the challenge, is it allowd to change the format of model mentioned. from pytorch to tensorflow / onnx ?

alinapark commented 1 year ago

@khushpatel2002 yes, feel free to use either

khushpatel2002 commented 1 year ago

@alinapark Thank you for your reply.

mizykov commented 1 year ago

@alinapark, hello! Is this challenge still relevant?