To fully take advantage of this repo, we recommend you to clone it and try out the service locally.
This requires Python3.8+ and pip
installed.
git clone https://github.com/bentoml/transformers-nlp-service.git && cd transformers-nlp-service
pip install -r requirements/tests.txt
bentoml serve
You can then open your browser at http://127.0.0.1:3000 and interact with the service through Swagger UI.
We also provide two pre-built container to run on CPU and GPU respectively. This requires any container engine, such as docker, podman, ... You can then quickly try out the service by running the container:
# cpu
docker run -p 3000:3000 ghcr.io/bentoml/transformers-nlp-service:cpu
# gpu
docker run --gpus all -p 3000:3000 ghcr.io/bentoml/transformers-nlp-service:gpu
Note that to run with GPU, you will need to have nvidia-docker setup.
One can also use the BentoML Python API to serve their models.
Run the following to build a Bento within the Bento Store:
bentoml build
Then, start a server with bentoml.HTTPServer
:
import bentoml
# Retrieve Bento from Bento Store
bento = bentoml.get("transformers-nlp-service")
server = bentoml.HTTPServer(bento, port=3000)
server.start(blocking=True)
If you wish to use gRPC, this project also include gRPC support. Run the following:
bentoml serve-grpc
To run the container with gRPC, do
docker run -p 3000:3000 -p 3001:3001 ghcr.io/bentoml/nlp:cpu serve-grpc
To find more information about gRPC with BentoML, refer to our documentation
The default mode of BentoML's model serving is via HTTP server. Here, we showcase a few examples of how one can interact with the service:
The following example shows how to send a request to the service to summarize a text via cURL:
curl -X 'POST' \
'http://0.0.0.0:3000/summarize' \
-H 'accept: text/plain' \
-H 'Content-Type: text/plain' \
-d 'The three words that best describe Hunter Schafer'\''s Vanity Fair Oscars party look? Less is more.
Dressed in a bias-cut white silk skirt, a single ivory-colored feather and — crucially — nothing else, Schafer was bound to raise a few eyebrows. Google searches for the actor and model skyrocketed on Sunday night as her look hit social media. On Twitter, pictures of Schafer immediately received tens of thousands of likes, while her own Instagram post has now been liked more than 2 million times.
Look of the Week: Zendaya steals the show at Louis Vuitton in head-to-toe tiger print
But more than just creating a headline-grabbing moment, Schafer'\''s ensemble was clearly considered. Fresh off the Fall-Winter 2023 runway, the look debuted earlier this month at fashion house Ann Demeulemeester'\''s show in Paris. It was designed by Ludovic de Saint Sernin, the label'\''s creative director since December.
Celebrity fashion works best when there'\''s a story behind a look. For example, the plausible Edie Sedgwick reference in Kendall Jenner'\''s Bottega Veneta tights, or Paul Mescal winking at traditional masculinity in a plain white tank top.
For his first Ann Demeulemeester collection, De Saint Sernin was inspired by "fashion-making as an authentic act of self-involvement." It was a love letter — almost literally — to the Belgian label'\''s founder, with imagery of "authorship and autobiography" baked into the clothes (Sernin called his feather bandeaus "quills" in the show notes).
Hunter Schafer'\''s barely-there Oscars after party look was more poetic than it first seemed.
These ideas of self-expression, self-love and self-definition took on new meaning when worn by Schafer. As a trans woman whose ascent to fame was inextricably linked to her gender identity — her big break was playing trans teenager Jules in HBO'\''s "Euphoria" — Schafer'\''s body is subjected to constant scrutiny online. The comment sections on her Instagram posts often descend into open forums, where users feel entitled (and seemingly compelled) to ask intimate questions about the trans experience or challenge Schafer'\''s womanhood.
Fittingly, there is a long lineage of gender-defying sentiments stitched into Schafer'\''s outfit. Founded in 1985 by Ann Demeulemeester and her husband Patrick Robyn, the brand boasts a long legacy of gender-non-conforming fashion.
"I was interested in the tension between masculine and feminine, but also the tension between masculine and feminine within one person," Demeulemeester told Vogue ahead of a retrospective exhibition of her work in Florence, Italy, last year. "That is what makes every person really interesting to me because everybody is unique."
In his latest co-ed collection, De Saint Sernin — who is renowned in the industry for his eponymous, gender-fluid label — brought his androgynous world view to Ann Demeulemeester with fitted, romantic menswear silhouettes and sensual fabrics for all (think skin-tight mesh tops, leather, and open shirts made from a translucent organza material).
Celebrity stylist Law Roach on dressing Zendaya and '\''faking it '\''till you make it'\''
A quill strapped across her chest, Schafer let us know she is still writing her narrative — and defining herself on her own terms. There'\''s an entire story contained in those two garments. As De Saint Sernin said in the show notes: "Thirty-six looks, each one a heartfelt sentence."
The powerful ensemble may become one of Law Roach'\''s last celebrity styling credits. Roach announced over social media on Tuesday that he would be retiring from the industry after 14 years of creating conversation-driving looks for the likes of Zendaya, Bella Hadid, Anya Taylor-Joy, Ariana Grande and Megan Thee Stallion.'
To send requests in Python, one can use bentoml.client.Client
to send requests to the service:
if __name__ == "__main__":
import bentoml
client = bentoml.client.Client.from_url(f"http://{host}:3000")
print("Summarized text from the article:", client.summarize(SAMPLES))
print("Categories prediction of the article:", client.categorize({'text': SAMPLES, 'categories': CATEGORIES}))
Run python client.py
to see it in action.
Checkout the
client.py
file for more details.
Note that all API endpoints defined in service.py
can be access through client through its sync and async methods. For example, the service.py
contains three endpoints: /summarize
, /categorize
and /make_analysis
, and hence the following
methods are available on the client instance:
client.async_summarize
| client.summarize
client.async_categorize
| client.categorize
client.async_make_analysis
| client.make_analysis
You can also send requests to this service with axios
in JS.
The following example sends a request to make analysis on a given text and categories:
import axios from 'axios'
var TEXT = `...`
var CATEGORIES = [ 'world', 'politics', 'technology', 'defence', 'parliament' ]
const client = axios.create({
baseURL: 'http://localhost:3000',
timeout: 3000,
})
client
.post('/make_analysis', {
text: TEXT,
categories: CATEGORIES.join(', '),
})
.then(function (r) {
console.log('Full analysis:', r.data)
})
Run the client.js
with yarn run client
or npm run client
, and it should yield the following result
Full analysis: {
summary: " Actor and model Hunter Schafer wore a barely-there Oscars after party look . The look debuted
earlier this month at fashion house Ann Demeulemeester's Fall-Winter 2023 runway . It was designed by des
igner Ludovic de Saint Sernin, who is renowned for his eponymous label .",
categories: {
entertainment: 0.4694322943687439,
healthcare: 0.4245288372039795,
defence: 0.42102956771850586,
world: 0.416515976190567,
}
}
Checkout the
client.js
for more details.
This project is designed to be used with different NLP tasks and its corresponding models:
You can add more tasks and models by editing the download_model.py
file.
Pre/post processing logics can be set in the service.py
file.
BentoML supports Transformers models out of the box. You can find more details in the BentoML support for Transformers.
Effortlessly transition your project into a production-ready application using BentoCloud, the production-ready platform for managing and deploying machine learning models.
Start by creating a BentoCloud account. Once you've signed up, log in to your BentoCloud account using the command:
bentoml cloud login --api-token <your-api-token> --endpoint <bento-cloud-endpoint>
Note: Replace
<your-api-token>
and<bento-cloud-endpoint>
with your specific API token and the BentoCloud endpoint respectively.
Next, build your BentoML service using the build
command:
bentoml build
Then, push your freshly-built Bento service to BentoCloud using the push
command:
bentoml push <name:version>
Lastly, deploy this application to BentoCloud with a single bentoml deployment create
command following the deployment instructions.
BentoML offers a number of options for deploying and hosting online ML services into production, learn more at Deploying a Bento.
BentoML has a thriving open source community where thousands of ML/AI practitioners are contributing to the project, helping other users and discussing the future of AI. 👉 Pop into our Slack community!