IBM / text-generation-inference

IBM development fork of https://github.com/huggingface/text-generation-inference
Apache License 2.0
57 stars 30 forks source link

:bug: disallow downloads for fast tokenizer conversion #63

Closed joerunde closed 7 months ago

joerunde commented 7 months ago

Motivation

Some of our models are failing internal integration tests where the AutoTokenizer tries to download files even though they exist locally in the cache.

Modifications

This adds the local_files_only kwarg to the .from_pretrained() call in the fast tokenizer conversion in the launcher, to prevent hf from attempting to download tokenizer files

Result

Local tests show this allows the failing models to successfully load up tokenizers