Open natolambert opened 3 weeks ago
cc @mgoin
@natolambert Could you help narrow this down to a specific model checkpoint and prompt? This will help directly compare with some reference
@mgoin we were using your checkpoints (FP8) versions, both instruct and base. I can find the links if needed to be more specific.
Your current environment
The python environment
```text [project] name = "birr" version = "0.2.0" description = "Pipeline for rewriting Web Documents" authors = [{ name = "Allen AI", email = "contact@allenai.org" }] license = { text = "Apache-2.0" } readme = "README.md" requires-python = ">=3.8" dependencies = [ "boto3", "pyaml", "pydantic>=2,<3", "pydantic-settings", "smart_open[s3]", "zstandard", ] classifiers = ["Development Status :: 3 - Alpha", "Typing :: Typed"] [project.urls] Homepage = "https://github.com/allenai/refine" Source = "https://github.com/allenai/refine" Tracker = "https://github.com/allenai/refine/issues" [project.optional-dependencies] dev = [ "black[jupyter]>=22.6.0", "flake8>=5.0", "flake8-pyi>=22.8.1", "Flake8-pyproject>=1.1.0", "ipdb>=0.13.0", "ipython>=8.4.0", "isort>=5.10.1", "mypy>=0.971", "pytest>=5.2", "moto[s3,sqs]", ] batch_inference = [ "bitsandbytes", "datadog", "numpy", "psutil", "pynvml==11.4.1", "ray[default]", "sentencepiece", "torch>=2.2.0", "transformers>=4.41.1", ] # vllm is separated because it cannot be installed on macs vllm = [ "vllm==0.5.4", "vllm-flash-attn==2.6.1" ] ```🐛 Describe the bug
Nemotron 340b models implemented in #6611 are not generating the EOS token correctly. Some more information.
Parameters:
Basic code:
Example:
Before submitting a new issue...