mosaicml / llm-foundry

LLM training code for Databricks foundation models
https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm
Apache License 2.0
3.99k stars 525 forks source link

Add fp32 to the set of valid inputs to attention layer #1347

Closed j316chuck closed 2 months ago

j316chuck commented 2 months ago

Description

Add fp32 to the set of valid inputs for attention layer.

Note:

Tests:

j316chuck commented 2 months ago

Tested manually, there is no unit test for this.

dakinggg commented 2 months ago

@j316chuck I don't think this is correct. Flash attention does not support fp32 (unless that changed recently?)