Open yhyu13 opened 7 months ago
Use transformer implementation of flash attention2
related issue https://github.com/OpenBMB/ToolBench/issues/201
Use transformer implementation of flash attention2
related issue https://github.com/OpenBMB/ToolBench/issues/201