From 653cc1cd87ffbcacf44032643b4c52e13a26e918 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 2 Dec 2024 18:59:28 +0000 Subject: [PATCH] :arrow_up: Update flash-attn requirement from ~=2.6.3 to >=2.6.3,<2.8.0 Updates the requirements on [flash-attn](https://github.com/Dao-AILab/flash-attention) to permit the latest version. - [Release notes](https://github.com/Dao-AILab/flash-attention/releases) - [Commits](https://github.com/Dao-AILab/flash-attention/compare/v2.6.3...v2.7.0.post2) --- updated-dependencies: - dependency-name: flash-attn dependency-type: direct:production ... Signed-off-by: dependabot[bot] --- pyproject.toml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pyproject.toml b/pyproject.toml index 553ae9f..5a34d10 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -41,7 +41,7 @@ dependencies = [ "accelerate>=0.33,<1.1", "sentencepiece~=0.2.0", "peft>=0.12,<0.14", - "flash-attn~=2.6.3; sys_platform != 'darwin'", + "flash-attn>=2.6.3,<2.8.0; sys_platform != 'darwin'", "einops~=0.8.0", "timm~=1.0.9", "typer~=0.12.5",