gradio==5.38.2 diffusers==0.34.0 torch==2.5.1 transformers==4.45.2 einops==0.8.0 timm==1.0.19 torchvision sentencepiece https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.2.post1/flash_attn-2.7.2.post1+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl