Skip to content

Flex attention + refactor #34809

@ArthurZucker

Description

@ArthurZucker

Opening this to add support for all models following #34282

Lets bring support for flex attention to more models! 🤗

  • Gemma2

It would be great to add the support for more architectures such as

  • Qwen2
  • Llama
  • Gemma
  • QwenVl
  • Mistral
  • Clip

... and many more

For anyone who wants to contribute just open a PR and link it to this issue, and ping me for a review!! 🤗

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions