-
Notifications
You must be signed in to change notification settings - Fork 7k
add SP support for flash_varlen_hub backend
#13479
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Changes from all commits
df8994d
2d12f46
003fa34
bc61551
86fec43
5034b2b
e05bb28
534fdc1
76414aa
1cd670b
db7b8d4
99e1660
3d8cbf4
1b39db4
04a1bf5
37a6db5
849062a
b042eb0
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -25,6 +25,7 @@ | |
| AttentionTesterMixin, | ||
| BaseModelTesterConfig, | ||
| BitsAndBytesTesterMixin, | ||
| ContextParallelAttentionBackendsTesterMixin, | ||
| ContextParallelTesterMixin, | ||
| LoraHotSwappingForModelTesterMixin, | ||
| LoraTesterMixin, | ||
|
|
@@ -253,6 +254,15 @@ class TestQwenImageTransformerContextParallel(QwenImageTransformerTesterConfig, | |
| """Context Parallel inference tests for QwenImage Transformer.""" | ||
|
|
||
|
|
||
| class TestQwenImageTransformerContextParallelAttnBackends( | ||
| QwenImageTransformerTesterConfig, ContextParallelAttentionBackendsTesterMixin | ||
| ): | ||
| """Context Parallel inference x attention backends tests for QwenImage Transformer""" | ||
|
|
||
| # flash_hub and _flash_3_hub do not support attn_mask | ||
| unsupported_attn_backends = ["flash_hub", "_flash_3_hub"] | ||
|
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Any not varlen attention backend would fail no? If so, I would rather do something like if "varlen" not in attention_backend:
pytest.skip(...)
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. like FluxPipeline, it can also support I’m not sure what the most suitable place is to put this |
||
|
|
||
|
|
||
| class TestQwenImageTransformerLoRA(QwenImageTransformerTesterConfig, LoraTesterMixin): | ||
| """LoRA adapter tests for QwenImage Transformer.""" | ||
|
|
||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should
varlentests get their own testing mixin class?There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think the
varlenkernel can handle all the cases supported by thenon-varlenkernel. Personally, I prefer to put them together.