-
Notifications
You must be signed in to change notification settings - Fork 24.7k
[cp] dispatch flex_attention on DTensor to cp implementation #151900
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: gh/XilunWu/136/base
Are you sure you want to change the base?
Conversation
[ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/151900
Note: Links to docs will display an error until the docs builds have been completed. ❌ 3 New Failures, 1 Unrelated FailureAs of commit 11c155d with merge base b7c7000 ( NEW FAILURES - The following jobs have failed:
BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
ghstack-source-id: 939ab50 Pull Request resolved: #151497 DTensor HOP call in TorchDispatchMode ghstack-source-id: 939ab50 Pull Request resolved: #151685 [cp] Add e2e flex_attention test w/ causal masking ghstack-source-id: 939ab50 Pull Request resolved: #151900 [cp] Context Parallel: dispatch flex_attention to CP impl in TorchDispatchMode ghstack-source-id: 939ab50 Pull Request resolved: #151903
ghstack-source-id: 9f052d6 Pull Request resolved: #151497 DTensor HOP call in TorchDispatchMode ghstack-source-id: 9f052d6 Pull Request resolved: #151685 [cp] Add e2e flex_attention test w/ causal masking ghstack-source-id: 9f052d6 Pull Request resolved: #151900 [cp] Context Parallel: dispatch flex_attention to CP impl in TorchDispatchMode ghstack-source-id: 9f052d6 Pull Request resolved: #151903
ghstack-source-id: e098b23 Pull Request resolved: #151497 DTensor HOP call in TorchDispatchMode ghstack-source-id: e098b23 Pull Request resolved: #151685 [cp] Add e2e flex_attention test w/ causal masking ghstack-source-id: e098b23 Pull Request resolved: #151900 [cp] Context Parallel: dispatch flex_attention to CP impl in TorchDispatchMode ghstack-source-id: e098b23 Pull Request resolved: #151903
ghstack-source-id: 4524be1 Pull Request resolved: #151497 DTensor HOP call in TorchDispatchMode ghstack-source-id: 4524be1 Pull Request resolved: #151685 [cp] Add e2e flex_attention test w/ causal masking ghstack-source-id: 4524be1 Pull Request resolved: #151900 [cp] Context Parallel: dispatch flex_attention to CP impl in TorchDispatchMode ghstack-source-id: 4524be1 Pull Request resolved: #151903
Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as |
ghstack-source-id: af830b8 Pull Request resolved: pytorch/pytorch#151900
Stack from ghstack (oldest at bottom):
cc @H-Huang @awgu @wanchaol @fegin @fduwjj @wz337 @wconstab @d4l3k