Skip to content

Conversation

@YangKai0616
Copy link
Contributor

What does this PR do?

This PR:

  1. Refactors ModernBERT's Flash Attention implementation using the built-in interfaces from Transformers.
  2. Fixes some UT issues in the tests/models/modernbert/test_modeling_modernbert.py, tests/models/pe_video/test_modeling_pe_video.py, and tests/models/pe_audio/test_modeling_pe_audio.py test files.

Background PR: #42952.

@github-actions
Copy link
Contributor

[For maintainers] Suggested jobs to run (before merge)

run-slow: modernbert, pe_audio, pe_video

@YangKai0616
Copy link
Contributor Author

Hey @vasqu , could you please review it again? Thank you very much!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant