Define ST_F8_E8M0#3448
Merged
zcbenz merged 1 commit intoml-explore:mainfrom May 5, 2026
Merged
Conversation
No conversion is performed, this just allows safetensors files to be loaded. I think this should help load the native DeepSeek checkpoints without conversion, as the attention scales use F8_E8M0.
This was referenced Apr 24, 2026
zcbenz
approved these changes
Apr 24, 2026
Collaborator
zcbenz
left a comment
There was a problem hiding this comment.
Looks good to me!
Adding test is tricky because the from_fp8/to_fp8 ops assume e4m3, we could extend the APIs to e8m0 but I think maybe we should just add fp8 support instead. So I'm good with no test for this one.
Contributor
Author
👍 Sounds good! I searched for type mentions in the docs but I found nothing to update either. Let me know if anything else is missing. |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Opening for discussion. I saw #3374, so perhaps this should go there instead. The goal is just to be able to load the native DeepSeek V4 safetensors files, since the attention scales use F8_E8M0.
If this direction is ok, I'm happy to add tests or update docs as needed.
Proposed changes
Define ST_F8_E8M0 so loading safetensor files that include this type succeed. Handling would be deferred to
sanitizefunctions inmlx_lmand other user code.Checklist
Put an
xin the boxes that apply.pre-commit run --all-filesto format my code / installed pre-commit prior to committing changes