Skip to content

fix(m3): make bf16 inference/train loading safe and unify numpy conversion#1567

Merged
hanhainebula merged 2 commits intoFlagOpen:masterfrom
lnxtree:feat-fix-embedder_use_bf16
Mar 26, 2026
Merged

fix(m3): make bf16 inference/train loading safe and unify numpy conversion#1567
hanhainebula merged 2 commits intoFlagOpen:masterfrom
lnxtree:feat-fix-embedder_use_bf16

Conversation

@lnxtree
Copy link
Contributor

@lnxtree lnxtree commented Mar 26, 2026

  • replace m3 embedder .cpu().numpy() paths with base _convert_to_numpy(...) for bf16-safe conversion
  • add torch_dtype plumbing in m3 runner model loading (AutoModel/colbert_linear/sparse_linear) to keep dtype behavior consistent

lnxtree added 2 commits March 26, 2026 15:19
…rsion

  - replace m3 embedder .cpu().numpy() paths with base _convert_to_numpy(...) for bf16-safe conversion
  - add torch_dtype plumbing in m3 runner model loading (AutoModel/colbert_linear/sparse_linear) to keep dtype behavior consistent
@hanhainebula hanhainebula merged commit 4bcfc6e into FlagOpen:master Mar 26, 2026
1 check passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants