You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
support fp16 dtypes for input weight and bias (pytorch#1017)
Summary:
X-link: pytorch#3931
Pull Request resolved: facebookresearch/FBGEMM#1017
ATT, this diff
- Supports fp16 inputs for fp8 quantization and gemm
- add addmm fallback to cpu for fp8gemm, s.t. the module is compatible with publish time processings
Reviewed By: sijiac
Differential Revision: D72479579
fbshipit-source-id: db9615095b1f246578fec39b92be6c2238ded5da
0 commit comments