Skip to content

support fp16 dtypes for input weight and bias #3931

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

zhaozhul
Copy link
Contributor

@zhaozhul zhaozhul commented Apr 4, 2025

Summary:
X-link: https://github.com/facebookresearch/FBGEMM/pull/1017

ATT, this diff

  • Supports fp16 inputs for fp8 quantization and gemm
  • add addmm fallback to cpu for fp8gemm, s.t. the module is compatible with publish time processings

Differential Revision: D72479579

Summary:
X-link: facebookresearch/FBGEMM#1017

ATT, this diff
- Supports fp16 inputs for fp8 quantization and gemm
- add addmm fallback to cpu for fp8gemm, s.t. the module is compatible with publish time processings

Differential Revision: D72479579
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D72479579

Copy link

netlify bot commented Apr 4, 2025

Deploy Preview for pytorch-fbgemm-docs ready!

Name Link
🔨 Latest commit 1b04d46
🔍 Latest deploy log https://app.netlify.com/sites/pytorch-fbgemm-docs/deploys/67f04a1d51a7330008dad41f
😎 Deploy Preview https://deploy-preview-3931--pytorch-fbgemm-docs.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

@zhaozhul
Copy link
Contributor Author

zhaozhul commented Apr 4, 2025

@pytorchbot merge

Copy link

pytorch-bot bot commented Apr 4, 2025

Mergebot is not configured for this repository. Please use the merge button provided by GitHub.

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in e4905d3.

q10 pushed a commit to q10/FBGEMM that referenced this pull request Apr 10, 2025
Summary:
X-link: pytorch#3931

Pull Request resolved: facebookresearch/FBGEMM#1017

ATT, this diff
- Supports fp16 inputs for fp8 quantization and gemm
- add addmm fallback to cpu for fp8gemm, s.t. the module is compatible with publish time processings

Reviewed By: sijiac

Differential Revision: D72479579

fbshipit-source-id: db9615095b1f246578fec39b92be6c2238ded5da
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants