Skip to content

Makes use_fast_accum configurable. #3829

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

levendlee
Copy link
Member

Summary:
[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

  • Exposes fast accumulation as a configurable.
  • Not enable it by default. No change in default behavior.
  • No additional tuning regarding to use_fast_accum=True.

Differential Revision: D71290596

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

Copy link

netlify bot commented Mar 17, 2025

Deploy Preview for pytorch-fbgemm-docs ready!

Name Link
🔨 Latest commit c848446
🔍 Latest deploy log https://app.netlify.com/sites/pytorch-fbgemm-docs/deploys/67d854628c9e5b0008bf35be
😎 Deploy Preview https://deploy-preview-3829--pytorch-fbgemm-docs.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable. 
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable. 
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:
Pull Request resolved: pytorch#3829

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:
Pull Request resolved: pytorch#3829

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable. 
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable. 
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:
Pull Request resolved: pytorch#3829

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:
Pull Request resolved: pytorch#3829

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Differential Revision: D71290596
@levendlee levendlee force-pushed the export-D71290596 branch 2 times, most recently from dc8074c to 58ba3e8 Compare March 17, 2025 16:43
levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable. 
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Reviewed By: htyu

Differential Revision: D71290596
levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable. 
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Reviewed By: htyu

Differential Revision: D71290596
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

levendlee added a commit to levendlee/FBGEMM that referenced this pull request Mar 17, 2025
Summary:
Pull Request resolved: pytorch#3829

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Reviewed By: htyu

Differential Revision: D71290596
Summary:
Pull Request resolved: pytorch#3829

X-link: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Reviewed By: htyu

Differential Revision: D71290596
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D71290596

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 067b63c.

liligwu pushed a commit to ROCm/FBGEMM that referenced this pull request Mar 19, 2025
Summary:
Pull Request resolved: pytorch#3829

X-link: https://github.com/facebookresearch/FBGEMM/pull/913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Reviewed By: htyu

Differential Revision: D71290596

fbshipit-source-id: 8e2a20899f301f861d8d72f6290e573e23288e63
q10 pushed a commit to q10/FBGEMM that referenced this pull request Apr 10, 2025
Summary:
X-link: pytorch#3829

Pull Request resolved: facebookresearch/FBGEMM#913

[Public to OSS]

Thanks htyu for pointing out the issue. Looking forward to warp specialization support on Nvidia!

- Exposes fast accumulation as a configurable.
- Not enable it by default. No change in default behavior.
- No additional tuning regarding to `use_fast_accum=True`.

W/ HIP backend, the semantics of `c += tl.dot(a, b)` and `c = tl.dot(a,b,c)` seems to be the same.

Reviewed By: htyu

Differential Revision: D71290596

fbshipit-source-id: 8e2a20899f301f861d8d72f6290e573e23288e63
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants