You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Fix handling of dynamic FP8 grouped gemm on Nvidia (pytorch#3616)
Summary:
X-link: https://github.com/facebookresearch/FBGEMM/pull/695
Pull Request resolved: pytorch#3616
This diff is the nvidia mirror of D68686266, which changes dynamic grouped gemm to return a tensor of shape [total_M, N] when zero_start_index_M isnt provided. We also add appropriate tests to make sure the behavior doesnt break going forward.
Reviewed By: jasonjk-park, jianyuh, jiawenliu64
Differential Revision: D68689077
fbshipit-source-id: f60b533e6ec90b753dc15f2136c7cef6e162bf1c
0 commit comments