Skip to content

[mlir][quant] Bump up the MaxStorageBits from 32 to 64. #91584

@dansuh17

Description

@dansuh17

64-bit storage type for quantized type is often used for certain settings like on-device or when dealing with models that are quality-sensitive. For example, a TFLite micro kernel supports 64-bit quantized types for 16x8 quantized operations.

Currently the Quant dialect allows up to 32 bit storage bits: MaxStorageBits. Bump this limit up to 64.

Activity

dansuh17

dansuh17 commented on May 9, 2024

@dansuh17
Author

(first time contributing directly on github - how do I assign myself to this issue? I'm planning on working on this)

bviyer

bviyer commented on May 15, 2024

@bviyer
Contributor

(first time contributing directly on github - how do I assign myself to this issue? I'm planning on working on this)

I did it for you.

dansuh17

dansuh17 commented on May 16, 2024

@dansuh17
Author

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

    Development

    No branches or pull requests

      Participants

      @EugeneZelenko@dansuh17@bviyer

      Issue actions

        [mlir][quant] Bump up the MaxStorageBits from 32 to 64. · Issue #91584 · llvm/llvm-project