Quantized op implementation: quantized_decomposed::dequantize_per_channel_group #7676
Labels
module: kernels
Issues related to kernel libraries, e.g. portable kernels and optimized kernels
module: quantization
rfc
Request for comment and feedback on a post, proposal, etc.
🚀 The feature, motivation and pitch
The op
quantized_decomposed::dequantize_per_channel_group
implementation would help in using executorch's quantized CPU ops in models like llama3Came across this issue when trying to build llama3 without XNNPack. More details here #6975
Alternatives
No response
Additional context
No response
RFC (Optional)
No response
The text was updated successfully, but these errors were encountered: