diff --git a/backends/cadence/aot/quantizer/quantizer.py b/backends/cadence/aot/quantizer/quantizer.py index 69c9518166b..761b2bf8d31 100644 --- a/backends/cadence/aot/quantizer/quantizer.py +++ b/backends/cadence/aot/quantizer/quantizer.py @@ -193,7 +193,6 @@ def get_cadence_default_quantizers() -> List[Quantizer]: CadenceAtenQuantizer(BmmPattern(), qconfig_A8W8), CadenceAtenQuantizer(Conv1dPattern(), qconfig_A8W8sym), CadenceAtenQuantizer(Conv2dPattern(), qconfig_A8W8sym), - CadenceAtenQuantizer(LayerNormPattern(), qconfig_A8W8), CadenceAtenQuantizer(LinearPattern(), qconfig_A8W8), CadenceAtenQuantizer(MatmulPattern(), qconfig_A8W8), CadenceAtenQuantizer(ReluPattern0(), qconfig_A8W8), @@ -236,9 +235,21 @@ def __init__( super().__init__([]) +class CadenceWithLayerNormQuantizer(CadenceQuantizer): + """ + Quantizer including layer norm + """ + + def __init__(self, quantizers: Optional[list[Quantizer]] = None) -> None: + if quantizers is None: + quantizers = get_cadence_default_quantizers() + quantizers.append(CadenceAtenQuantizer(LayerNormPattern(), qconfig_A8W8)) + super().__init__(quantizers) + + class CadenceWakeWordQuantizer(CadenceQuantizer): """ - Quantizer for WakeWord, including add + Quantizer for WakeWord, including add and cat """ def __init__(self, quantizers: Optional[list[Quantizer]] = None) -> None: