Skip to content

Int8DynActInt4WeightQuantizer quantize() not propagating the scales_precision argument #2571

@HaroldBenoit

Description

@HaroldBenoit

Hello,

In Int8DynActInt4WeightQuantizer, the scales_precision argument of replace_linear_8da4w is set to self.precision instead of self.scales_precision.

This introduces an inconsistency between calling _create_quantized_state_dict(model) and calling quantize(model).state_dict().

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions