[0.6.1] InternLM SmoothQuant does not work #705
Labels
Low Precision
Issue about lower bit quantization, including int8, int4, fp8
triaged
Issue has been triaged by maintainers
I am running 0.6.1 with InternLM model, with the following configurations
The conversion is finished successfully. However, when you start build the engine:
This error started to appear
I have printed out the layer
It seemed that the
bias
property are not passed correctly when converting to smoothquant. This caused the bias object are initialized as None.Environment information
The text was updated successfully, but these errors were encountered: