Release Notes

MindSpore Golden Stick 0.6.0 Release Notes

Major Features and Improvements

  • The RoundToNearest supports Mindformers' KVCache int8 quantization now, i.e. PagedAttentionMgr class, mainly for Llama2 networks.

  • Added Post-Training Quantization algorithm named PTQ which supports SmoothQuant, A16W8, KVCacheInt8 and their combinations, such as A16W8 combined with KVCacheInt8, SmoothQuant combined with KVCacheInt8, etc., and the corresponding algorithm capabilities can be obtained by configuring PTQConfig. The algorithm is mainly supports ParallelLlama2 network from the MindFormers community.

API Change

  • PTQConfig adds the following three parameters:

    • act_quant_dtype: The data type is mindspore.dtype. The default value is None. The options and meanings are as follows:

      act_quant_dtype

      mindspore.dtype.int8

      None(default)

      meanings

      quantize input to int8

      does not quantize input

    • weight_quant_dtype: The data type is mindspore.dtype. The default value is mindspore.dtype.int8. The options and meanings are as follows:

      weight_quant_dtype

      mindspore.dtype.int8(default)

      None

      meanings

      quantize weights to int8

      does not quantize weights

    • kvcache_quant_dtype: The data type is mindspore.dtype. The default value is None. The options and meanings are as follows:

      kvcache_quant_dtype

      mindspore.dtype.int8

      None(default)

      meanings

      quantize kvcache to int8

      does not quantize kvcache

    • outliers_suppression: The data type is OutliersSuppressionType. The default value is OutliersSuppressionType.NONE. The options and meanings are as follows:

      outliers_suppression

      OutliersSuppressionType.SMOOTH

      OutliersSuppressionType.NONE(default)

      meanings

      employ smooth approach to suppress outliers in activation and weight

      does not suppress outliers

Contributors

Thanks goes to these wonderful people:

ccsszz, yyyyrf, hangangqiang

Contributions of any kind are welcome!