intel / neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
https://intel.github.io/neural-compressor/
Apache License 2.0
2.23k stars 257 forks source link

Fix to_json_file function in BaseConfig #1953

Closed yuwenzho closed 3 months ago

yuwenzho commented 3 months ago

Type of Change

bug fix API changed or not

Description

Fix to_json_file function in BaseConfig

How has this PR been tested?

local test

Dependency Change?

no

chensuyue commented 3 months ago

Will fixed in internal repo first.