diff --git a/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_benchmark.sh b/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_benchmark.sh index 97754097b..64706c474 100644 --- a/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_benchmark.sh +++ b/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_benchmark.sh @@ -14,19 +14,19 @@ function init_params { do case $var in --input_model=*) - input_model=$(echo $var |cut -f2 -d=) + input_model=$(echo "$var" |cut -f2 -d=) ;; --batch_size=*) - batch_size=$(echo $var |cut -f2 -d=) + batch_size=$(echo "$var" |cut -f2 -d=) ;; --tokenizer=*) - tokenizer=$(echo $var |cut -f2 -d=) + tokenizer=$(echo "$var" |cut -f2 -d=) ;; --mode=*) - mode=$(echo $var |cut -f2 -d=) + mode=$(echo "$var" |cut -f2 -d=) ;; --intra_op_num_threads=*) - intra_op_num_threads=$(echo $var |cut -f2 -d=) + intra_op_num_threads=$(echo "$var" |cut -f2 -d=) ;; esac done diff --git a/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_quant.sh b/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_quant.sh index 255611b22..0ddf19804 100644 --- a/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_quant.sh +++ b/examples/nlp/huggingface_model/text_generation/quantization/weight_only/run_quant.sh @@ -12,25 +12,25 @@ function init_params { do case $var in --input_model=*) - input_model=$(echo $var |cut -f2 -d=) + input_model=$(echo "$var" |cut -f2 -d=) ;; --output_model=*) - output_model=$(echo $var |cut -f2 -d=) + output_model=$(echo "$var" |cut -f2 -d=) ;; --batch_size=*) - batch_size=$(echo $var |cut -f2 -d=) + batch_size=$(echo "$var" |cut -f2 -d=) ;; --dataset=*) - dataset=$(echo $var |cut -f2 -d=) + dataset=$(echo "$var" |cut -f2 -d=) ;; --tokenizer=*) - tokenizer=$(echo $var |cut -f2 -d=) + tokenizer=$(echo "$var" |cut -f2 -d=) ;; --algorithm=*) - algorithm=$(echo $var |cut -f2 -d=) + algorithm=$(echo "$var" |cut -f2 -d=) ;; --quant_format=*) - quant_format=$(echo $var |cut -f2 -d=) + quant_format=$(echo "$var" |cut -f2 -d=) ;; esac done diff --git a/onnx_neural_compressor/quantization/config.py b/onnx_neural_compressor/quantization/config.py index 01acd7eaf..20c58fd32 100644 --- a/onnx_neural_compressor/quantization/config.py +++ b/onnx_neural_compressor/quantization/config.py @@ -200,6 +200,7 @@ class ExampleAlgorithmConfig: return config_registry.register_config_impl(algo_name=algo_name, priority=priority) + class Encoder(json.JSONEncoder): def default(self, o): if isinstance(o, quantization.QuantType): @@ -209,7 +210,8 @@ def default(self, o): if isinstance(o, quantization.CalibrationMethod): return getattr(o, "name") return super().default(o) - + + class BaseConfig(ABC): """The base config for all algorithm configs.""" diff --git a/test/quantization/test_algorithm_utility.py b/test/quantization/test_algorithm_utility.py index 28525eeca..4ab8fc5db 100644 --- a/test/quantization/test_algorithm_utility.py +++ b/test/quantization/test_algorithm_utility.py @@ -41,5 +41,6 @@ def test_is_B_transposed(self): ) self.assertFalse(quant_utils.is_B_transposed(node)) + if __name__ == "__main__": unittest.main()