Relax onnxruntime requirement for conversion (#12)
This commit is contained in:
parent
a65808b35a
commit
3399f8feb3
|
@ -142,7 +142,6 @@ def quantize(models_name_or_path, model_type):
|
|||
model_output=model,
|
||||
per_channel=True,
|
||||
reduce_range=True, # should be the same as per_channel
|
||||
activation_type=QuantType.QUInt8,
|
||||
|
||||
weight_type=weight_type,
|
||||
optimize_model=False,
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
torch
|
||||
transformers
|
||||
optimum[onnxruntime]>=1.7.1
|
||||
onnxruntime==1.10.0
|
||||
onnxruntime
|
||||
tqdm
|
||||
|
|
Loading…
Reference in New Issue