PyTorch |
Mixed Precision |
pytorch_amp |
PyTorch |
Channels Last |
pytorch_channels_last |
PyTorch |
JIT (Just-In-Time) Script/Trace & optimize_for_inference |
pytorch_jit_script , pytorch_jit_trace , pytorch_jit_script_ofi , pytorch_jit_trace_ofi |
PyTorch |
JIT with TorchDynamo |
pytorch_torchdynamo_jit_script , pytorch_torchdynamo_jit_trace , pytorch_torchdynamo_jit_script_ofi , pytorch_torchdynamo_jit_trace_ofi |
PyTorch |
Intel Neural Compressor (INC) Mixed Precision |
pytorch_inc_bf16 |
PyTorch |
INC INT8 Static Quantization (FX/IPEX) |
pytorch_inc_static_quant_fx , pytorch_inc_static_quant_ipex |
PyTorch |
INC INT8 Dynamic Quantization |
pytorch_inc_dynamic_quant |
PyTorch |
Intel Extension for PyTorch (FP32, BF16, INT8 Static/Dynamic Quantization) |
pytorch_ipex_fp32 , pytorch_ipex_bf16 , pytorch_ipex_int8_static_quant , pytorch_ipex_int8_dynamic_quant |
PyTorch |
Alibaba Blade-DISC |
pytorch_aliblade |
PyTorch Lightning |
Mixed Precision |
pytorch_lightning_bf16_cpu |
TensorFlow |
Mixed Precision |
tensorflow_amp |
Keras |
Mixed Precision |
keras_amp |
TensorFlow/Keras Model |
INC Quantization |
tensorflow_inc |
Keras Script |
INC Quantization |
keras_inc |
ONNX Runtime |
INC Static Quantization (QLinear) |
onnx_inc_static_quant_qlinear |
ONNX Runtime |
INC Static Quantization (QDQ) |
onnx_inc_static_quant_qdq |
ONNX Runtime |
INC Dynamic Quantization |
onnx_inc_dynamic_quant |
HuggingFace Optimum-Intel |
INC Quantization |
pytorch_inc_huggingface_optimum_static , pytorch_inc_huggingface_optimum_dynamic |
Intel Extension for Transformers |
INC Quantization |
intel_extension_for_transformers |
BigDL Nano |
Optimization List |
nano_ + specific alias |
Auto-Detect |
INC Quantization |
inc_auto |