Objective
Introduction
In terms of evaluating the status of a specific model during tuning, we should have general objectives to measure the status of different models.
Intel Extension for Transformers supports optimized low-precision recipes for deep learning models to achieve optimal product objectives like inference performance and memory usage with expected accuracy criteria.
Supported Objectives Matrix:
|Argument |Type |Description |Default value |
|:———-|:———-|:———————————————–|:—————-|
|name |string |a objective name in Intel Neural Compressor. Like “performance”, “modelsize”,……and so on| / |
|greater_is_better|bool |used to describe the usage of the objective, like: greater is better for performance, but lower is better for modelsize| True |
|weight_ratio|float |used when there are multiple objective.
for example: different weight proportion on performance and modelsize.| None |
Examples:
There are two built-in objective instances: performance, modelsize. Users can also build their own objective as below:
from intel_extension_for_transformers.objectives import performance, modelsize
or
from intel_extension_for_transformers.transformers import objectives
performance = objectives.Objective(name="performance", greater_is_better=True, weight_ratio=None)