tlt.models.hf_model.HFModel

class tlt.models.hf_model.HFModel(model_name: str, framework: FrameworkType, use_case: UseCaseType)[source]

Base class to represent a Hugging Face model

__init__(model_name: str, framework: FrameworkType, use_case: UseCaseType)[source]

Class constructor

Methods

__init__(model_name, framework, use_case)

Class constructor

benchmark(dataset[, saved_model_dir, ...])

Use Intel Neural Compressor to benchmark the model with the dataset argument.

evaluate(dataset)

Evaluate the model using the specified dataset.

export(output_dir)

Export the serialized model to an output directory

load_from_directory(model_dir)

Load a model from a directory

optimize_graph(output_dir[, overwrite_model])

Performs FP32 graph optimization using the Intel Neural Compressor on the model and writes the inference-optimized model to the output_dir. Graph optimization includes converting variables to constants, removing training-only operations like checkpoint saving, stripping out parts of the graph that are never reached, removing debug operations like CheckNumerics, folding batch normalization ops into the pre-calculated weights, and fusing common operations into unified versions. :param output_dir: Writable output directory to save the optimized model :type output_dir: str :param overwrite_model: Specify whether or not to overwrite the output_dir, if it already exists (default: False) :type overwrite_model: bool.

quantize(output_dir, dataset[, config, ...])

Performs post training quantization using the Intel Neural Compressor on the model using the dataset.

train(dataset, output_dir[, epochs, ...])

Train the model using the specified dataset

Attributes

framework

Framework with which the model is compatible

learning_rate

Learning rate for the model

model_name

Name of the model

preprocessor

Preprocessor for the model

use_case

Use case (or category) to which the model belongs