# Getting Started ## Installation Prebuilt wheel files are released for multiple Python versions. You can install them simply with the following pip command. ```bash python -m pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cpu python -m pip install intel_extension_for_pytorch ``` You can run a simple sanity test to double confirm if the correct version is installed, and if the software stack can get correct hardware information onboard your system. ```bash python -c "import torch; import intel_extension_for_pytorch as ipex; print(torch.__version__); print(ipex.__version__);" ``` More detailed instructions can be found at [Installation Guide](./installation.md). ## Coding IntelĀ® Extension for PyTorch\* doesn't require complex code changes to get it working. Usage is as simple as several-line code change. In general, APIs invocation should follow orders below. 1. `import intel_extension_for_pytorch as ipex` 2. Invoke `optimize()` function to apply optimizations. 3. Convert the imperative model to a graph model. - For TorchScript, invoke `torch.jit.trace()` and `torch.jit.freeze()`. - For TorchDynamo, invoke `torch.compile(model, backend="ipex")`. (*Experimental feature*, FP32 ONLY) **Note:** It is highly recommended to `import intel_extension_for_pytorch` right after `import torch`, prior to importing other packages. ```python import torch ############## import ipex ############### import intel_extension_for_pytorch as ipex ########################################## model = Model() model.eval() data = ... ############## TorchScript ############### model = ipex.optimize(model, dtype=torch.bfloat16) with torch.no_grad(), torch.cpu.amp.autocast(): model = torch.jit.trace(model, data) model = torch.jit.freeze(model) model(data) ########################################## ############## TorchDynamo ############### model = ipex.optimize(model) model = torch.compile(model, backend="ipex") with torch.no_grad(): model(data) ########################################## ``` More examples, including training and usage of low precision data types are available at [Examples](./examples.md).