Full Publications/Events (52)¶
2022 (34)¶
#MLefficiency — Optimizing transformer models for efficiency(Dec 2022)
One-Click Acceleration of Hugging Face Transformers with Intel’s Neural Coder(Dec 2022)
One-Click Quantization of Deep Learning Models with the Neural Coder Extension(Dec 2022)
Accelerate Stable Diffusion with Intel Neural Compressor(Dec 2022)
Intel Neural Compressor for TF Virtual Appliance packaged by Bitnami (Nov 2022)
Neural Compressor: an open-source Python library for network compression (Nov 2022)
Personalized Stable Diffusion with Few-Shot Fine-Tuning (Nov 2022)
Fast DistilBERT on CPUs (Oct 2022)
Meet the Innovation of Intel AI Software: Intel® Extension for TensorFlow* (Oct 2022)
PyTorch* Inference Acceleration with Intel® Neural Compressor (Oct 2022)
Neural Coder, a new plug-in for Intel Neural Compressor was covered by Twitter, LinkedIn, and Intel Developer Zone from Intel, and Twitter and LinkedIn from Hugging Face. (Oct 2022)
Intel Neural Compressor successfully landed on GCP, AWS, and Azure marketplace. (Oct 2022)
Alibaba Cloud and Intel Neural Compressor Deliver Better Productivity for PyTorch Users [Chinese version] (Sep 2022)
Efficient Text Classification with Intel Neural Compressor (Sep 2022)
Dynamic Neural Architecture Search with Intel Neural Compressor (Sep 2022)
Easy Quantization in PyTorch Using Fine-Grained FX (Sep 2022)
One-Click Enabling of Intel Neural Compressor Features in PyTorch Scripts (Aug 2022)
Deep learning inference optimization for Address Purification (Aug 2022)
Accelerate AI Inference without Sacrificing Accuracy (Jun 2022)
PyTorch Inference Acceleration with Intel® Neural Compressor (Jun 2022)
Intel and Hugging Face Partner to Democratize Machine Learning Hardware Acceleration (Jun 2022)
Intel® Neural Compressor oneAPI (Jun 2022)
INC as PT ecosystem project (Apr 2022)
Dynamic Quantization with Intel Neural Compressor and Transformers (Mar 2022)
Quantizing ONNX Models using Intel® Neural Compressor (Feb 2022)
Quantize AI Model by Intel® oneAPI AI Analytics Toolkit on Alibaba Cloud (Feb 2022)
AI Performance and Productivity with Intel® Neural Compressor (Jan 2022)
Ease-of-use quantization for PyTorch with Intel® Neural Compressor (Jan 2022)
2021 (14)¶
Intel Neural Compressor Tutorial on BiliBili (Dec 2021)
Faster AI/ML Results With Intel Neural Compressor (Dec 2021)
Prune Once for All: Sparse Pre-Trained Language Models (Nov 2021)
Faster, Easier Optimization with Intel® Neural Compressor (Nov 2021)
Accelerate Deep Learning with Intel® Extension for TensorFlow* (Oct 2021)
Intel® Neural Compressor: A Scalable Quantization Tool for ONNX Models (Oct 2021)
Optimize TensorFlow Pre-trained Model for Inference (Jun 2021)
3D Digital Face Reconstruction Solution enabled by 3rd Gen Intel® Xeon® Scalable Processors (Apr 2021)
MLPerf™ Performance Gains Abound with latest 3rd Generation Intel® Xeon® Scalable Processors (Apr 2021)
Using Low-Precision Optimizations for High-Performance DL Inference Applications (Apr 2021)
Quantization support for ONNX using LPOT (Low precision optimization tool) (Mar 2021)
2018 - 2020 (4)¶
Reduced Precision Strategies for Deep Learning: 3DGAN Use Case - presentation on 4th IML Machine Learning Workshop (Oct 2020)
Intel Neural Compressor (Sep 2020)
Lower Numerical Precision Deep Learning Inference and Training (May 2018)
Highly Efficient 8-bit Low Precision Inference of Convolutional Neural Networks with IntelCaffe (May 2018)