Model Compression based on Quantization, Pruning, Knowledge Distillation, and AutoML
Opt-AI’s optimization solution compresses AI models using state-of-the-art techniques including quantization, pruning, knowledge distillation, and AutoML.
Quantization
Pruning
Knowledge Distillation
Moreover, AI models are accelerated on any target AI chipset including CPUs, GPUs, ASICs and FPGAs, with the Opt-AI’s hardware-aware automated profiling solution
Inference Speedup
Accuracy Preserving
Hardware Full-Awareness
Model size & Memory Reduction
Opt-AI Inc. HQ: B209 Sciences and Technology Building Ⅱ at Hanyang University, 55, Hanyangdaehak-ro, Sangnok-gu, Ansan-si, Gyeonggi-do Lab: SuperStartLab 06, LG Science Park E10, 70, Magokjungang 10-ro , Gangseo-gu, Seoul