Sparsity-aware deep learning inference runtime for CPUs
Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes
ML model optimization product to accelerate inference.