Olive-ai 0.3.1
Examples
The following examples are added
- Red Pajama Optimization with Optimum
- Stable Diffusion XL Optimization with DirectML
- GPT-J Optimization Using Intel® Neural Compressor
- BERT example using Intel Neural Compressor SmoothQuant
- Whisper example using Intel Neural Compressor
- Open LLaMA workflow example
Passes (optimization techniques)
- Introduce TorchTRTConversion
- Introduce SparseGPT pass for one-shot model pruning on large GPT like models using the algorithm proposed in https://arxiv.org/abs/2301.00774.
Systems
- Add AzureML sku support for AMLSystem
Evaluator
- Add metric_func config to custom metric. Olive will run the inference for custom eval func for user. User doesn't need to do inference by themselves.
- Add RawDataContainer:
SNPE evaluation and quantization now accept generic dataloaders such as torch dataloader
Metrics
- Add Perplexity metric for text-generation task
Engine
- Provide the interface to let user set the multi pass flows to run in save olive workflow