.. Copyright (c) 2021 - present / Neuralmagic, Inc. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. =================== SparseZoo |version| =================== Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes .. raw:: html
GitHub Documentation GitHub release Contributor Covenant
Overview ======== SparseZoo is a constantly-growing repository of highly sparse and sparse-quantized models with matching sparsification recipes for neural networks. It simplifies and accelerates your time-to-value in building performant deep learning models with a collection of inference-optimized models and recipes to prototype from. Available via API and hosted in the cloud, the SparseZoo contains both baseline and models optimized to different degrees of inference performance vs. baseline loss recovery. Recipe-driven approaches built around sparsification algorithms allow you to take the models as given, transfer-learn from the models onto private datasets, or transfer the recipes to your architectures. `This repository `_ contains the Python API code to handle the connection and authentication to the cloud. Sparsification ============== Sparsification is the process of taking a trained deep learning model and removing redundant information from the overprecise and over-parameterized network resulting in a faster and smaller model. Techniques for sparsification are all encompassing including everything from inducing sparsity using `pruning `_ and `quantization `_ to enabling naturally occurring sparsity using `activation sparsity `_ or `winograd/FFT `_. When implemented correctly, these techniques result in significantly more performant and smaller models with limited to no effect on the baseline metrics. For example, pruning plus quantization can give noticeable improvements in performance while recovering to nearly the same baseline accuracy. The Deep Sparse product suite builds on top of sparsification enabling you to easily apply the techniques to your datasets and models using recipe-driven approaches. Recipes encode the directions for how to sparsify a model into a simple, easily editable format. - Download a sparsification recipe and sparsified model from the `SparseZoo `_. - Alternatively, create a recipe for your model using `Sparsify `_. - Apply your recipe with only a few lines of code using `SparseML `_. - Finally, for GPU-level performance on CPUs, deploy your sparse-quantized model with the `DeepSparse Engine `_. **Full Deep Sparse product flow:** .. raw:: html Resources and Learning More =========================== - `SparseML Documentation `_ - `Sparsify Documentation `_ - `DeepSparse Documentation `_ - `Neural Magic Blog `_, `Resources `_, `Website `_ Release History =============== Official builds are hosted on PyPI - stable: `sparsezoo `_ - nightly (dev): `sparsezoo-nightly `_ Additionally, more information can be found via `GitHub Releases `_. .. toctree:: :maxdepth: 3 :caption: General source/quicktour source/installation source/models source/recipes .. toctree:: :maxdepth: 2 :caption: API api/sparsezoo .. toctree:: :maxdepth: 3 :caption: Connect Online Bugs, Feature Requests Support, General Q&A Forums Deep Sparse Community Slack Neural Magic Docs