Products
DeepSparse Engine
Sparsity-aware neural network inference engine for GPU-class performance on CPUs.
SparseML
Open-source libraries for applying sparsification recipes to neural networks with a few lines of code.
SparseZoo
Open-source model repository for sparse and sparse-quantized models.
Use Cases
Natural Language Processing (NLP)
Get started with Sparse NLP use cases. Apply your data and redeploy on CPUs with ease.
Compute Vision
Get started with sparse computer vision use cases. Apply your data and deploy on CPUs with ease.
Community
Deep Sparse Slack Community
Engage, discuss, and get direct access to our engineering teams.
Code
Explore our code on GitHub.
Docs
Get support or get started.
Labs
Labs by Neural Magic
Partner with experts to ensure your ML training and deployment pipelines are optimized with SOTA research and industry best practices.
Resources
Blogs & News
Stay current & discover SOTA performance techniques.
Research Papers
Explore our published research papers that are driving ML simplicity & performance.
Upcoming Events
Explore upcoming events Neural Magic will be attending. Come say hello!
On-Demand Videos
Listen to on-demand ML performance content on your time.
Our Technology
Learn about the "magic" behind Neural Magic!
Subscribe
Sing up for the regular Neural Magic email updates.
Company
Team & Investors
Explore our team and our investors.
Careers
Discover open positions. Apply or refer a friend.
Contact
Reach out to us directly.
Legal
View Neural Magic's legal documentation.
menu-icon
Products
DeepSparse Engine
SparseML
SparseZoo
Use Cases
Natural Language Processing (NLP)
Compute Vision
Community
Deep Sparse Slack Community
Code
Docs
Labs
Labs by Neural Magic
Resources
Blogs & News
Research Papers
Upcoming Events
On-Demand Videos
Our Technology
Subscribe
Company
Team & Investors
Careers
Contact
Legal
Home
Home
Optimize for Inference
Deploy on CPUs
Quick Tour
Get Started
Installation
DeepSparse Community
DeepSparse Enterprise
SparseML
SparseZoo
Use a Model
NLP Text Classification
CV Object Detection
Custom Use Case
Transfer a Sparsified Model
NLP Text Classification
CV Object Detection
Sparsify a Model
Supported Integrations
Custom Integrations
Deploy a Model
NLP Text Classification
CV Object Detection
Use Cases
Natural Language Processing
Question Answering
Text Classification
Token Classification
Deploying
Image Classification
Sparsifying
Deploying
Object Detection
Sparsifying
Deploying
Embedding Extraction
User Guides
Sparsification
Recipes
Creating
Enabling Pipelines
ONNX Export
DeepSparse
Supported Hardware
Inference Types
Benchmarking
Diagnostics/Debugging
numactl Utility
DeepSparse Logging
Layer 1
Deploying DeepSparse
DeepSparse Server
Amazon SageMaker
AWS Lambda
Google Cloud Run
Products
DeepSparse
DeepSparse Community
CLI
Python API
C++ API
DeepSparse Enterprise
CLI
Python API
C++ API
SparseML
CLI
Python API
SparseZoo
Models
CLI
Python API
Details
Research Papers
Glossary
FAQs
User Guides
Deploying DeepSparse
Edit on GitHub
Ask on Slack
User Guides For Deploying DeepSparse
The user guides offers more information for deploying DeepSparse.
Guides
DeepSparse Server
Deploy DeepSparse as a model service endpoint.
Amazon SageMaker
Deploy DeepSparse with Amazon SageMaker
AWS Lambda
Deploy DeepSparse with AWS Lambda.
Google Cloud Run
Deploy DeepSparse with GCP Cloud Run.
DeepSparse Logging
Deploying with DeepSparse Server