Deep learning lifecycle made simple,
with outstanding performance.

Choose Your Plan


For developers looking to boost performance and shorten development time.

  • Model repo & management hub
  • Performance benchmarking
  • Up to 5 runtime optimizations
  • Cloud cost-to-serve calculator
  • Deployment tools
  • Model trainer (Coming soon)
  • API access
  • Community support


For deep learning teams looking to quickly achieve outstanding accuracy & inference performance.

  • Everything included in Community
  • DeciNets models catalog
  • Unlimited runtime optimizations
  • 8-Bit quantization support
  • Extended hardware support
  • Additional frameworks included
  • Standard support


For deep learning experts looking to meet specific performance goals for highly customized use cases.

  • Everything included in Professional
  • Custom NAS-based model optimization
  • Any target hardware
  • Any deep learning framework
  • On-premise optimizations
  • Custom legal terms
  • Premium support





Model Hub
?Open source SoTA pre-trained models.
DeciNets Catalog
?Achieve the best accuracy-latency trade-off with pre-trained models generated for a wide range of target inference hardware and tasks (powered by AutoNAC).
Performance Benchmarking
?Easily compare inference performance across multiple models, hardware types and batch sizes.


Runtime Optimization
?Runtime optimizations based on automated compilation & quantization tools.
5 Optimizations
NAS Based Model Performance Optimization
?Leverage Deci's proprietary AutoNAC technology to optimize your deep learning models' performance and achieve unparalleled accuracy and inference speed on any hardware.
On-Premise Optimizations
?The Deci platform is hosted on Deci’s servers. An on-premise deployment is available for our enterprise solution.

Hardware & Frameworks

Supported Target Hardware
Common Cloud Machines
A Wide Range Of Hardware
Any Hardware
Supported Frameworks
ONNX, TensorFlow, Keras
Common Frameworks
Any Framework


Infery Runtime Engine
?A Python packaged inference engine
RTiC (Runtime inference Container)
?Scale up your inference workloads & maximize the utilization of your hardware with Deci’s runtime inference container.
API Access

Support & Legal

Community Support
Standard Support
Premium Support
Legal Terms


For developers looking to boost performance and shorten development time.

View Details


For deep learning teams looking to quickly achieve outstanding accuracy & inference performance.

View Details


For deep learning experts looking to meet specific performance goals for highly customized use cases.

View Details


  • Can I start with the community plan and upgrade to professional?

    Absolutely. The Community plan allows you to quickly start with your own deep learning models or start with a pre-loaded model from our ModelHub. You can upgrade your plan at any time by contacting us.

  • Is there a trial period for the community plan?

    No. It’s free forever. Knock yourself out.

  • Do I need a credit card?

    No. Credit card is not needed.

  • Is there a limit to the number of models I can upload to the platform?

    No limit, go for it.

  • Is there a limit to the number of model optimizations I can carry out with the platform?

    There is no limit to the numbers of optimizations you can carry out. Keep in mind that if you are in the Community plan, you are limited to using any 5 of those optimized models.

  • What is AutoNAC inference optimization?

    AutoNAC, short for Automated Neural Architecture Construction, is Deci’s proprietary optimization technology. It is a Neural Architecture Search (NAS) algorithm that provides you with end-to-end accuracy-preserving hardware-aware inference acceleration. AutoNAC considers and leverages all components in the inference stack, including compilers, pruning, and quantization. Learn more about our technology. Learn more about the kind of results you can get.

  • What are Deci's inference runtime engines?

    Deci's deep learning development platform enables you to push models from the Deci Lab, optimized or not, directly to one of our inference engines for seamless deployment and ultra-fast serving. Deci offers 2 types of runtime engines: 1. Quickly run inference on your models with Infery, a Python inference runtime engine that simplifies deep learning model inference across multiple frameworks and hardware using only 3 lines of code. 2. Scale your workloads with RTiC (Runtime Inference Container), a high-performance container-based inference server. RTiC fits best with microservice environments and has built-in performance optimizations.

  • Is there a limit to the usage of optimized models?

    You can upload and optimize as many deep learning models as you want. The Community plan limits you to using any 5 of those optimized models for inference. Once you see how the different optimizations add value to the original model, you can choose which of the optimized models you want to deploy. The Professional plan includes unlimited usage of optimized models. Get in touch to get a quote.

  • Can I invite a co-worker to collaborate with me on the platform?

    Absolutely. You can invite an unlimited number of co-workers using the ‘Invite’ button within the platform. All of you will be able to collaborate on the same workspace.

  • What data is leaving my perimeter and goes to Deci?

    Your data’s confidentiality and privacy are our highest priority. We comply with strict information security practices, and are ISO270001 and ISO 27799 compliant. When you optimize your own model, it is uploaded to our server. We never use your model for any purpose other than its optimization on the platform. For the Community and Professional plans, you do not need to upload any data aside from the model itself (e.g., dataset). Alternatively, you can use a pre-loaded model from our ModelHub. For the Enterprise plan, your data will serve as another input to our AutoNAC optimization engine to ensure that you get the full value Deci has to offer. The process can be done entirely on your premises, without the data ever leaving your site. To learn more, please refer to our Privacy Policy.

  • Will Deci work with my current tools and CI/CD?

    Absolutely. You can easily integrate Deci’s deep learning acceleration platform using our API access. Read more about our API access.

AutoNAC White Paper

Explore the Deep Learning Development Platform

With Deci’s deep learning development platform, developers can accelerate inference performance by up to 15x, for any hardware, without compromising accuracy; cut up to 80% on compute costs; and reduce time-to-production of models.

Request Free Trial

"By collaborating with Deci, we aim to help our customers accelerate AI innovation and deploy AI solutions everywhere using our industry-leading platforms, from data centers to edge systems that accelerate high-throughput inference."

Arti Garg

Head of Advanced AI Solutions & Technologies, HPE

“At RingCentral, we strive to provide our customers with the best AI-based experiences. With Deci’s platform, we were able to exceed our deep learning performance goals while shortening our development cycles. Working with Deci allows us to launch superior products faster.”

Vadim Zhuk

Senior Vice President R&D, RingCentral

“Deci exceeded our expectations in achieving aggressive targets for reducing latency and memory usage on TPU edge hardware in a test environment, all while preserving BriefCam’s high standard for accuracy.”

Yogev Wallach

Sr. Product Manager, BriefCam

“Deci delivers optimized deep learning inference on Intel processors as highlighted in MLPerf, allowing our customers to meet performance SLAs, reduce cost, decrease time to deployment, and gives them the ability to effectively scale.”

Monica Livingston

AI Solutions and Sales Director, Intel

"With Deci we were able to get 6.4x higher throughput on our detection model. Its easy-to-use platform allowed us to quickly optimize and benchmark for choosing our best configuration of hardware and parameters."

Santiago Tamagno

CTO, UNX Digital by Grupo Prominente

"The classification model I uploaded and integrated using Infery achieved a 33% performance boost, which is very cool for 10 minutes of work!"

Amir Zait

Algorithm Developer, Sight Diagnostics

"We are excited to be working with Deci's platform - it provided amazing results and achieved 4.6x acceleration on a model we ran in production and helped us provide faster service to our customers.”

Daniel Shichman

CEO, WSC Sports Technologies

“With Deci, we increased by 2.6x the inference throughput of one of our multiclass classification models running on V100 machines - without losing accuracy. Deci can cut 50% off the inference costs in our cloud deployments worldwide.”

Chaim Linhart

CTO and Co-Founder, Ibex Medical Analytics

Selected as one of the top 100 AI startups in the world

Recognized as a tech innovator for edge AI

Achieved 11.8x acceleration in collaboration with Intel

Named a leader among AI optimization solutions

Discover What Makes Deci Unique

How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

Blog Post

How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

An Introduction to the Inference Stack and Acceleration Techniques

Blog Post

How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

Accelerate Deep Neural Network Inference with AutoNAC

Blog Post


Unleash Your
Deep Learning Models