Platform
Pricing

Deep learning lifecycle made simple,
with outstanding performance.

Choose Your Plan

Basic

For developers looking to accelerate inference and deploy to production in no time.

TALK TO SALES
Features
  • Starting from 5 users
  • Unlimited inference optimization (8/16-Bit post-training quantization, compilation)
  • Model repo & management hub
  • Performance benchmarking on various hardware (even edge devices)
  • Deployment tools including a Python inference runtime engine
  • API access
Support
  • Support for NVIDIA, Intel edge and cloud chips
  • Support Team / under SLA

Professional

For deep learning teams looking to achieve better than SOTA accuracy & inference performance.

TALK TO SALES
Features
  • Starting from 5 users
  • Everything included in Basic
  • AutoNAC engine (starts at 2 runs per year)
  • 8-Bit quantization aware training
Support
  • Dedicated DL expert / under SLA

Enterprise

For deep learning experts looking to meet specific performance goals for highly customized use cases.

TALK TO SALES
Features
  • 10+ users
  • Everything included in Professional
  • Any target hardware
  • Any deep learning framework
  • On-premise optimizations
  • Custom legal terms
Support
  • Dedicated DL expert / under SLA
Details
 

Basic

Professional

Enterprise

Runtime Layer Optimization

Hardware Benchmarking
?Easily compare inference performance across multiple models, hardware types and batch sizes.
Compilation & Post Training Quantization (FP16 / INT8)
Inference Engine

Algorithmic Layer Optimization

AutoNAC Engine Runs
?Achieve unparalleled accuracy and inference speed by building custom, hardware aware model architectures with Deci's proprietary AutoNAC engine.
-
Quantization Aware Training
-

Hardware & Frameworks

Supported Target Hardware
A wide range of hardware including edge devices
A wide range of hardware including edge devices
Any Hardware
Supported Frameworks
Standard
Standard
Any Framework

Support & Legal

Support
Support Team / 24 hours
Dedicated DL expert / 12 hours
Dedicated DL expert / custom SLA
Legal Terms
Standard
Standard
Custom

Basic

For developers looking to boost performance and shorten development time.

View Details

Professional

For deep learning teams looking to quickly achieve outstanding accuracy & inference performance.

View Details

Enterprise

For deep learning experts looking to meet specific performance goals for highly customized use cases.

View Details

Q&A

  • Is it an annual subscription?

    Yes, the Basic, Professional, and Enterprise Plans are all annual subscriptions.

  • Can I start with the basic plan and upgrade to professional?

    Absolutely. The basic plan allows you to quickly start optimizing your deep learning models. You can upgrade your plan at any time by contacting us.

  • What is AutoNAC inference optimization?

    AutoNAC, short for Automated Neural Architecture Construction, is Deci’s proprietary optimization technology. It is a Neural Architecture Search (NAS) algorithm that provides you with end-to-end accuracy-preserving hardware-aware inference acceleration. AutoNAC considers and leverages all components in the inference stack, including compilers, pruning, and quantization.

  • What is Deci's inference runtime engines?

    Deci's deep learning development platform enables you to push models from the Deci Lab, optimized or not, directly to our inference engine for seamless deployment and ultra-fast serving. Quickly run inference on your models with Infery, a Python inference runtime engine that simplifies deep learning model inference across multiple frameworks and hardware using only 3 lines of code.

  • Will Deci work with my current tools and CI/CD?

    Absolutely. You can easily integrate Deci’s deep learning acceleration platform using our API access. Read more about our API access.

AutoNAC White Paper

Explore the Deep Learning Development Platform

With Deci’s deep learning development platform, developers can accelerate inference performance by up to 5x, for any hardware, without compromising accuracy; cut up to 80% on compute costs; and reduce time-to-production of models.

TALK TO AN EXPERT

"By collaborating with Deci, we aim to help our customers accelerate AI innovation and deploy AI solutions everywhere using our industry-leading platforms, from data centers to edge systems that accelerate high-throughput inference."

Arti Garg

Head of Advanced AI Solutions & Technologies, HPE

“At RingCentral, we strive to provide our customers with the best AI-based experiences. With Deci’s platform, we were able to exceed our deep learning performance goals while shortening our development cycles. Working with Deci allows us to launch superior products faster.”

Vadim Zhuk

Senior Vice President R&D, RingCentral

“Deci exceeded our expectations in achieving aggressive targets for reducing latency and memory usage on TPU edge hardware in a test environment, all while preserving BriefCam’s high standard for accuracy.”

Yogev Wallach

Sr. Product Manager, BriefCam

“Deci delivers optimized deep learning inference on Intel processors as highlighted in MLPerf, allowing our customers to meet performance SLAs, reduce cost, decrease time to deployment, and gives them the ability to effectively scale.”

Monica Livingston

AI Solutions and Sales Director, Intel

"With Deci we were able to get 6.4x higher throughput on our detection model. Its easy-to-use platform allowed us to quickly optimize and benchmark for choosing our best configuration of hardware and parameters."

Santiago Tamagno

CTO, UNX Digital by Grupo Prominente

"The classification model I uploaded and integrated using Infery achieved a 33% performance boost, which is very cool for 10 minutes of work!"

Amir Zait

Algorithm Developer, Sight Diagnostics

"We are excited to be working with Deci's platform - it provided amazing results and achieved 4.6x acceleration on a model we ran in production and helped us provide faster service to our customers.”

Daniel Shichman

CEO, WSC Sports Technologies

“With Deci, we increased by 2.6x the inference throughput of one of our multiclass classification models running on V100 machines - without losing accuracy. Deci can cut 50% off the inference costs in our cloud deployments worldwide.”

Chaim Linhart

CTO and Co-Founder, Ibex Medical Analytics

Selected as one of the top 100 AI startups in the world

Recognized as a tech innovator for edge AI

Achieved 11.8x acceleration in collaboration with Intel

Named a leader among AI optimization solutions

Discover What Makes Deci Unique

How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

Blog Post

READ MORE
How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

An Introduction to the Inference Stack and Acceleration Techniques

Blog Post

READ MORE
How Deci and Intel Hit 11.8x Inference Acceleration at MLPerf

Accelerate Deep Neural Network Inference with AutoNAC

Blog Post

DOWNLOAD NOW

Unleash Your
Deep Learning Models