Model Serving

Share

Model serving refers to the way trained models are made available for others to use. Choosing a model serving strategy can be the first step in model deployment, where factors such as user expectations, production requirements, business rules, and existing technologies are considered. There are four common model serving tactics, namely, batch inference, model as a service, online model as a service, and edge deployment.

Filter terms by

Glossary Alphabetical filter

Related resources

TensorRT-framework-overview
Deployment
deci-pytorch-coreml-blog
Deployment
deci-infery-updates-blog-featured
Deployment
Share
Add Your Heading Text Here
				
					from transformers import AutoFeatureExtractor, AutoModelForImageClassification

extractor = AutoFeatureExtractor.from_pretrained("microsoft/resnet-50")

model = AutoModelForImageClassification.from_pretrained("microsoft/resnet-50")