Nneji123/Serving-Machine-Learning-Models
This repository contains instructions, template source code and examples on how to serve/deploy machine learning models using various frameworks and applications such as Docker, Flask, FastAPI, BentoML, Streamlit, MLflow and even code on how to deploy your machine learning model as an android app.
No commits in the last 6 months.
Stars
54
Forks
8
Language
CSS
License
MIT
Category
Last pushed
Feb 15, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/mlops/Nneji123/Serving-Machine-Learning-Models"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
combust/mleap
MLeap: Deploy ML Pipelines to Production
ml-tooling/opyrator
🪄 Turns your machine learning code into microservices with web API, interactive GUI, and more.
ebhy/budgetml
Deploy a ML inference service on a budget in less than 10 lines of code.
jpmorganchase/inference-server
Deploy your AI/ML model to Amazon SageMaker for Real-Time Inference and Batch Transform using...
SocAIty/APIPod
Create web-APIs for long-running tasks. Job based task handling. Get the result with the job id...