You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Deploy Modular, Data-centric AI applications at scale
π‘ About
Seldon Core 2 is an MLOps and LLMOps framework for deploying, managing and scaling AI systems in Kubernetes - from singular models, to modular and data-centric applications. With Core 2 you can deploy in a standardized way across a wide range of model types, on-prem or in any cloud, and production-ready out of the box.
To reach out to Seldon regarding commercial use, visit our website.
π Documentation
The Seldon Core 2 Docs can be found here. For most specific sections, see here:
Pipelines: Deploy composable AI applications, leveraging Kafka for realtime data streaming between components
Autoscaling for models and application components based on native or custom logic
Multi-Model Serving: Save infrastructure costs by consolidating multiple models on shared inference servers
Overcommit: Deploy more models than available memory allows, saving infrastructure costs for unused models
Experiments: Route data between candidate models or pipelines, with support for A/B tests and shadow deployments
Custom Components: Implement custom logic, drift & outlier detection, LLMs and more through plug-and-play integrate with the rest of Seldon's ecosytem of ML/AI products!
π¬ Research
These features are influenced by our position paper on the next generation of ML model serving frameworks:
Seldon is distributed under the terms of the The Business Source License. A complete version of the license is available in the LICENSE file in this repository. Any contribution made to this project will be licensed under the Business Source License.
About
An MLOps framework to package, deploy, monitor and manage thousands of production machine learning models