• 设为首页
  • 点击收藏
  • 手机版
    手机扫一扫访问
    迪恩网络手机版
  • 关注官方公众号
    微信扫一扫关注
    公众号

mlrun/mlrun: Machine Learning automation and tracking

原作者: [db:作者] 来自: 网络 收藏 邀请

开源软件名称(OpenSource Name):

mlrun/mlrun

开源软件地址(OpenSource Url):

https://github.com/mlrun/mlrun

开源编程语言(OpenSource Language):

Python 99.2%

开源软件介绍(OpenSource Introduction):

Build Status License PyPI version fury.io Documentation Code style: black GitHub commit activity GitHub release (latest SemVer) Join MLOps Live

MLRun logo

MLRun - The Open Source MLOps Orchestration Framework

MLRun enables production pipeline design using a modular strategy, where the different parts contribute to a continuous, automated, and far simpler path from research and development to scalable production pipelines, without refactoring code, adding glue logic, or spending significant efforts on data and ML engineering.

MLRun uses Serverless Function technology: write the code once, using your preferred development environment and simple “local” semantics, and then run it as-is on different platforms and at scale. MLRun automates the build process, execution, data movement, scaling, versioning, parameterization, outputs tracking, CI/CD integration, deployment to production, monitoring, and more.

Those easily developed data or ML “functions” can then be published or loaded from a marketplace and used later to form offline or real-time production pipelines with minimal engineering efforts.

mlrun-flow


Data preparation, model development, model and application delivery, and end to end monitoring are tightly connected: they cannot be managed in silos. This is where MLRun MLOps orchestration comes in. ML, data, and DevOps/MLOps teams collaborate using the same set of tools, practices, APIs, metadata, and version control.

MLRun simplifies & accelerates the time to production.

Architecture

pipeline

MLRun is composed of the following layers:

  • Feature Store — collects, prepares, catalogs, and serves data features for development (offline) and real-time (online) usage for real-time and batch data. See also Feature store: data ingestion and Feature store: data retrieval, as well as the Feature Store tutorials.
  • ML CI/CD pipeline — automatically trains, tests, optimizes, and deploys or updates models using a snapshot of the production data (generated by the feature store) and code from the source control (Git).
  • Real-Time Serving Pipeline — Rapid deployment of scalable data and ML pipelines using real-time serverless technology, including the API handling, data preparation/enrichment, model serving, ensembles, driving and measuring actions, etc.
  • Real-Time monitoring and retraining — monitors data, models, and production components and provides a feedback loop for exploring production data, identifying drift, alerting on anomalies or data quality issues, triggering re-training jobs, measuring business impact, etc.

Get started

It's easy to start using MLRun:

  1. Install the MLRun service locally using Docker or over Kubernetes Cluster. Alternatively, you can use Iguazio's managed MLRun service
  2. Set up your client environment to work with the service.
  3. Follow the Quick Start tutorial and Additional Tutorials and Examples to learn how to use MLRun to develop and deploy machine learning applications to production.

For hands-on learning, try the MLRun Katakoda Scenarios. And you can watch the Tutorial on Youtube to see the flow in action.

MLRun documentation

Read more in the MLRun documentation, including:




鲜花

握手

雷人

路过

鸡蛋
该文章已有0人参与评论

请发表评论

全部评论

专题导读
热门推荐
阅读排行榜

扫描微信二维码

查看手机版网站

随时了解更新最新资讯

139-2527-9053

在线客服(服务时间 9:00~18:00)

在线QQ客服
地址:深圳市南山区西丽大学城创智工业园
电邮:jeky_zhao#qq.com
移动电话:139-2527-9053

Powered by 互联科技 X3.4© 2001-2213 极客世界.|Sitemap