You can orchestrate an end-to-end flow of assets from creation through deployment on a graphical canvas with the Pipelines editor. You can assemble and configure a pipeline to create, train, deploy, and update machine learning models and Python
scripts.
Design a pipeline by dragging nodes onto the canvas, specifying objects and parameters, then running and monitoring the pipeline.
Automating the path to production
Copy link to section
Putting a model into a product is a multi-step process. Data must be loaded and processed, models must be trained and evaluated before they are deployed and tested. AI experiments, data analysis and machine learning models all require more observation,
evaluation, and updating over time to avoid bias or drift.
The following graphic shows one such example of a model lifecycle that you can automate out of many possible flows that you can create.
You can automate the pipeline to:
load and process data securely from a wide range of internal sources and external connections.
get the results that you want by building, running, evaluating, and deploying models or run scripts in a cohesive way.
make it simple to run the paths of your flow by creating branches and collect results with direct visuals.
Pipelines can run experiments including but not limited to:
AutoAI experiments
Jupyter Notebook jobs
Data Refinery jobs
SPSS Modeler jobs
To shorten the time from conception to production, you can assemble the pipeline and rapidly update and test modifications. The Pipelines canvas provides tools to visualize the pipeline, customize it at run time with pipeline parameter variables,
and then run it as a trial job or on a schedule.
Use the Pipelines editing tools for more cohesive collaboration between a data scientist and a ModelOps engineer. A data scientist can create and train a model. A ModelOps engineer can then automate the process of training, deploying, and evaluating
the model after it is published to a production environment.
Use cases and tutorials
Copy link to section
You can include IBM Orchestration Pipelines in your data fabric solution to manage and automate your data and AI lifecycle. For more information on how Data Fabric can support your machine learning goals and operations in practical ways, see
Use cases.
Data science and MLOps use case describes how to manage data, build a model building and deployment pipeline, and evaluate model fairness and performance.
About cookies on this siteOur websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising.For more information, please review your cookie preferences options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.