You can orchestrate an end-to-end flow of assets from creation through deployment on a graphical canvas with the Pipelines editor. You can assemble and configure a pipeline to create, train, deploy, and update machine learning models and Python
scripts.
Design a pipeline by dragging nodes onto the canvas, specifying objects and parameters, then running and monitoring the pipeline.
Automating the path to production
Copy link to section
Putting a model into a product is a multi-step process. Data must be loaded and processed, models must be trained and evaluated before they are deployed and tested. AI experiments, data analysis and machine learning models all require more observation,
evaluation, and updating over time to avoid bias or drift.
The following graphic shows one such example of a model lifecycle that you can automate out of many possible flows that you can create.
You can automate the pipeline to:
load and process data securely from a wide range of internal sources and external connections.
get the results that you want by building, running, evaluating, and deploying models or run scripts in a cohesive way.
make it simple to run the paths of your flow by creating branches and collect results with direct visuals.
Pipelines can run experiments including but not limited to:
AutoAI experiments
Jupyter Notebook jobs
Data Refinery jobs
SPSS Modeler jobs
To shorten the time from conception to production, you can assemble the pipeline and rapidly update and test modifications. The Pipelines canvas provides tools to visualize the pipeline, customize it at run time with pipeline parameter variables,
and then run it as a trial job or on a schedule.
Use the Pipelines editing tools for more cohesive collaboration between a data scientist and a ModelOps engineer. A data scientist can create and train a model. A ModelOps engineer can then automate the process of training, deploying, and evaluating
the model after it is published to a production environment.
Use cases and tutorials
Copy link to section
Orchestration Pipelines can also be integrated into the IBM watsonx.ai experience side by side with AI solutions that incorporate usage of natural language. For real-world use cases and tutorials for using Orchestration Pipelines to orchestrate
AI solutions, see:
About cookies on this siteOur websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising.For more information, please review your cookie preferences options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.