TL;DR
- set of re-usable coarse-grained components (just a bunch of code)
- think of tasks, not functions (e.g., read from a database, transform data, train model, deploy model)
- write once, runs everywhere (export to Kubeflow, Apache Airflow, Apache Nifi)
- just use python - no other skills required (no Kubeflow component YAML, maven, Java)
- 1st class citizen in JupyterLab and the Elyra Pipeline Editor (creating a low code / no code IDE for data science)
- upstream repository to IBM Watson Studio Pipelines contributed components in IBM Cloud Pak for Data
CLAIMED is a component library for artificial intelligence, machine learning, "extract, transform, load" processes, and data science. The goal is to enable low-code/no-code rapid prototyping. The library provides ready-made components for various business domains, supports multiple computer languages, works on different data flow editors, and runs various execution engines. To demonstrate its utility, we constructed a workflow composed exclusively of this library's components. To display the capabilities of this library, we made use of a publicly available Computed Tomography (CT) scan dataset [covidata]. We created a deep learning model, which is supposed to classify exams as either COVID-19 positive or negative. We built the pipeline with Elyra's Pipeline Visual Editor, with support for local, Airflow, and Kubeflow execution https://arxiv.org/abs/2103.03281.
Low Code / No Code pipeline creation tool for data science
Bring the latest and greatest libraries to the hands of everybody.
AIX360/LIME highlights a poor deep learning covid classification model looking at bones only
Components of this library can be exported as:
- Kubeflow pipeline components
- Apache Airflow components
- Standalone graphical components for the Elyra pipeline editor
- Standalone components to be run from the command line
Visually create pipelines from notebooks and run them everywhere
Each notebook is following a similar format.
- The first cell contains a description of the component itself.
- The second cell installs all dependencies using pip3.
- The third cell imports all dependencies.
- The fourth cell contains a list of dependencies, input parameters, and return values as Python comments
- The fifth cell reads the input parameters from environment variables.
Export notebooks and files as runtime components for different engines
To learn more on how this library works in practice, please have a look at the following video
[covidata] Joseph Paul Cohen et al. COVID-19 Image Data Collection: Prospective Predictions Are the Future, arXiv:2006.11988, 2020
We welcome your questions, ideas, and feedback. Please create an issue or a discussion thread.
Interested in helping make the NLP editor better? We encourage you to take a look at our Contributing page.