pipelines/components
Alexey Volkov 4339e70962 Components - Added AutoML Tables components and tests (#2174)
* Components - Added AutoML Tables components

* Added the sample - AutoML Tables - Retail product stockout prediction

* Replaced the project ID with dummy placeholder

* Fixed the description parameter passing

* Replaced pip with pip3 and changed quotes

* Added licenses

* Updated the component links

* Revert "Replaced pip with pip3"

This reverts commit 65ed0a7fc6. (part of it)

Here, `pip` is not the name of executable. It's the module name which is
just `pip`, not `pip3`.

* Changed quotes to single quotes

* Moved the components to the gcp folder

* Switched container images to python:3.7

* Updated component versions in sample
2019-09-25 13:42:06 -07:00
..
arena Make wget quieter (#2069) 2019-09-09 14:32:54 -07:00
aws Lint Python code for undefined names (#1721) 2019-08-21 15:04:31 -07:00
deprecated Remove dataflow components (#2161) 2019-09-23 11:12:27 -07:00
gcp Components - Added AutoML Tables components and tests (#2174) 2019-09-25 13:42:06 -07:00
ibm-components Update Watson Machine Learning auth with IAM (#1737) 2019-08-07 19:17:18 -07:00
kubeflow Release 57d9f7f1cf (#2184) 2019-09-23 00:19:55 +08:00
local Release 57d9f7f1cf (#2184) 2019-09-23 00:19:55 +08:00
nuclio add nuclio components (to build/deploy, delete, invoke functions) (#1295) 2019-05-08 01:58:33 -07:00
sample/keras/train_classifier Marked all scripts as executable (#1177) 2019-04-23 16:12:00 -07:00
OWNERS clean up owner file (#1928) 2019-08-22 15:29:19 -07:00
README.md move old gcp components to deprecated folder (#2031) 2019-09-06 16:29:20 -07:00
build_image.sh common build image script (#815) 2019-02-13 10:37:19 -08:00
license.sh Initial commit of the kubeflow/pipeline project. 2018-11-02 14:02:31 -07:00
release.sh Remove dataflow components (#2161) 2019-09-23 11:12:27 -07:00
test_load_all_components.sh Test loading all component.yaml definitions (#1045) 2019-04-02 12:25:18 -07:00
third_party_licenses.csv Components - Added the pymongo license URL (#1740) 2019-08-05 19:49:53 -07:00

README.md

Kubeflow pipeline components

Kubeflow pipeline components are implementations of Kubeflow pipeline tasks. Each task takes one or more artifacts as input and may produce one or more artifacts as output.

Example: XGBoost DataProc components

Each task usually includes two parts:

Client code The code that talks to endpoints to submit jobs. For example, code to talk to Google Dataproc API to submit a Spark job.

Runtime code The code that does the actual job and usually runs in the cluster. For example, Spark code that transforms raw data into preprocessed data.

Container A container image that runs the client code.

Note the naming convention for client code and runtime code—for a task named "mytask":

  • The mytask.py program contains the client code.
  • The mytask directory contains all the runtime code.

See how to use the Kubeflow Pipelines SDK and build your own components.