Skip to content

Commit

Permalink
chore(release): bumped version to 1.5.0-rc.2
Browse files Browse the repository at this point in the history
  • Loading branch information
Bobgy committed Apr 2, 2021
1 parent 29676ce commit f870570
Show file tree
Hide file tree
Showing 72 changed files with 120 additions and 105 deletions.
15 changes: 15 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,20 @@
# Changelog

## [1.5.0-rc.2](https://github.com/kubeflow/pipelines/compare/1.5.0-rc.1...1.5.0-rc.2) (2021-04-02)


### Bug Fixes

* **frontend:** Avoid crashing, and show error message for invalid trigger date/time format. ([\#5418](https://github.com/kubeflow/pipelines/issues/5418)) ([19667e1](https://github.com/kubeflow/pipelines/commit/19667e1086593075bb0db3168b4500b5a89e73fe))
* **manifests:** Multi-User manifests fixes for 1.3 ([\#5416](https://github.com/kubeflow/pipelines/issues/5416)) ([6033718](https://github.com/kubeflow/pipelines/commit/6033718786574cd64b1ebd59e0246e7c5ba1fc88))
* **sdk:** Support returning plain tuples in lightweight function components v2. ([\#5412](https://github.com/kubeflow/pipelines/issues/5412)) ([a6a7255](https://github.com/kubeflow/pipelines/commit/a6a725560c30a769da3c49de739c5e008e761a4f))
* **sdk.v2:** fix bug for component input parameter/artifact referencing ([\#5419](https://github.com/kubeflow/pipelines/issues/5419)) ([eb55842](https://github.com/kubeflow/pipelines/commit/eb558423ecdca969f2dea55e0d1b0197fafbcb13))


### Other Pull Requests

* metadata-writer: Fix multi-user mode ([\#5417](https://github.com/kubeflow/pipelines/issues/5417)) ([684d639](https://github.com/kubeflow/pipelines/commit/684d6392d35f9d847caec373ff13b947243da424))

## [1.5.0-rc.1](https://github.com/kubeflow/pipelines/compare/1.5.0-rc.0...1.5.0-rc.1) (2021-04-01)


Expand Down
2 changes: 1 addition & 1 deletion VERSION
Original file line number Diff line number Diff line change
@@ -1 +1 @@
1.5.0-rc.1
1.5.0-rc.2
4 changes: 2 additions & 2 deletions backend/api/python_http_client/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ This file contains REST API specification for Kubeflow Pipelines. The file is au

This Python package is automatically generated by the [OpenAPI Generator](https://openapi-generator.tech) project:

- API version: 1.5.0-rc.1
- Package version: 1.5.0-rc.1
- API version: 1.5.0-rc.2
- Package version: 1.5.0-rc.2
- Build package: org.openapitools.codegen.languages.PythonClientCodegen
For more information, please visit [https://www.google.com](https://www.google.com)

Expand Down
2 changes: 1 addition & 1 deletion backend/api/python_http_client/kfp_server_api/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@

from __future__ import absolute_import

__version__ = "1.5.0-rc.1"
__version__ = "1.5.0-rc.2"

# import apis into sdk package
from kfp_server_api.api.experiment_service_api import ExperimentServiceApi
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ def __init__(self, configuration=None, header_name=None, header_value=None,
self.default_headers[header_name] = header_value
self.cookie = cookie
# Set default User-Agent.
self.user_agent = 'OpenAPI-Generator/1.5.0-rc.1/python'
self.user_agent = 'OpenAPI-Generator/1.5.0-rc.2/python'
self.client_side_validation = configuration.client_side_validation

def __enter__(self):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -351,8 +351,8 @@ def to_debug_report(self):
return "Python SDK Debug Report:\n"\
"OS: {env}\n"\
"Python Version: {pyversion}\n"\
"Version of the API: 1.5.0-rc.1\n"\
"SDK Package Version: 1.5.0-rc.1".\
"Version of the API: 1.5.0-rc.2\n"\
"SDK Package Version: 1.5.0-rc.2".\
format(env=sys.platform, pyversion=sys.version)

def get_host_settings(self):
Expand Down
2 changes: 1 addition & 1 deletion backend/api/python_http_client/setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@
from setuptools import setup, find_packages # noqa: H301

NAME = "kfp-server-api"
VERSION = "1.5.0-rc.1"
VERSION = "1.5.0-rc.2"
# To install the library, run the following
#
# python setup.py install
Expand Down
2 changes: 1 addition & 1 deletion backend/api/swagger/kfp_api_single_file.swagger.json

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

40 changes: 20 additions & 20 deletions components/README.md

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/to_CSV/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ outputs:
type: CSV
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.bigquery, query,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/to_gcs/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.bigquery, query,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/to_table/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@ Note: The following sample code works in an IPython notebook or directly in Pyth
import kfp.components as comp

bigquery_query_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/bigquery/query/to_table/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/bigquery/query/to_table/component.yaml')
help(bigquery_query_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/to_table/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.bigquery, query,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/container/component_sdk/python/setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
from setuptools import setup

PACKAGE_NAME = 'kfp-component'
VERSION = '1.5.0-rc.1'
VERSION = '1.5.0-rc.2'

setup(
name=PACKAGE_NAME,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_python/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ The steps to use the component in a pipeline are:
```python
import kfp.components as comp

dataflow_python_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataflow/launch_python/component.yaml')
dataflow_python_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataflow/launch_python/component.yaml')
help(dataflow_python_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_python/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataflow, launch_python,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_python/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@
"import kfp.components as comp\n",
"\n",
"dataflow_python_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataflow/launch_python/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataflow/launch_python/component.yaml')\n",
"help(dataflow_python_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_template/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ Follow these steps to use the component in a pipeline:
import kfp.components as comp

dataflow_template_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataflow/launch_template/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataflow/launch_template/component.yaml')
help(dataflow_template_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_template/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataflow, launch_template,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_template/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@
"import kfp.components as comp\n",
"\n",
"dataflow_template_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataflow/launch_template/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataflow/launch_template/component.yaml')\n",
"help(dataflow_template_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/create_cluster/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_create_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/create_cluster/component.yaml')
dataproc_create_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/create_cluster/component.yaml')
help(dataproc_create_cluster_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/create_cluster/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, create_cluster,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/create_cluster/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_create_cluster_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/create_cluster/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/create_cluster/component.yaml')\n",
"help(dataproc_create_cluster_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/delete_cluster/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_delete_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/delete_cluster/component.yaml')
dataproc_delete_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/delete_cluster/component.yaml')
help(dataproc_delete_cluster_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/delete_cluster/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ inputs:
type: Integer
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
kfp_component.google.dataproc, delete_cluster,
--project_id, {inputValue: project_id},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/delete_cluster/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_delete_cluster_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/delete_cluster/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/delete_cluster/component.yaml')\n",
"help(dataproc_delete_cluster_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hadoop_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_hadoop_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_hadoop_job/component.yaml')
dataproc_submit_hadoop_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_hadoop_job/component.yaml')
help(dataproc_submit_hadoop_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hadoop_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_hadoop_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hadoop_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_hadoop_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_hadoop_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_hadoop_job/component.yaml')\n",
"help(dataproc_submit_hadoop_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hive_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -72,7 +72,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_hive_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_hive_job/component.yaml')
dataproc_submit_hive_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_hive_job/component.yaml')
help(dataproc_submit_hive_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hive_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_hive_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hive_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_hive_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_hive_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_hive_job/component.yaml')\n",
"help(dataproc_submit_hive_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pig_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_pig_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_pig_job/component.yaml')
dataproc_submit_pig_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_pig_job/component.yaml')
help(dataproc_submit_pig_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pig_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_pig_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pig_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_pig_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_pig_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_pig_job/component.yaml')\n",
"help(dataproc_submit_pig_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pyspark_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_pyspark_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_pyspark_job/component.yaml')
dataproc_submit_pyspark_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_pyspark_job/component.yaml')
help(dataproc_submit_pyspark_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pyspark_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_pyspark_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pyspark_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_pyspark_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_pyspark_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_pyspark_job/component.yaml')\n",
"help(dataproc_submit_pyspark_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_spark_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -94,7 +94,7 @@ Follow these steps to use the component in a pipeline:
import kfp.components as comp

dataproc_submit_spark_job_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_spark_job/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_spark_job/component.yaml')
help(dataproc_submit_spark_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_spark_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_spark_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_spark_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_spark_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_spark_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_spark_job/component.yaml')\n",
"help(dataproc_submit_spark_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_sparksql_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_sparksql_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_sparksql_job/component.yaml')
dataproc_submit_sparksql_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_sparksql_job/component.yaml')
help(dataproc_submit_sparksql_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_sparksql_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.5.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_sparksql_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_sparksql_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_sparksql_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.1/components/gcp/dataproc/submit_sparksql_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.5.0-rc.2/components/gcp/dataproc/submit_sparksql_job/component.yaml')\n",
"help(dataproc_submit_sparksql_job_op)"
]
},
Expand Down
Loading

0 comments on commit f870570

Please sign in to comment.