You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: CHANGELOG.md
+18-1
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,21 @@
2
2
3
3
## [Unreleased]
4
4
5
+
## [0.12.0] - 2025-03-11
6
+
7
+
- Support for python 3.11 & 3.12 added, dropped support for python 3.8
8
+
- Brought back the Vertex AI Pipelines scheduling capability
9
+
- Migrated to kfp 2 (locked to <2.9.0 due to pipeline spec changes)
10
+
- Removed `image_pull_policy` parameter from configuration, as it only applies to Kubernetes backend and not Vertex AI,
11
+
and it's only available in `kfp-kubernetes` extension package
12
+
- Removed `--timeout-seconds` parameter from `run-once` command for now, as in the old version of the plugin exceeding the specified time
13
+
didn't alter the remote pipeline execution, and only escaped the local Python processs. The timeout funcionality will be added later on,
14
+
with the proper remote pipeline execution handling, and possibly per-task timeout enabled by [the new kfp feature](https://github.com/kubeflow/pipelines/pull/10481).
15
+
- Assign pipelines to Vertex AI experiments
16
+
- Migrated `pydantic` library to v2
17
+
- Custom dataset that creates Vertex AI artifact
18
+
- Added pipeline parametrization
19
+
5
20
## [0.11.1] - 2024-07-01
6
21
7
22
## [0.11.0] - 2024-03-22
@@ -83,7 +98,9 @@
83
98
84
99
- Initial version of **kedro-vertexai** plugin extracted from [kedro-kubeflow v0.6.0](https://github.com/getindata/kedro-kubeflow/tree/0.6.0)
# Pull policy to be used for the steps. Use Always if you push the images
14
-
# on the same tag, or Never if you use only local images
15
-
image_pull_policy: IfNotPresent
16
-
17
13
# Location of Vertex AI GCS root
18
14
root: bucket_name/gcs_suffix
19
15
20
-
# Name of the kubeflow experiment to be created
16
+
# Name of the Vertex AI experiment to be created
21
17
experiment_name: MyExperiment
22
18
19
+
# Optional description of the Vertex AI experiment to be created
20
+
# experiment_description: "My experiment description."
21
+
23
22
# Name of the scheduled run, templated with the schedule parameters
24
23
scheduled_run_name: MyExperimentRun
25
24
@@ -98,6 +97,23 @@ run_config:
98
97
# client_id: iam-client-id
99
98
100
99
dynamic_config_providers: []
100
+
101
+
# Schedules configuration
102
+
schedules:
103
+
default_schedule:
104
+
cron_expression: "0 * * * *"
105
+
timezone: Etc/UTC
106
+
# Optional. Timestamp after which the first run can be scheduled. If unspecified, it defaults to the schedule creation timestamp.
107
+
start_time: null
108
+
# Optional. Timestamp after which no more runs will be scheduled. If unspecified, then runs will be scheduled indefinitely.
109
+
end_time: null
110
+
# Optional. Whether new scheduled runs can be queued when max_concurrent_runs limit is reached.
111
+
allow_queueing: false
112
+
# Optional. Maximum run count of the schedule. If specified, The schedule will be completed when either started_run_count >= max_run_count or when end_time is reached. Must be positive and <= 2^63-1.
113
+
max_run_count: null
114
+
# Optional. Maximum number of runs that can be started concurrently for this PipelineJobSchedule.
Copy file name to clipboardexpand all lines: docs/source/03_getting_started/01_quickstart.md
+6-3
Original file line number
Diff line number
Diff line change
@@ -108,17 +108,14 @@ Adjusted `catalog.yml` should look like this (note: remove the rest of the entri
108
108
companies:
109
109
type: pandas.CSVDataSet
110
110
filepath: data/01_raw/companies.csv
111
-
layer: raw
112
111
113
112
reviews:
114
113
type: pandas.CSVDataSet
115
114
filepath: data/01_raw/reviews.csv
116
-
layer: raw
117
115
118
116
shuttles:
119
117
type: pandas.ExcelDataSet
120
118
filepath: data/01_raw/shuttles.xlsx
121
-
layer: raw
122
119
```
123
120
124
121
All intermediate and output data will be stored in the location with the following pattern:
@@ -180,3 +177,9 @@ As you can see, the pipeline was compiled and started in Vertex AI Pipelines. Wh
180
177

181
178
182
179
180
+
## Log datasets to Vertex AI Metadata
181
+
182
+
The plugin implements custom `kedro_vertexai.vertex_ai.datasets.KedroVertexAIMetadataDataset` dataset that creates an Vertex AI Artifact.
183
+
It allows to specify any Kedro dataset in the `base_dataset` argument, and it uses its `_save` and `_load` methods for the io.
184
+
The base dataset arguments are passed in `base_dataset_args` argument as dictionary. The created artifact is associated with Vertex AI run id and job name as metadata, and additional metadata can be specified in the `metadata` argument.
185
+
The `dispaly_name` and `schema` arguments are used for the artifact creation, please reference [Vertex AI docs](https://cloud.google.com/vertex-ai/docs/ml-metadata/tracking#create-artifact) to learn more about them.
0 commit comments