From 2515818837a004c9768474781c93562443bdb44b Mon Sep 17 00:00:00 2001 From: Fabio Buso Date: Thu, 5 Dec 2024 14:48:38 +0100 Subject: [PATCH] Fix link to job configuration documentation (#425) (#426) --- python/hsfs/feature_group.py | 6 +++--- python/hsfs/feature_view.py | 14 +++++++------- python/hsfs/training_dataset.py | 4 ++-- 3 files changed, 12 insertions(+), 12 deletions(-) diff --git a/python/hsfs/feature_group.py b/python/hsfs/feature_group.py index 42e3a208e..06ced9ee6 100644 --- a/python/hsfs/feature_group.py +++ b/python/hsfs/feature_group.py @@ -2694,7 +2694,7 @@ def save( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to write data into the feature group. * key `wait_for_job` and value `True` or `False` to configure @@ -2883,7 +2883,7 @@ def insert( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to write data into the feature group. * key `wait_for_job` and value `True` or `False` to configure @@ -3046,7 +3046,7 @@ def multi_part_insert( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to write data into the feature group. * key `wait_for_job` and value `True` or `False` to configure diff --git a/python/hsfs/feature_view.py b/python/hsfs/feature_view.py index 5d3151b18..d7e0161c3 100644 --- a/python/hsfs/feature_view.py +++ b/python/hsfs/feature_view.py @@ -1424,7 +1424,7 @@ def create_training_data( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -1703,7 +1703,7 @@ def create_train_test_split( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -1979,7 +1979,7 @@ def create_train_validation_test_split( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -2103,7 +2103,7 @@ def recreate_training_dataset( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -2222,7 +2222,7 @@ def training_data( * key `"arrow_flight_config"` to pass a dictionary of arrow flight configurations. For example: `{"arrow_flight_config": {"timeout": 900}}`. * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. Defaults to `{}`. spine: Spine dataframe with primary key, event time and @@ -2385,7 +2385,7 @@ def train_test_split( * key `"arrow_flight_config"` to pass a dictionary of arrow flight configurations. For example: `{"arrow_flight_config": {"timeout": 900}}` * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. Defaults to `{}`. spine: Spine dataframe with primary key, event time and @@ -2588,7 +2588,7 @@ def train_validation_test_split( * key `"arrow_flight_config"` to pass a dictionary of arrow flight configurations. For example: `{"arrow_flight_config": {"timeout": 900}}` * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. Defaults to `{}`. spine: Spine dataframe with primary key, event time and diff --git a/python/hsfs/training_dataset.py b/python/hsfs/training_dataset.py index 7d9e89ec8..81dfbaa34 100644 --- a/python/hsfs/training_dataset.py +++ b/python/hsfs/training_dataset.py @@ -624,7 +624,7 @@ def save( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -690,7 +690,7 @@ def insert( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the insert call should return only