diff --git a/python/hsfs/feature_group.py b/python/hsfs/feature_group.py index d020d64a1..8562924a8 100644 --- a/python/hsfs/feature_group.py +++ b/python/hsfs/feature_group.py @@ -2703,7 +2703,7 @@ def save( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to write data into the feature group. * key `wait_for_job` and value `True` or `False` to configure @@ -2892,7 +2892,7 @@ def insert( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to write data into the feature group. * key `wait_for_job` and value `True` or `False` to configure @@ -3055,7 +3055,7 @@ def multi_part_insert( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to write data into the feature group. * key `wait_for_job` and value `True` or `False` to configure diff --git a/python/hsfs/feature_view.py b/python/hsfs/feature_view.py index f5d53eda9..faeb57105 100644 --- a/python/hsfs/feature_view.py +++ b/python/hsfs/feature_view.py @@ -1421,7 +1421,7 @@ def create_training_data( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -1686,7 +1686,7 @@ def create_train_test_split( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -1947,7 +1947,7 @@ def create_train_validation_test_split( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -2059,7 +2059,7 @@ def recreate_training_dataset( * key `use_spark` and value `True` to materialize training dataset with Spark instead of [Hopsworks Feature Query Service](https://docs.hopsworks.ai/latest/setup_installation/common/arrow_flight_duckdb/). * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -2178,7 +2178,7 @@ def training_data( * key `"arrow_flight_config"` to pass a dictionary of arrow flight configurations. For example: `{"arrow_flight_config": {"timeout": 900}}`. * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. Defaults to `{}`. spine: Spine dataframe with primary key, event time and @@ -2341,7 +2341,7 @@ def train_test_split( * key `"arrow_flight_config"` to pass a dictionary of arrow flight configurations. For example: `{"arrow_flight_config": {"timeout": 900}}` * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. Defaults to `{}`. spine: Spine dataframe with primary key, event time and @@ -2544,7 +2544,7 @@ def train_validation_test_split( * key `"arrow_flight_config"` to pass a dictionary of arrow flight configurations. For example: `{"arrow_flight_config": {"timeout": 900}}` * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. Defaults to `{}`. spine: Spine dataframe with primary key, event time and diff --git a/python/hsfs/training_dataset.py b/python/hsfs/training_dataset.py index 7d9e89ec8..81dfbaa34 100644 --- a/python/hsfs/training_dataset.py +++ b/python/hsfs/training_dataset.py @@ -624,7 +624,7 @@ def save( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the save call should return only @@ -690,7 +690,7 @@ def insert( When using the `python` engine, write_options can contain the following entries: * key `spark` and value an object of type - [hsfs.core.job_configuration.JobConfiguration](../job_configuration) + [hsfs.core.job_configuration.JobConfiguration](../jobs/#jobconfiguration) to configure the Hopsworks Job used to compute the training dataset. * key `wait_for_job` and value `True` or `False` to configure whether or not to the insert call should return only