From 3f87a453a912e17426e834c84415ad9e325b0c67 Mon Sep 17 00:00:00 2001 From: Eric Prestat Date: Sun, 17 Dec 2023 17:16:19 +0000 Subject: [PATCH 1/2] Make warning of "beta" state of lazy signals more specific --- doc/user_guide/big_data.rst | 21 +++++++++++---------- 1 file changed, 11 insertions(+), 10 deletions(-) diff --git a/doc/user_guide/big_data.rst b/doc/user_guide/big_data.rst index 17bbdf6712..a03feeb435 100644 --- a/doc/user_guide/big_data.rst +++ b/doc/user_guide/big_data.rst @@ -3,16 +3,6 @@ Working with big data ********************* -.. warning:: All the features described in this chapter are in beta state. - - Although most of them work as described, their operation may not always - be optimal, well-documented and/or consistent with their in-memory counterparts. - - Therefore, although efforts will be taken to minimise major disruptions, - the syntax and features described here may change in patch and minor - HyperSpy releases. If you experience issues with HyperSpy's lazy features - please report them to the developers. - .. versionadded:: 1.2 HyperSpy makes it possible to analyse data larger than the available memory by @@ -123,6 +113,11 @@ operations are only performed lazily, use the Machine learning ---------------- +.. warning:: The machine learning features are in beta state. + + Although most of them work as described, their operation may not always + be optimal, well-documented and/or consistent with their in-memory counterparts. + :ref:`mva.decomposition` algorithms for machine learning often perform large matrix manipulations, requiring significantly more memory than the data size. To perform decomposition operation lazily, HyperSpy provides access to several "online" @@ -498,6 +493,12 @@ Dask has two schedulers available for single machines. Distributed Scheduler ^^^^^^^^^^^^^^^^^^^^^ +.. warning:: Distributed computing is not supported for all file formats. + + Distributed computing is limited to a few file formats, see the list of + :external+rsciio:ref:`supported file format ` in + RosettaSciIO documentation. + The recommended way to use dask is with the distributed scheduler. This allows you to scale your computations to a cluster of machines. The distributed scheduler can be used on a single machine as well. ``dask-distributed`` also gives you access to the dask dashboard which allows you to monitor your computations. From e84820085b8d5cdd49e175b8a96cd1e8f6d9d9c8 Mon Sep 17 00:00:00 2001 From: Eric Prestat Date: Mon, 18 Dec 2023 22:53:49 +0000 Subject: [PATCH 2/2] Add changelog entry --- upcoming_changes/3282.doc.rst | 1 + 1 file changed, 1 insertion(+) create mode 100644 upcoming_changes/3282.doc.rst diff --git a/upcoming_changes/3282.doc.rst b/upcoming_changes/3282.doc.rst new file mode 100644 index 0000000000..060eebdc23 --- /dev/null +++ b/upcoming_changes/3282.doc.rst @@ -0,0 +1 @@ +Update warning of "beta" state in big data section to be more specific. \ No newline at end of file