Skip to content

Commit fdbe302

Browse files
committed
move distribution module to private for now
1 parent d2695f7 commit fdbe302

8 files changed

Lines changed: 6 additions & 14 deletions

File tree

docs/source/user_guide.rst

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -149,13 +149,6 @@ Guide Sections
149149
Using the simple ``.search()``
150150
method.
151151

152-
.. grid-item-card:: Distributed Evaluation
153-
:link: user_guide/distributed
154-
:link-type: doc
155-
156-
Parallel evaluation across
157-
multiple workers.
158-
159152
.. grid-item-card:: Evaluation Storage
160153
:link: user_guide/storage
161154
:link-type: doc
@@ -226,7 +219,6 @@ GFO provides 22 optimization algorithms organized into four categories:
226219
user_guide/memory
227220
user_guide/stopping_conditions
228221
user_guide/search_interface
229-
user_guide/distributed
230222
user_guide/storage
231223

232224
.. toctree::

src/gradient_free_optimizers/distributed/__init__.py renamed to src/gradient_free_optimizers/_distributed/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
Use a distribution backend as a decorator on your objective function
44
to enable parallel batch evaluation during optimization::
55
6-
from gradient_free_optimizers.distributed import Multiprocessing
6+
from gradient_free_optimizers._distributed import Multiprocessing
77
88
@Multiprocessing(n_workers=4).distribute
99
def objective(para):
File renamed without changes.

src/gradient_free_optimizers/distributed/_dask.py renamed to src/gradient_free_optimizers/_distributed/_dask.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ class Dask(BaseDistribution):
3737
--------
3838
Local cluster (created automatically)::
3939
40-
from gradient_free_optimizers.distributed import Dask
40+
from gradient_free_optimizers._distributed import Dask
4141
4242
@Dask(n_workers=4).distribute
4343
def objective(para):

src/gradient_free_optimizers/distributed/_joblib.py renamed to src/gradient_free_optimizers/_distributed/_joblib.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ class Joblib(BaseDistribution):
3333
--------
3434
::
3535
36-
from gradient_free_optimizers.distributed import Joblib
36+
from gradient_free_optimizers._distributed import Joblib
3737
3838
@Joblib(n_workers=4).distribute
3939
def objective(para):

src/gradient_free_optimizers/distributed/_multiprocessing.py renamed to src/gradient_free_optimizers/_distributed/_multiprocessing.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ class Multiprocessing(BaseDistribution):
4040
--------
4141
::
4242
43-
from gradient_free_optimizers.distributed import Multiprocessing
43+
from gradient_free_optimizers._distributed import Multiprocessing
4444
4545
@Multiprocessing(n_workers=4).distribute
4646
def objective(para):

src/gradient_free_optimizers/distributed/_ray.py renamed to src/gradient_free_optimizers/_distributed/_ray.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ class Ray(BaseDistribution):
3737
--------
3838
Local usage::
3939
40-
from gradient_free_optimizers.distributed import Ray
40+
from gradient_free_optimizers._distributed import Ray
4141
4242
@Ray(n_workers=4).distribute
4343
def objective(para):

tests/test_main/test_distributed.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@
1414
PowellsMethod,
1515
RandomSearchOptimizer,
1616
)
17-
from gradient_free_optimizers.distributed import (
17+
from gradient_free_optimizers._distributed import (
1818
BaseDistribution,
1919
Joblib,
2020
Multiprocessing,

0 commit comments

Comments
 (0)