File tree Expand file tree Collapse file tree
src/gradient_free_optimizers/_distributed Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -149,13 +149,6 @@ Guide Sections
149149 Using the simple ``.search() ``
150150 method.
151151
152- .. grid-item-card :: Distributed Evaluation
153- :link: user_guide/distributed
154- :link-type: doc
155-
156- Parallel evaluation across
157- multiple workers.
158-
159152 .. grid-item-card :: Evaluation Storage
160153 :link: user_guide/storage
161154 :link-type: doc
@@ -226,7 +219,6 @@ GFO provides 22 optimization algorithms organized into four categories:
226219 user_guide/memory
227220 user_guide/stopping_conditions
228221 user_guide/search_interface
229- user_guide/distributed
230222 user_guide/storage
231223
232224.. toctree ::
Original file line number Diff line number Diff line change 33Use a distribution backend as a decorator on your objective function
44to enable parallel batch evaluation during optimization::
55
6- from gradient_free_optimizers.distributed import Multiprocessing
6+ from gradient_free_optimizers._distributed import Multiprocessing
77
88 @Multiprocessing(n_workers=4).distribute
99 def objective(para):
File renamed without changes.
Original file line number Diff line number Diff line change @@ -37,7 +37,7 @@ class Dask(BaseDistribution):
3737 --------
3838 Local cluster (created automatically)::
3939
40- from gradient_free_optimizers.distributed import Dask
40+ from gradient_free_optimizers._distributed import Dask
4141
4242 @Dask(n_workers=4).distribute
4343 def objective(para):
Original file line number Diff line number Diff line change @@ -33,7 +33,7 @@ class Joblib(BaseDistribution):
3333 --------
3434 ::
3535
36- from gradient_free_optimizers.distributed import Joblib
36+ from gradient_free_optimizers._distributed import Joblib
3737
3838 @Joblib(n_workers=4).distribute
3939 def objective(para):
Original file line number Diff line number Diff line change @@ -40,7 +40,7 @@ class Multiprocessing(BaseDistribution):
4040 --------
4141 ::
4242
43- from gradient_free_optimizers.distributed import Multiprocessing
43+ from gradient_free_optimizers._distributed import Multiprocessing
4444
4545 @Multiprocessing(n_workers=4).distribute
4646 def objective(para):
Original file line number Diff line number Diff line change @@ -37,7 +37,7 @@ class Ray(BaseDistribution):
3737 --------
3838 Local usage::
3939
40- from gradient_free_optimizers.distributed import Ray
40+ from gradient_free_optimizers._distributed import Ray
4141
4242 @Ray(n_workers=4).distribute
4343 def objective(para):
Original file line number Diff line number Diff line change 1414 PowellsMethod ,
1515 RandomSearchOptimizer ,
1616)
17- from gradient_free_optimizers .distributed import (
17+ from gradient_free_optimizers ._distributed import (
1818 BaseDistribution ,
1919 Joblib ,
2020 Multiprocessing ,
You can’t perform that action at this time.
0 commit comments