We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 76ca370 commit 1f53590Copy full SHA for 1f53590
stubs/networkx/algorithms/approximation/dominating_set.pyi
@@ -1,10 +1,6 @@
1
from ...classes.graph import Graph
2
-from ...utils import not_implemented_for
3
-from ..matching import maximal_matching
4
5
__all__ = ["min_weighted_dominating_set", "min_edge_dominating_set"]
6
7
-# TODO Why doesn't this algorithm work for directed graphs?
8
-
9
def min_weighted_dominating_set(G: Graph, weight: str | None = None) -> set: ...
10
def min_edge_dominating_set(G: Graph) -> set: ...
stubs/networkx/algorithms/clique.pyi
-from collections import defaultdict, deque
-from itertools import chain, combinations, islice
from numpy.typing import ArrayLike
from ..classes.graph import Graph
-from ..utils import not_implemented_for
__all__ = [
"find_cliques",
@@ -22,8 +18,6 @@ __all__ = [
22
18
23
19
def enumerate_all_cliques(G: Graph): ...
24
20
def find_cliques(G: Graph, nodes=None): ...
25
26
-# TODO Should this also be not implemented for directed graphs?
27
21
def find_cliques_recursive(G: Graph, nodes=None): ...
28
def make_max_clique_graph(G: Graph, create_using=None): ...
29
def make_clique_bipartite(G: Graph, fpos: bool | None = None, create_using=None, name=None): ...
stubs/networkx/algorithms/cuts.pyi
@@ -1,4 +1,3 @@
-from itertools import chain
from typing import Any
@@ -14,19 +13,11 @@ __all__ = [
14
13
"volume",
15
]
16
17
-# TODO STILL NEED TO UPDATE ALL THE DOCUMENTATION!
def cut_size(G: Graph, S, T=None, weight: Any = None): ...
def volume(G: Graph, S, weight: Any = None): ...
def normalized_cut_size(G: Graph, S, T=None, weight: Any = None): ...
def conductance(G: Graph, S, T=None, weight: Any = None): ...
def edge_expansion(G: Graph, S, T=None, weight: Any = None): ...
def mixing_expansion(G: Graph, S, T=None, weight: Any = None): ...
-# TODO What is the generalization to two arguments, S and T? Does the
-# denominator become `min(len(S), len(T))`?
def node_expansion(G: Graph, S): ...
30
31
32
def boundary_expansion(G: Graph, S): ...
stubs/networkx/algorithms/distance_regular.pyi
@@ -3,8 +3,6 @@ from collections.abc import Iterable
-from .distance_measures import diameter
"is_distance_regular",
@@ -16,7 +14,4 @@ __all__ = [
def is_distance_regular(G: Graph) -> bool: ...
def global_parameters(b: ArrayLike, c: ArrayLike) -> Iterable: ...
def intersection_array(G: Graph): ...
-# TODO There is a definition for directed strongly regular graphs.
def is_strongly_regular(G: Graph) -> bool: ...
stubs/networkx/algorithms/tree/branchings.pyi
@@ -1,27 +1,3 @@
-# TODO: Implement method from Gabow, Galil, Spence and Tarjan:
-#
-# @article{
-# year={1986},
-# issn={0209-9683},
-# journal={Combinatorica},
-# volume={6},
-# number={2},
-# doi={10.1007/BF02579168},
-# title={Efficient algorithms for finding minimum spanning trees in
11
-# undirected and directed graphs},
12
-# url={https://doi.org/10.1007/BF02579168},
-# publisher={Springer-Verlag},
-# keywords={68 B 15; 68 C 05},
-# author={Gabow, Harold N. and Galil, Zvi and Spencer, Thomas and Tarjan,
-# Robert E.},
-# pages={109-122},
-# language={English}
-# }
-import string
-from dataclasses import dataclass, field
-from enum import Enum
-from operator import itemgetter
-from queue import PriorityQueue
from typing import Literal
from ...classes.digraph import DiGraph
stubs/networkx/conftest.pyi
@@ -1,14 +1,8 @@
-import sys
-import warnings
-import networkx
import pytest
def pytest_addoption(parser): ...
def pytest_configure(config): ...
def pytest_collection_modifyitems(config, items): ...
-# TODO: The warnings below need to be dealt with, but for now we silence them.
@pytest.fixture(autouse=True)
def set_warnings(): ...
stubs/sklearn/cluster/_agglomerative.pyi
@@ -49,8 +49,7 @@ class AgglomerativeClustering(ClusterMixin, BaseEstimator):
49
self,
50
n_clusters: None | int = 2,
51
*,
52
- affinity: str | Callable = "deprecated", # TODO(1.4): Remove
53
- metric: None | str | Callable = None, # TODO(1.4): Set to "euclidean"
+ metric: str | Callable = "euclidean",
54
memory: None | Memory | str = None,
55
connectivity: None | ArrayLike | Callable = None,
56
compute_full_tree: Literal["auto"] | bool = "auto",
@@ -77,8 +76,7 @@ class FeatureAgglomeration(ClassNamePrefixFeaturesOutMixin, AgglomerativeCluster
77
76
78
79
80
81
82
83
84
stubs/sklearn/covariance/_robust_covariance.pyi
@@ -7,17 +7,6 @@ from numpy.random import RandomState
from .._typing import Float, Int, MatrixLike
from . import EmpiricalCovariance
-# Author: Virgile Fritsch <virgile.fritsch@inria.fr>
-# License: BSD 3 clause
-# Minimum Covariance Determinant
-# Implementing of an algorithm by Rousseeuw & Van Driessen described in
-# (A Fast Algorithm for the Minimum Covariance Determinant Estimator,
-# 1999, American Statistical Association and the American Society
-# for Quality, TECHNOMETRICS)
-# XXX Is this really a public function? It's not listed in the docs or
-# exported by sklearn.covariance. Deprecate?
def c_step(
X: MatrixLike,
n_support: Int,
stubs/sklearn/datasets/_samples_generator.pyi
@@ -110,9 +110,6 @@ def make_low_rank_matrix(
110
tail_strength: Float = 0.5,
111
random_state: RandomState | None | Int = None,
112
) -> ndarray: ...
113
114
-# TODO(1.3): Change argument `data_transposed` default from True to False.
115
-# TODO(1.3): Deprecate data_transposed, always return data not transposed.
116
def make_sparse_coded_signal(
117
n_samples: Int,
118
stubs/sklearn/ensemble/_gb.pyi
@@ -70,7 +70,6 @@ class GradientBoostingClassifier(ClassifierMixin, BaseGradientBoosting):
70
feature_importances_: ndarray = ...
71
n_estimators_: int = ...
72
73
- # TODO(1.3): remove "deviance"
74
_parameter_constraints: ClassVar[dict] = ...
75
def __init__(
0 commit comments