We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent c9ff9db commit d4797fcCopy full SHA for d4797fc
1 file changed
docs/source/examples/grouping.rst
@@ -1,10 +1,9 @@
1
Grouping
2
========
3
4
-When applying a conflict-resolving aggregator such as :class:`~torchjd.aggregation.GradVac` in
5
-multi-task learning, the cosine similarities between task gradients can be computed at different
6
-granularities. The [Gradient Vaccine paper](https://arxiv.org/pdf/2010.05874) introduces four
7
-strategies, each partitioning the shared parameter vector differently:
+The aggregation can be made independently on groups of parameters, at different granularities. The
+[Gradient Vaccine paper](https://arxiv.org/pdf/2010.05874) introduces four strategies to partition
+the parameters:
8
9
1. **Together** (baseline): one group covering all shared parameters. Corresponds to the
10
`whole_model` stategy in the paper.
0 commit comments