You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CHANGELOG_DEV.md
+23-1Lines changed: 23 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -163,4 +163,26 @@ Some HF tokenisers such as `xlm-roberta-large` add special tokens (e.g., eod tok
163
163
This side-effect in the transformers library has lead to the eod token being appended twice when tokenizing / packing our data. We added a check for this and only append the eod token once now:
Additionally, I added a script that verifies the consistency of the indexation and tokenization of a given JSONL file. We run the indexation and tokenization routines in modalities and compare it to tokenized JSONL file to which we applied the HF tokenizer directly.
166
+
Additionally, I added a script that verifies the consistency of the indexation and tokenization of a given JSONL file. We run the indexation and tokenization routines in modalities and compare it to tokenized JSONL file to which we applied the HF tokenizer directly.
167
+
168
+
## PR #379 Instruction Tuning Support
169
+
170
+
* New entry point `apply_chat_template` to form chats and create index and pbin files of it
171
+
* A wrapper for collate functions to include tokens in the loss which appear between indicator tokens
172
+
* A new parameter for the PackedMemMapDatasetContinuous to allow not to re-use the last target token
173
+
* A tutorial how to apply instruction-tuning on a Huggingface Model
174
+
175
+
176
+
## PR #359 Activation Checkpoint with FSDP2
177
+
178
+
This PR adds activation checkpointing (AC) support for FSDP2.
179
+
There are now three AC variants:
180
+
* Full AC (same as before, where entire complete modules get ACed, leading to the largest memory footprint reduction)
181
+
* Selective Layer AC (only very nth layer or module is ACed)
182
+
* Selective OP Ac (only certain OPs, typically low memory but compute intense, are checkpointed)
183
+
184
+
## PR #374 Tensor Parallelism Support
185
+
186
+
* adds support for Tensor Parallelism (including Sequence Parallelism).
187
+
* adds a debugging toolkit to track the input and output tensors during a forward pass, gradients during the backward pass and weight tensors.
# for developer: use [tests,linting] and install pre-commit hooks
87
+
uv pip install -e .[tests,linting]
88
+
pre-commit install --install-hooks
89
+
```
90
+
77
91
## Usage
78
92
Modalities provides several entry points to interact with the framework. The following section lists the available entry points and their respective functionalities.
0 commit comments