feat: support convert_lora_to_hf without merge #2234
Conversation
43f4642 to
0642c7a
Compare
|
/ok to test eeef66c |
|
/ok to test 9de1880 |
|
/ok to test 64a8679 |
64a8679 to
ed1bbdd
Compare
|
/ok to test ed1bbdd |
|
/ok to test f24be7a |
yuki-97
left a comment
There was a problem hiding this comment.
thanks for fixing and adding the support! lgtm
|
@terrykong could you take a review as well? |
|
/ok to test 5396d85 |
terrykong
left a comment
There was a problem hiding this comment.
Nice work on this PR! The refactoring to extract _build_megatron_model_with_lora as a shared context manager is clean, the bug fix for adapter loading (using dist_checkpointing.load with a filtered state dict via apply_peft_adapter_filter_to_state_dict) correctly addresses the KeyError from the old _load_model_weights_from_checkpoint path, and the docs clearly present both export modes with examples. Nice work addressing the community request (#2190).
Two minor suggestions below — neither is a blocker.
Generated by Claude Code
terrykong
left a comment
There was a problem hiding this comment.
lgtm docstring was just a nit, btw looks like sphinx build failure
…pters in HuggingFace PEFT format Signed-off-by: ruit <ruit@nvidia.com>
Signed-off-by: ruit <ruit@nvidia.com>
Signed-off-by: ruit <ruit@nvidia.com>
Signed-off-by: ruit <ruit@nvidia.com>
…options Signed-off-by: ruit <ruit@nvidia.com>
Signed-off-by: ruit <ruit@nvidia.com>
5396d85 to
2db3e7d
Compare
|
/ok to test 2db3e7d |
Summary
parameters
What does this PR do ?
Add a one line overview of what this PR aims to accomplish.
Issues
closes #2190
Usage
# Add a code snippet demonstrating how to use thisBefore your PR is "Ready for review"
Pre checks:
Additional Information