Modify remote_function decorators in multi_lora_transformers#173
Merged
tastelikefeet merged 1 commit intomodelscope:mainfrom Apr 21, 2026
Merged
Modify remote_function decorators in multi_lora_transformers#173tastelikefeet merged 1 commit intomodelscope:mainfrom
tastelikefeet merged 1 commit intomodelscope:mainfrom
Conversation
Updated remote_function decorators to specify collection methods.
Contributor
There was a problem hiding this comment.
Code Review
This pull request updates the MultiLoraTransformers class by adding collection strategies to remote function decorators. Specifically, it configures calculate_loss to use a 'mean' collection strategy for aggregating losses across ranks and get_state_dict to use a 'first' collection strategy for efficient state retrieval. I have no feedback to provide as the review comments were explanatory in nature.
tastelikefeet
approved these changes
Apr 21, 2026
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Updated remote_function decorators to specify collection methods.
PR type
PR information
When I'm using the self-host mode for LoRA SFT training, during the eval phase,
the following error occurs when executing the code below:
I found that the
calculate_lossmethod inMultiLoraTransformersModelalters the base class's distributed semantics, causing incorrect calculations under multi-GPU DP distributed training.Paste your experiment result here(if needed).