Extending the LLPR module to be compatible with scalar targets where num_subtarget > 1
#607
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
As explained in the title. Archetypal case is for DOS learning, and indeed, this is something that I am rolling out for @HowWeiBin's DOS models (and hopefully my future ones 😅 ).
I'm already making this draft PR live so that my work is aware and no time is lost with multiple people working on the same extension.
I will coordinate with @Luthaf and @frostedoyster on some of the TODO's I've already place-marked. A decision has to be made on how to handle the agonistic loss used in DOS training in this case, which I will coordinate with Wei Bin first, then discuss with core mtt devs.
I'm also tagging @ppegolo to take a look so that we can already start thinking about accommodating for more general targets, but this PR will only prioritize this specific case of scalar targets with
num_subtargets > 1
.Contributor (creator of pull-request) checklist
Reviewer checklist