perf(transforms): Use stop_gradient in custom_function VJP #2951
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Proposed changes
This PR optimizes the fallback VJP logic within custom_function in mlx/transforms.cpp. Previously, the VJP computation tracked gradients for all inputs, even those not specified in argnums. This PR introduces stop_gradient for inputs that are not part of the differentiation request. This change:
Prevents wasteful computation of gradients for constant or irrelevant inputs.
Ensures that argnums is fully respected, aligning with the expected behavior of efficient gradient tracking.
Checklist
Put an
xin the boxes that apply.pre-commit run --all-filesto format my code / installed pre-commit prior to committing changes