Skip to content

Conversation

priyakasimbeg
Copy link
Contributor

@priyakasimbeg priyakasimbeg commented Jun 18, 2025

Make the dropout rate an argument of the model call functions such that the dropout rate can be in calls to model function. Also remove logic to read dropout rate from hyperparameter config for model initialization.
Previously, the dropout rate was set during model initialization, which led to the following issues:

  1. Inconsistency between external tuning ruleset where users had the ability to set the dropout rate through initialization while self-tuning ruleset only used default dropout values.
  2. No ability to change dropout value during training, which would enable "sequential hyperparameter" type submissions.

Addresses: #753

@priyakasimbeg priyakasimbeg requested a review from a team as a code owner June 18, 2025 23:33
Copy link

github-actions bot commented Jun 18, 2025

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@priyakasimbeg priyakasimbeg merged commit f9fbbab into dev Jun 25, 2025
31 checks passed
@github-actions github-actions bot locked and limited conversation to collaborators Jun 25, 2025
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants