Skip to content

Add Mixture of Experts override to the UI#7527

Open
Quiet-Joker wants to merge 14 commits intooobabooga:mainfrom
Quiet-Joker:main
Open

Add Mixture of Experts override to the UI#7527
Quiet-Joker wants to merge 14 commits intooobabooga:mainfrom
Quiet-Joker:main

Conversation

@Quiet-Joker
Copy link
Copy Markdown

@Quiet-Joker Quiet-Joker commented May 1, 2026

Checklist:

-Add the ability to override the number of experts on models such as Gemma, GLM, Mistral, etc. Without having to manually write the command in the extra flags of the llama server.

image

Quiet-Joker added 10 commits May 1, 2026 05:08
Added handling for MoE expert overrides and reset logic for model loading.
Added MoE expert information variables for model loading.
Refactor MoE expert settings to use shared variables instead of model_settings.
-Removed spurious f prefix from the no-placeholder string in update_gpu_layers_and_vram

-Refactored the VRAM formula into named intermediate variables (kv_term, layer_term) to sidestep the W503/W504 conflict (those two rules are mutually exclusive, so restructuring is the only clean resolution)
-Both except Exception as e: blocks captured e but then used traceback.format_exc() instead, making e dead. Changed to except Exception: in both places.
@Quiet-Joker Quiet-Joker changed the title Add Mixture of Experts override to the UI and clean up lint Add Mixture of Experts override to the UI May 1, 2026
Removed the row_split checkbox from the UI model menu.
Stops gather_interface_values from reading a stale True when the user clicks Load after switching models.
Tighten the guard to also require moe_total_experts > 0, so even if shared.args.moe_experts_override_enabled is somehow left True, a non-MoE model (which has no expert_count key in its GGUF) can never trigger the override. This is the defensive backstop.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant