Skip to content

Add Graph of Tiered Experts Architecture environment#410

Open
signabuilder wants to merge 5 commits intoNousResearch:mainfrom
signabuilder:community/moe-routing
Open

Add Graph of Tiered Experts Architecture environment#410
signabuilder wants to merge 5 commits intoNousResearch:mainfrom
signabuilder:community/moe-routing

Conversation

@signabuilder
Copy link

@signabuilder signabuilder commented Mar 11, 2026

Graph of Tiered Experts Architecture Environment

Trains a language model to act as a routing policy over frozen Hermes-series expert tiers.

What it does

  • 7 frozen Hermes-series experts distributed across 3B, 8B, and 70B tiers
  • 120 routing scenarios (8 query templates x 15 topics) covering triage, synthesis, validation, execution, simulation, classification, and research
  • 3-component reward: ideal match (Jaccard similarity, 50%), capability alignment (30%), cost efficiency (10%)
  • Expert selection via JSON output, e.g. ["a0", "v0"]

Why it matters

Standard MoE trains experts and gates jointly inside a single model architecture. This environment instead learns post-hoc routing across frozen Hermes-series tiers, making the routing problem trainable without retraining the experts themselves.

Architecture

Expert Tier Role
g0 DeepHermes 3 3B Triage
g1 DeepHermes 3 3B Classification
a0 DeepHermes 3 8B Synthesis
a1 DeepHermes 3 8B Adversarial challenge
v0 Hermes 3 70B Validation
b0 Hermes 3 70B Execution
q0 DeepHermes 3 8B Quorum simulation

Quick start

vllm serve NousResearch/DeepHermes-3-Llama-3-8B-Preview --port 8378
python environments/community/moe_routing/moe_routing_env.py serve --port 8332
python environments/community/moe_routing/moe_routing_env.py process --num_trajectories 100

Research applications

  • Graph of Tiered Experts Architecture routing
  • LM-as-router via RL
  • Cost-aware expert selection
  • Federated routing policy optimization via Psyche/DisTrO

Thomas Perry added 2 commits March 11, 2026 10:21
Trains a language model to act as a gating network for heterogeneous
Mixture-of-Experts inference. 7 frozen experts (0.8B-35B), 120 query
scenarios, 3-component reward (ideal match + capability + cost).
- Remove unused `Any` import (flake8 F401)
- Add `pragma: allowlist secret` for detect-secrets false positive on api_key="local"
- Remove dead Chameleon concept paper link from README
@signabuilder signabuilder force-pushed the community/moe-routing branch from fdbf99d to af7b2a9 Compare March 12, 2026 17:24
Copy link
Author

@signabuilder signabuilder left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

n

@dmahan93
Copy link
Collaborator

This being called MoE conflicts with the common nomenclature of referring to a specific model architecture, usually i see this referred to as agent routing or similar

@signabuilder signabuilder changed the title Add MoE routing community environment Add Graph of Tiered Experts Architecture environment Mar 22, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants