Mixture of Experts (MOE)
Model divides into specialized “experts”; router activates relevant subset per input.
From 10-AI-Concepts
- Huge params, but efficient (speed/cost).
- Scales intelligence without cost explosion.
Model divides into specialized “experts”; router activates relevant subset per input.
From 10-AI-Concepts