WizardLM-2 8x22B is Microsoft's latest Mixture-of-Experts model, using 8 expert networks of 22B parameters each. This 176B total parameter architecture delivers competitive performance in knowledge-intensive tasks.
anthropic