Sponsor of the Day:
Jerkmate
https://huggingface.co/papers/2503.07137
Paper page - A Comprehensive Survey of Mixture-of-Experts: Algorithms, Theory, and Applications
Join the discussion on this paper page
comprehensive surveymixture expertsalgorithms theorypaperapplications
https://www.ibm.com/think/podcasts/mixture-of-experts/12-days-open-ai-neurips-arc-prize-llama-3-3-70b
12 Days of OpenAI, NeurIPS, ARC Prize and Llama 3.3 70B | Mixture of Experts | IBM
Tune in to episode 33 of Mixture of Experts to know more about 12 Days of OpenAI, NeurIPS, ARC Prize and Llama 3.3 70B
llama 3 70b12 daysarc prizemixture expertsopenai
https://arxiv.org/abs/2503.07137
[2503.07137] A Comprehensive Survey of Mixture-of-Experts: Algorithms, Theory, and Applications
Abstract page for arXiv paper 2503.07137: A Comprehensive Survey of Mixture-of-Experts: Algorithms, Theory, and Applications
comprehensive surveymixture expertsalgorithms theory2503applications
https://huggingface.co/papers/2604.19835
Paper page - Expert Upcycling: Shifting the Compute-Efficient Frontier of Mixture-of-Experts
Join the discussion on this paper page
compute efficientmixture expertspaperupcyclingshifting
https://www.ibm.com/think/podcasts/mixture-of-experts/google-gemini-3-ai-agents-reasoning-search-mode
Google’s Gemini 3: AI agents, reasoning and search mode | Mixture of Experts | IBM
This week on Mixture of Experts we discuss Google’s Gemini 3: AI agents, reasoning and search mode.
gemini 3 aiexperts ibmagentsreasoningsearch
https://allenai.org/blog/bar
Train separately, merge together: Modular post-training with mixture-of-experts | Ai2
BAR is a recipe for post-training language models one capability at a time—train domain experts independently, merge them into a single mixture-of-experts...
post trainingseparatelymergetogethermodular
https://arxiv.org/abs/2303.06318
[2303.06318] A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts...
Abstract page for arXiv paper 2303.06318: A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts Training
expert data2303hybridtensorparallelism
https://www.ibm.com/think/podcasts/mixture-of-experts/claude-opus-4-7-apple-ai-glasses-workplace-ai-adoption-deep-mind-manipulation-research
Claude Opus 4.7, Apple’s AI glasses and Allbirds AI pivot | Mixture of Experts | IBM
Claude Opus 4.7, Apple's AI glasses strategy, workplace AI adoption stats and DeepMind's manipulation research. Tune in to this week's Mixture of Experts.
claude opus 4ai glassesallbirds pivotexperts ibm7
https://www.ibm.com/think/podcasts/mixture-of-experts/ai-year-review-trends-2026
AI year in review: Trends shaping 2026 | Mixture of Experts | IBM
Our experts review 2025's AI breakthroughs and predict 2026 trends. AI hardware scarcity, open source wins, super agents and multimodal evolution discussed.
trends shaping 2026experts ibmaiyearreview
https://www.ibm.com/think/podcasts/mixture-of-experts?utm=XFLatestEpMoE
Mixture of Experts | IBM
Mixture of Experts is a weekly news podcast, recapping the latest trends and innovations in the artificial intelligence industry.
experts ibmmixture
https://arxiv.org/abs/2303.07226
[2303.07226] Scaling Vision-Language Models with Sparse Mixture of Experts
Abstract page for arXiv paper 2303.07226: Scaling Vision-Language Models with Sparse Mixture of Experts
vision language models2303scalingsparsemixture
https://stackoverflow.blog/mixture-of-experts/
mixture of experts - Stack Overflow
stack overflowmixtureexperts
https://47zzz.github.io/MoVE/
MoVE: Translating Laughter and Tears via Mixture of Vocalization Experts in S2ST
MoVE: Mixture-of-LoRA-Experts architecture for emotion-preserving Speech-to-Speech Translation. Interspeech 2026 (Under Review).
movetranslatinglaughtertearsvia
https://www.ibm.com/think/podcasts/mixture-of-experts/nvidia-nemoclaw-openai-shopify-agents
NVIDIA NemoClaw, OpenAI’s pivot and Shopify agents | Mixture of Experts | IBM
nvidia nemoclawexperts ibmpivotshopifyagents
https://blogs.nvidia.com/blog/mixture-of-experts-frontier-models/
Mixture of Experts Powers the Most Intelligent Frontier Models | NVIDIA Blog
Mar 3, 2026 - Kimi K2 Thinking, DeepSeek-R1, Mistral Large 3 and others run 10x faster on NVIDIA GB200 NVL72.
frontier modelsnvidia blogmixtureexpertspowers
https://www.ibm.com/think/podcasts/mixture-of-experts/searchgpt-naptime-big-sleep-github-octoverse
SearchGPT, From Naptime to Big Sleep, And GitHub Octoverse Updates | Mixture of Experts | IBM
Could AI wipe out software engineers? In Episode 28 of Mixture of Experts, host Tim Hwang is joined by Chris Hay, Kaoutar El Maghraoui, and Shobhit Varshney to...
experts ibmsearchgptnaptimebigsleep
https://journals.plos.org:443/ploscompbiol/article?id=10.1371/journal.pcbi.1013397
MoCETSE: A mixture-of-convolutional experts and transformer-based model for predicting...
Author summary Secreted effector proteins are a class of key virulence factors in Gram-negative bacteria. After being injected into host cells, they interfere...
transformer basedmixtureconvolutionalexpertsmodel
https://arxiv.org/abs/2603.24930
[2603.24930] CROSS: A Mixture-of-Experts Reinforcement Learning Framework for Generalizable...
Abstract page for arXiv paper 2603.24930: CROSS: A Mixture-of-Experts Reinforcement Learning Framework for Generalizable Large-Scale Traffic Signal Control
reinforcement learning2603crossmixtureexperts
https://www.ibm.com/think/topics/mixture-of-experts
What is mixture of experts? | IBM
Nov 17, 2025 - Mixture of experts (MoE) is a machine learning approach, diving an AI model into multiple “expert” models, each specializing in a subset of the input data.
experts ibmmixture
https://www.liquid.ai/blog/lfm2-8b-a1b-an-efficient-on-device-mixture-of-experts
LFM2-8B-A1B: An Efficient On-device Mixture-of-Experts | Liquid AI
Oct 24, 2025 - We are releasing LFM2-8B-A1B, our first on-device Mixture-of-Experts (MoE) with 8.3B total parameters and 1.5B active parameters per token. By activating only...
liquid ailfm28befficientdevice
https://deepgram.com/ai-glossary/mixture-of-experts
Mixture of Experts
Are you curious about how Mixture of Experts stands apart in the world of artificial intelligence and what it could mean for the future of machine learning?...
mixtureexperts
https://www.ibm.com/think/podcasts/mixture-of-experts?lnk=thinkhpsppi6us
Mixture of Experts | IBM
Mixture of Experts is a weekly news podcast, recapping the latest trends and innovations in the artificial intelligence industry.
experts ibmmixture
https://www.ibm.com/think/podcasts/mixture-of-experts
Mixture of Experts | IBM
Mixture of Experts is a weekly news podcast, recapping the latest trends and innovations in the artificial intelligence industry.
experts ibmmixture