The race to construct the Prime basis forecasting mannequin is on!
Salesforce’s MOIRAI, one of many early basis fashions, achieved excessive benchmark outcomes and was open-sourced together with its pretraining dataset, LOTSA.
We extensively analyzed how MOIRAI works right here — and constructed an end-to-end challenge evaluating MOIRAI with well-liked statistical fashions.
Salesforce has now launched an upgraded model — MOIRAI-MOE — with important enhancements, significantly the addition of Combination-of-Specialists (MOE). We briefly mentioned MOE when one other mannequin, Time-MOE, additionally used a number of specialists.
On this article, we’ll cowl:
- How MOIRAI-MOE works and why it’s a strong mannequin.
- Key variations between MOIRAI and MOIRAI-MOE.
- How MOIRAI-MOE’s use of Combination-of-Specialists enhances accuracy.
- How Combination-of-Specialists typically solves frequency variation points in basis time-series fashions.
Let’s get began.
✅ I’ve launched AI Horizon Forecast, a publication specializing in time-series and modern…