This repo provides the model weights released in the paper Towards Neural Scaling Laws for Time Series Foundation Models.

The models have varying sizes, ranging from 1M to 1B parameters, and were trained on datasets spanning from 10M to 16B time points.

Code: https://github.com/Qingrenn/TSFM-ScalingLaws

Dataset: https://huggingface.co/datasets/Qingren/TSFM-ScalingLaws-Dataset


Figure1: Scaling laws for NLL in relation to model size, compute, and dataset size. The blue lines represent ID performance, while the red and green lines show OOD performance on LSF subset and Monash subset.


Figure2: Prediction results of models with sizes 1B, 300M, 100M, and 10M.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.