vijaye12 commited on
Commit
7a087fa
1 Parent(s): c54c504

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -21,13 +21,12 @@ can be easily fine-tuned for your target data. Refer to our [paper](https://arxi
21
  - *SimMTM (NeurIPS 23) by 17% in few-shot forecasting*.
22
  - *Time-LLM (ICLR 24) by 8% in few-shot (5%) forecasting*
23
  - *UniTime (WWW 24) by 27% in zero-shot forecasting.*
24
- -
25
  - Zero-shot results of TTM surpass the *few-shot results of many popular SOTA approaches* including
26
  PatchTST (ICLR 23), PatchTSMixer (KDD 23), TimesNet (ICLR 23), DLinear (AAAI 23) and FEDFormer (ICML 22).
27
  - TTM (1024-96, released in this model card with 1M parameters) outperforms pre-trained MOIRAI (Small, 14M parameters) by 10%, MOIRAI (Base, 91M parameters) by 4% and
28
- MOIRAI (Large, 311M parameters) by 3% on forecast length 96.
29
  - TTM quick fine-tuning also outperforms the hard statistical baselines (Statistical ensemble and S-Naive) in
30
- M4-hourly dataset which pretrained TS models are finding hard to outperform.
31
  - TTM takes only a *few seconds for zeroshot/inference* and a *few minutes for finetuning* in 1 GPU machine, as
32
  opposed to long timing-requirements and heavy computing infra needs of other existing pretrained models.
33
 
 
21
  - *SimMTM (NeurIPS 23) by 17% in few-shot forecasting*.
22
  - *Time-LLM (ICLR 24) by 8% in few-shot (5%) forecasting*
23
  - *UniTime (WWW 24) by 27% in zero-shot forecasting.*
 
24
  - Zero-shot results of TTM surpass the *few-shot results of many popular SOTA approaches* including
25
  PatchTST (ICLR 23), PatchTSMixer (KDD 23), TimesNet (ICLR 23), DLinear (AAAI 23) and FEDFormer (ICML 22).
26
  - TTM (1024-96, released in this model card with 1M parameters) outperforms pre-trained MOIRAI (Small, 14M parameters) by 10%, MOIRAI (Base, 91M parameters) by 4% and
27
+ MOIRAI (Large, 311M parameters) by 3% on forecast length 96. (TODO: add notebook)
28
  - TTM quick fine-tuning also outperforms the hard statistical baselines (Statistical ensemble and S-Naive) in
29
+ M4-hourly dataset which pretrained TS models are finding hard to outperform. (TODO: add notebook)
30
  - TTM takes only a *few seconds for zeroshot/inference* and a *few minutes for finetuning* in 1 GPU machine, as
31
  opposed to long timing-requirements and heavy computing infra needs of other existing pretrained models.
32