Update README.md
Browse files
README.md
CHANGED
|
@@ -145,30 +145,31 @@ For a step-by-step guide on running inferences with Toto, please refer to our [G
|
|
| 145 |
|
| 146 |
- For optimal inference speed install [xformers](https://github.com/facebookresearch/xformers?tab=readme-ov-file#installing-xformers) and [flash-attention](https://github.com/Dao-AILab/flash-attention?tab=readme-ov-file#installation-and-features)
|
| 147 |
|
| 148 |
-
## Training Details
|
| 149 |
|
| 150 |
### Pre-Training Data
|
| 151 |
-
| Dataset |
|
| 152 |
-
|----------------------------------------------------------------------------------|
|
| 153 |
-
| [GiftEval Pretrain](https://huggingface.co/datasets/Salesforce/GiftEvalPretrain) |
|
| 154 |
-
| [Chronos](https://huggingface.co/datasets/autogluon/chronos_datasets) (Note: we use a subset of the Chronos dataset to avoid contamination with the GiftEval benchmark.) |
|
| 155 |
-
| Synthetic |
|
| 156 |
-
| Observability (**Note: No customer data was used in the training of this model**) |
|
| 157 |
|
|
|
|
| 158 |
|
| 159 |
-
|
| 160 |
|
|
|
|
| 161 |
|
| 162 |
-
|
| 163 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 164 |
If you use Toto in your research or applications, please cite us using the following:
|
| 165 |
|
| 166 |
-
```
|
| 167 |
-
|
| 168 |
-
title={TOTO: Time Series Optimized Transformer for Observability},
|
| 169 |
-
author={Your Author Names Here},
|
| 170 |
-
journal={arXiv preprint arXiv:XXXX.XXXXX},
|
| 171 |
-
year={2025},
|
| 172 |
-
url={https://arxiv.org/abs/XXXX.XXXXX}
|
| 173 |
-
}
|
| 174 |
```
|
|
|
|
| 145 |
|
| 146 |
- For optimal inference speed install [xformers](https://github.com/facebookresearch/xformers?tab=readme-ov-file#installing-xformers) and [flash-attention](https://github.com/Dao-AILab/flash-attention?tab=readme-ov-file#installation-and-features)
|
| 147 |
|
| 148 |
+
## Training Details
|
| 149 |
|
| 150 |
### Pre-Training Data
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
|
| 152 |
+
Toto was trained on a massive and diverse mixture of time series datasets:
|
| 153 |
|
| 154 |
+
#### Observability Data
|
| 155 |
|
| 156 |
+
The largest portion of pretraining data comes from a dataset of approximately 1 trillion time series points collected from Datadog metrics. These metrics are generated from Datadog's monitoring of internal systems, and **do not** include any customer data. They cover a diverse array of software stacks and types of services, and span wide variety of domains within observability, including application performance, infrastructure, networking, security, databases, and more.
|
| 157 |
|
| 158 |
+
#### Public Datasets
|
| 159 |
|
| 160 |
+
To improve the performance of Toto on general-purpose time series forecasting across many domains, we include publcly availa
|
| 161 |
+
- [GiftEval Pretrain](https://huggingface.co/datasets/Salesforce/GiftEvalPretrain)
|
| 162 |
+
- [Chronos pretraining data](https://huggingface.co/datasets/autogluon/chronos_datasets) (Note: only a subset of this dataset was used to avoid leakage with the GiftEval benchmark)
|
| 163 |
+
|
| 164 |
+
#### Synthetic Data
|
| 165 |
+
To improve robustness, approximately 1/3 of the pretraining data mix consists of synthetically-generated time series.
|
| 166 |
+
|
| 167 |
+
For more details about the pretraining data and preprocessing steps, please refer to the [paper](#TODO-Link-to-Paper).
|
| 168 |
+
|
| 169 |
+
|
| 170 |
+
## Citation
|
| 171 |
If you use Toto in your research or applications, please cite us using the following:
|
| 172 |
|
| 173 |
+
```bibtex
|
| 174 |
+
TODO
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 175 |
```
|