ben-cohen-datadog commited on
Commit
7fb145d
·
verified ·
1 Parent(s): af7a04c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -17
README.md CHANGED
@@ -145,30 +145,31 @@ For a step-by-step guide on running inferences with Toto, please refer to our [G
145
 
146
  - For optimal inference speed install [xformers](https://github.com/facebookresearch/xformers?tab=readme-ov-file#installing-xformers) and [flash-attention](https://github.com/Dao-AILab/flash-attention?tab=readme-ov-file#installation-and-features)
147
 
148
- ## Training Details - TODO keep or remove?
149
 
150
  ### Pre-Training Data
151
- | Dataset |
152
- |----------------------------------------------------------------------------------|
153
- | [GiftEval Pretrain](https://huggingface.co/datasets/Salesforce/GiftEvalPretrain) |
154
- | [Chronos](https://huggingface.co/datasets/autogluon/chronos_datasets) (Note: we use a subset of the Chronos dataset to avoid contamination with the GiftEval benchmark.) |
155
- | Synthetic |
156
- | Observability (**Note: No customer data was used in the training of this model**) |
157
 
 
158
 
159
- For more details about the pretraining data and preprocessing steps, please refer to the [paper](#TODO-Link-to-Paper) or the [GitHub repository](https://github.com/DataDog/toto.git).
160
 
 
161
 
162
- ## Citation - TODO
163
 
 
 
 
 
 
 
 
 
 
 
 
164
  If you use Toto in your research or applications, please cite us using the following:
165
 
166
- ```
167
- @article{Toto-Open-Base-1.0,
168
- title={TOTO: Time Series Optimized Transformer for Observability},
169
- author={Your Author Names Here},
170
- journal={arXiv preprint arXiv:XXXX.XXXXX},
171
- year={2025},
172
- url={https://arxiv.org/abs/XXXX.XXXXX}
173
- }
174
  ```
 
145
 
146
  - For optimal inference speed install [xformers](https://github.com/facebookresearch/xformers?tab=readme-ov-file#installing-xformers) and [flash-attention](https://github.com/Dao-AILab/flash-attention?tab=readme-ov-file#installation-and-features)
147
 
148
+ ## Training Details
149
 
150
  ### Pre-Training Data
 
 
 
 
 
 
151
 
152
+ Toto was trained on a massive and diverse mixture of time series datasets:
153
 
154
+ #### Observability Data
155
 
156
+ The largest portion of pretraining data comes from a dataset of approximately 1 trillion time series points collected from Datadog metrics. These metrics are generated from Datadog's monitoring of internal systems, and **do not** include any customer data. They cover a diverse array of software stacks and types of services, and span wide variety of domains within observability, including application performance, infrastructure, networking, security, databases, and more.
157
 
158
+ #### Public Datasets
159
 
160
+ To improve the performance of Toto on general-purpose time series forecasting across many domains, we include publcly availa
161
+ - [GiftEval Pretrain](https://huggingface.co/datasets/Salesforce/GiftEvalPretrain)
162
+ - [Chronos pretraining data](https://huggingface.co/datasets/autogluon/chronos_datasets) (Note: only a subset of this dataset was used to avoid leakage with the GiftEval benchmark)
163
+
164
+ #### Synthetic Data
165
+ To improve robustness, approximately 1/3 of the pretraining data mix consists of synthetically-generated time series.
166
+
167
+ For more details about the pretraining data and preprocessing steps, please refer to the [paper](#TODO-Link-to-Paper).
168
+
169
+
170
+ ## Citation
171
  If you use Toto in your research or applications, please cite us using the following:
172
 
173
+ ```bibtex
174
+ TODO
 
 
 
 
 
 
175
  ```