fwimbauer commited on
Commit
70dbbc3
·
verified ·
1 Parent(s): 72bf896

Upload config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.yaml +387 -0
config.yaml ADDED
@@ -0,0 +1,387 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ training:
2
+ optimizer:
3
+ type: adam
4
+ args:
5
+ lr: 1.0e-05
6
+ betas:
7
+ - 0.9
8
+ - 0.999
9
+ eps: 1.0e-08
10
+ weight_decay: 0.0
11
+ amsgrad: false
12
+ scheduler:
13
+ type: step
14
+ step_size: 100000
15
+ gamma: 0.1
16
+ num_epochs: 1000
17
+ continue: true
18
+ checkpoint_every: 2500
19
+ log_every_iters: 100
20
+ n_saved: 4
21
+ stop_iteration: 250000
22
+ from_pretrained: out/post_submission/anycam_baseline_fc2_backend-nccl-2_1311123/training_checkpoint_247500.pt
23
+ dataset_cfgs:
24
+ sintel:
25
+ type: sintel
26
+ data_path_training: data/Sintel/training
27
+ data_path_testing: data/Sintel/training
28
+ split: null
29
+ image_size: 384
30
+ frame_count: 2
31
+ preprocessed_path_training: data/unimatch_flows/sintel/training
32
+ preprocessed_path_testing: data/unimatch_flows/sintel/training
33
+ return_depth: false
34
+ return_flow: true
35
+ sintel_gt:
36
+ type: sintel-gt
37
+ data_path_training: data/Sintel/training
38
+ data_path_testing: data/Sintel/training
39
+ split: null
40
+ image_size: 384
41
+ frame_count: 2
42
+ preprocessed_path_training: data/unimatch_flows/sintel/training
43
+ preprocessed_path_testing: data/unimatch_flows/sintel/training
44
+ return_depth: true
45
+ return_flow: true
46
+ waymo:
47
+ type: waymo
48
+ data_path_training: data/waymo/training
49
+ data_path_testing: data/waymo/testing
50
+ split: null
51
+ image_size: 384
52
+ frame_count: 2
53
+ preprocessed_path_training: data/unimatch_flows/waymo/training
54
+ preprocessed_path_testing: data/unimatch_flows/waymo/testing
55
+ return_depth: false
56
+ return_flow: true
57
+ re10k_eval_seqs:
58
+ type: re10k
59
+ data_path_training: ${data_root}/re10k/data/test.pickle
60
+ data_path_testing: ${data_root}/re10k/data/test.pickle
61
+ split: anycam/datasets/realestate10k/splits/eval_seqs
62
+ image_size: 384
63
+ frame_count: 2
64
+ preprocessed_path_training: ${data_root}/re10k/flows_1
65
+ preprocessed_path_testing: ${data_root}/re10k/flows_1
66
+ return_depth: false
67
+ return_flow: true
68
+ re10k_eval_seqs_clean:
69
+ type: re10k
70
+ data_path_training: ${data_root}/re10k/data/test.pickle
71
+ data_path_testing: ${data_root}/re10k/data/test.pickle
72
+ split: anycam/datasets/realestate10k/splits/eval_seqs_clean
73
+ image_size: 384
74
+ frame_count: 2
75
+ preprocessed_path_training: ${data_root}/re10k/flows_1
76
+ preprocessed_path_testing: ${data_root}/re10k/flows_1
77
+ return_depth: false
78
+ return_flow: true
79
+ tumrgbd_eval_seqs:
80
+ type: tumrgbd
81
+ data_path_training: data/TUM_RGBD
82
+ data_path_testing: data/TUM_RGBD
83
+ split: anycam/datasets/tum_rgbd/splits/dynamic_seqs
84
+ image_size: 384
85
+ frame_count: 2
86
+ dilation: 10
87
+ return_depth: false
88
+ return_flow: false
89
+ tumrgbd_eval_seqs_64:
90
+ type: tumrgbd
91
+ data_path_training: data/TUM_RGBD
92
+ data_path_testing: data/TUM_RGBD
93
+ split: anycam/datasets/tum_rgbd/splits/dynamic_seqs_64
94
+ image_size: 384
95
+ frame_count: 2
96
+ dilation: 10
97
+ return_depth: false
98
+ return_flow: false
99
+ waymo_eval_seqs_2_64:
100
+ type: waymo
101
+ data_path_training: data/waymo/training
102
+ data_path_testing: data/waymo/testing
103
+ split: anycam/datasets/waymo/splits/eval_seqs_2_64
104
+ image_size: 384
105
+ frame_count: 2
106
+ preprocessed_path_training: data/unimatch_flows/waymo/training
107
+ preprocessed_path_testing: data/unimatch_flows/waymo/testing
108
+ return_depth: false
109
+ return_flow: true
110
+ re10k:
111
+ type: re10k
112
+ data_path_training: ${data_root}/re10k/data/test.pickle
113
+ data_path_testing: ${data_root}/re10k/data/test.pickle
114
+ split: null
115
+ image_size: 384
116
+ frame_count: 2
117
+ dilation: 1
118
+ preprocessed_path_training: ${data_root}/re10k/flows_1
119
+ preprocessed_path_testing: ${data_root}/re10k/flows_1
120
+ return_depth: false
121
+ return_flow: true
122
+ re10k_n2:
123
+ type: re10k
124
+ data_path_training: ${data_root}/re10k/data/test.pickle
125
+ data_path_testing: ${data_root}/re10k/data/test.pickle
126
+ split: null
127
+ image_size: 384
128
+ frame_count: 2
129
+ dilation: 2
130
+ preprocessed_path_training: ${data_root}/re10k/flows_2
131
+ preprocessed_path_testing: ${data_root}/re10k/flows_2
132
+ return_depth: false
133
+ return_flow: true
134
+ re10k_n3:
135
+ type: re10k
136
+ data_path_training: ${data_root}/re10k/data/test.pickle
137
+ data_path_testing: ${data_root}/re10k/data/test.pickle
138
+ split: null
139
+ image_size: 384
140
+ frame_count: 2
141
+ dilation: 3
142
+ preprocessed_path_training: ${data_root}/re10k/flows_3
143
+ preprocessed_path_testing: ${data_root}/re10k/flows_3
144
+ return_depth: false
145
+ return_flow: true
146
+ youtube_vos:
147
+ type: youtubevos
148
+ data_path_training: ${data_root}/youtubevos/data/train
149
+ data_path_testing: ${data_root}/youtubevos/data/valid
150
+ split: null
151
+ image_size: 384
152
+ frame_count: 2
153
+ preprocessed_path_training: ${data_root}/youtubevos/flows_1/train
154
+ preprocessed_path_testing: ${data_root}/youtubevos/flows_1/valid
155
+ return_depth: false
156
+ return_flow: false
157
+ youtube_vos_n2:
158
+ type: youtubevos
159
+ data_path_training: ${data_root}/youtubevos/data/train
160
+ data_path_testing: ${data_root}/youtubevos/data/train
161
+ split: null
162
+ image_size: 384
163
+ frame_count: 2
164
+ preprocessed_path_training: ${data_root}/youtubevos/flows_2
165
+ preprocessed_path_testing: ${data_root}/youtubevos/flows_2
166
+ return_depth: false
167
+ return_flow: false
168
+ dilation: 2
169
+ youtube_vos_n3:
170
+ type: youtubevos
171
+ data_path_training: ${data_root}/youtubevos/data/train
172
+ data_path_testing: ${data_root}/youtubevos/data/train
173
+ split: null
174
+ image_size: 384
175
+ frame_count: 2
176
+ preprocessed_path_training: ${data_root}/youtubevos/flows_3
177
+ preprocessed_path_testing: ${data_root}/youtubevos/flows_3
178
+ return_depth: false
179
+ return_flow: false
180
+ dilation: 3
181
+ opendv:
182
+ type: opendv
183
+ data_path_training: ${data_root}/opendv/data
184
+ data_path_testing: ${data_root}/opendv/data
185
+ split: null
186
+ image_size: 384
187
+ frame_count: 2
188
+ preprocessed_path_training: ${data_root}/opendv/flows_1
189
+ preprocessed_path_testing: ${data_root}/opendv/flows_1
190
+ return_depth: false
191
+ return_flow: true
192
+ opendv_n2:
193
+ type: opendv
194
+ data_path_training: ${data_root}/opendv/data
195
+ data_path_testing: ${data_root}/opendv/data
196
+ split: null
197
+ image_size: 384
198
+ frame_count: 2
199
+ preprocessed_path_training: ${data_root}/opendv/flows_2
200
+ preprocessed_path_testing: ${data_root}/opendv/flows_2
201
+ return_depth: false
202
+ return_flow: true
203
+ dilation: 2
204
+ opendv_n3:
205
+ type: opendv
206
+ data_path_training: ${data_root}/opendv/data
207
+ data_path_testing: ${data_root}/opendv/data
208
+ split: null
209
+ image_size: 384
210
+ frame_count: 2
211
+ preprocessed_path_training: ${data_root}/opendv/flows_3
212
+ preprocessed_path_testing: ${data_root}/opendv/flows_3
213
+ return_depth: false
214
+ return_flow: true
215
+ dilation: 3
216
+ walkingtours:
217
+ type: walkingtours
218
+ data_path_training: ${data_root}/walkingtours/data
219
+ data_path_testing: ${data_root}/walkingtours/data
220
+ split: null
221
+ image_size: 384
222
+ frame_count: 2
223
+ preprocessed_path_training: ${data_root}/walkingtours/flows_1
224
+ preprocessed_path_testing: ${data_root}/walkingtours/flows_1
225
+ return_depth: false
226
+ return_flow: true
227
+ walkingtours_n2:
228
+ type: walkingtours
229
+ data_path_training: ${data_root}/walkingtours/data
230
+ data_path_testing: ${data_root}/walkingtours/data
231
+ split: null
232
+ image_size: 384
233
+ frame_count: 2
234
+ preprocessed_path_training: ${data_root}/walkingtours/flows_2
235
+ preprocessed_path_testing: ${data_root}/walkingtours/flows_2
236
+ return_depth: false
237
+ return_flow: true
238
+ dilation: 2
239
+ walkingtours_n3:
240
+ type: walkingtours
241
+ data_path_training: ${data_root}/walkingtours/data
242
+ data_path_testing: ${data_root}/walkingtours/data
243
+ split: null
244
+ image_size: 384
245
+ frame_count: 2
246
+ preprocessed_path_training: ${data_root}/walkingtours/flows_3
247
+ preprocessed_path_testing: ${data_root}/walkingtours/flows_3
248
+ return_depth: false
249
+ return_flow: true
250
+ dilation: 3
251
+ epickitchens:
252
+ type: epickitchens
253
+ data_path_training: ${data_root}/epickitchens/data
254
+ data_path_testing: ${data_root}/epickitchens/data
255
+ split: null
256
+ image_size: 384
257
+ frame_count: 2
258
+ preprocessed_path_training: ${data_root}/epickitchens/flows_1
259
+ preprocessed_path_testing: ${data_root}/epickitchens/flows_1
260
+ return_depth: false
261
+ return_flow: true
262
+ epickitchens_n2:
263
+ type: epickitchens
264
+ data_path_training: ${data_root}/epickitchens/data
265
+ data_path_testing: ${data_root}/epickitchens/data
266
+ split: null
267
+ image_size: 384
268
+ frame_count: 2
269
+ preprocessed_path_training: ${data_root}/epickitchens/flows_2
270
+ preprocessed_path_testing: ${data_root}/epickitchens/flows_2
271
+ return_depth: false
272
+ return_flow: true
273
+ dilation: 2
274
+ epickitchens_n3:
275
+ type: epickitchens
276
+ data_path_training: ${data_root}/epickitchens/data
277
+ data_path_testing: ${data_root}/epickitchens/data
278
+ split: null
279
+ image_size: 384
280
+ frame_count: 2
281
+ preprocessed_path_training: ${data_root}/epickitchens/flows_3
282
+ preprocessed_path_testing: ${data_root}/epickitchens/flows_3
283
+ return_depth: false
284
+ return_flow: true
285
+ dilation: 3
286
+ data_root: ${oc.env:ANYCAM_DATA, data_new}
287
+ training_type: anycam_training
288
+ seed: 0
289
+ backend: nccl
290
+ nproc_per_node: 2
291
+ with_amp: true
292
+ name: anycam_seq8
293
+ batch_size: 4
294
+ num_workers: 8
295
+ dataset:
296
+ - re10k
297
+ - youtube_vos
298
+ - opendv
299
+ - walkingtours
300
+ - epickitchens
301
+ val_dataset:
302
+ - re10k_eval_seqs
303
+ - re10k_eval_seqs_clean
304
+ dataset_params:
305
+ frame_count: 8
306
+ return_flow: true
307
+ image_size: 336
308
+ sequential: true
309
+ dataloading:
310
+ epoch_length: 80000
311
+ output:
312
+ path: pretrained_models/anycam_seq8
313
+ unique_id: 5051794
314
+ loss:
315
+ - type: pose_loss
316
+ lambda_dist: 0
317
+ pose_token_weight_decay: 0.01
318
+ lambda_fwd_bwd_consistency: 1
319
+ lambda_label_scale: 100
320
+ model:
321
+ depth_predictor:
322
+ type: unidepth
323
+ pose_predictor:
324
+ type: anycam
325
+ focal_parameterization: linlog-candidates
326
+ focal_min: 0.2
327
+ focal_max: 7
328
+ rotation_parameterization: axis-angle
329
+ separate_pose_candidates: true
330
+ separate_uncertainty_candidates: true
331
+ pose_token_partial_dropout: 0.5
332
+ depth_aligner:
333
+ type: identity
334
+ flow_model: unimatch
335
+ use_provided_flow: true
336
+ use_provided_proj: false
337
+ train_directions: both
338
+ perform_subsampled_pose_pass: false
339
+ subsampling_drop_n: 1
340
+ single_focal_warmup_iters: 0
341
+ z_near: 0.1
342
+ z_far: 100
343
+ validation:
344
+ validation:
345
+ batch_size: 1
346
+ subset:
347
+ type: range
348
+ args:
349
+ start: 0
350
+ end: 512
351
+ custom_validator: anycam.video_validator.video_validator
352
+ fit_video_config: anycam/configs/eval_cfgs/train_eval.yaml
353
+ log_loss: false
354
+ global_step:
355
+ type: trainer iteration
356
+ events:
357
+ - type: ITERATION_COMPLETED
358
+ args:
359
+ every: 5000
360
+ - type: COMPLETED
361
+ args: null
362
+ visualization:
363
+ metrics: []
364
+ subset:
365
+ type: range
366
+ args:
367
+ start: 0
368
+ end: 1
369
+ visualize:
370
+ input_imgs: null
371
+ depth: null
372
+ occlusions: null
373
+ rendered_flow: null
374
+ gt_flow: null
375
+ predicted_occlusions: null
376
+ uncertainty: null
377
+ log_loss: false
378
+ global_step:
379
+ type: trainer iteration
380
+ events:
381
+ - type: ITERATION_COMPLETED
382
+ args:
383
+ every: 2500
384
+ - type: COMPLETED
385
+ args: null
386
+ master_port: 2222
387
+ cuda device name: NVIDIA A100-SXM4-40GB