forked from openvinotoolkit/open_model_zoo
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathdataset_definitions.yml
1359 lines (1203 loc) · 41.5 KB
/
dataset_definitions.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
datasets:
- name: ms_coco_mask_rcnn
annotation_conversion:
converter: mscoco_mask_rcnn
annotation_file: instances_val2017.json
has_background: True
sort_annotations: True
annotation: mscoco_mask_rcnn.pickle
dataset_meta: mscoco_mask_rcnn.json
data_source: val2017
- name: ms_coco_mask_rcnn_short_80_classes
annotation_conversion:
converter: mscoco_mask_rcnn
annotation_file: instances_val2017_short.json
has_background: True
sort_annotations: True
annotation: mscoco_mask_rcnn_short_80.pickle
dataset_meta: mscoco_mask_rcnn_short_80.json
data_source: val2017
- name: ms_coco_mask_rcnn_short_80_classes_without_background
annotation_conversion:
converter: mscoco_mask_rcnn
annotation_file: instances_val2017.json
has_background: False
sort_annotations: True
annotation: mscoco_mask_rcnn_short_80_without_bkgr.pickle
dataset_meta: mscoco_mask_rcnn_short_80_without_bkgr.json
data_source: val2017
- name: ms_coco_mask_rcnn_short_91_classes
annotation_conversion:
converter: mscoco_mask_rcnn
annotation_file: instances_val2017_short.json
has_background: True
sort_annotations: True
use_full_label_map: True
annotation: mscoco_mask_rcnn_short_91.pickle
dataset_meta: mscoco_mask_rcnn_short_91.json
data_source: val2017
preprocessing:
- type: resize
aspect_ratio_scale: fit_to_window
dst_height: 800
dst_width: 1365
- type: padding
dst_height: 800
dst_width: 1365
pad_type: right_bottom
postprocessing:
- type: faster_rcnn_postprocessing_resize
dst_height: 800
dst_width: 1365
- name: ms_coco_detection_91_classes
annotation_conversion:
converter: mscoco_detection
annotation_file: instances_val2017.json
has_background: True
sort_annotations: True
use_full_label_map: True
annotation: mscoco_det_91.pickle
dataset_meta: mscoco_det_91.json
data_source: val2017
preprocessing:
- type: resize
aspect_ratio_scale: fit_to_window
dst_height: 600
dst_width: 1024
- type: padding
dst_height: 600
dst_width: 1024
pad_type: right_bottom
postprocessing:
- type: faster_rcnn_postprocessing_resize
dst_height: 600
dst_width: 1024
- name: ms_coco_detection_80_class_without_background
data_source: val2017
annotation_conversion:
converter: mscoco_detection
annotation_file: instances_val2017.json
has_background: False
sort_annotations: True
use_full_label_map: False
annotation: mscoco_det_80.pickle
dataset_meta: mscoco_det_80.json
- name: ms_coco_detection_80_class_with_background
data_source: val2017
annotation_conversion:
converter: mscoco_detection
annotation_file: instances_val2017.json
has_background: True
sort_annotations: True
use_full_label_map: False
annotation: mscoco_det_80_bkgr.pickle
dataset_meta: mscoco_det_80_bkgr.json
- name: ms_coco_detection_90_class_without_background
data_source: MSCOCO/val2017
annotation_conversion:
converter: mscoco_detection
annotation_file: MSCOCO/annotations/instances_val2017.json
has_background: False
sort_annotations: True
use_full_label_map: True
annotation: mscoco_det_90.pickle
dataset_meta: mscoco_det_90.json
- name: ms_coco_keypoints
data_source: val2017
annotation_conversion:
converter: mscoco_keypoints
annotation_file: person_keypoints_val2017.json
sort_key: image_size
annotation: mscoco_keypoints.pickle
dataset_meta: mscoco_keypoints.json
metrics:
- name: AP
type: coco_precision
max_detections: 20
- name: ms_coco_val2017_keypoints
data_source: val2017
annotation_conversion:
converter: mscoco_keypoints
annotation_file: person_keypoints_val2017.json
remove_empty_images: True
sort_annotations: True
sort_key: image_size
images_dir: val2017
annotation: mscoco_val2017_keypoints.pickle
dataset_meta: mscoco_val2017_keypoints.json
metrics:
- name: AP
type: coco_orig_keypoints_precision
- name: ms_coco_val2017_keypoints_5k_images
data_source: val2017
annotation_conversion:
converter: mscoco_keypoints
annotation_file: person_keypoints_val2017.json
remove_empty_images: False
sort_annotations: True
sort_key: image_size
images_dir: val2017
annotation: mscoco_val2017_keypoints_5k_images.pickle
dataset_meta: mscoco_val2017_keypoints_5k_images.json
metrics:
- name: AP
type: coco_orig_keypoints_precision
- name: imagenet_1000_classes
annotation_conversion:
converter: imagenet
annotation_file: val.txt
annotation: imagenet1000.pickle
data_source: ILSVRC2012_img_val
metrics:
- name: accuracy@top1
type: accuracy
top_k: 1
- name: accuracy@top5
type: accuracy
top_k: 5
- name: imagenet_1000_classes_2015
annotation_conversion:
converter: imagenet
annotation_file: val15.txt
annotation: imagenet1000_2015.pickle
data_source: ILSVRC2012_img_val
metrics:
- name: accuracy@top1
type: accuracy
top_k: 1
- name: accuracy@top5
type: accuracy
top_k: 5
- name: imagenet_1001_classes
annotation_conversion:
converter: imagenet
annotation_file: val.txt
has_background: True
annotation: imagenet1001.pickle
data_source: ILSVRC2012_img_val
metrics:
- name: accuracy@top1
type: accuracy
top_k: 1
- name: accuracy@top5
type: accuracy
top_k: 5
- name: VOC2012
annotation_conversion:
converter: voc_detection
annotations_dir: VOCdevkit/VOC2012/Annotations
images_dir: VOCdevkit/VOC2012/JPEGImages
imageset_file: VOCdevkit/VOC2012/ImageSets/Main/val.txt
data_source: VOCdevkit/VOC2012/JPEGImages
annotation: voc12.pickle
dataset_meta: voc12.json
postprocessing:
- type: resize_prediction_boxes
metrics:
- type: map
integral: 11point
ignore_difficult: True
presenter: print_scalar
- name: VOC2012_without_background
annotation_conversion:
converter: voc_detection
annotations_dir: VOCdevkit/VOC2012/Annotations
images_dir: VOCdevkit/VOC2012/JPEGImages
imageset_file: VOCdevkit/VOC2012/ImageSets/Main/val.txt
has_background: False
data_source: VOCdevkit/VOC2012/JPEGImages
annotation: voc12_without_background.pickle
dataset_meta: voc12_without_background.json
postprocessing:
- type: resize_prediction_boxes
metrics:
- type: map
integral: 11point
ignore_difficult: True
presenter: print_scalar
- name: VOC2012_Segmentation
annotation_conversion:
converter: voc_segmentation
imageset_file: VOCdevkit/VOC2012/ImageSets/Segmentation/val.txt
images_dir: VOCdevkit/VOC2012/JPEGImages/
mask_dir: VOCdevkit/VOC2012/SegmentationClass/
data_source: VOCdevkit/VOC2012/JPEGImages/
segmentation_masks_source: VOCdevkit/VOC2012/SegmentationClass/
annotation: voc2012_segmentation.pickle
dataset_meta: voc2012_segmentation.json
- name: mapillary_20
annotation_conversion:
converter: mapillary_20
data_dir: Mapillary_20
annotation: mapillary_20.pickle
dataset_meta: mapillary_20.json
- name: mapillary_vistas
data_source: mapillary_vistas/validation/images
segmentation_masks_source: mapillary_vistas/validation/labels
annotation_conversion:
converter: mapillary_vistas
data_dir: mapillary_vistas/validation
annotation: mapillary_vistas.pickle
dataset_meta: mapillary_vistas.json
- name: wider
data_source: WIDER_val/images
annotation_conversion:
converter: wider
annotation_file: wider_face_split/wider_face_val_bbx_gt.txt
annotation: wider.pickle
dataset_meta: wider.json
- name: wider_without_bkgr
data_source: WIDER_val/images
annotation_conversion:
converter: wider
annotation_file: wider_face_split/wider_face_val_bbx_gt.txt
label_start: 0
annotation: wider_0.pickle
dataset_meta: wider_0.json
- name: facial_landmarks_35
data_source: VOCdevkit/VOC2012/JPEGImages
annotation_conversion:
converter: cvat_facial_landmarks
annotation_file: 3632_OMZ_task3_facial_landmarks_35_adas.xml
annotation: facial_landmarks_35.pickle
preprocessing:
- type: resize
size: 60
postprocessing:
- type: normalize_landmarks_points
- name: emotions_recognition
data_source: VOCdevkit/VOC2012/JPEGImages
annotation_conversion:
converter: cvat_attributes_recognition
annotation_file: 3631_OMZ_task2_emotions_recognition.xml
label: face
annotation: emotions_recognition.pickle
dataset_meta: emotions_recognition.json
preprocessing:
- type: extend_around_rect
augmentation_param: 0.3
- type: crop_rect
- type: resize
size: 64
- name: age_gender
data_source: ILSVRC2012_img_val
annotation_conversion:
converter: cvat_age_gender
annotation_file: 3630_OMZ_task1_age_gender.xml
annotation: age_gender.pickle
dataset_meta: age_gender.json
- name: vehicle_attributes
data_source: val2017
annotation_conversion:
converter: cvat_attributes_recognition
annotation_file: 3634_OMZ_task8_vehicle_attributes_recognition_barrier_0039.xml
label: vehicle
annotation: vehicle_attributes.pickle
dataset_meta: vehicle_attributes.json
- name: vehicle_attributes_0042
data_source: val2017
annotation_conversion:
converter: cvat_attributes_recognition
annotation_file: 3634_OMZ_task8_vehicle_attributes_recognition_barrier_0039.xml
label: vehicle
annotation: vehicle_attributes.pickle
dataset_meta: vehicle_attributes.json
- name: person_8_attributes
data_source: ILSVRC2012_img_val
annotation_conversion:
converter: cvat_multilabel_binary_attributes_recognition
annotation_file: 3640_OMZ_task6_person_attributes_recognition_crossroad_0230.xml
label: person
annotation: person_8_attributes.pickle
dataset_meta: person_8_attributes.json
- name: vehicle_license_plate_detection
data_source: ILSVRC2012_img_val
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3638_OMZ_task13_vehicle_license_plate_detection_barrier_0106.xml
has_background: True
annotation: vlpd.pickle
dataset_meta: vlpd.json
- name: action_detection_dataset_3_classes
data_source: WIDER_val/images/44--Aerobics
annotation_conversion:
converter: cvat_person_detection_action_recognition
use_case: common_3_actions
annotation_file: 3766_OMZ_task14_person-detection-raisinghand-recognition-0001.xml
annotation: action_detection_3classes.pickle
dataset_meta: action_detection_3classes.json
- name: action_detection_dataset_6_classes
data_source: WIDER_val/images/44--Aerobics
annotation_conversion:
converter: cvat_person_detection_action_recognition
use_case: common_6_actions
annotation_file: 3766_OMZ_task14_person-detection-raisinghand-recognition-0001.xml
annotation: action_detection_6classes.pickle
dataset_meta: action_detection_6classes.json
- name: action_detection_dataset_teacher
data_source: WIDER_val/images/44--Aerobics
annotation_conversion:
converter: cvat_person_detection_action_recognition
use_case: teacher
annotation_file: 3766_OMZ_task14_person-detection-raisinghand-recognition-0001.xml
annotation: action_detection_teacher.pickle
dataset_meta: action_detection_teacher.json
- name: action_detection_dataset_raising_hand
data_source: WIDER_val/images/44--Aerobics
annotation_conversion:
converter: cvat_person_detection_action_recognition
use_case: raising_hand
annotation_file: 3766_OMZ_task14_person-detection-raisinghand-recognition-0001.xml
annotation: action_detection_raising_hand.pickle
dataset_meta: action_detection_raising_hand.json
- name: person_detection
data_source: val2017
annotation_conversion:
converter: mscoco_detection
annotation_file: person_keypoints_val2017.json
has_background: True
sort_annotations: True
use_full_label_map: True
annotation: mscoco_person_detection.pickle
dataset_meta: mscoco_person_detection.json
- name: mscoco_person_detection
data_source: val2017
annotation_conversion:
converter: mscoco_detection
annotation_file: person_keypoints_val2017.json
has_background: True
sort_annotations: True
use_full_label_map: True
annotation: mscoco_person_detection.pickle
dataset_meta: mscoco_person_detection.json
- name: crossroad_dataset_1016
data_source: val2017
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3637_OMZ_task12_person_vehicle_bike_detection_crossroad_0078.xml
labels_file: person-vehicle-bike-detection-crossroad-1016-labels.json
has_background: True
annotation: crossroad-1016.pickle
dataset_meta: crossroad-1016.json
- name: crossroad_dataset_0078
data_source: val2017
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3637_OMZ_task12_person_vehicle_bike_detection_crossroad_0078.xml
labels_file: person-vehicle-bike-detection-crossroad-0078-labels.json
has_background: True
annotation: crossroad-0078.pickle
dataset_meta: crossroad-0078.json
- name: crossroad_extra_untagged
data_source: crossroad_extra_untagged/val_data
annotation_conversion:
converter: mscoco_detection
annotation_file: crossroad_extra_untagged/val_coco_no_bg.json
has_background: False
sort_annotations: True
use_full_label_map: False
annotation: crossroad_extra_untagged.pickle
dataset_meta: crossroad_extra_untagged.json
- name: crossroad_extra_untagged_person
data_source: crossroad_extra_untagged/val_data
annotation_conversion:
converter: mscoco_detection
annotation_file: crossroad_extra_untagged/val_coco_no_bg_person.json
has_background: False
sort_annotations: True
use_full_label_map: False
annotation: crossroad_extra_untagged_person.pickle
dataset_meta: crossroad_extra_untagged_person.json
- name: crossroad_extra_untagged_person_hb
data_source: crossroad_extra_untagged/val_data
annotation_conversion:
converter: mscoco_detection
annotation_file: crossroad_extra_untagged/val_coco_no_bg_person.json
has_background: True
sort_annotations: True
use_full_label_map: False
annotation: crossroad_extra_untagged_person_hb.pickle
dataset_meta: crossroad_extra_untagged_person_hb.json
- name: crossroad_extra_untagged_vehicle
data_source: crossroad_extra_untagged/val_data
annotation_conversion:
converter: mscoco_detection
annotation_file: crossroad_extra_untagged/val_coco_no_bg_vehicle.json
has_background: False
sort_annotations: True
use_full_label_map: False
annotation: crossroad_extra_untagged_vehicle.pickle
dataset_meta: crossroad_extra_untagged_vehicle.json
- name: crossroad_extra_untagged_vehicle_labels_from_1
data_source: crossroad_extra_untagged/val_data
annotation_conversion:
converter: mscoco_detection
annotation_file: crossroad_extra_untagged/val_coco_no_bg_vehicle.json
has_background: True
annotation: crossroad_extra_untagged_vehicle_labels_from_1.pickle
dataset_meta: crossroad_extra_untagged_vehicle_labels_from_1.json
- name: pedestrian_and_vehicle_dataset
data_source: val2017
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3636_OMZ_task11_pedestrian_and_vehicle_detector_adas_0001.xml
labels_file: pedestrian-and-vehicle-labels.json
has_background: True
annotation: pedestrian_and_vehicle.pickle
dataset_meta: pedestrian_and_vehicle.json
- name: pedestrian_detection_dataset
data_source: val2017
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3636_OMZ_task11_pedestrian_and_vehicle_detector_adas_0001.xml
labels_file: pedestrian-detection-labels.json
has_background: True
annotation: pedestrian_detection.pickle
dataset_meta: pedestrian_detection.json
- name: vehicle_detection_dataset
data_source: val2017
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3636_OMZ_task11_pedestrian_and_vehicle_detector_adas_0001.xml
labels_file: vehicle-detection-labels.json
has_background: True
annotation: vehicle_detection.pickle
dataset_meta: vehicle_detection.json
- name: barrier_vehicle_detection_dataset_index_class_2
data_source: val2017
annotation_conversion:
converter: cvat_object_detection
annotation_file: 3636_OMZ_task11_pedestrian_and_vehicle_detector_adas_0001.xml
labels_file: vehicle-detection-labels.json
has_background: True
annotation: vehicle_detection.pickle
dataset_meta: vehicle_detection.json
- name: synthetic_chinese_license_plates
data_source: Synthetic_Chinese_License_Plates
annotation_conversion:
converter: lpr_txt
annotation_file: Synthetic_Chinese_License_Plates/annotation
decoding_dictionary_file: dict
annotation: lpr.pickle
dataset_meta: lpr.json
preprocessing:
- type: resize
dst_width: 94
dst_height: 24
- name: image_retrieval
data_source: textile_crops
annotation_conversion:
converter: image_retrieval
data_dir: textile_crops
gallery_annotation_file: textile_crops/gallery/gallery.txt
queries_annotation_file: textile_crops/queries/quieries.txt
annotation: textile.pickle
dataset_meta: textile.json
preprocessing:
- type: resize
size: 224
- name: lfw
data_source: LFW/lfw
annotation_conversion:
converter: lfw
pairs_file: LFW/annotation/pairs.txt
landmarks_file: LFW/annotation/lfw_landmark.txt
annotation: lfw.pickle
metrics:
- type: pairwise_accuracy_subsets
subset_number: 2
- name: lfw_mtcnn_align
# converted from LFW public dataset with using
# script https://github.com/davidsandberg/facenet/blob/master/src/align/align_dataset_mtcnn.py
data_source: lfw_mtcnnpy_160
annotation_conversion:
converter: lfw
pairs_file: LFW/annotation/pairs.txt
extension: png
annotation: lfw_mtcnn_align.pickle
- name: ICDAR2015
data_source: ICDAR15_DET/ch4_test_images
annotation_conversion:
converter: icdar_detection
data_dir: ICDAR15_DET/gt
annotation: icdar15_detection.pickle
- name: ICDAR2015_recognition
data_source: ic15_1811
annotation_conversion:
converter: lmdb_text_recognition_database
lower_case: true
data_dir: ic15_1811
annotation: icdar15_recognition.pickle
reader:
type: lmdb_reader
- name: ICDAR03_recognition
data_source: ic03_867
annotation_conversion:
converter: lmdb_text_recognition_database
lower_case: true
data_dir: ic03_867
annotation: icdar03_recognition.pickle
reader:
type: lmdb_reader
- name: SVT_recognition
data_source: svt_647
annotation_conversion:
converter: lmdb_text_recognition_database
lower_case: true
data_dir: svt_647
annotation: svt_recognition.pickle
reader:
type: lmdb_reader
- name: IIIT5K
data_source: IIIT5K_3000
annotation_conversion:
converter: lmdb_text_recognition_database
lower_case: true
data_dir: IIIT5K_3000
annotation: IIIT5K_3000.pickle
reader:
type: lmdb_reader
- name: ICDAR2015_word_spotting
data_source: ICDAR15_DET/ch4_test_images
annotation_conversion:
converter: icdar_detection
word_spotting: True
data_dir: ICDAR15_DET/gt_proper
annotation: icdar15_detection.pickle
- name: ICDAR2013_detection
data_source: ICDAR13_DET
annotation_conversion:
converter: icdar_detection
data_dir: ICDAR13_DET/gt
annotation: icdar13_detection.pickle
- name: ICDAR2013
data_source: ICDAR13_REC/Challenge2_Test_Task3_Images
annotation_conversion:
converter: icdar13_recognition
annotation_file: ICDAR13_REC/gt/gt.txt.fixed.alfanumeric
annotation: icdar13_recognition.pickle
dataset_meta: icdar13_recognition.json
- name: ICDAR2013_detection
data_source: ICDAR13_DET
annotation_conversion:
converter: icdar_detection
data_dir: ICDAR13_DET/gt
annotation: icdar13_detection.pickle
- name: im2latex_medium_rendered
data_source: im2latex_medium_rendered/images_processed
annotation_conversion:
converter: im2latex_formula_recognition
data_dir: im2latex_medium_rendered
images_dir: im2latex_medium_rendered/images_processed
formula_file: im2latex_medium_rendered/formulas.norm.lst
split_file: im2latex_medium_rendered/validate_filter.lst
vocab_file: im2latex_medium_rendered/vocab.json
annotation: im2latex_medium_rendered.pickle
dataset_meta: im2latex_medium_rendered.json
- name: im2latex_medium_photographed
data_source: im2latex_medium_photographed/images_processed
annotation_conversion:
converter: im2latex_formula_recognition
data_dir: im2latex_medium_photographed
images_dir: im2latex_medium_photographed/images_processed
formula_file: im2latex_medium_photographed/formulas.norm.lst
split_file: im2latex_medium_photographed/test_filter.lst
vocab_file: im2latex_medium_photographed/vocab.pkl
annotation: im2latex_medium_photographed.pickle
dataset_meta: im2latex_medium_photographed.json
- name: im2latex_polynomials_handwritten
data_source: im2latex_polynomials_handwritten/images_processed
annotation_conversion:
converter: im2latex_formula_recognition
data_dir: im2latex_polynomials_handwritten
images_dir: im2latex_polynomials_handwritten/images_processed
formula_file: im2latex_polynomials_handwritten/formulas.norm.lst
split_file: im2latex_polynomials_handwritten/validate_filter.lst
vocab_file: im2latex_polynomials_handwritten/vocab.json
annotation: im2latex_polynomials_handwritten.pickle
dataset_meta: im2latex_polynomials_handwritten.json
- name: market1501
data_source: Market-1501-v15.09.15
annotation_conversion:
converter: market1501_reid
data_dir: Market-1501-v15.09.15
annotation: market1501_reid.pickle
- name: veri776
data_source: VeRi-776
annotation_conversion:
converter: veri776_reid
data_dir: VeRi-776
annotation: veri776_reid.pickle
- name: reid_dataset
data_source: Market-1501-v15.09.15
annotation_conversion:
converter: market1501_reid
data_dir: Market-1501-v15.09.15
annotation: market1501_reid.pickle
- name: vgg2face
data_source: VGGFaces2/test
annotation_conversion:
converter: vgg_face
landmarks_csv_file: VGGFaces2/bb_landmark/loose_landmark_test.csv
bbox_csv_file: VGGFaces2/bb_landmark/loose_bb_test.csv
annotation: vggfaces2.pickle
dataset_meta: vggfaces2.json
- name: semantic_segmentation_adas
data_source: segmentation/images
segmentation_masks_source: segmentation/mask_segmentation_adas
annotation_conversion:
converter: common_semantic_segmentation
images_dir: segmentation/images
masks_dir: segmentation/mask_segmentation_adas
image_postfix: .JPEG
mask_postfix: .png
dataset_meta: segmentation/mask_segmentation_adas/dataset_meta.json
annotation: semantic_segmentation_adas.pickle
dataset_meta: semantic_segmentation_adas.json
preprocessing:
- type: resize
dst_height: 1024
dst_width: 2048
postprocessing:
- type: encode_segmentation_mask
apply_to: annotation
- type: resize_segmentation_mask
apply_to: annotation
dst_height: 1024
dst_width: 2048
- name: road_segmentation
data_source: segmentation/images
segmentation_masks_source: segmentation/mask_road_segmentation
annotation_conversion:
converter: common_semantic_segmentation
images_dir: segmentation/images
masks_dir: segmentation/mask_road_segmentation
image_postfix: .JPEG
mask_postfix: .png
dataset_meta: segmentation/mask_road_segmentation/dataset_meta.json
annotation: road_segmentation.pickle
dataset_meta: road_segmentation.json
preprocessing:
- type: resize
dst_height: 512
dst_width: 896
postprocessing:
- type: encode_segmentation_mask
apply_to: annotation
- type: resize_segmentation_mask
apply_to: annotation
dst_height: 512
dst_width: 896
metrics:
- type: mean_iou
presenter: print_vector
- type: mean_accuracy
presenter: print_vector
- name: super_resolution_x3
data_source: super_resolution
annotation_conversion:
converter: super_resolution
data_dir: super_resolution
lr_suffix: lr_x3
upsample_suffix: upsample_x3
hr_suffix: hr
two_streams: True
annotation: super_resolution_x3.pickle
preprocessing:
- type: auto_resize
postprocessing:
- type: resize
apply_to: prediction
metrics:
- type: psnr
scale_border: 4
presenter: print_vector
- name: super_resolution_x4
data_source: super_resolution
annotation_conversion:
converter: super_resolution
data_dir: super_resolution
lr_suffix: lr_x4
upsample_suffix: upsample_x4
hr_suffix: hr
two_streams: True
annotation: super_resolution_x4.pickle
preprocessing:
- type: auto_resize
postprocessing:
- type: resize
apply_to: prediction
metrics:
- type: psnr
scale_border: 4
presenter: print_vector
- name: text_super_resolution_x3
data_source: super_resolution
annotation_conversion:
converter: super_resolution
data_dir: super_resolution
lr_suffix: lr_x3
hr_suffix: hr_gray
annotation: text_super_resolution_x3.pickle
preprocessing:
- type: bgr_to_gray
- type: auto_resize
postprocessing:
- type: resize
apply_to: prediction
metrics:
- type: psnr
scale_border: 4
presenter: print_vector
- name: head_pose
data_source: WIDER_val/images/16--Award_Ceremony
annotation: head_pose.pickle
preprocessing:
- type: crop_rect
- type: resize
size: 60
- name: gaze_estimation_dataset
data_source: gaze_estimation
annotation: gaze_estimation.pickle
reader:
type: combine_reader
scheme:
".*.png": opencv_imread
".*.json":
type: json_reader
key: head_pose_angles
- name: handwritten_score_recognition
data_source: ILSVRC2012_img_val
annotation: handwritten_score_recognition.pickle
dataset_meta: handwritten_score_recognition.json
- name: cmu_panoptic_keypoints
data_source: cmu_panoptic_subset
annotation_conversion:
converter: cmu_panoptic_keypoints
data_dir: cmu_panoptic_subset
annotation: cmu_panoptic_keypoints.pickle
- name: kinetics-400
data_source: kinetics/frames_val
annotation_conversion:
converter: clip_action_recognition
annotation_file: kinetics/kinetics_400.json
data_dir: kinetics/frames_val
annotation: kinetics_action_recognition.pickle
dataset_meta: kinetics_action_recognition.json
- name: kinetics-400-frames-79
data_source: kinetics/frames_val
annotation_conversion:
converter: clip_action_recognition
annotation_file: kinetics/kinetics_400.json
data_dir: kinetics/frames_val
clip_duration: 79
annotation: kinetics_action_recognition_79.pickle
dataset_meta: kinetics_action_recognition_79.json
- name: kinetics-400-frames-79-400
data_source: kinetics/frames_val
annotation_conversion:
converter: clip_action_recognition
annotation_file: kinetics/kinetics_400.json
data_dir: kinetics/frames_val
clip_duration: 79
temporal_stride: 1
num_samples: 400
annotation: kinetics_action_recognition_79_400.pickle
dataset_meta: kinetics_action_recognition_79_400.json
- name: driver_action_recognition_dataset
data_source: kinetics/frames_val
annotation_conversion:
converter: clip_action_recognition
annotation_file: kinetics/driver_action_recognition.json
data_dir: kinetics/frames_val
annotation: driver_action_recognition.pickle
dataset_meta: driver_action_recognition.json
- name: BraTS
data_source: BraTS
reader: numpy_reader
annotation_conversion:
converter: brats_numpy
data_dir: BraTS
ids_file: BraTS/val_ids.p
labels_file: BraTS/labels
annotation: brats.pickle
dataset_meta: brats.json
- name: BraTS_2017
data_source: BraTS_2017
reader:
type: nifti_reader
channels_first: True
annotation_conversion:
converter: brats
data_dir: BraTS_2017
labels_file: BraTS_2017/labels
mask_channels_first: True
annotation: brats2017.pickle
dataset_meta: brats2017.json
- name: product_detection
annotation: product_detection.pickle
dataset_meta: product_detection.json
data_source: product_detection
metrics:
- type: coco_precision
- name: ms_coco_single_keypoints
data_source: val2017
annotation_conversion:
converter: mscoco_single_keypoints
annotation_file: person_keypoints_val2017.json
annotation: mscoco_single_keypoints.pickle
dataset_meta: mscoco_single_keypoints.json
metrics:
- name: AP
type: coco_orig_keypoints_precision
- name: CamVid
data_source: CamVid
annotation_conversion:
converter: camvid
annotation_file: CamVid/val.txt
annotation: camvid.pickle
dataset_meta: camvid.json
- name: msasl-100
data_source: msasl/global_crops
annotation_conversion:
converter: continuous_clip_action_recognition
annotation_file: msasl/msasl_100.txt
data_dir: msasl/global_crops
out_fps: 15
clip_length: 16
annotation: msasl_action_recognition.pickle
- name: jester
data_source: jester/global_crops
annotation_conversion:
converter: continuous_clip_action_recognition
annotation_file: jester/val.txt
data_dir: jester/global_crops
out_fps: 15
clip_length: 8
annotation: jester_action_recognition.pickle
- name: common_sign_language_gestures
data_source: common_sign_language_gestures/global_crops
annotation_conversion:
converter: continuous_clip_action_recognition
annotation_file: common_sign_language_gestures/test.txt
data_dir: common_sign_language_gestures/global_crops
out_fps: 15
clip_length: 8
img_prefix: ''
annotation: common_sign_language_recognition.pickle
- name: ReDWeb_V1
data_source: ReDWeb_V1
annotation_conversion:
converter: redweb
data_dir: ReDWeb_V1
annotation_file: ReDWeb_validation_360.txt