pain commited on
Commit
7866cb1
·
verified ·
1 Parent(s): 822f7c2

Training in progress epoch 0

Browse files
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ step_*
2
+ epoch_*
README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Code: [More Information Needed]
9
+ - Paper: [More Information Needed]
10
+ - Docs: [More Information Needed]
best_model/README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Code: [More Information Needed]
9
+ - Paper: [More Information Needed]
10
+ - Docs: [More Information Needed]
best_model/best_model_info.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_epoch": 0,
3
+ "best_metric": 0.06449081003665924,
4
+ "metric_name": "map",
5
+ "all_metrics": {
6
+ "map": 0.06449081003665924,
7
+ "map_50": 0.135758638381958,
8
+ "map_75": 0.05395117774605751,
9
+ "map_small": 0.0023179275449365377,
10
+ "map_medium": 0.02037796564400196,
11
+ "map_large": 0.15389712154865265,
12
+ "mar_1": 0.08930881321430206,
13
+ "mar_10": 0.1069440096616745,
14
+ "mar_100": 0.10772563517093658,
15
+ "mar_small": 0.0025136969052255154,
16
+ "mar_medium": 0.0381893515586853,
17
+ "mar_large": 0.2603810429573059,
18
+ "map_per_class": [
19
+ 0.07780038565397263,
20
+ 0.0017598657868802547,
21
+ 0.007928945124149323,
22
+ 0.03651401400566101,
23
+ 0.011985496617853642,
24
+ 0.004355185199528933,
25
+ 0.05744529142975807,
26
+ 0.028387103229761124,
27
+ 0.20311343669891357,
28
+ 0.060909710824489594,
29
+ 0.06694461405277252,
30
+ 0.21674565970897675
31
+ ],
32
+ "mar_100_per_class": [
33
+ 0.17803029716014862,
34
+ 0.03787878900766373,
35
+ 0.025114620104432106,
36
+ 0.07053571194410324,
37
+ 0.02178817056119442,
38
+ 0.009627403691411018,
39
+ 0.09702970087528229,
40
+ 0.050704225897789,
41
+ 0.27803030610084534,
42
+ 0.11810719221830368,
43
+ 0.12135417014360428,
44
+ 0.28450703620910645
45
+ ],
46
+ "classes": [
47
+ 0,
48
+ 1,
49
+ 2,
50
+ 3,
51
+ 4,
52
+ 5,
53
+ 6,
54
+ 7,
55
+ 8,
56
+ 9,
57
+ 10,
58
+ 11
59
+ ]
60
+ }
61
+ }
best_model/config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dinov3_model_name": "facebook/dinov3-vits16-pretrain-lvd1689m",
3
+ "expected_channels": [
4
+ 96,
5
+ 192,
6
+ 384,
7
+ 768
8
+ ],
9
+ "freeze_adapter": false,
10
+ "freeze_backbone": true,
11
+ "freeze_encoder": true,
12
+ "freeze_pixel_level": true,
13
+ "hub_token": null,
14
+ "id2label": {
15
+ "0": "Crosswalk - Plain",
16
+ "1": "Driveway",
17
+ "2": "Person",
18
+ "3": "Bicyclist",
19
+ "4": "Street Light",
20
+ "5": "Pole",
21
+ "6": "Traffic Sign Frame",
22
+ "7": "Bicycle",
23
+ "8": "Bus",
24
+ "9": "Car",
25
+ "10": "Motorcycle",
26
+ "11": "Truck"
27
+ },
28
+ "label2id": {
29
+ "Bicycle": 7,
30
+ "Bicyclist": 3,
31
+ "Bus": 8,
32
+ "Car": 9,
33
+ "Crosswalk - Plain": 0,
34
+ "Driveway": 1,
35
+ "Motorcycle": 10,
36
+ "Person": 2,
37
+ "Pole": 5,
38
+ "Street Light": 4,
39
+ "Traffic Sign Frame": 6,
40
+ "Truck": 11
41
+ },
42
+ "use_gradient_checkpointing": false,
43
+ "use_improved_adapter": true
44
+ }
best_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1657d8627ce7d688c0aaedfb47a7788492fdf201f30a7f7a9271576a6f23067c
3
+ size 975804864
best_model/preprocessor_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": false,
5
+ "device": null,
6
+ "disable_grouping": null,
7
+ "do_center_crop": null,
8
+ "do_convert_rgb": null,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_reduce_labels": true,
12
+ "do_rescale": true,
13
+ "do_resize": true,
14
+ "ignore_index": 12,
15
+ "image_mean": [
16
+ 0.48500001430511475,
17
+ 0.4560000002384186,
18
+ 0.4059999883174896
19
+ ],
20
+ "image_processor_type": "Mask2FormerImageProcessorFast",
21
+ "image_std": [
22
+ 0.2290000021457672,
23
+ 0.2239999920129776,
24
+ 0.22499999403953552
25
+ ],
26
+ "input_data_format": null,
27
+ "num_labels": 12,
28
+ "pad_size": null,
29
+ "reduce_labels": true,
30
+ "resample": 2,
31
+ "rescale_factor": 0.00392156862745098,
32
+ "return_tensors": null,
33
+ "size": {
34
+ "height": 512,
35
+ "width": 512
36
+ },
37
+ "size_divisor": 32
38
+ }
config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dinov3_model_name": "facebook/dinov3-vits16-pretrain-lvd1689m",
3
+ "expected_channels": [
4
+ 96,
5
+ 192,
6
+ 384,
7
+ 768
8
+ ],
9
+ "freeze_adapter": false,
10
+ "freeze_backbone": true,
11
+ "freeze_encoder": true,
12
+ "freeze_pixel_level": true,
13
+ "hub_token": null,
14
+ "id2label": {
15
+ "0": "Crosswalk - Plain",
16
+ "1": "Driveway",
17
+ "2": "Person",
18
+ "3": "Bicyclist",
19
+ "4": "Street Light",
20
+ "5": "Pole",
21
+ "6": "Traffic Sign Frame",
22
+ "7": "Bicycle",
23
+ "8": "Bus",
24
+ "9": "Car",
25
+ "10": "Motorcycle",
26
+ "11": "Truck"
27
+ },
28
+ "label2id": {
29
+ "Bicycle": 7,
30
+ "Bicyclist": 3,
31
+ "Bus": 8,
32
+ "Car": 9,
33
+ "Crosswalk - Plain": 0,
34
+ "Driveway": 1,
35
+ "Motorcycle": 10,
36
+ "Person": 2,
37
+ "Pole": 5,
38
+ "Street Light": 4,
39
+ "Traffic Sign Frame": 6,
40
+ "Truck": 11
41
+ },
42
+ "use_gradient_checkpointing": false,
43
+ "use_improved_adapter": true
44
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1657d8627ce7d688c0aaedfb47a7788492fdf201f30a7f7a9271576a6f23067c
3
+ size 975804864
preprocessor_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": false,
5
+ "device": null,
6
+ "disable_grouping": null,
7
+ "do_center_crop": null,
8
+ "do_convert_rgb": null,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_reduce_labels": true,
12
+ "do_rescale": true,
13
+ "do_resize": true,
14
+ "ignore_index": 12,
15
+ "image_mean": [
16
+ 0.48500001430511475,
17
+ 0.4560000002384186,
18
+ 0.4059999883174896
19
+ ],
20
+ "image_processor_type": "Mask2FormerImageProcessorFast",
21
+ "image_std": [
22
+ 0.2290000021457672,
23
+ 0.2239999920129776,
24
+ 0.22499999403953552
25
+ ],
26
+ "input_data_format": null,
27
+ "num_labels": 12,
28
+ "pad_size": null,
29
+ "reduce_labels": true,
30
+ "resample": 2,
31
+ "rescale_factor": 0.00392156862745098,
32
+ "return_tensors": null,
33
+ "size": {
34
+ "height": 512,
35
+ "width": 512
36
+ },
37
+ "size_divisor": 32
38
+ }