praeclarumjj3 commited on
Commit
3049ce4
1 Parent(s): 28e2d10
Files changed (1) hide show
  1. config.json +45 -56
config.json CHANGED
@@ -2107,51 +2107,19 @@
2107
  "use_bfloat16": false,
2108
  "window_size": 12
2109
  },
2110
- "decoder_config": {
2111
- "common_stride": 4,
2112
- "conv_dim": 256,
2113
- "decoder_layers": 10,
2114
- "dim_feedforward": 2048,
2115
- "dropout": 0.1,
2116
- "encoder_feedforward_dim": 1024,
2117
- "encoder_layers": 6,
2118
- "enforce_input_proj": false,
2119
- "hidden_dim": 256,
2120
- "mask_dim": 256,
2121
- "norm": "GN",
2122
- "num_heads": 8,
2123
- "pre_norm": false,
2124
- "query_dec_layers": 2,
2125
- "use_task_norm": true
2126
- },
2127
- "general_config": {
2128
- "class_weight": 2.0,
2129
- "contrastive_temperature": 0.07,
2130
- "contrastive_weight": 0.5,
2131
- "deep_supervision": true,
2132
- "dice_weight": 5.0,
2133
- "ignore_value": 255,
2134
- "importance_sample_ratio": 0.75,
2135
- "init_std": 0.02,
2136
- "init_xavier_std": 1.0,
2137
- "is_train": false,
2138
- "layer_norm_eps": 1e-05,
2139
- "mask_weight": 5.0,
2140
- "no_object_weight": 0.1,
2141
- "num_classes": 133,
2142
- "num_queries": 150,
2143
- "output_auxiliary_logits": true,
2144
- "oversample_ratio": 3.0,
2145
- "strides": [
2146
- 4,
2147
- 8,
2148
- 16,
2149
- 32
2150
- ],
2151
- "train_num_points": 12544,
2152
- "use_auxiliary_loss": true
2153
- },
2154
- "hidden_size": 256,
2155
  "id2label": {
2156
  "0": "person",
2157
  "1": "bicycle",
@@ -2287,8 +2255,11 @@
2287
  "131": "wall-other-merged",
2288
  "132": "rug-merged"
2289
  },
 
 
2290
  "init_std": 0.02,
2291
  "init_xavier_std": 1.0,
 
2292
  "label2id": {
2293
  "airplane": 4,
2294
  "apple": 47,
@@ -2424,21 +2395,39 @@
2424
  "wine glass": 40,
2425
  "zebra": 22
2426
  },
 
 
 
 
2427
  "model_type": "oneformer",
 
 
2428
  "num_attention_heads": 8,
 
2429
  "num_hidden_layers": 10,
 
2430
  "output_attentions": true,
 
2431
  "output_hidden_states": true,
2432
- "text_encoder_config": {
2433
- "max_seq_len": 77,
2434
- "task_seq_len": 77,
2435
- "text_encoder_context_length": 77,
2436
- "text_encoder_n_ctx": 16,
2437
- "text_encoder_num_layers": 6,
2438
- "text_encoder_proj_layers": 2,
2439
- "text_encoder_vocab_size": 49408,
2440
- "text_encoder_width": 256
2441
- },
 
 
 
 
 
 
2442
  "torch_dtype": "float32",
2443
- "transformers_version": null
 
 
 
2444
  }
 
2107
  "use_bfloat16": false,
2108
  "window_size": 12
2109
  },
2110
+ "class_weight": 2.0,
2111
+ "common_stride": 4,
2112
+ "contrastive_temperature": 0.07,
2113
+ "contrastive_weight": 0.5,
2114
+ "conv_dim": 256,
2115
+ "decoder_layers": 10,
2116
+ "dice_weight": 5.0,
2117
+ "dim_feedforward": 2048,
2118
+ "dropout": 0.1,
2119
+ "encoder_feedforward_dim": 1024,
2120
+ "encoder_layers": 6,
2121
+ "enforce_input_proj": false,
2122
+ "hidden_dim": 256,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2123
  "id2label": {
2124
  "0": "person",
2125
  "1": "bicycle",
 
2255
  "131": "wall-other-merged",
2256
  "132": "rug-merged"
2257
  },
2258
+ "ignore_value": 255,
2259
+ "importance_sample_ratio": 0.75,
2260
  "init_std": 0.02,
2261
  "init_xavier_std": 1.0,
2262
+ "is_training": false,
2263
  "label2id": {
2264
  "airplane": 4,
2265
  "apple": 47,
 
2395
  "wine glass": 40,
2396
  "zebra": 22
2397
  },
2398
+ "layer_norm_eps": 1e-05,
2399
+ "mask_dim": 256,
2400
+ "mask_weight": 5.0,
2401
+ "max_seq_len": 77,
2402
  "model_type": "oneformer",
2403
+ "no_object_weight": 0.1,
2404
+ "norm": "GN",
2405
  "num_attention_heads": 8,
2406
+ "num_classes": 133,
2407
  "num_hidden_layers": 10,
2408
+ "num_queries": 150,
2409
  "output_attentions": true,
2410
+ "output_auxiliary_logits": true,
2411
  "output_hidden_states": true,
2412
+ "oversample_ratio": 3.0,
2413
+ "pre_norm": false,
2414
+ "query_dec_layers": 2,
2415
+ "strides": [
2416
+ 4,
2417
+ 8,
2418
+ 16,
2419
+ 32
2420
+ ],
2421
+ "task_seq_len": 77,
2422
+ "text_encoder_context_length": 77,
2423
+ "text_encoder_n_ctx": 16,
2424
+ "text_encoder_num_layers": 6,
2425
+ "text_encoder_proj_layers": 2,
2426
+ "text_encoder_vocab_size": 49408,
2427
+ "text_encoder_width": 256,
2428
  "torch_dtype": "float32",
2429
+ "train_num_points": 12544,
2430
+ "transformers_version": null,
2431
+ "use_auxiliary_loss": true,
2432
+ "use_task_norm": true
2433
  }