aboutsummaryrefslogtreecommitdiff
path: root/cfg/bouncingballs/bb-level1-run1.json
blob: 1e6f21b785e1b603c45b6e1d0db5b62617053fe2 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
{
   "model_path": "bb_baseline",
   "datatype": "bouncingballs",
   "dataset": "BOUNCINGBALLS_OCCLUSION",
   "scenario": "occlusion",
   "num_updates": 0,
   "max_epochs": 1000,
   "max_updates": 400000,
   "learning_rate": {
    "lr": 0.0004,
    "deacrease_lr_every": 100000,
    "deacrease_lr_factor": 0.75
   },
   "blackout": {
    "blackout_start_timestep": 10,
    "blackout_rate": 0.1,
    "blackout_rate_max": 0.45,
    "blackout_increase_every": 10000,
    "blackout_increase_rate": 0.01
   },
   "phases": {
    "start_inner_loop": 40000,
    "shufleslots_end": 3000000,
    "entity_pretraining_phase2_end": 30000,
    "entity_pretraining_phase1_end": 15000,
    "background_pretraining_end": 0
   },
   "defaults": {
    "num_workers": 2,
    "prefetch_factor": 2,
    "statistics_offset": 10,
    "load_optimizers": false,
    "teacher_forcing": 5,
    "skip_frames": 1,
    "error_dropout": 0.1,
    "seed": 67
   },
   "bptt": {
    "bptt_start_timestep": 0,
    "bptt_steps": 1,
    "bptt_steps_max": 3,
    "increase_bptt_steps_every": 100000
   },
   "model": {
    "level": 2,
    "batch_size": 128,
    "num_objects": 3,
    "img_channels": 3,
    "input_size": [
     64,
     64
    ],
    "latent_size": [
     4,
     4
    ],
    "gestalt_size": 12,
    "bottleneck": "binar",
    "position_regularizer": 0.01,
    "time_regularizer": 0.25,
    "encoder_regularizer": 0.333333,
    "latent_regularizer": 0.0,
    "inner_loop_enabled": false,
    "latent_loss_enabled": false,
    "encoder": {
     "channels": 24,
     "level1_channels": 12,
     "num_layers": 2,
     "reg_lambda": 1e-10
    },
    "predictor": {
     "heads": 2,
     "layers": 3,
     "channels_multiplier": 2,
     "reg_lambda": 1e-10,
     "transformer_type": "shared"
    },
    "decoder": {
     "channels": 24,
     "level1_channels": 12,
     "num_layers": 2
    },
    "update_module": {
     "reg_lambda": 1e-05
    }
   }
  }