diff --git a/cnn_training.py b/cnn_training.py index f69c795..8ff0eee 100644 --- a/cnn_training.py +++ b/cnn_training.py @@ -76,6 +76,7 @@ def main( conv_0_enable_softmax=bool(config["conv_0_enable_softmax"]), conv_0_power_softmax=float(config["conv_0_power_softmax"]), conv_0_meanmode_softmax=bool(config["conv_0_meanmode_softmax"]), + conv_0_no_input_mode_softmax=bool(config["conv_0_no_input_mode_softmax"]), scale_data=int(config["scale_data"]), use_scheduler=bool(config["use_scheduler"]), use_adam=bool(config["use_adam"]), @@ -120,6 +121,7 @@ def run_network( conv_0_enable_softmax: bool, conv_0_power_softmax: float, conv_0_meanmode_softmax: bool, + conv_0_no_input_mode_softmax: bool, scale_data: float, use_scheduler: bool, use_adam: bool, @@ -249,6 +251,7 @@ def run_network( l_relu_negative_slope=leak_relu_negative_slope, conv_0_power_softmax=conv_0_power_softmax, conv_0_meanmode_softmax=conv_0_meanmode_softmax, + conv_0_no_input_mode_softmax=conv_0_no_input_mode_softmax, ).to(device) logger.info(model) diff --git a/config.json b/config.json index f47f0b4..6edeb9e 100644 --- a/config.json +++ b/config.json @@ -29,10 +29,15 @@ "minimum_learning_rate": 1e-10, "learning_rate": 1e-4, // <- LR Scheduler + // type of pooling layer -> "pooling_type": "max", // (max), average, none + // <- type of pooling layer + // convolution layer 0 softmax parameters -> "conv_0_enable_softmax": true, // true, (false) "conv_0_power_softmax": 0.0, // (0.0) = Exp "conv_0_meanmode_softmax": false, // true, false + "conv_0_no_input_mode_softmax": false, // true, (false) + // <- convolution layer 0 softmax parameters "use_adam": true, // (true) => adam, false => SGD "condition": "Coignless", "scale_data": 255.0, // (255.0)