From 019ef4b06370203cb067d6593ff4375eb88224f4 Mon Sep 17 00:00:00 2001
From: tomrink <rink@ssec.wisc.edu>
Date: Tue, 6 Dec 2022 16:40:43 -0600
Subject: [PATCH] snapshot..

---
 modules/deeplearning/srcnn_l1b_l2.py | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git a/modules/deeplearning/srcnn_l1b_l2.py b/modules/deeplearning/srcnn_l1b_l2.py
index c8d8d643..7b47664d 100644
--- a/modules/deeplearning/srcnn_l1b_l2.py
+++ b/modules/deeplearning/srcnn_l1b_l2.py
@@ -106,12 +106,12 @@ y_2 = y_67
 
 
 def build_residual_conv2d_block(conv, num_filters, block_name, activation=tf.nn.relu, padding='SAME',
-                                kernel_initializer='he_uniform', scale=None,
+                                kernel_initializer='he_uniform', scale=None, kernel_size=3,
                                 do_drop_out=True, drop_rate=0.5, do_batch_norm=False):
 
     with tf.name_scope(block_name):
-        skip = tf.keras.layers.Conv2D(num_filters, kernel_size=3, padding=padding, kernel_initializer=kernel_initializer, activation=activation)(conv)
-        skip = tf.keras.layers.Conv2D(num_filters, kernel_size=3, padding=padding, activation=None)(skip)
+        skip = tf.keras.layers.Conv2D(num_filters, kernel_size=kernel_size, padding=padding, kernel_initializer=kernel_initializer, activation=activation)(conv)
+        skip = tf.keras.layers.Conv2D(num_filters, kernel_size=kernel_size, padding=padding, activation=None)(skip)
 
         if scale is not None:
             skip = tf.keras.layers.Lambda(lambda x: x * scale)(skip)
-- 
GitLab