From 71200f13553cc8bebf253b202a0427a8bce1b80f Mon Sep 17 00:00:00 2001
From: tomrink <rink@ssec.wisc.edu>
Date: Wed, 6 Apr 2022 10:52:06 -0500
Subject: [PATCH] minor

---
 modules/deeplearning/unet.py | 22 ++++++++++------------
 1 file changed, 10 insertions(+), 12 deletions(-)

diff --git a/modules/deeplearning/unet.py b/modules/deeplearning/unet.py
index d0c23939..53cfff79 100644
--- a/modules/deeplearning/unet.py
+++ b/modules/deeplearning/unet.py
@@ -533,7 +533,8 @@ class UNET:
 
         input_2d = self.inputs[0]
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=5, strides=1, padding=padding, activation=None)(input_2d)
-        print(conv.shape)
+        print('Contracting Branch')
+        print('input: ', conv.shape)
         skip = conv
 
         if NOISE_TRAINING:
@@ -545,7 +546,6 @@ class UNET:
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=5, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.MaxPool2D(padding=padding)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
 
         skip = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=None)(skip)
         skip = tf.keras.layers.MaxPool2D(padding=padding)(skip)
@@ -553,7 +553,7 @@ class UNET:
 
         conv = conv + skip
         conv = tf.keras.layers.LeakyReLU()(conv)
-        print(conv.shape)
+        print('1d: ', conv.shape)
         # -----------------------------------------------------------------------------------------------------------
 
         conv_2 = conv
@@ -562,7 +562,6 @@ class UNET:
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.MaxPool2D(padding=padding)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
 
         skip = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=None)(skip)
         skip = tf.keras.layers.MaxPool2D(padding=padding)(skip)
@@ -570,7 +569,7 @@ class UNET:
 
         conv = conv + skip
         conv = tf.keras.layers.LeakyReLU()(conv)
-        print(conv.shape)
+        print('2d: ', conv.shape)
         # ----------------------------------------------------------------------------------------------------------
 
         conv_3 = conv
@@ -579,7 +578,6 @@ class UNET:
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.MaxPool2D(padding=padding)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
 
         skip = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=None)(skip)
         skip = tf.keras.layers.MaxPool2D(padding=padding)(skip)
@@ -587,6 +585,7 @@ class UNET:
 
         conv = conv + skip
         conv = tf.keras.layers.LeakyReLU()(conv)
+        print('3d: ', conv.shape)
         # -----------------------------------------------------------------------------------------------------------
 
         conv_4 = conv
@@ -595,7 +594,6 @@ class UNET:
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.MaxPool2D(padding=padding)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
 
         skip = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=None)(skip)
         skip = tf.keras.layers.MaxPool2D(padding=padding)(skip)
@@ -603,6 +601,7 @@ class UNET:
 
         conv = conv + skip
         conv = tf.keras.layers.LeakyReLU()(conv)
+        print('4d: ', conv.shape)
 
         # Expanding (Decoding) branch -------------------------------------------------------------------------------
         print('expanding branch')
@@ -612,26 +611,25 @@ class UNET:
         conv = tf.keras.layers.concatenate([conv, conv_4])
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
+        print('5: ', conv.shape)
 
         num_filters /= 2
         conv = tf.keras.layers.Conv2DTranspose(num_filters, kernel_size=3, strides=2, padding=padding)(conv)
         conv = tf.keras.layers.concatenate([conv, conv_3])
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
+        print('6: ', conv.shape)
 
         num_filters /= 2
         conv = tf.keras.layers.Conv2DTranspose(num_filters, kernel_size=3, strides=2, padding=padding)(conv)
         conv = tf.keras.layers.concatenate([conv, conv_2])
         conv = tf.keras.layers.Conv2D(num_filters, kernel_size=3, strides=1, padding=padding, activation=activation)(conv)
         conv = tf.keras.layers.BatchNormalization()(conv)
-        print(conv.shape)
+        print('7: ', conv.shape)
 
         num_filters /= 2
         conv = tf.keras.layers.Conv2DTranspose(num_filters, kernel_size=3, strides=2, padding=padding)(conv)
-        conv = tf.keras.layers.concatenate([conv, conv_1])
-        print(conv.shape)
+        print('8: ', conv.shape)
 
         if NumClasses == 2:
             activation = tf.nn.sigmoid  # For binary
-- 
GitLab