Parcourir la source

New svd metric model

Jerome Buisine il y a 5 ans
Parent
commit
61e39389c4

+ 294 - 0
classification_cnn_keras_svd_img.py

@@ -0,0 +1,294 @@
+'''This script goes along the blog post
+"Building powerful image classification models using very little data"
+from blog.keras.io.
+```
+data/
+    train/
+        final/
+            final001.png
+            final002.png
+            ...
+        noisy/
+            noisy001.png
+            noisy002.png
+            ...
+    validation/
+        final/
+            final001.png
+            final002.png
+            ...
+        noisy/
+            noisy001.png
+            noisy002.png
+            ...
+```
+'''
+import sys, os, getopt
+import json
+
+from keras.preprocessing.image import ImageDataGenerator
+from keras.models import Sequential
+from keras.layers import Conv2D, MaxPooling2D, AveragePooling2D
+from keras.layers import Activation, Dropout, Flatten, Dense, BatchNormalization
+from keras.optimizers import Adam
+from keras.regularizers import l2
+from keras import backend as K
+from keras.utils import plot_model
+
+import tensorflow as tf
+import numpy as np
+
+from modules.model_helper import plot_info
+from modules.image_metrics import svd_metric
+
+# preprocessing of images
+from path import Path
+from PIL import Image
+import shutil
+import time
+
+##########################################
+# Global parameters (with default value) #
+#### ######################################
+img_width, img_height = 100, 100
+
+train_data_dir = 'data_svd_**img_size**/train'
+validation_data_dir = 'data_svd_**img_size**/validation'
+nb_train_samples = 7200
+nb_validation_samples = 3600
+epochs = 50
+batch_size = 16
+
+input_shape = (3, img_width, img_height)
+
+###########################################
+
+def init_directory(img_size, generate_data):
+
+    img_size_str = str(img_size)
+
+    svd_data_folder = str('data_svd_' + img_size_str)
+
+    if os.path.exists(svd_data_folder) and 'y' in generate_data:
+        print("Removing all previous data...")
+
+        shutil.rmtree(svd_data_folder)
+
+    if not os.path.exists(svd_data_folder):
+        print("Creating new data... Just take coffee... Or two...")
+        os.makedirs(str(train_data_dir.replace('**img_size**', img_size_str) + '/final'))
+        os.makedirs(str(train_data_dir.replace('**img_size**', img_size_str) + '/noisy'))
+
+        os.makedirs(str(validation_data_dir.replace('**img_size**', img_size_str) + '/final'))
+        os.makedirs(str(validation_data_dir.replace('**img_size**', img_size_str) + '/noisy'))
+    
+        for f in Path('./data').walkfiles():
+            if 'png' in f:
+                img = Image.open(f)
+                new_img = svd_metric.get_s_model_data_img(img)
+                new_img_path = f.replace('./data', str('./' + svd_data_folder))
+                new_img.save(new_img_path)
+                print(new_img_path)
+
+
+'''
+Method which returns model to train
+@return : DirectoryIterator
+'''
+def generate_model():
+
+    model = Sequential()
+
+    model.add(Conv2D(100, (2, 2), input_shape=input_shape))
+    model.add(Activation('relu'))
+    model.add(BatchNormalization())
+    model.add(MaxPooling2D(pool_size=(2, 2)))
+
+    model.add(Conv2D(60, (2, 2), input_shape=input_shape))
+    model.add(Activation('relu'))
+    model.add(BatchNormalization())
+    model.add(MaxPooling2D(pool_size=(2, 2)))
+
+    model.add(Conv2D(40, (2, 2)))
+    model.add(Activation('relu'))
+    model.add(MaxPooling2D(pool_size=(2, 2)))
+
+    model.add(Conv2D(30, (2, 2)))
+    model.add(Activation('relu'))
+    model.add(MaxPooling2D(pool_size=(2, 2)))
+
+    model.add(Flatten())
+    model.add(Dense(150, kernel_regularizer=l2(0.01)))
+    model.add(BatchNormalization())
+    model.add(Activation('relu'))
+    model.add(Dropout(0.2))
+
+    model.add(Dense(120, kernel_regularizer=l2(0.01)))
+    model.add(BatchNormalization())
+    model.add(Activation('relu'))
+    model.add(Dropout(0.2))
+
+    model.add(Dense(80, kernel_regularizer=l2(0.01)))
+    model.add(BatchNormalization())
+    model.add(Activation('relu'))
+    model.add(Dropout(0.2))
+
+    model.add(Dense(40, kernel_regularizer=l2(0.01)))
+    model.add(BatchNormalization())
+    model.add(Activation('relu'))
+    model.add(Dropout(0.2))
+
+    model.add(Dense(20, kernel_regularizer=l2(0.01)))
+    model.add(BatchNormalization())
+    model.add(Activation('relu'))
+    model.add(Dropout(0.1))
+
+    model.add(Dense(1))
+    model.add(Activation('sigmoid'))
+
+    model.compile(loss='binary_crossentropy',
+                  optimizer='rmsprop',
+                  metrics=['accuracy'])
+
+    return model
+
+'''
+Method which loads train data
+@return : DirectoryIterator
+'''
+def load_train_data():
+
+    # this is the augmentation configuration we will use for training
+    train_datagen = ImageDataGenerator(
+        rescale=1. / 255,
+        #shear_range=0.2,
+        #zoom_range=0.2,
+        #horizontal_flip=True,
+        #preprocessing_function=svd_metric.get_s_model_data_img
+        )
+
+    train_generator = train_datagen.flow_from_directory(
+        train_data_dir,
+        target_size=(img_width, img_height),
+        batch_size=batch_size,
+        class_mode='binary')
+
+    return train_generator
+
+'''
+Method which loads validation data
+@return : DirectoryIterator
+'''
+def load_validation_data():
+
+    # this is the augmentation configuration we will use for testing:
+    # only rescaling
+    test_datagen = ImageDataGenerator(
+        rescale=1. / 255,
+        #preprocessing_function=svd_metric.get_s_model_data_img
+        )
+
+    validation_generator = test_datagen.flow_from_directory(
+        validation_data_dir,
+        target_size=(img_width, img_height),
+        batch_size=batch_size,
+        class_mode='binary')
+
+    return validation_generator
+
+def main():
+
+    # update global variable and not local
+    global batch_size
+    global epochs   
+    global input_shape
+    global train_data_dir
+    global validation_data_dir
+    global nb_train_samples
+    global nb_validation_samples 
+
+    if len(sys.argv) <= 1:
+        print('Run with default parameters...')
+        print('classification_cnn_keras_svd.py --directory xxxx --output xxxxx --batch_size xx --epochs xx --img xx --generate (y/n)')
+        sys.exit(2)
+    try:
+        opts, args = getopt.getopt(sys.argv[1:], "ho:d:b:e:i:g", ["help", "output=", "directory=", "batch_size=", "epochs=", "img=", "generate="])
+    except getopt.GetoptError:
+        # print help information and exit:
+        print('classification_cnn_keras_svd.py --directory xxxx --output xxxxx --batch_size xx --epochs xx --img xx --generate (y/n)')
+        sys.exit(2)
+    for o, a in opts:
+        if o == "-h":
+            print('classification_cnn_keras_svd.py --directory xxxx --output xxxxx --batch_size xx --epochs xx --img xx --generate (y/n)')
+            sys.exit()
+        elif o in ("-o", "--output"):
+            filename = a
+        elif o in ("-b", "--batch_size"):
+            batch_size = int(a)
+        elif o in ("-e", "--epochs"):
+            epochs = int(a)
+        elif o in ("-d", "--directory"):
+            directory = a
+        elif o in ("-i", "--img"):
+            image_size = int(a)
+        elif o in ("-g", "--generate"):
+            generate_data = a
+        else:
+            assert False, "unhandled option"
+
+    # 3 because we have 3 color canals
+    if K.image_data_format() == 'channels_first':
+        input_shape = (3, img_width, img_height)
+    else:
+        input_shape = (img_width, img_height, 3)
+
+    img_str_size = str(image_size)
+    train_data_dir = str(train_data_dir.replace('**img_size**', img_str_size))
+    validation_data_dir = str(validation_data_dir.replace('**img_size**', img_str_size))
+
+    # configuration
+    with open('config.json') as json_data:
+        d = json.load(json_data)
+
+        try:
+            nb_train_samples = d[str(image_size)]['nb_train_samples']
+            nb_validation_samples = d[str(image_size)]['nb_validation_samples']
+        except:
+             print("--img parameter missing of invalid (--image_width xx --img_height xx)")
+             sys.exit(2)
+
+
+    init_directory(image_size, generate_data)
+    # load of model
+    model = generate_model()
+    model.summary()
+
+    if(directory):
+        print('Your model information will be saved into %s...' % directory)
+
+    history = model.fit_generator(
+        load_train_data(),
+        steps_per_epoch=nb_train_samples // batch_size,
+        epochs=epochs,
+        validation_data=load_validation_data(),
+        validation_steps=nb_validation_samples // batch_size)
+
+    # if user needs output files
+    if(filename):
+
+        # update filename by folder
+        if(directory):
+            # create folder if necessary
+            if not os.path.exists(directory):
+                os.makedirs(directory)
+            filename = directory + "/" + filename
+
+        # save plot file history
+        plot_info.save(history, filename)
+
+        plot_model(model, to_file=str(('%s.png' % filename)), show_shapes=True)
+        model.save_weights(str('%s.h5' % filename))
+
+
+if __name__ == "__main__":
+    main()

+ 1 - 1
generate_dataset.py

@@ -62,7 +62,7 @@ def main():
         sys.exit(2)
     for o, a in opts:
 
-        if o == "-h":
+        if o == "--help":
             print('generate_dataset.py --nb xxxx')
             print('20x20 : 1600')
             print('40x40 : 400')

+ 24 - 0
modules/image_metrics/svd_metric.py

@@ -2,12 +2,16 @@
 
 from numpy.linalg import svd
 from PIL import Image
+import matplotlib.pyplot as plt
 from scipy import misc
 
 import time
+
 import numpy as np
 from sklearn import preprocessing
 
+import modules.model_helper.image_conversion as img_c 
+
 '''
 Method which extracts SVD features from image and returns 's' vector
 @return 's' vector
@@ -23,6 +27,26 @@ def get_s_model_data(image):
 
     return result
 
+def get_s_model_data_img(image):
+    fig_size = plt.rcParams["figure.figsize"]
+    fig_size[0] = 1
+    fig_size[1] = 1
+    plt.rcParams["figure.figsize"] = fig_size
+
+    U, s, V = svd(image, full_matrices=False)
+    
+    plt.figure()   # create a new figure
+      
+    plt.plot(s[:, 0])
+    plt.plot(s[:, 1])
+    plt.plot(s[:, 2])
+
+    img = img_c.fig2img(plt.gcf())
+
+    plt.close('all')
+
+    return img
+
 def get(image):
     return svd(image, full_matrices=False)
 

+ 32 - 0
modules/model_helper/image_conversion.py

@@ -0,0 +1,32 @@
+from PIL import Image
+
+import numpy as np
+
+def fig2data(fig):
+    """
+    @brief Convert a Matplotlib figure to a 4D numpy array with RGBA channels and return it
+    @param fig a matplotlib figure
+    @return a numpy 3D array of RGBA values
+    """
+    # draw the renderer
+    fig.canvas.draw()
+ 
+    # Get the RGBA buffer from the figure
+    w,h = fig.canvas.get_width_height()
+    buf = np.fromstring(fig.canvas.tostring_rgb(), dtype=np.uint8)
+    buf.shape = (w, h, 3)
+ 
+    # canvas.tostring_argb give pixmap in ARGB mode. Roll the ALPHA channel to have it in RGBA mode
+    buf = np.roll(buf, 3, axis=2)
+    return buf
+    
+def fig2img(fig):
+    """
+    @brief Convert a Matplotlib figure to a PIL Image in RGBA format and return it
+    @param fig a matplotlib figure
+    @return a Python Imaging Library (PIL) image : default size (480,640,3)
+    """
+    # put the figure pixmap into a numpy array
+    buf = fig2data(fig)
+    w, h, d = buf.shape
+    return Image.frombytes("RGB", (w, h), buf.tostring())

+ 1 - 1
modules/model_helper/plot_info.py

@@ -2,7 +2,7 @@
 
 # avoid tk issue
 import matplotlib
-matplotlib.use('agg')
+#matplotlib.use('agg')
 import matplotlib.pyplot as plt
 
 '''

+ 1 - 0
requirements.txt

@@ -5,3 +5,4 @@ sklearn
 image_slicer
 pydot
 matplotlib
+path.py