Synthesis Images noise detection metrics developed including all approaches using SVD or others compression methods

Jérôme BUISINE fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
analysis 914e4bc50c Refactoring of the whole project 5 anni fa
data_processing fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
display 914e4bc50c Refactoring of the whole project 5 anni fa
generate fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
modules @ 139aa3c231 4a684e884e Refactoring of all display script 5 anni fa
others fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
prediction 1a06c5d683 Update of prediction scripts path used 5 anni fa
run fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
simulation fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
.gitignore 914e4bc50c Refactoring of the whole project 5 anni fa
.gitmodules 4a684e884e Refactoring of all display script 5 anni fa
LICENSE ee12267d9c Update repo for new organization 5 anni fa
README.md fab9fb5111 Change metric to feature key words in some scripts 5 anni fa
__init__.py 914e4bc50c Refactoring of the whole project 5 anni fa
custom_config.py 28240cb45e Add of new computed feature 5 anni fa
data_attributes.py 28240cb45e Add of new computed feature 5 anni fa
dataset 4a684e884e Refactoring of all display script 5 anni fa
models.py 914e4bc50c Refactoring of the whole project 5 anni fa
requirements.txt 5911c7da90 New models creation 6 anni fa
train_keras_svd.py 4a684e884e Refactoring of all display script 5 anni fa
train_model.py 914e4bc50c Refactoring of the whole project 5 anni fa

README.md

Noise detection using SVM

Requirements

pip install -r requirements.txt

Generate all needed data for each features (which requires the the whole dataset. In order to get it, you need to contact us).

python generate/generate_all_data.py --feature all

For noise detection, many features are available:

  • lab
  • mscn
  • mscn_revisited
  • low_bits_2
  • low_bits_4
  • low_bits_5
  • low_bits_6
  • low_bits_4_shifted_2

You can also specify feature you want to compute and image step to avoid some images:

python generate/generate_all_data.py --feature mscn --step 50
  • step: keep only image if image id % 50 == 0 (assumption is that keeping spaced data will let model better fit).

How to use

Multiple directories and scripts are available:

  • fichiersSVD_light/*: all scene files information (zones of each scene, SVD descriptor files information and so on...).
  • train_model.py: script which is used to run specific model available.
  • data/*: folder which will contain all .train & .test files in order to train model.
  • saved_models/*.joblib: all scikit learn models saved.
  • models_info/*: all markdown files generated to get quick information about model performance and prediction.
  • results: This folder contains model_comparisons.csv obtained after running runAllmaxwell*.sh script.
  • modules/*: contains all modules usefull for the whole project (such as configuration variables)

Scripts for generating data files

Two scripts can be used for generating data in order to fit model:

  • generate_data_model.py: zones are specified and stayed fixed for each scene
  • generate_data_model_random.py: zones are chosen randomly (just a number of zone is specified)
  • generate_data_model_random_maxwell.py: zones are chosen randomly (just a number of zone is specified). Only maxwell scene are used.

Remark: Note here that all python script have --help command.

python generate/generate_data_model.py --help

python generate/generate_data_model.py --output xxxx --interval 0,20  --kind svdne --scenes "A, B, D" --zones "0, 1, 2" --percent 0.7 --sep: --rowindex 1 --custom custom_min_max_filename

Parameters explained:

  • output: filename of data (which will be split into two parts, .train and .test relative to your choices).
  • interval: the interval of data you want to use from SVD vector.
  • kind: kind of data ['svd', 'svdn', 'svdne']; not normalize, normalize vector only and normalize together.
  • scenes: scenes choice for training dataset.
  • zones: zones to take for training dataset.
  • percent: percent of data amount of zone to take (choose randomly) of zone
  • sep: output csv file seperator used
  • rowindex: if 1 then row will be like that 1:xxxxx, 2:xxxxxx, ..., n:xxxxxx
  • custom: specify if you want your data normalized using interval and not the whole singular values vector. If it is, the value of this parameter is the output filename which will store the min and max value found. This file will be usefull later to make prediction with model (optional parameter).

Train model

This is an example of how to train a model

python train_model.py --data 'data/xxxxx.train' --output 'model_file_to_save' --choice 'model_choice'

Expected values for the choice parameter are ['svm_model', 'ensemble_model', 'ensemble_model_v2'].

Predict image using model

Now we have a model trained, we can use it with an image as input:

python prediction/predict_noisy_image_svd.py --image path/to/image.png --interval "x,x" --model saved_models/xxxxxx.joblib --feature 'lab' --mode 'svdn' --custom 'min_max_filename'
  • feature: feature choice need to be one of the listed above.
  • custom: specify filename with custom min and max from your data interval. This file was generated using custom parameter of one of the generate_data_model*.py script (optional parameter).

The model will return only 0 or 1:

  • 1 means noisy image is detected.
  • 0 means image seem to be not noisy.

All SVD features developed need:

  • Name added into feature_choices_labels global array variable of modules/utils/config.py file.
  • A specification of how you compute the feature into get_svd_data method of modules/utils/data_type.py file.

Predict scene using model

Now we have a model trained, we can use it with an image as input:

python prediction/prediction_scene.py --data path/to/xxxx.csv --model saved_model/xxxx.joblib --output xxxxx --scene xxxx

Remark: scene parameter expected need to be the correct name of the Scene.

Visualize data

All scripts with names display/display_*.py are used to display data information or results.

Just use --help option to get more information.

Simulate model on scene

All scripts named predict_seuil_expe*.py are used to simulate model prediction during rendering process. Do not forget the custom parameter filename if necessary.

Once you have simulation done. Checkout your threshold_map/%MODEL_NAME%/simulation_curves_zones_*/ folder and use it with help of display_simulation_curves.py script.

Others scripts

Test model on all scene data

In order to see if a model well generalized, a bash script is available:

bash others/testModelByScene.sh '100' '110' 'saved_models/xxxx.joblib' 'svdne' 'lab'

Parameters list:

  • 1: Begin of interval of data from SVD to use
  • 2: End of interval of data from SVD to use
  • 3: Model we want to test
  • 4: Kind of data input used by trained model
  • 5: feature used by model

Get treshold map

Main objective of this project is to predict as well as a human the noise perception on a photo realistic image. Human threshold is available from training data. So a script was developed to give the predicted treshold from model and compare predicted treshold from the expected one.

python prediction/predict_seuil_expe.py --interval "x,x" --model 'saved_models/xxxx.joblib' --mode ["svd", "svdn", "svdne"] --feature ['lab', 'mscn', ...] --limit_detection xx --custom 'custom_min_max_filename'

Parameters list:

  • model: mode file saved to use
  • interval: the interval of data you want to use from SVD vector.
  • mode: kind of data ['svd', 'svdn', 'svdne']; not normalize, normalize vector only and normalize together.
  • limit_detection: number of not noisy images found to stop and return threshold (integer).
  • custom: custom filename where min and max values are stored (optional parameter).

Display model performance information

Another script was developed to display into Mardown format the performance of a model.

The content will be divised into two parts:

  • Predicted performance on all scenes
  • Treshold maps obtained from model on each scenes

The previous script need to already have ran to obtain and display treshold maps on this markdown file.

python others/save_model_result_in_md.py --interval "xx,xx" --model saved_models/xxxx.joblib --mode ["svd", "svdn", "svdne"] --feature ['lab', 'mscn']

Parameters list:

  • model: mode file saved to use
  • interval: the interval of data you want to use from SVD vector.
  • mode: kind of data ['svd', 'svdn', 'svdne']; not normalize, normalize vector only and normalize together.

Markdown file with all information is saved using model name into models_info folder.

Others...

All others bash scripts are used to combine and run multiple model combinations...

License

The MIT license