Red Analyze Focused Supervised

Overview of Red Analyze Architectures

The Red Analyze tool is used to inspect images, find defect pixels, and distinguish defective images from normal ones or vice versa. There are 2 types of architectures in Red Analyze Tool: Focused (Focused Supervised, Focused Unsupervised) and High Detail.

 

  • The Red Analyze Focused Supervised is used to segment specific regions such as defects or other areas of interest. Be it blowholes in cast metal or bruised vegetables on a conveyor; the Red Analyze tool in Supervised mode can identify all of these and many more problems simply by learning the varying appearance of the defect or target region. To train the Red Supervised tool, all you need to provide are images of the type of defective regions you are looking for.

  • The Red Analyze Focused Unsupervised is used to detect anomalies and aesthetic defects. Be it scratches on a decorated surface, incomplete or improper assemblies or even weaving problems in textiles; the Red Analyze tool can identify all of these and many more problems simply by learning the normal appearance of an object including its significant but tolerable variations. To train the Red Unsupervised tool, all you need to provide are images of good objects.

  • The Red Analyze High Detail is the reinforced version of Red Analyze Focused – Supervised in terms of segmentation performance but costs some drop in training and processing speed. Its higher performance compared to Red Analyze Focused – Supervised comes from its unique training architecture in segmentation tasks.

 

About Red Analyze Focused Supervised

With Red Analyze Focused Supervised, Red Analyze tool is taught the appearance of defects. As such it does not (at least explicitly) form a model of the inspected part and as a consequence is much less dependent on part configuration, type or the conditions during image acquisition. However, the Red Analyze tool in Focused - Supervised mode will need to form an explicit model of the different types of defects. Thus, it needs both good and bad samples to train. In particular, it requires a representative collection of the latter (which are often quite challenging to obtain).

Also, while in Focused - Supervised mode, a Red Analyze tool can be used to search for differences other than defects. The tool can be used to identify different regions in an image that are intended to be there. A Red Analyze tool in Focused - Supervised mode could be used to label the targeted regions on the training images in order to be able find and mark similar type of regions on untrained images after training.

When the Red Analyze tool is in Focused Supervised, the focus is on teaching the network what defects look like. During training, only images that are labeled will be considered during training. Images which contain labeled defect regions are trained to learn the defects. In addition, the tool also learns the parts of the labeled images which do not include defect regions. In general, when you are training a Red Analyze tool in Supervised mode, you are training the tool to detect and respond to defects. Therefore, the most import guideline for labeling images is that if an image is labeled as containing a defect, each and every defect in the image must be labeled.

When you label an image as Good in Supervised mode (in other words, an image that does not contain any defect regions), the tool will also use that image for training. In particular, the tool will attempt to train the network so that images labeled Good do not generate any defect responses. Adding labeled, defect-free Good images to your Training Image Set can help you validate the performance of the tool in classifying good and bad images.

When selecting images, you will want to ensure that you have a training image set that includes all of the defects you expect to encounter during runtime, as well as defect-free images. If you did not teach the tool what a particular defect looks like, the tool will not find those defects. For example, if you only teach the tool what stains look like, the tool will not find scratches.

 

Architectures: Red Analyze Focused Supervised vs Red Analyze High Detail

High Detail mode uses an architecture which is different from that of Focused mode. Due to its architectural difference, it doesn't have Sampling Parameters in Tool Parameters pane as it samples from the entire view. Because of this, High Detail mode takes more time for Training/Processing than the Focused mode, but you can get more accurate or detailed results in pixel levels. The way you label images and create a model in High Detail mode is basically the same as Focused mode but there are some differences in tool parameters. As well as Red Analyze Focused Supervised, only the binary classes (Good/Bad) are supported in Red Analyze High Detail. It means that multiple classes are not supported.

 

  Focused - Supervised High Detail (Supervised)
Training/Processing Time Short Long
Results

Accurate

More Accurate

Image Dataset Composition Training Set, Test Set Training Set, Validation Set, Test Set

 

Supported Features vs Architectures

Features \ Architectures

Red Analyze Focused Supervised,

Red Analyze Focused Unsupervised

Red Analyze High Detail
Loss Inspector Not supported Supported
Validation Set Not used in training Used in training
VisionPro Deep Learning Tool Parameters Less parameters

More parameters for control*,
No sampling parameters

* More training and perturbation parameters for detailed training control and granular adjustment

Note: To exploit all functions of each tool, go to Help and enable Expert Mode.

 

Training Workflow for Red Analyze Focused Supervised

When a Red Analyze tool is in Red Analyze Focused Supervised mode, the training workflow of the tool is:

 

  1. Launch VisionPro Deep Learning.
  2. Create a new workspace or import an existing one into VisionPro Deep Learning.
  3. Collect images and load them into VisionPro Deep Learning.
  4. Define ROI (Region of Interest) to construct Views.
    1. If a pose from a Blue Locate tool is being used to transform the orientation of the View being used as an input to the Red Analyze tool, process the images (press the Scissors icon) before opening the Red Analyze tool. For more details, see ROI Options Following a Blue Locate Tool.
    2. If necessary, adjust the Region of Interest (ROI). Within the Display Area, right-click and select Edit ROI from the menu.

    3. After adjusting the ROI, press the Apply button and the adjusted ROI will be applied to all of the images.
    4. Press the Close button on the toolbar to continue.
  5. If there is extraneous information in the image, add appropriate masks to exclude those areas of the image. Within the Display Area, right-click and select Edit Mask from the menu.
    1. From the Mask toolbar, select and edit the appropriate masks.

    2. After adding the necessary mask(s), press the Apply button and the mask will be applied to the current image.

      If you click Apply All and click Yes button on the ensuing Apply Mask dialog after adding the necessary masks, the same mask will be applied on all the images. If you click No on the dialog, the mask will not be applied and you go back to the Edit Mask window.

    3. Press the Close button on the toolbar to continue.
  6. Go through all of the images and label the defects in the images. Make sure that all of the images have been labeled. See Create Label (Labeling) for the details of labeling.
  7. Split the entire images into the training images and the test images. Utilize image sets to properly divide them into the training and test group. Add images to the training set.
    1. Select the images in the View Browser, and on the right-click pop-up menu click Add views to training set. To select multiple images in the View Browser, use the Shift + Left Mouse Button.
    2. Or, use Display Filters to show the desired images for the training only and add them to the training set by clicking Actions for ... views → Add views to training set.
  8. Prior to training, you need to set parameters in Tool Parameters. You can configure Training, Sampling, and Perturbation parameters or just use the default values of these. See Configure Tool Parameters for the details of the supported parameters.
    1. Ensure that the Feature Size parameter in Sampling parameters has been set. The Feature Size parameter gives the network a clue to the size of the defects that you are interested in detecting. So, if the Feature Size parameter setting is larger than the defects in your application, there is a good chance the tool will not identify the defects in your image.
    2. You can set the Feature Size by either manually adjusting the parameter value, or graphically by re-sizing the interactive Feature Size graphic.
    3. If you want more granular control over training or processing, turn on Expert Mode on Help menu to initialize additional parameters in Tool Parameters.
  9. Train the tool by pressing the Brain icon.
    1. If you stop training in the middle of it by pressing the Stop icon, you can stop training but you will lose the current tool so far trained.
  10. After training, review the results. Open Database Overview panel and review the Scores/ROC graph, the Confusion Matrix, and its F1 Score with switching over between categories in Count drop-down list. Review Precision, Recall, F-Score (Region Area Metrics) to understand the results in pixel level. See Interpret Results for the details of interpreting results.

  11. After reviewing the results, go through all of the images and see how the tool correctly or incorrectly marked the defects in the image.
    1. If the tool has correctly marked the features, right-click the image and select Accept View.
    2. If the tool has incorrectly marked the defects, or failed to identify a present defect:
      1. Right-click the image again and select Clear Markings and Labels.
      2. Manually label the defective pixels.
    3. If you want to tune the result by affecting the processing, you can manipulate Processing parameters and re-process by clicking Magnifier icon to get the changed results. For example, to change the decision boundary which is determined by T1 and T2, you can manipulate the Threshold parameter.
    4. If you encountered the scenario in (a.), you are ready to use the tool. If you encountered the scenario in (b.), you will need to retrain the tool and repeat the steps 8 ~ 11.

 

The details of each step are explained in each subsection of Training Red Analyze Focused Supervised.