SkillAgentSearch skills...

PySODEvalToolkit

PySODEvalToolkit: A Python-based Evaluation Toolbox for Salient Object Detection and Camouflaged Object Detection

Install / Use

/learn @lartpang/PySODEvalToolkit

README

A Python-based image grayscale/binary segmentation evaluation toolbox.

中文文档

TODO

  • More flexible configuration script.
    • [ ] Use the yaml file that meets matplotlib requirements to control the drawing format.
    • [ ] Replace the json with a more flexible configuration format, such as yaml or toml.
  • [ ] Add test scripts.
  • [ ] Add more detailed comments.
  • Optimize the code for exporting evaluation results.
    • [x] Implement code to export results to XLSX files.
    • [ ] Optimize the code for exporting to XLSX files.
    • [ ] Consider if using a text format like CSV would be better? It can be opened as a text file and also organized using Excel.
  • [ ] Replace os.path with pathlib.Path.
  • [x] Improve the code for grouping data, supporting tasks like CoSOD, Video Binary Segmentation, etc.
  • [x] Support concurrency strategy to speed up computation. Retained support for multi-threading, removed the previous multi-process code.
    • [ ] Currently, due to the use of multi-threading, there is an issue with extra log information being written, which needs more optimization.
  • [X] Separate USVOS code into another repository PyDavis16EvalToolbox.
  • [X] Use more rapid and accurate metric code PySODMetrics as the evaluation benchmark.

[!tip]

  • Some methods provide result names that do not match the original dataset's ground truth names.
    • [Note] (2021-11-18) Currently, support is provided for both prefix and suffix names, so users generally do not need to change the names themselves.
    • [Optional] The provided script tools/rename.py can be used to rename files in bulk. Please use it carefully to avoid data overwriting.
    • [Optional] Other tools, such as rename on Linux, and Microsoft PowerToys on Windows.

Features

  • Benefiting from PySODMetrics, it supports a richer set of metrics. For more details, see utils/recorders/metric_recorder.py.
    • Supports evaluating grayscale images, such as predictions from saliency object detection (SOD) and camouflaged object detection (COD) tasks.
      • MAE
      • Emeasure
      • Smeasure
      • Weighted Fmeasure
      • Maximum/Average/Adaptive Fmeasure
      • Maximum/Average/Adaptive Precision
      • Maximum/Average/Adaptive Recall
      • Maximum/Average/Adaptive IoU
      • Maximum/Average/Adaptive Dice
      • Maximum/Average/Adaptive Specificity
      • Maximum/Average/Adaptive BER
      • Fmeasure-Threshold Curve (run eval.py with the metric fmeasure)
      • Emeasure-Threshold Curve (run eval.py with the metric em)
      • Precision-Recall Curve (run eval.py with the metrics precision and recall, this is different from previous versions as the calculation of precision and recall has been separated from fmeasure)
    • Supports evaluating binary images, such as common binary segmentation tasks.
      • Binary Fmeasure
      • Binary Precision
      • Binary Recall
      • Binary IoU
      • Binary Dice
      • Binary Specificity
      • Binary BER
  • Richer functions.
    • Supports evaluating models according to the configuration.
    • Supports drawing PR curves, F-measure curves and E-measure curves based on configuration and evaluation results.
    • Supports exporting results to TXT files.
    • Supports exporting results to XLSX files (re-supported on January 4, 2021).
    • Supports exporting LaTeX table code from generated .npy files, and marks the top three methods with different colors.
    • … :>.

How to Use

Installing Dependencies

Install the required libraries: pip install -r requirements.txt

The metric evaluation is based on my another project: PySODMetrics. Bug reports are welcome!

Configuring Paths for Datasets and Method Predictions

This project relies on json files to store data. Examples for dataset and method configurations are provided in ./examples: config_dataset_json_example.json and config_method_json_example.json. You can directly modify them for subsequent steps.

[!note]

  • Please note that since this project relies on OpenCV to read images, ensure that the path strings do not contain non-ASCII characters.
  • Make sure that the name of the dataset in the dataset configuration file matches the name of the dataset in the method configuration file. After preparing the json files, it is recommended to use the provided tools/check_path.py to check if the path information in the json files is correct.
<details> <summary> More Details on Configuration </summary>

Example 1: Dataset Configuration

Note, "image" is not necessary here. The actual evaluation only reads "mask".

{
    "LFSD": {
        "image": {
            "path": "Path_Of_RGBDSOD_Datasets/LFSD/Image",
            "prefix": "some_gt_prefix",
            "suffix": ".jpg"
        },
        "mask": {
            "path": "Path_Of_RGBDSOD_Datasets/LFSD/Mask",
            "prefix": "some_gt_prefix",
            "suffix": ".png"
        }
    }
}

Example 2: Method Configuration

{
    "Method1": {
        "PASCAL-S": {
            "path": "Path_Of_Method1/PASCAL-S",
            "prefix": "some_method_prefix",
            "suffix": ".png"
        },
        "ECSSD": {
            "path": "Path_Of_Method1/ECSSD",
            "prefix": "some_method_prefix",
            "suffix": ".png"
        },
        "HKU-IS": {
            "path": "Path_Of_Method1/HKU-IS",
            "prefix": "some_method_prefix",
            "suffix": ".png"
        },
        "DUT-OMRON": {
            "path": "Path_Of_Method1/DUT-OMRON",
            "prefix": "some_method_prefix",
            "suffix": ".png"
        },
        "DUTS-TE": {
            "path": "Path_Of_Method1/DUTS-TE",
            "suffix": ".png"
        }
    }
}

Here, path represents the directory where image data is stored. prefix and suffix refer to the prefix and suffix outside the common part in the names of the predicted images and the actual ground truth images.

During the evaluation process, the matching of method predictions and dataset ground truths is based on the shared part of the file names. Their naming patterns are preset as [prefix]+[shared-string]+[suffix]. For example, if there are predicted images like method1_00001.jpg, method1_00002.jpg, method1_00003.jpg and ground truth images gt_00001.png, gt_00002.png, gt_00003.png, then we can configure it as follows:

Example 3: Dataset Configuration

{
    "dataset1": {
        "mask": {
            "path": "path/Mask",
            "prefix": "gt_",
            "suffix": ".png"
        }
    }
}

Example 4: Method Configuration

{
    "method1": {
        "dataset1": {
            "path": "path/dataset1",
            "prefix": "method1_",
            "suffix": ".jpg"
        }
    }
}
</details>

Running the Evaluation

  • Once all the previous steps are correctly completed, you can begin the evaluation. For usage of the evaluation script, refer to the output of the command python eval.py --help.
  • Add configuration options according to your needs and execute the command. If there are no exceptions, it will generate result files with the specified filename.
    • If not all files are specified, it will directly output the results, as detailed in the help information of eval.py.
    • If --curves-npy is specified, the metrics information related to drawing will be saved in the corresponding .npy file.
  • [Optional] You can use tools/converter.py to directly export the LaTeX table code from the generated npy files.

Plotting Curves for Grayscale Image Evaluation

You can use plot.py to read the .npy file to organize and draw PR, F-measure, and E-measure curves for specified methods and datasets as needed. The usage of this script can be seen in the output of python plot.py --help. Add configuration items as per your requirement and execute the command.

The most basic instruction is to specify the values in the figure.figsize item in the configuration file according to the number of subplots reasonably.

A Basic Execution Process

Here I'll use the RGB SOD configuration in my local configs folder as an example (necessary modifications should be made according to the actual situation).

# Check Configuration Files
python tools/check_path.py --method-jsons configs/methods/rgb-sod/rgb_sod_methods.json --dataset-jsons configs/datasets/rgb_sod.json

# After ensuring there's nothing unreasonable in the output information, you can begin the evaluation with the following commands:
# --dataset-json: Set `configs/datasets/rgb_sod.json` as dataset configuration file
# --method-json: Set `configs/methods/rgb-sod/rgb_sod_methods.json` as method configuration file
# --metric-npy: Set `output/rgb_sod/metrics.npy` to store the metrics information in npy format
# --curves-npy: Set `output/rgb_sod/curves.npy` to store the curves information in npy format
# --record-txt: Set `output/rgb_sod/results.txt` to store the results information in text format
# --record-xlsx: Set `output/rgb_sod/results.xlsx` to store the results information in Excel format
# --metric-names: Specify `fmeasure em precision recall` as the metrics to be calculated
# --include-methods: Specify the methods from `configs/methods/rgb-sod/rgb_sod_methods.json` to be evaluated
# --include-datasets: Specify the datasets from `configs/datasets/rgb_sod.json` to be evaluated
python eval.py --dataset-json configs/datasets/rgb_sod.json --method-json configs/methods/rgb-sod/rgb_sod_methods.json --metric-npy output/rgb_sod/metrics.npy --curves-npy output/rgb_sod/curves.npy -
View on GitHub
GitHub Stars352
CategoryDevelopment
Updated4d ago
Forks33

Languages

Python

Security Score

100/100

Audited on Mar 23, 2026

No findings