Scippy

GCG

Branch-and-Price & Column Generation for Everyone

The Visualization Suite

On this page, we present and guide through the main visualization reporting capabilities of GCG. If you are looking for visualizations of decompositions please visit the guide on The Explore Menu.

In general, we want to visualize runtime and algorithmic behaviour, either for single test runs (Test Set Report) or for comparisons between different runs (Comparison Report). The scripts can automatically conduct test(s), generate all visualizations and afterwards compile them into a captioned report.


A Remark for CMake Users
Generally, we don't recommend to do testing and experiments using a CMake build, which is due to make check not generating the usual runtime data, see Manual Installation for more information). Therefore, automatic test set report generation is currently not supported under CMake builds. When using the manual mode instead, you have to give runtime data generated with a make installation using make test or with a CMake build using the cmake gcg_cluster target.

Reporting Functionalities

Note: For the generation of the report PDFs, the package latexmk is required.

In this section, we present the two main reporting functionalities of the Visualization Suite. Note that they are only tested under Linux operating systems.

Test Set Report

The test set report generates a PDF that includes all visualizations, both for all single instances and for aggregated statistics for the test set, to give you a compiled and captioned overview of visualizations of different aspects of GCG's algorithmics.
The report can be generated in two ways:

  1. Execute a test and generate the report (continue here)
  2. Have runtime data collected already and generate the report afterwards (continue here)

Automatically generate a Test Set Report

The automatic generation is very simple and straightforward and recommended, especially for beginners.
However, it will always (re-)generate all runtime data, which might be undesirable, in particular for very big test sets. For a different method, see here.

Preparation

  • For a full report, it is required to have compiled GCG with make STATISTICS=true. Otherwise, GCG will not print out extensive pricing and bounds statistics for the respective visualizations.
  • All other requirements listed on the visualization script page have to be fulfilled (e.g. correctly configured python).
  • The script will not print any warnings (e.g. if your python is not configured correctly) unless you set DEBUG=true in the settings file (see below).
  • You can create a script settings file, e.g. settings.vset where you can define which plots to generate and which arguments to generate them with. Furthermore, if you have a big testset or just want to check if the feature works, you can also enable a draft mode that will generate a reduced version of the report. A full list of possible settings can be found here.
  • To use the script settings, execute the test with VISUSETTINGS=settings.vset, while having settings.vset lying in the GCG root directory.

Generation

When wanting to generate the report automatically, you can work with the usual make test. It will support all flags, just as you are used to when doing tests, and the data generated during the test will remain in the usual place. To tell GCG to generate the report after the test has been executed, please add the flag VISU to your command, i.e.

make test VISU=true

This will make the test script call the report generator script afterwards, already with the correct arguments.

Output

An example of the generated Test Set Report can be found on this page:

Preview


As output, the script will create a folder called testsetreport_<testsetname>_<settingsname>_<timestamp> inside check/reports, unless defined otherwise inside the script settings file. Inside the report folder, you will find a folder logs including the out-file, res-file and vbc-folder of your test run. Next, you can find the folder plots, where, subdivided into the plots types, the visualizations are located. Finally, the report is inside the directory as a (LaTeX-generated) PDF, already compiled for you and opened automatically.

Manually generate a Test Set Report

The manual generation requires you to have the runtime data available already.
You can either give outfile, resfile and vbc folder via the script settings file, else you will be asked for them during runtime.

Preparation

  • For a full report, it is required to have generated your runtime data with both, a GCG compiled with make STATISTICS=true and a test executed with statistics make test STATISTICS=true. Otherwise, GCG will not print out extensive pricing and bounds statistics for the respective visualizations. Furthermore, for detection visualizations, DETECTIONSTATISTICS=true must be set during testing.
  • All other requirements listed on the visualization script page have to be fulfilled (e.g. correctly configured python).
  • The script will not print any warnings (e.g. if your python is not configured correctly) unless you set DEBUG=true in the settings file (see below).
  • You have to create a script settings file, e.g. settings.vset, where you have to define parameters of your given test run. A full list of possible settings can be found here, where it is also described which flags have to be defined for the scripts to yield results at all and those that can be defined for the report to have full information. Note that for make, none of the optional flags have to be set - they will be found automatically.
  • To use the script settings, execute the test with VISUSETTINGS=settings.vset, while having settings.vset lying in the GCG root directory.

Generation

In order to generate a test set report without testing, you can call the visualization target. This can be done using

make visu VISUSETTINGS=settings.vset DATADIR=<folder containing .res and .out file and vbc file folder>

The data directory can also be given via the settings file. If you give none, you will be asked for it when the script is called.

Output

You will get the same output as with the automated report generation. If visualizations are missing, please set the debug flag

DEBUG=true

inside your script settings file to troubleshoot.

Comparison Report

The comparison report generates a PDF with a comparison table and all comparing visualizations, to give you a compiled and captioned overview of visualizations of different aspects of your two or more test runs.
Note: We do not support an automated mode natively, check the version comparison script for a possibility to automatically generate the runtime data.

Preparation

  • For a full report, it is required to have generated your runtime data with both, a GCG compiled with make STATISTICS=true and a test executed with statistics make test STATISTICS=true. Otherwise, GCG will not print out extensive bounds statistics for the respective visualizations. Furthermore, for detection visualizations, DETECTIONSTATISTICS=true must be set during testing.
  • All other requirements listed on the visualization script page have to be fulfilled (e.g. correctly configured python), and additionally, you have to have the python package tikzplotlib.
  • The script will not print any warnings (e.g. if your python is not configured correctly) unless you set DEBUG=true in the settings file (see below).
  • You can create a script settings file, e.g. settings.vset where you can define which plots to generate and which arguments to generate them with. Furthermore, if you have a big testset or just want to check if the feature works, you can also enable a draft mode that will generate a reduced version of the report. A full list of possible settings can be found here.
  • To use the script settings, execute the test with VISUSETTINGS=settings.vset, while having settings.vset lying in the GCG root directory. By default, the script will assume that you conducted the tests with the flag that was set when you last compiled this binary. You can overwrite that by giving LAST_STATISTICS=true in the script settings file.

Generation

In order to generate a comparison report (always with already existing runtime data), you can call the visualization target, but with a data directory that contains multiple runs:

make visu VISUSETTINGS=settings.vset DATADIR=<folder containing multiple .res and .out files>

The data directory can also be given via the settings file. If you give none, you will be asked for it when the script is called.

Output

An example of the generated Comparison Report can be found on this page:

Preview


As output, the script will create a folder called comparisonreport_<timestamp> inside check/reports, unless defined otherwise inside the script settings file. Inside the report folder, you will find a folder logs including the out-files and res-files of your test runs, as well as a pickle folder that includes dataframes with your logs, but parsed. Next, you can find the folder plots, where, subdivided into the plots types, the comparing visualizations are located. Finally, the report is inside the directory as a (LaTeX-generated) PDF, already compiled for you and opened automatically.

If visualizations are missing, please set the debug flag

DEBUG=true

inside your script settings file to troubleshoot.

Manual Visualization Generation

Using the scripts "by hand" is the most versatile, yet complicated approach. For beginners, we recommend using the notebook (see next section).

With the GCG Visualization Suite, you can generate general, pricing, bounds, detection and tree statistics visualizations. Guides concerning the manual execution of these scripts can be found here.

Visualization Notebook

The notebook is a very easy entry into the GCG world of visualizations. You basically only need Python installed and can get started with exploring facets of the solving process of your experiment data created using the guide How to conduct experiments with GCG.

A preview of the Visualization Notebook can be found on this page:

Preview


Using our shipped Python-based Jupyter Notebook, you can easily find out more about how GCG solved your problems. In this section, we will explain how to get started with it.

Notebook Functionality (Changelog)

In general, the notebooks do not (yet) offer all the advanced functionality and plots that the scripts itself would offer.
However, they do offer easy customization through good documentation and hints at locations where one can add own code.

To get a preview of the notebook, click here.

Version 1.0

  • Configuration Framework Wrapper with descriptions
  • Plotting:
    • Time Distribution Plot (similar to the manual plotter in the Visualization Suite), includes time inclusion calculations
    • Tree Statistics Plot (imported from the Visualization Suite) that shows the explored nodes on each level of the branch-and-bound tree
    • Bounds Plot (imported from the Visualization Suite) visualizing the development of the primal and dual bound in the root node
  • Viewing:
    • Plot Export to a png file to share or publish
    • Web View with tooltips to explore runtime features
    • Interactive Filtering within the notebook (with automatic connection to the web view)

Using the Notebook

Preparations

Apart from a working installation of Python 3.6+ and Jupyter (we furthermore recommend an IDE that can display Jupyter notebooks such as VSCode), the following Python packages (installable via pip3 install) are required minimally:

numpy pandas matplotlib IPython subprocess

Getting Started

The visualization notebook is located in the stats folder. It will access the scripts that are in other subfolders of the stats directory. Everything that is required to get started with the notebook is already shipped with it, including descriptions in the markdown boxes of the notebook.

Troubleshooting

Currently, no bugs are known. If you encounter any, please contact us.