Skip to content

Postprocessing is extermely slow for large numbers of repetitions #60

@brockho

Description

@brockho

When running the postprocessing on data, produced on the bbob-constrained suite with the latest cocoex code (2.8.1), a large(r) number of repetitions is envoked for the algorithm, I ran (e.g. about 25 on $f_7$). This causes the postprocessing to be extremely slow: postprocessing two of those large-repetition algorithms together with some from the data archive takes >5 hours to run, in particular the aggregated ECDFs take in the order of 30+ minutes per plot.

The output warns me about this:

[...]
  Will generate output data in folder ppdata\constrained_10000_10kRu_AL1-C_AL2-C_AL3-C_AL4-C_COBYL_et_al_101609h1557
    this might take several minutes.
ECDF graphs per noise group...
C:\Users\brockhof\AppData\Local\anaconda3\Lib\site-packages\cocopp\compall\pprldmany.py:700: UserWarning: Sample size equals 19530 which may take very long. This is likely to be unintended, hence a bug.
  warnings.warn(warntxt)
  done (Thu Oct 16 12:50:23 2025).
ECDF graphs per function group...
C:\Users\brockhof\AppData\Local\anaconda3\Lib\site-packages\cocopp\compall\pprldmany.py:700: UserWarning: Sample size equals 15960 which may take very long. This is likely to be unintended, hence a bug.
  warnings.warn(warntxt)
C:\Users\brockhof\AppData\Local\anaconda3\Lib\site-packages\cocopp\compall\pprldmany.py:700: UserWarning: Sample size equals 19530 which may take very long. This is likely to be unintended, hence a bug.
  warnings.warn(warntxt)
C:\Users\brockhof\AppData\Local\anaconda3\Lib\site-packages\cocopp\compall\pprldmany.py:700: UserWarning: Sample size equals 11280 which may take very long. This is likely to be unintended, hence a bug.
  warnings.warn(warntxt)
C:\Users\brockhof\AppData\Local\anaconda3\Lib\site-packages\cocopp\compall\pprldmany.py:700: UserWarning: Sample size equals 19530 which may take very long. This is likely to be unintended, hence a bug.
  warnings.warn(warntxt)
[...]

Together with @nikohansen, we nailed down the origin of the problem in the experiment itself where the large number of possible repetitions was 100 in my case (the default in the example experiment in python). In any case, the postprocessing should be able to deal with it---hence, I opened the issue here.

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions