=Paper=
{{Paper
|id=Vol-2827/KBS-Paper_2
|storemode=property
|title=Paintings-from-Polygons: Simulated Annealing
|pdfUrl=https://ceur-ws.org/Vol-2827/KBS-Paper_2.pdf
|volume=Vol-2827
|authors=Redouane Dahmani,Sven Boogmans,Arne Meijs,Daan van den Berg
}}
==Paintings-from-Polygons: Simulated Annealing==
Paintings-from-Polygons: Simulated Annealing
Redouane Dahmania , Sven Boogmansb , Arne Meijsc and Daan van den Bergd
a
Institute for Interdisciplinary Studies (IIS), University of Amsterdam
b
Master Information Studies, University of Amsterdam
c
Minor Programmeren, University of Amsterdam
d
Informatics Institute, University of Amsterdam
Abstract
Paintings-from-polygons is an optimization problem in which the objective is to approximate a target
bitmap by optimally arranging a fixed number of semi-opaque colored polygons. In a recent report,
two optimization algorithms showed strong performance on the task, but the third, simulated anneal-
ing, failed miserably. The authors conjectured its poor performance to be due to the specific cooling
schedule used. In this study, we use the same problem instances but parameterize simulated annealing
with eight new cooling schedules known from literature: Geman & Geman with different π-parameter
settings, geometric, linear, cosine, linear reheat, sigmoid, and staircase. We find that the previous con-
jecture was right: the performance of simulated annealing on this problem critically relies on its cool-
ing schedule, and can be quite succesful once parameterized properly. Moreover, we find a consistent
βperformance hierarchyβ in which most of the cooling schedulesβ performance appears related to their
average temperatures only.
Keywords
Paintings-from-Polygons, Computational creativity, Evolutionary art, Paintings, Polygons,
Simulated annealing, Cooling schedules, NP-hard, Optimization, Heuristic algorithm
1. Introduction
As a discipline, computational creativity has flourished in recent history. Subjects such as ca-
sual creation, photo quality classification, music genre detection, evolutionary architecture and
visual imagery have gone through substantial development in the past decade [1][2][3][4][5][6].
Many of these are in some way related to optimization algorithms such as hillclimbing, plant
propagation, simulated annealing or genetic algorithms, which have all proven successful in
both theoretical and practical settings [7][8][9][10].
Being located on the intersection of computational creativity and algorithmic optimization,
the paintings-from-polygons (PFP) problemβs objective is to approximate a given target bitmap
by arranging a fixed number of semi-opaque colored polygons. On initialization, the polygon
constellationβs properties are assigned random values, but the numbers of polygon and vertices
Joint Proceedings of the ICCC 2020 Workshops (ICCC-WS 2020), September 7-11 2020, Coimbra (PT) / Online
" redouane.dahmani@student.uva.nl (R. Dahmani); sven.boogmans@student.uva.nl (S. Boogmans);
arne.meijs@student.uva.nl (A. Meijs); D.vandenBerg@uva.nl (D.v.d. Berg)
~ https://github.com/RedRedouane/Paintings-from-Polygons-Simulated-Annealing (R. Dahmani);
http://www.heuristieken.nl (D.v.d. Berg)
0000-0002-9422-3023 (R. Dahmani); 0000-0002-9422-3023 (S. Boogmans); 0000-0003-1747-730X (A. Meijs);
0000-0001-5060-3342 (D.v.d. Berg)
Β© 2020 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
Workshop
Proceedings
http://ceur-ws.org
ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org)
Figure 1: The choice of cooling schedule greatly influences the performance of simulated annealing
on paintings-from-polygons. Top to bottom: linear reheat, Geman & Geman (c=1), sigmoidal, and
geometric cooling schedules; the fifth box shows the MSE for the algorithmic approximation directly
above it. All constellations contain 1000 vertices in 250 polygons, and completed a run of 106 iterations
(eq.: evaluations). A video clip of the process is publicly available[11].
are fixed (to 250 and 1000 in this study) throughout an experiment [12]. From there, heuris-
tic algorithms can change a polygonβs color, move one of its vertices, transfer a vertex from
one polygon to the next, or mutate the drawing index of the polygons, a parameter that de-
termines the rendering sequence when generating an output bitmap. By iteratively applying
these mutations to the polygon constellation and rendering afterwards, a target bitmap, usually
a painting, is approximated ever closer (Fig. 1). The objective value for each rendered bitmap
is the difference to the target bitmap, expressed as the mean squared error (MSE) over the color
channels:
π₯β
π¦β
3
(π
ππππππππ β ππππππ‘π )2
πππΈ = β (1)
π=1 π₯ β
π¦
Here, π₯ and π¦ are the dimensions of the target bitmap in pixels, π
ππππππππ is the value of a
red, green or blue (RGB) channel in the bitmap rendered from the polygon constellation, and
ππππππ‘π is the value of the corresponding RGB-channel for the target bitmap. Since the maxi-
mum difference between two RGB channels is 255, and there are three channels in one pixel,
the maximum MSE is 2552 β
3 = 195075. The minimum MSE of zero is reached iff the rendered
polygon constellation and the target bitmap are identical, a situation that is highly unlikely for
any realistic number of vertices. A simplified form of PFP was proven to be NP-hard in 2020
[13]. Although this technically speaking guarantees nothing about the full PFP-problem, it has
no known subexponential exact methods, and thereby provides a suitable testing ground for
heuristic algorithms. It also gives rise to some interesting questions about the complexity1 of
visual imagery, and produces aesthetically pleasing results.
There are 2563 β 7.93 β
10312,107 different existable bitmaps of 180 by 240 pixels, the
(180β
240)
default size in the experiments. A maximum of 180 β
240 β
4 = 172, 800 vertices is therefore suffi-
cient to exactly reproduce any given target bitmap of aformentioned dimensions[12]. However,
Paauw and Van den Berg also give a lower bound of 39,328 vertices, the lowest number that
can give rise to more different constellations than the number of possible bitmaps of 180 by 240
pixels. These numbers are a long way beyond computational efficiency, and assuming the state
space is not completely convex, the problem could well be untractable, and heuristic algorithms
such as simulated annealing, could provide a feasible way forward.
In their seminal study, Paauw and Van den Berg used a stochastic hillclimber, simulated
annealing, and plant propagation on paintings-from-poygons [12]. These three algorithms
all retained their best found objective values (or βfitnessβ) throughout the runs of 106 function
evaluations (Fig. 1, lower part). The hillclimber and plant propagation achieved relatively good
results, but the same could not be said for simulated annealing. In this study however, weβll
take care of that.
2. Simulated Annealing
Introduced by Kirkpatrick, Gelatt & Vecchi in 1983 and independently by ΔernΓ½ in 1985, the
simulated annealing algorithm draws inspiration directly from statistical mechanics, the cen-
tral discipline of condensed matter physics [7][14][15]. In the physical annaealing process, a
desirable low energy state of a metal is reached by optimizing the spatial configuration of the
atoms. To reach such a configuration, dislocated atoms need to move to vacant sites which
can be facilitated by cooling the liquid metal slowly. Higher temperatures increase the atomsβ
movement, but also increases the number of expected vacant sites due to the increase of mixing
entropy. Simulated annealing emulates this process for combinatorial optimization problems,
1
The term is intentionally loosely applied here; it could mean βKolmogorov complexityβ, or βRLE-complexityβ,
amongst others.
even up to the exact Boltzmann factor (Eq.2) used for βmovementβ through a combinatorial
state space. As such, it is one of the few heuristic algorithms in the field that is more than just
a metaphor; it truly pries at the barrier between physics and information science. Good solu-
tions to combinatorial problems such as timetabling, the wiring of electronic systems, the job
shop scheduling problem and the traveling salesman problem have been found earlier using
simulated annealing [14][7][16][17][18].
For many heuristic algorithms, parameterization is βessential for good algorithm perfor-
manceβ [19], and simulated annealing is no exception. It has proven quite sensitive to param-
eter changes in an early study by Christopher SkiΕcim and Bruce Golden, who also illustrated
the difficulty of cooling down slowly on the one hand, but not too slowly on the other [20]. A
great number of cooling schedules have been tested throughout history, varying from linear
temperature decrease to schedules that also allow reheating [21]. The logarithmic schedule by
Geman & Geman however, is the only one that has theoretically been proven to reach a global
optimum [22][23]. But the proof critically relies on the number of iterations going to infinity,
so aptly elaborated on by Kenneth Boese and Andrew Kahng: a βwhere-you-areβ schedule like
Geman & Gemanβs, which needs infinite computation time, is practically useless [21]. This
is the critical mistake Paauw and Van den Berg made on paintings-from-polygons, and even
though they did retain the βbest-so-farβ constellations throughout the runs, their intermediate
and final results from simulated annealing were of abominable quality compared to their other
two algorithms[12]. According to the authors, its failure is βeasily explained from its high tem-
perature", a claim that immediately warrants a follow-up investigation. In this study, we will
try nine different cooling schedules, see to what extend the authors were right, and whether
any improvement is possible.
For now, we specifically do not address the other two algorithms from the earlier study, and
with good reason: PPA has shown to be largely parameter insensitive and unbiased ([24][25][26],
ongoing work), and the stochastic hillclimberβs behaviour is so closely akin to simulated an-
nealing with Geman&Geman on π = 1, that a separate quantitative parametrization analysis
seems somewhat superfluous. A future study in qualitative parameters however (i.e. the algo-
rithmic components), is still an open avenue waiting to be explored.
3. Simulated Annealing on Paintings-from-Polygons
A simulated annealing run on the paintings-from-polygons problem starts with randomly dis-
tributing π£ vertices over π polygons (π£ = 1000, π = 250 in this study). Each polygon is first
assigned three vertices, after which the remaining vertices are randomly distributed. Then, all
the vertices are assigned random coordinate values within the dimensions of the target bitmap.
Every polygon is randomly assigned color and opacity (eq.: βalphaβ) values within the bounds
of its four RGBA-channels. Each iteration, a randomly selected mutation is performed on the
polygon constellation. There are four different mutations types, each of which has equal prob-
ability of occurring:
1. The change color mutation randomly chooses one polygon and one of its RGBA chan-
nels, and assigns it a new random value 0 β€ π β€ 255.
2. The move vertex mutation randomly chooses one vertex from one polygon and assigns
it new random values 0 β€ π₯ < π₯πππ₯ and 0 β€ π¦ < π¦πππ₯.
3. The transfer vertex mutation randomly chooses two polygons, π1 and π2, in which π1
is not a triangle. A randomly selected vertex is then deleted from π1, and a new vertex
is placed somewhere on the line between two neighboring vertices in π2.
4. The change drawing index mutation randomly chooses a polygon and assigns its draw-
ing index a new random value 0 β€ π < π. If the new index is lower than the current index,
all indexes of the in-between polygons will be increased by one. If the new index is higher
than the current index, all indexes of the in-between polygons above the new index will
be decreased by one.
After each iteration, the new constellation is rendered and the new MSE is calculated. If the
new constellation has a lower MSE (equiv.: a smaller pixel-by-pixel difference with the target
bitmap), the mutation is accepted and this new constellation will be the starting point for the
next iteration. Mutations that lead to a constellation with a higher MSE are not immediately
rejected though. Analogous to the annealing metaphor, they might still be accepted, depending
on the magnitude of the difference and the temperature. The acceptance probability: π(ππππππ‘)
is given by the Boltzmann factor
( βΞπππΈ
(2)
π )
π(ππππππ‘) = π π
in which ΞπππΈ is the difference in MSE between the new and the old constellation, and ππ
is the temperature at iteration π retrieved from the cooling schedule. Notice the correlation
between the MSE difference and the acceptance probability: a greater increase in MSE is still
acceptable as long as the temperature ππ is still high. Correct parameterization of the cooling
schedule is therefore of critical importance to both annealing and simulated annealing. For
PFP however, it appears as average the temperature alone predicts an upper bound on the
performance, irrespective of the actual trajectory a schedule might traverse.
4. Cooling Schedules
The first three cooling schedules used in this study are of the Geman & Geman type, in which
the temperature at iteration π is given by
π
ππ = . (3)
πππ(π + 1)
In the seminal study, its parameter was set to π = 195075, as it is the maximal existable error
barrier that separates two local minima in this problem [12]. In this study, we replicate the
experiment with its original settings, but also add two new parameter settings for this schedule:
π = 50 and π = 1. The number of iterations π = 106 , lowering the temperature from π1 β 648025
to π106 β 32512 for π = 195075. For π = 50, the temperature descends from π1 β 166 to π106 β 8,
and for π = 1, it ranges from π1 β 3 to π106 β 0.17 (Fig. 2).
In the linear cooling schedule, the temperature ππ is defined as:
Figure 2: The course of the temperature for different cooling schedules. The redder the figure, the
higher its average temperature. Note that the vertical scale for the top row is much larger, due to the
extremities in the Geman & Geman schedule.
π
ππ = π1 β
(1 β ) (4)
ππππ₯
where in this study ππππ₯ = 106 and π1 = 1000, resulting in a linear decrease of temperature
throughout the iterations (Fig. 2). The staircase cooling schedule also starts π1 = 1000, but
drops its temperature once every 105 iterations, resulting in a temperature of 0 for the final 105
iterations. The motivation for the staircase cooling schedule stems from a 2016-publication by
Strobl and Barker, who observed that temperatures should be kept constant for some duration
of time in order to reach thermal equilibrium [27].
In the sigmoidal cooling schedule, temperature follows a sigmoidal decrease given by
1000
ππ = β5 5 (5)
1 + π 10 β
(πβ5β
10 )
This cooling schedule decreases slowly in temperature during the run, with its steepest decline,
its maximum absolute derivative, at 5 β
105 iterations. It comes from a website with cooling
schedules once maintained by Brian T. Luke. It appears to have gone offline2 , but the schedule
(and the author) can still be found in a set of public slides at Stanford University [29]. The
geometric cooling schedule has a history of giving βsatisfactory resultsβ on a wide variety of
optimization problems [30]. Its temperature ππ is defined as:
ππ+1 = π½ β
ππ (6)
where in this study, π½ is fixed at 0.99999, with π1 = 1000.
In a study on best-so-far solutions by Boese and Kahng, optimal cooling schedules were
shown to be non-monotonically decreasing [21]. As the simulated annealing algorithm used
on PFP study also retains its best-so-far value, periodical reheat might provide an alluring al-
ternative to more traditional schedules. The linear reheat cooling schedule linearly lowers its
temperature to zero in epochs of 105 iterations, but then reheats to half the previous epochβs
temperature, as can be seen in figure 2. This results in nine reheating cycles, with their re-
spective peaks being ππ = {1000, 500, ..., 3.90625} at π = {0, 105 , ..., 9 β
105 }. The cosine cooling
schedule follows a simple cosine function, defined as
π
ππ = 500 β
πππ ( ) + 500. (7)
16753
The fraction 16753 is used to ensure 9.5 cycles in 106 iterations (notice that 16753 β 2πβ9.5 ) and
π 10 6
the constant term +500 at the end of the equation ensures the oscillating function starts at
π1 = 1000 to and ends at π106 = 0.
5. Experiments and Results
The performance of the different cooling schedules is evaluated by performing five runs of
simulated arnealing with each cooling schedule on each of the nine different 180x240 target
bitmaps: βSalvator Mundiβ, βLady with an Ermineβ and βMona Lisaβ painted by Leonardo Da
Vinci between 1490 and 1503, βThe Starry Nightβ by Van Gogh in 1889, Portrait of composer Jo-
hann Sebastian Bach by Elias Gottlieb Haussmann in 1746, βThe Kissβ by Gustav Klimt in 1908,
βComposition with Red, Blue and Yellowβ by Mondriaan in 1930, βThe Persistence of Memoryβ
by Dali in 1931, and βConvergenceβ by Pollock in 1952. Each run consisted of 106 iterations3 ,
with a nonchanging total of 250 polygons and 1000 vertices (the experimentβs source code and
data are publicly accessible [32]). For every combination of cooling schedule and target bitmap,
the normalized average performance is calculated as the mean final MSE of the five runs
2
BTLβs cooling schedules can still be found through the waybackmachine [28].
3
β[Function evaluations are the gold standard for measuring the performance of heuristic algorithms]β [31], but
in our case, the number iterations corresponds exactly to the number of evaluations, which is typical for simulated
annealing.
Figure 3: Normalized mean end results for all nine cooling schedules reveal a consistent, painting
invariant performance hierarchy for simulated annealing cooling schedules on the paintings-from-
polygons problem. The red bars are the logarithmically plotted average temperatures of the corre-
sponding schedules on the horizontal axes.
π₯ β π₯πππ
π₯ππππππππ§ππ = (8)
π₯πππ₯ β π₯πππ
in which π₯ is the average MSE of the five runs after 106 iterations, π₯πππ the lowest MSE of the
five, and π₯πππ₯ the highest. Notice that a low normalized MSE corresponds to a high perfor-
mance.
After completing all 405 runs, a firm pattern in the performance of the cooling schedules
emerges (Fig. 3). The cooling schedule of Geman & Geman with π = 1 performs best on all
paintings, immediately followed by the geometric, linear reheat and Geman & Geman with
π = 50. These top four best performing cooling schedules form a group with relatively similar
performance when compared to the other cooling schedules. Fifth, sixth and seventh are the
sigmoidal, linear and cosine cooling schedules, followed by the poor performance of the stair-
case. Very much in last place is the high π-valued Geman & Geman cooling schedule used in
the study of Paauw and Van den Berg [12]. It is by far the worst possible choice out of these
Figure 4: Opacity per polygon drawing index averaged over 45 runs (nine paintings, five runs each)
per schedule. Remarkably enough, the three best performing cooling schedules show a mean parabolic
curvature increase, suggesting that good approximations have low-opacity polygons on top.
nine.
The performance hierarchy of the cooling schedules can be characterized more rigorously
then by its final MSE order alone. It turns out that seven of nine cooling schedulesβ average log
temperatures and final MSEs are bound by a linear equation4 for each painting, with the error
of fit between 0.11 and 0.04 on all paintings. Two schedules however, are different. These are
the linear reheat schedule, with an error of fit 6 to 11 times greater, and the geometric cooling
schedule with an error of fit 6 to 44 times greater, depending on the specific painting. The
remarkable thing is however, these errors are larger in downward direction.
In other words, when the final MSE and the temperature are compared, all schedules show
roughly the same performance, independent of their actual shape. The only two exceptions
are the linear reheat and geometric cooling schedules, which perform better than other sched-
ules with the same average temperature. A future experiment, with all cooling schedules nor-
malized to the same average temperature, should either confirm or disconfirm whether these
4
polynomial equations on nonlog data yield comparable results.
differences are substantial.
Another remarkable phenomenon can be witnessed in the structural development of the
polygon constellations of this experiment. After random initialization, a typical constellation5
holds about 111 triangles, 71 quadrangles, 40 pentagons, 18 hexagons and smaller numbers of
7-gons to to 12-gons. But a typical end constellation has up to 28% fewer quadrangles, penta-
and hexagons, and very small numbers of larger polygons, up to as many as 35 vertices. Most
striking however, is that the number of triangles increases by about 16%, to roughly 128. It
is unknown what causes this increase, but it is thought that triangles βfacilitateβ optimization
through their relative mobility in vertex position. This however, still requires further confir-
mation.
A final interesting phenomenon is that the three best performing cooling schedules appear
to exert an influence on the relation between a polygonβs opacity and its drawing index. For
the three best performing cooling schedules, a higher polygon drawing index corresponds to a
lower opacity (Fig. 4). When fitting a parabola, its steepness increases 3 to 24 times for the best
three cooling schedules, suggesting that succesful optimization runs produce more transparent
polygons on top. Again, further analysis could (dis)confirm the consistency of this effect .
6. Conclusion
It appears that Paauw and Van den Berg were right in their assessment that the poor perfor-
mance of their simulated annealing was due to the high temperatures used at the time. As
shown in this study, the success of the algorithm depends largely on the average temperature
of the schedule only, a few interesting exceptions aside. Furthermore, successful runs seem to
increase the number triangles in a constellation, and decrease the opacity of its lastly drawn
polygons. To what extent these properties are more typical to the problem or to simulated
annealing itself still remains a subject for future endeavours.
References
[1] S. Colton, J. McCormack, S. Berns, E. Petrovskaya, M. Cook, Adapting and enhancing
evolutionary art for casual creation, in: International Conference on Computational In-
telligence in Music, Sound, Art and Design (Part of EvoStar), Springer, 2020, pp. 17β34.
[2] S. L. Banal, V. Ciesielski, A deep learning neural network for classifying good and bad
photos, in: International Conference on Computational Intelligence in Music, Sound, Art
and Design (Part of EvoStar), Springer, 2020, pp. 1β16.
[3] F. Heerde, I. Vatolkin, G. Rudolph, Comparing fuzzy rule based approaches for music
genre classification, in: International Conference on Computational Intelligence in Music,
Sound, Art and Design (Part of EvoStar), Springer, 2020, pp. 35β48.
[4] M. Muehlbauer, J. Burry, A. Song, An aesthetic-based fitness measure and a framework
for guidance of evolutionary design in architecture, in: International Conference on Com-
putational Intelligence in Music, Sound, Art and Design (Part of EvoStar), Springer, 2020,
pp. 134β149.
5
Here, βa typical constellationβ means: values averaged over all constellations in the experiment.
[5] A. Neumann, B. Alexander, F. Neumann, Evolutionary image transition and painting
using random walks, Evolutionary Computation (2020) 1β34.
[6] J. Berg, N. G. A. Berggren, S. A. Borgeteien, C. R. A. Jahren, A. Sajid, S. Nichele, Evolved art
with transparent, overlapping, and geometric shapes, in: Symposium of the Norwegian
AI Society, Springer, 2019, pp. 3β15.
[7] S. Kirkpatrick, C. D. Gelatt, M. Vecchi, Optimization by simulated annealing., Science
220(4598) (1983) 671β680. doi:10.1126/science.220.4598.671.
[8] J. Sleegers, D. van den Berg, Plant propagation & hard hamiltonian graphs, Evo* 2020
(2020) 10.
[9] M. Sulaiman, A. Salhi, A. Khan, S. Muhammad, W. Khan, On the theoretical analysis of
the plant propagation algorithms, Mathematical Problems in Engineering 2018 (2018).
[10] A. E. Eiben, J. E. Smith, Introduction to Evolutionary Computing, first ed., Springer, Berlin,
2003. doi:10.1007/978-3-662-05094-1.
[11] JSB, Paintings-from-polygons movie clip, 2020. https://www.youtube.com/watch?v=
u91gRGY8ElQ (Last accessed on May 23rd, 2020).
[12] M. Paauw, D. Van den Berg, Paintings, polygons and plant propagation, in: Interna-
tional Conference on Computational Intelligence in Music, Sound, Art and Design (Part
of EvoStar), Springer, 2019, pp. 84β97.
[13] D. van den Berg, Simplified paintings-from-polygons is np-hard, Evo* 2020 (2020) 15.
[14] V. ΔernyΜ, Thermodynamical approach to the traveling salesman problem: An efficient
simulation algorithm, Journal of optimization theory and applications 45 (1985) 41β51.
[15] D. S. Johnson, C. R. Aragon, L. A. McGeoch, C. Schevon, Optimization by simulated
annealing: An experimental evaluation; part i, graph partitioning, Operations research
37 (1989) 865β892.
[16] D. Abramson, M. Krishnamoorthy, H. Dang, et al., Simulated annealing cooling schedules
for the school timetabling problem, Asia Pacific Journal of Operational Research 16 (1999)
1β22.
[17] P. J. Van Laarhoven, E. H. Aarts, J. K. Lenstra, Job shop scheduling by simulated annealing,
Operations research 40 (1992) 113β125.
[18] R. A. Rutenbar, Simulated annealing algorithms: An overview, IEEE Circuits and Devices
magazine 5 (1989) 19β26.
[19] S. K. Smit, A. E. Eiben, Comparing parameter tuning methods for evolutionary algorithms,
in: 2009 IEEE congress on evolutionary computation, IEEE, 2009, pp. 399β406.
[20] C. C. SkiΕcim, B. L. Golden, Optimization by simulated annealing: A preliminary compu-
tational study for the tsp, in: Proceedings of the 15th conference on Winter Simulation-
Volume 2, IEEE Press, 1983, pp. 523β535.
[21] K. D. Boese, A. B. Kahng, Best-so-far vs. where-you-are: implications for optimal finite-
time annealing, Systems & control letters 22 (1994) 71β78.
[22] S. Geman, D. Geman, Stochastic relaxation, gibbs distributions, and the bayesian restora-
tion of images, IEEE Transactions on pattern analysis and machine intelligence (1984)
721β741.
[23] Y. Nourani, B. Andresen, A comparison of simulated annealing cooling strategies., Journal
of Physics A: Mathematical and General 31 (1998) 8373β8385. doi:10.1088/0305-4470/
31/41/011.
[24] M. de Jonge, D. van den Berg, Plant propagation parameterization: Offspring & population
size, Evo* 2020 (2020) 19.
[25] M. de Jonge, D. van den Berg, Parameter sensitivity patterns in the plant propagation
algorithm, IJCCI 2020: Proceedings of the 12th International Joint Conference on Com-
putational Intelligence (2020).
[26] W. Vrielink, D. van den Berg, Fireworks algorithm versus plant propagation algorithm.,
in: IJCCI, 2019, pp. 101β112.
[27] M. A. Strobl, D. Barker, On simulated annealing phase transitions in phylogeny recon-
struction, Molecular phylogenetics and evolution 101 (2016) 46β55.
[28] B. T. Luke, [via wayback machine] simulated annealing cooling schedules, 2019. http:
//web.archive.org/web/20190620071317/http://www.btluke.com/simanf1.html (Last ac-
cessed on May 23rd, 2020).
[29] S. University, Stats 318: Lecture, 2020. https://statweb.stanford.edu/~candes/teaching/
stats318/Handouts/Annealing.pdf (Last accessed on May 23rd, 2020).
[30] P. N. Strenski, S. Kirkpatrick, Analysis of finite length annealing schedules, Algorithmica
6 (1991) 346β366.
[31] T. Bartz-Beielstein, C. Doerr, J. Bossek, S. Chandrasekaran, T. Eftimov, A. Fischbach,
P. Kerschke, M. Lopez-Ibanez, K. M. Malan, J. H. Moore, et al., Benchmarking in opti-
mization: Best practice and open issues, arXiv preprint arXiv:2007.03488 (2020).
[32] R. Dahmani, Reddyβs repo, 2020. https://github.com/RedRedouane/
Paintings-from-Polygons-Simulated-Annealing (Last accessed on May 23rd, 2020).