Yasser Gonzalez-Fernandez’s research while affiliated with York University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (19)


Fig. 1. Convergence vs. stall in PSO. In d = 3 dimensions (a), the swarm has converged after the last red circle representing successful exploration. In d = 30 dimensions (b), large amounts of (failed) exploration continue to occur in more promising regions of the search space as indicated by the blue dots. However, swarm progress has largely stalled as indicated by the lack of red circles which represent successful exploration.
Fig. 4. Examples of line searches for Rastrigin. Although random lines (c) often do not identify all optima (see (a) and (b) for examples), useful estimates on the size of attraction basins are still possible.
Fig. 5. Examples of line searches on the CEC2013 benchmark functions. In addition to the spacing between the local optima, insights into the structure of the search space can be gleaned. Among the multi-modal functions, three major profiles emerge: globally convex (a), deceptive (b), and noisy (c).
Particle Swarm Optimization with pbest Perturbations
  • Conference Paper
  • Full-text available

September 2020

·

349 Reads

·

18 Citations

·

Imran Abdulselam

·

·

Yasser Gonzalez-Fernandez

Restarts are a popular remedy to address (premature) convergence in metaheuristics. In Particle Swarm Optimization , it has been observed that swarms often "stall" as opposed to "converge". A stall occurs when all of the forward progress that could occur is instead rejected as failed exploration. Since the swarm is in a good region of the search space with the potential to make more progress, a (random) restart could be counter productive. We instead introduce a method to address the stall mechanism. The introduction of perturbations to the pbest positions leads to significant improvements in the performance of standard Particle Swarm Optimization.

Download

Efficient elicitation of software configurations using crowd preferences and domain knowledge

March 2019

·

287 Reads

·

5 Citations

Automated Software Engineering

As software systems grow in size and complexity, the process of configuring them to meet individual needs becomes more and more challenging. Users, especially those that are new to a system, are faced with an ever increasing number of configuration possibilities, making the task of choosing the right one more and more daunting. However, users are rarely alone in using a software system. Crowds of other users or the designers themselves can provide with examples and rules as to what constitutes a meaningful configuration. We introduce a technique for designing optimal interactive configuration elicitation dialogs, aimed at utilizing crowd and expert information to reduce the amount of manual configuration effort. A repository of existing user configurations supplies us with information about popular ways to complete an existing partial configuration. Designers augment this information with their own constraints. A Markov decision process (MDP) model is then created to encode configuration elicitation dialogs that maximize the automatic configuration decisions based on the crowd and the designers’ information. A genetic algorithm is employed to solve the MDP when problem sizes prevent use of common exact techniques. In our evaluation with various configuration models we show that the technique is feasible, saves configuration effort and scales for real problem sizes of a few hundreds of features.




Fig. 2: Efficacy of exploration as a random solution approaches its local optimum: the deeper it gets in the basin, the less likely that it can move to a better basin by finding a better solution.
Fig. 3: Height in the basin of the best solution found so far in terms of the function evaluations on Rastrigin's function in 30 dimensions. The best solutions move deeper into their respective attraction basins as the search progresses.
Fig. 5: Height in the basin of the best solution in the two populations of Leaders and Followers in terms of the function evaluations on Rastrigin's function in 30 dimensions. Many samples occur above the 200 height mark, where the sampling rate for finding better solutions in better basins is more favourable.  
Leaders and Followers – A New Metaheuristic to Avoid the Bias of Accumulated Information

May 2015

·

845 Reads

·

45 Citations

Finding good solutions on multi-modal optimization problems depends mainly on the efficacy of exploration. However, many search techniques applied to multi-modal problems were initially conceptualized with unimodal functions in mind, prioritizing exploitation over exploration. In this paper, we perform a study on the efficacy of exploration under random sampling, which leads to the identification of an important comparison bias that occurs when a solution which has benefited from local search is compared to the first (random) solution in a new search area. With the goal of eliminating this bias and improving the efficacy of exploration, we have developed a new search technique explicitly designed for multi-modal search spaces. “Leaders and Followers” aims to eliminate the negative effects of information accumulation and at the same time use the information from the best solutions in a way that they have controlled influence over the newly-sampled solutions. The proposed metaheuristic outperforms both Particle Swarm Optimization and Differential Evolution across a broad range of multi-modal optimization problems.


Figure 2: As a random sample solution is moved closer towards its local optimum, the probability of successful exploration rapidly approaches 0. 
Figure 3: Concurrent exploration and exploitation is shown in PSO. During exploration, improvements to the dashed line occur. However, exploitation also occurs-as indicated by the convergence of the solid and dashed lines. 
A Review of Thresheld Convergence

May 2015

·

334 Reads

·

17 Citations

A multi-modal search space can be defined as having multiple attraction basins – each basin has a single local optimum which is reached from all points in that basin when greedy local search is used. Optimization in multi-modal search spaces can then be viewed as a two-phase process. The first phase is exploration in which the most promising attraction basin is identified. The second phase is exploitation in which the best solution (i.e. the local optimum) within the previously identified attraction basin is attained. The goal of thresheld convergence is to improve the performance of search techniques during the first phase of exploration. The effectiveness of thresheld convergence has been demonstrated through applications to existing metaheuristics such as particle swarm optimization and differential evolution, and through the development of novel metaheuristics such as minimum population search and leaders and followers.



Modeling with Copulas and Vines in Estimation of Distribution Algorithms

January 2015

·

411 Reads

·

15 Citations

The aim of this work is studying the use of copulas and vines in numerical optimization with Estimation of Distribution Algorithms (EDAs). Two EDAs built around the multivariate product and normal copulas, and other two based on pair-copula decomposition of vine models are studied. We analyze empirically the effect of both marginal distributions and dependence structure in order to show that both aspects play a crucial role in the success of the optimization process. The results show that the use of copulas and vines opens new opportunities to a more appropriate modeling of search distributions in EDAs.


Figure 2: Results of PSO on Rastrigin for n = 30. The plot shows the average performance over 51 trials: mean f (x) of gbest is 63.8.  
Figure 3: Results of five restarts of PSO-with 1/5 of the allotted FEs each-on Rastrigin for n = 30. The plot shows the average performance over 51 trials: mean f (x) of gbest is 50.0. The discordance between this value and the chart is caused by the line plot averaging the individual restarts and not the best global result of each trial.
Figure 4: Average and overall minimum distance between the centroids of every pair of the k ∈ [2, 50] clusters obtained from the population of pbests. The horizontal lines show the minimum and maximum distance between adjacent local optima in Rastrigin for n = 30.
Figure 5: Results of the final version of multi-start PSO with thresheld convergence-including normal resets and improved local search-on Rastrigin for n = 30. The plot shows the average performance over 51 trials: mean f (x) of gbest is 27.9.
Identifying and Exploiting the Scale of a Search Space in Particle Swarm Optimization

July 2014

·

768 Reads

·

10 Citations

Multi-modal optimization involves two distinct tasks: identifying promising attraction basins and finding the local optima in these basins. Unfortunately, the second task can interfere with the first task if they are performed simultaneously. Specifically, the promise of an attraction basin is often estimated by the fitness of a single sample solution, so an attraction basin represented by a random sample solution can appear to be less promising than an attraction basin represented by its local optimum. The goal of thresheld convergence is to prevent these biased comparisons by disallowing local search while global search is still in progress. Ideally, thresheld convergence achieves this goal by using a distance threshold that is correlated to the size of the attraction basins in the search space. In this paper, a clustering-based method is developed to identify the scale of the search space which thresheld convergence can then exploit. The proposed method employed in the context of a multi-start particle swarm optimization algorithm has led to large improvements across a broad range of multi-modal problems.


Identifying and Exploiting the Scale of a Search Space in Differential Evolution

July 2014

·

26 Reads

·

3 Citations

Optimisation in multimodal landscapes involves two distinct tasks: identifying promising regions and location of the (local) optimum within each region. Progress towards the second task can interfere with the first by providing a misleading estimate of a region's value. Thresheld convergence is a generally applicable "meta"-heuristic designed to control an algorithm's rate of convergence and hence which mode of search it is using at a given time. Previous applications of thresheld convergence in differential evolution (DE) have shown considerable promise, but the question of which threshold values to use for a given (unknown) function landscape remains open. This work explores the use of clustering-based method to infer the distances between local optima in order to set a series of decreasing thresholds in a multi-start DE algorithm. Results indicate that on those problems where normal DE converges, the proposed strategy can lead to sizable improvements.


Citations (13)


... Regardless, even when different types of search spaces have been identified (e.g. the globally convex, partially structured, deceptive, and noisy sub-categories [15] for the multi-modal search spaces in the CEC2013 benchmark [16]), this identification can be exceptionally difficult to perform accurately in real time (e.g. [17], [18]). ...

Reference:

Improvements to Dual Annealing in SciPy
Particle Swarm Optimization with pbest Perturbations

... Real-world applications that address both preferences from crowd and experts assembled in participatory models have a different domain of applications, and some examples include the election of software configuration (see [47]), e-cognocracy for new electronic democracy (see [48]), choosing winner in Eurovision Song musician contest (see [49]), deciding which art project to fund (see [50]), computational urban planning (see [51]), MicroTalk argumentation with 'true/false' answers (similar as approval voting) for improving accuracy of crowd workers (see [52]). ...

Efficient elicitation of software configurations using crowd preferences and domain knowledge

Automated Software Engineering

... In Fig. 7 the comparison of final version mCEDAf with the initial mCEDA and with other published algorithms -namely IPOP-CMA-ES [7], CMAES-RIS [1], PSO [9], LaF [2], SPAM-AOS [3] is reported. 1 The pairwise t-test is used instead of Wilcoxon ranked-sum test. 2 ...

Leaders and Followers on the CEC2013 Real-Parameter Optimization Benchmark Functions

... pbest positions in PSO and target solutions in DE) can experience early and rapid exploitation which leads them to become (near) local optima [7]. Other attempts to reduce exploitation of the reference solutions [9] or to move them away from their local optima [10] achieved some improvement in performance, but they did not meaningfully address the core flaw of Fitness-Based Selection which is that it was never designed to estimate the fitness of a region of the search space. ...

A Review of Thresheld Convergence

... [75] (measures, resampling strategies, tuning main processes and RS technique). The GA, PSO and EDA meta-heuristics were implemented using the GA 13 [76], pso 14 [77], and copulaedas 15 [78] R packages, respectively. The J48, CART and CTree algorithms were implemented using the RWeka 16 [79], rpart 17 [80] and party 18 [11] packages, respectively, wrapped into the mlr package. ...

copulaedas: An R Package for Estimation of Distribution Algorithms Based on Copulas

Journal of Statistical Software

... Attraction basins were used as features for automated algorithm selection as well [35]. Recently, attraction basins have received some attention for delineation of the exploration and exploitation phase [22,25,76,77]. This topic is found to be one of the key topics in the field of metaheuristics [37,39]. ...

Identifying and Exploiting the Scale of a Search Space in Particle Swarm Optimization

... The number of sub-tasks D and service candidates C varies from small to large, which consists of two cases: (1) D fixed to 100, C increased from 500 to 5000; (2) C fixed to 500, D increased from 100 to 1000. [46], and the parameters are set to the same as in the code linking to https://www.researchgate.net/publication/259643342_Source_code_for_an_implementation_of_Standard_Particle_Swarm_Optimization_--_revised. GL25 ...

Standard Particle Swarm Optimization on the CEC2013 Real‐Parameter Optimization Benchmark Functions -- revised