Welcome back to our series on Swarm Intelligence Metaheuristics for Optimization. On this post, we will focus on Particle Swarm Optimization. Recall we define Metaheuristics as a class of optimization algorithms which turn out to be very useful when the function being optimized is non-differentiable or does not have an analytical expression at all.
This post will focus on a class of metaheuristics known as Swarm Intelligence. The most amazing feature of these algorithms is their ability to solve complex problems by a set of cooperative agents posing very simple intelligence.
This is the second (and last) part of the series dealing with the formal comparison of Machine Learning (ML) algorithms from a statistical point of view. In this post, we examine how statistical tests are applied to performance data of ML algorithms.
In industry, when a practitioner (often a Data Scientist) uses a machine learning algorithm to build a predictive model to solve a real-world problem, they are interested in the performance when the model is deployed into a production environment…
The last chapter of the three-part series on Evolutionary Feature Selection with Big datasets. We will address some fundamental design aspects of a Genetic Algorithm (GA) and commonly chosen options, to then move on to the CHC algorithm and a distributed approach for Feature Selection.