A few months ago, I wrote a post about optimization using gradient descent, which involves searching for a model that best meets certain criteria by repeatedly making adjustments that improve things a little bit at a time. In many situations, this works quite well and will always or almost always finds the best solution. But in other cases, it’s possible for this approach to fall into a locally optimal solution that isn’t the overall best, but is better than any nearby solution. A common way to deal with this sort of situation is to add some randomness into the algorithm, making it possible to jump out of one of these locally optimal solutions into a slightly worse solution that is adjacent to a much better one. In this post, I want to explore one such approach, called a genetic algorithm (or an evolutionary algorithm), which I’ll illustrate with a specific type of genetic algorithm called symbolic regression. I first heard about this in an article about a small company in Somerville, MA called Nutonian that has built a whole data science platform around it.
-
Recent Posts
Recent Comments
Yuanqing on Configuration Spaces and the M… Mel on Recurrent Neural Networks Teradata Expands Cap… on K-modes Teradata Expands Cap… on K-modes Qu'est-ce qu… on Graphs and networks Archives
- December 2016
- November 2016
- October 2016
- June 2016
- April 2016
- January 2016
- November 2015
- October 2015
- July 2015
- June 2015
- May 2015
- January 2015
- September 2014
- June 2014
- May 2014
- March 2014
- February 2014
- January 2014
- December 2013
- October 2013
- September 2013
- August 2013
- July 2013
- June 2013
- May 2013
- April 2013
- March 2013
Categories
Meta