Distributed Parametric Optimization with the Geneva Library
Offered By: CppNow via YouTube
Course Description
Overview
Explore the Geneva library, an open-source collection of distributed optimization algorithms, in this conference talk from BoostCon 2011. Learn how Geneva leverages various Boost libraries to solve large-scale parametric optimization problems across multi-core systems, clusters, and cloud environments. Discover the library's support for gradient descents, evolutionary algorithms, and swarm algorithms, with plans to include simulated annealing. Understand how Geneva's unified data structures allow candidate solutions to move freely between optimization algorithms. Gain insights into the library's scalability, having been tested with hundreds of clients working simultaneously on optimization problems. Hear about the presenter's experiences using Boost libraries such as Serialization, Threads, Conversion, Date/Time, Function, and Bind from a user's perspective.
Syllabus
Distributed parametric optimization with the Geneva library - Ruediger Berlich [ BoostCon 2011 ]
Taught by
CppNow
Related Courses
Practical Predictive Analytics: Models and MethodsUniversity of Washington via Coursera Deep Learning Fundamentals with Keras
IBM via edX Introduction to Machine Learning
Duke University via Coursera Intro to Deep Learning with PyTorch
Facebook via Udacity Introduction to Machine Learning for Coders!
fast.ai via Independent