Asynchronous classification-based optimization

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Scopus citations

Abstract

Asynchronous parallelization is an effective way to accelerate optimization. While asynchronous parallelization can destroy the sequential structure of optimization algorithms, it has been found counter-intuitively that some optimization algorithms are proven to preserve their performance under asynchronous parallelization, including the stochastic gradient descent for first-order optimization of differentiable functions and Pareto optimization for zeroth-order optimization in binary space. Following this direction, in this paper, we show that the classification-based optimization, which is a recently developed framework for zeroth-order optimization in continuous space, can also enjoy the asynchronous parallelization. We implement ASRacos, an asynchronous version of a classification-based optimization algorithm SRacos, to accelerate the optimization through asynchronous parallelization. We theoretically provide the query complexity of ASRacos and further show that on certain conditions, ASRacos can achieve a better performance than SRacos even if using the same number of evaluations. Experiments on synthetic functions and controlling tasks in OpenAI Gym demonstrate that ASRacos can achieve almost linear speedup while preserving good solution quality.

Original languageEnglish
Title of host publicationProceedings of the 1st International Conference on Distributed Artificial Intelligence, DAI 2019
PublisherAssociation for Computing Machinery
ISBN (Electronic)9781450376563
DOIs
StatePublished - 13 Oct 2019
Externally publishedYes
Event1st International Conference on Distributed Artificial Intelligence, DAI 2019 - Beijing, China
Duration: 13 Oct 201915 Oct 2019

Publication series

NameACM International Conference Proceeding Series

Conference

Conference1st International Conference on Distributed Artificial Intelligence, DAI 2019
Country/TerritoryChina
CityBeijing
Period13/10/1915/10/19

Keywords

  • Derivative-free optimization
  • Distributed optimization
  • Hyper-parameter optimization
  • Non-convex optimization

Fingerprint

Dive into the research topics of 'Asynchronous classification-based optimization'. Together they form a unique fingerprint.

Cite this