
Open access
Date
2019Type
- Journal Article
Abstract
We consider the problem of estimating a probability distribution that maximizes the entropy while satisfying a finite number of moment constraints, possibly corrupted by noise. Based on duality of convex programming, we present a novel approximation scheme using a smoothed fast gradient method that is equipped with explicit bounds on the approximation error. We further demonstrate how the presented scheme can be used for approximating the chemical master equation through the zero-information moment closure method, and for an approximate dynamic programming approach in the context of constrained Markov decision processes with uncountable state and action spaces. Show more
Permanent link
https://doi.org/10.3929/ethz-b-000374034Publication status
publishedExternal links
Journal / series
Journal of Machine Learning ResearchVolume
Pages / Article No.
Publisher
MIT PressSubject
Entropy maximization; convex optimization; relative entropy minimization; fast gradient method; approximate dynamic programmingOrganisational unit
03751 - Lygeros, John / Lygeros, John
03781 - Renner, Renato / Renner, Renato
Related publications and datasets
Is new version of: http://hdl.handle.net/20.500.11850/220445
More
Show all metadata