{"file":[{"file_id":"7820","date_created":"2020-05-12T09:23:27Z","creator":"dernst","relation":"main_file","checksum":"89db06a0e8083524449cb59b56bf4e5b","content_type":"application/pdf","date_updated":"2020-07-14T12:45:45Z","file_name":"2018_PMLR_Kolmogorov.pdf","access_level":"open_access","file_size":408974}],"type":"conference","date_created":"2018-12-11T11:45:33Z","department":[{"_id":"VlKo"}],"publication_status":"published","has_accepted_license":"1","year":"2017","page":"228-249","quality_controlled":"1","volume":75,"user_id":"2DF688A6-F248-11E8-B48F-1D18A9856A87","license":"https://creativecommons.org/licenses/by/4.0/","intvolume":" 75","status":"public","author":[{"full_name":"Kolmogorov, Vladimir","id":"3D50B0BA-F248-11E8-B48F-1D18A9856A87","first_name":"Vladimir","last_name":"Kolmogorov"}],"day":"27","abstract":[{"lang":"eng","text":"We consider the problem of estimating the partition function Z(β)=∑xexp(−β(H(x)) of a Gibbs distribution with a Hamilton H(⋅), or more precisely the logarithm of the ratio q=lnZ(0)/Z(β). It has been recently shown how to approximate q with high probability assuming the existence of an oracle that produces samples from the Gibbs distribution for a given parameter value in [0,β]. The current best known approach due to Huber [9] uses O(qlnn⋅[lnq+lnlnn+ε−2]) oracle calls on average where ε is the desired accuracy of approximation and H(⋅) is assumed to lie in {0}∪[1,n]. We improve the complexity to O(qlnn⋅ε−2) oracle calls. We also show that the same complexity can be achieved if exact oracles are replaced with approximate sampling oracles that are within O(ε2qlnn) variation distance from exact oracles. Finally, we prove a lower bound of Ω(q⋅ε−2) oracle calls under a natural model of computation."}],"publisher":"PMLR","ddc":["510"],"conference":{"name":"COLT: Annual Conference on Learning Theory ","start_date":"2018-07-06","end_date":"2018-07-09"},"tmp":{"short":"CC BY (4.0)","image":"/images/cc_by.png","name":"Creative Commons Attribution 4.0 International Public License (CC-BY 4.0)","legal_code_url":"https://creativecommons.org/licenses/by/4.0/legalcode"},"file_date_updated":"2020-07-14T12:45:45Z","title":"A faster approximation algorithm for the Gibbs partition function","publist_id":"7628","_id":"274","publication":"Proceedings of the 31st Conference On Learning Theory","ec_funded":1,"date_published":"2017-12-27T00:00:00Z","oa_version":"Published Version","project":[{"_id":"25FBA906-B435-11E9-9278-68D0E5697425","name":"Discrete Optimization in Computer Vision: Theory and Practice","call_identifier":"FP7","grant_number":"616160"}],"oa":1,"article_processing_charge":"No","external_id":{"arxiv":["1608.04223"]},"month":"12","date_updated":"2021-01-12T06:59:23Z","language":[{"iso":"eng"}],"citation":{"short":"V. Kolmogorov, in:, Proceedings of the 31st Conference On Learning Theory, PMLR, 2017, pp. 228–249.","apa":"Kolmogorov, V. (2017). A faster approximation algorithm for the Gibbs partition function. In Proceedings of the 31st Conference On Learning Theory (Vol. 75, pp. 228–249). PMLR.","ama":"Kolmogorov V. A faster approximation algorithm for the Gibbs partition function. In: Proceedings of the 31st Conference On Learning Theory. Vol 75. PMLR; 2017:228-249.","ista":"Kolmogorov V. 2017. A faster approximation algorithm for the Gibbs partition function. Proceedings of the 31st Conference On Learning Theory. COLT: Annual Conference on Learning Theory vol. 75, 228–249.","chicago":"Kolmogorov, Vladimir. “A Faster Approximation Algorithm for the Gibbs Partition Function.” In Proceedings of the 31st Conference On Learning Theory, 75:228–49. PMLR, 2017.","mla":"Kolmogorov, Vladimir. “A Faster Approximation Algorithm for the Gibbs Partition Function.” Proceedings of the 31st Conference On Learning Theory, vol. 75, PMLR, 2017, pp. 228–49.","ieee":"V. Kolmogorov, “A faster approximation algorithm for the Gibbs partition function,” in Proceedings of the 31st Conference On Learning Theory, 2017, vol. 75, pp. 228–249."}}