**Reference**:- Ilya Baran, Erik D. Demaine, and Dmitriy A. Katz, “Optimally Adaptive Integration of Univariate Lipschitz Functions”, in
*Proceedings of the 7th Latin American Symposium on Theoretical Informatics (LATIN 2006)*, Valdivia, Chile, March 20–24, 2006, pages 142–153. **Abstract**:-
We consider the problem of approximately integrating a Lipschitz function
*f*(with a known Lipschitz constant) over an interval. The goal is to achieve an error of at most ε using as few samples of*f*as possible. We use the adaptive framework: on all problem instances an adaptive algorithm should perform almost as well as the best possible algorithm tuned for the particular problem instance. We distinguish between DOPT and ROPT, the performances of the best possible deterministic and randomized algorithms, respectively. We give a deterministic algorithm that uses*O*(DOPT(*f*, ε) ⋅ log (ε^{−1}/ DOPT(*f*, ε))) samples and show that an asymptotically better algorithm is impossible. However, any deterministic algorithm requires Ω(ROPT(*f*, ε)^{2}) samples on some problem instance. By combining a deterministic adaptive algorithm and Monte Carlo sampling with variance reduction, we give an algorithm that uses at most*O*(ROPT(*f*, ε)^{4/3}+ ROPT(*f*, ε) ⋅ log (1/ε)) samples. We also show that any algorithm requires Ω(ROPT(*f*, ε)^{4/3}+ ROPT(*f*, ε) ⋅ log (1/ε)) samples in expectation on some problem instance (*f*, ε), which proves that our algorithm is optimal. **Availability**:- The paper is available in PostScript (248k), gzipped PostScript (103k), and PDF (128k).
- See information on file formats.
- [Google Scholar search]
**Related papers**:- Integration_Algorithmica (Optimally Adaptive Integration of Univariate Lipschitz Functions)

See also other papers by Erik Demaine.

Last updated April 22, 2019 by Erik Demaine.