Skip to main content

Sampling Bounds for Stochastic Optimization

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 3624))

Abstract

A large class of stochastic optimization problems can be modeled as minimizing an objective function f that depends on a choice of a vector xX, as well as on a random external parameter ω∈ Ω given by a probability distribution π. The value of the objective function is a random variable and often the goal is to find an xX to minimize the expected cost E ω [f ω (x)]. Each ω is referred to as a scenario. We consider the case when Ω is large or infinite and we are allowed to sample from π in a black-box fashion. A common method, known as the SAA method (sample average approximation), is to pick sufficiently many independent samples from π and use them to approximate π and correspondingly E ω [f ω (x)]. This is one of several scenario reduction methods used in practice.

There has been substantial recent interest in two-stage stochastic versions of combinatorial optimization problems which can be modeled by the framework described above. In particular, we are interested in the model where a parameter λ bounds the relative factor by which costs increase if decisions are delayed to the second stage. Although the SAA method has been widely analyzed, the known bounds on the number of samples required for a (1+ε) approximation depend on the variance of π even when λ is assumed to be a fixed constant. Shmoys and Swamy [13,14] proved that a polynomial number of samples suffice when f can be modeled as a linear or convex program. They used modifications to the ellipsoid method to prove this.

In this paper we give a different proof, based on earlier methods of Kleywegt, Shapiro, Homem-De-Mello [6] and others, that a polynomial number of samples suffice for the SAA method. Our proof is not based on computational properties of f and hence also applies to integer programs. We further show that small variations of the SAA method suffice to obtain a bound on the sample size even when we have only an approximation algorithm to solve the sampled problem. We are thus able to extend a number of algorithms designed for the case when π is given explicitly to the case when π is given as a black-box sampling oracle.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Dhamhere, K., Ravi, R., Singh, M.: On two-stage Stochastic Minimum Spanning Trees. In: Jünger, M., Kaibel, V. (eds.) IPCO 2005. LNCS, vol. 3509, pp. 321–334. Springer, Heidelberg (2005)

    Chapter  Google Scholar 

  2. Flaxman, A., Frieze, A., Krivelevich, M.: On the random 2-stage minimum spanning tree. In: Proc. of SODA (2005)

    Google Scholar 

  3. a.Gupta, M. Pál, R. Ravi, and A. Sinha. Boosted sampling: Approximation algorithms for stochastic optimization. In: Proceedings of the 36th Annual ACM Symposium on Theory of Computing (2004)

    Google Scholar 

  4. Gupta, A., Ravi, R., Sinha, A.: An edge in time saves nine: Lp rounding approximation algorithms. In: Proceedings of the 45th Annual IEEE Symposium on Foundations of Computer Science (2004)

    Google Scholar 

  5. Immorlica, N., Karger, D., Minkoff, M., Mirrokni, V.: On the costs and benefits of procrastination: Approximation algorithms for stochastic combinatorial optimization problems. In: Proceedings of the 15th Annual ACM-SIAM Symposium on Discrete Algorithms (2004)

    Google Scholar 

  6. Kleywegt, J., Shapiro, A., Homem-De-Mello, T.: The sample average approximation method for stochastic discrete optimization. SIAM J. on Optimization 12, 479–502 (2001)

    Article  MATH  MathSciNet  Google Scholar 

  7. Mahdian, M.: Facility Location and the Analysis of Algorithms through Factor- Revealing Programs. Ph.D. Thesis, MIT (June 2004)

    Google Scholar 

  8. Motwani, R., Raghavan, P.: Randomized Algorithms. Cambridge University Press, Cambridge (1995)

    MATH  Google Scholar 

  9. Ruszczynski, A., Shapiro, A. (eds.): Stochastic Programming. Handbook in Operations Research and Management Science, vol. 10. Elsevier, Amsterdam (2003)

    MATH  Google Scholar 

  10. Shapiro, A.: Montecarlo sampling methods. In: Ruszczynski, A., Shapiro, A. (eds.) Stochastic Programming. Handbook in Operations Research and Management Science, vol. 10. Elsevier, Amsterdam (2003)

    Chapter  Google Scholar 

  11. Ravi, R., Sinha, A.: Hedging uncertainty: Approximation algorithms for stochastic optimization problems. In: Bienstock, D., Nemhauser, G.L. (eds.) IPCO 2004. LNCS, vol. 3064, pp. 101–115. Springer, Heidelberg (2004)

    Chapter  Google Scholar 

  12. Ravi, R., Singh, M.: Personal communication (February 2005)

    Google Scholar 

  13. Shmoys, D., Swamy, C.: Stochastic optimization is (almost) as easy as deterministic optimization. In: Proc. of FOCS (2004)

    Google Scholar 

  14. Shmoys, D., Swamy, C.: The Sample Average Approximation Method for 2-stage Stochastic Optimization (November 2004) (manuscript)

    Google Scholar 

  15. Shmoys, D., Swamy, C.: Sampling-based Approximation Algorithms for Multistage Stochastic Optimization (2005) (manuscript)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Charikar, M., Chekuri, C., Pál, M. (2005). Sampling Bounds for Stochastic Optimization. In: Chekuri, C., Jansen, K., Rolim, J.D.P., Trevisan, L. (eds) Approximation, Randomization and Combinatorial Optimization. Algorithms and Techniques. APPROX RANDOM 2005 2005. Lecture Notes in Computer Science, vol 3624. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11538462_22

Download citation

  • DOI: https://doi.org/10.1007/11538462_22

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-28239-6

  • Online ISBN: 978-3-540-31874-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics