Distributed Model-Free Stochastic Optimization in Wireless Sensor Networks

  • Daniel Yagan
  • Chen-Khong Tham
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4026)


With the improvement in computer electronics in terms of processing, memory and communication capabilities, it has become possible to scatter tiny embedded devices such as sensor nodes to monitor physical phenomena with greater flexibility. A large number of sensor nodes, communicating over the wireless medium, also allows information gathering with greater accuracy than current systems. This paper presents a new stochastic technique known as Incremental Simultaneous Perturbation Approximation(ISPA) for performing optimization in wireless sensor networks. The proposed algorithm is based on a combination of gradient-based decentralized incremental (GBDI) optimization and Simultaneous Perturbation Stochastic Approximation (SPSA) techniques. The former is based on Incremental Sub-Gradient Optimization (ISGO) techniques that allow the algorithm to be performed in a distributed and collaborative manner. The latter component addresses the limitations of the GBDI component especially in real-world sensor networks. Specifically, the SPSA component is a model-free technique that finds the optimal solution without requiring a functional model such as an input-output relationship and a cost gradient. Simulation results show that the proposed ISPA approach not only achieves distributed optimization in a stochastic environment, but can also be implemented in a practical manner for resource-constrained devices.


Cost Function Sensor Network Sensor Node Wireless Sensor Network Loss Function 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Rabbat, M., Nowak, R.: Distributed Optimization in Sensor Networks. In: Proc. Information Processing in Sensor Networks, Berkeley, CA, USA (2004)Google Scholar
  2. 2.
    Nedic, A., Bertsekas, D.: Incremental Subgradient Methods for Non-differentiable Optimization. Technical Report LIDS-P-2460, Massachusetts Institute of Technology, Cambridge, MA, USA (1999)Google Scholar
  3. 3.
    Spall, J.C.: Multivariate Stochastic Approximation Using a Simultaneous Perturbation Gradient Approximation. IEEE Trans. Automat. Contr. 37, 332–341 (1992)MATHCrossRefMathSciNetGoogle Scholar
  4. 4.
    Rabbat, M., Nowak, R.: Quantized Incremental Algorithms for Distributed Optimization. IEEE J. Select. Areas Commun. 23(4), 798–808 (2005)CrossRefGoogle Scholar
  5. 5.
    Nedic, A., Bertsekas, D.: Convergence rate of incremental subgradient algorithms. In: Uryasev, S., Pardalos, P. (eds.) Stochastic Optimization: Algorithms and Applications, pp. 263–304. Kluwer Academic Publishers, Dordrecht (2000)Google Scholar
  6. 6.
    Spall, J.C.: A Stochastic Approximation Algorithm for Large-Dimensional Systems in the Kiefer-Wolfowitz Setting. In: Proc. IEEE Conf. on Decision and Control., pp. 1544–1548 (1988)Google Scholar
  7. 7.
    Chin, D.: Comparative Study of Stochastic Algorithms for System Optimization Based on Gradient Approximations. IEEE Trans. Syst., Man, Cybern. B 27, 244–249 (1997)CrossRefGoogle Scholar
  8. 8.
    Sadegh, P., Spall, J.: Optimal Random Perturbations for Multivariate Stochastic Approximation Using a Simultaneous Perturbation Gradient Approximation. In: Proc. American Control Conf., pp. 3582–3586 (1997)Google Scholar
  9. 9.
    Spall, J.C., Cristion, J.A.: A Neural Network Controller for Systems with Unmodeled Dynamics with Applications to Wastewater Treatment. IEEE Trans. Syst., Man, Cybern. B 27, 369–375 (1997)CrossRefGoogle Scholar
  10. 10.
    Spall, J.C., Cristion, J.A.: Nonlinear Adaptive Control Using Neural Networks: Estimation Based on a Smoothed Form of Simultaneous Perturbation Gradient Approximation. Stat. Sinica 4, 1–27 (1994)MATHMathSciNetGoogle Scholar
  11. 11.
    Chin, D.C.: A More Efficient Global Optimization Algorithm Based on Styblinski and Tang. Neural Networks 7, 573–574 (1994)CrossRefGoogle Scholar
  12. 12.
    Maryak, J.L., Chin, D.C.: Efficient Global Optimization Using SPSA. In: Proc. American Control Conf., pp. 890–894 (1999)Google Scholar
  13. 13.
    Sadegh, P.: Constrained Optimization via Stochastic Approximation with a Simultaneous Perturbation Gradient Approximation. Automatica 33, 889–892 (1997)MATHCrossRefMathSciNetGoogle Scholar
  14. 14.
    Wang, I., Spall, J.: A Constrained Simultaneous Perturbation Stochastic Approximation Algorithm Based on Penalty Functions. In: Proc. American Control Conf., pp. 393–399 (1999)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2006

Authors and Affiliations

  • Daniel Yagan
    • 1
  • Chen-Khong Tham
    • 1
  1. 1.Department of Electrical and Computer EngineeringNational University of Singapore 

Personalised recommendations