Representation-burden Conservation Network Applied to Learning VQ (NPL270)

  • Authors:
  • Juang-Hua Wang;Chih-Ping Hsiao

  • Affiliations:
  • Dept of Electrical Engineering, National Taiwan Ocean University, 2 Peining Rd., Keelung 202, Taiwan;Dept of Electrical Engineering, National Taiwan Ocean University, 2 Peining Rd., Keelung 202, Taiwan

  • Venue:
  • Neural Processing Letters
  • Year:
  • 1997

Quantified Score

Hi-index 0.00

Visualization

Abstract

A self-creating network effective in learning vector quantization, called RCN (Representation-burden Conservation Network) is developed. Each neuron in RCN is characterized by a measure of representation-burden. Conservation is achieved by bounding the summed representation-burden of all neurons at constant 1, as representation-burden values of all neurons are updated after each input presentation. We show that RCN effectively fulfills the conscience principle [1] and achieves biologically plausible self-development capability. In addition, conservation in representation-burden facilitates systematic derivations of learning parameters, including the adaptive learning rate control useful in accelerating the convergence as well as in improving node-utilization. Because it is smooth and incremental, RCN can overcome the stability-plasticity dilemma. Simulation results show that RCN displays superior performance over other competitive learning networks in minimizing the quantization error.