Yin, H. and Allinson, N. M. (1998) A selforganising mixture network for density modelling. In: The 1998 IEEE International Joint Conference on Neural Networks, 49 May 1998, Anchorage, USA.
Documents 


PDF
00687216.pdf  Whole Document 440kB 
Official URL: http://dx.doi.org/10.1109/IJCNN.1998.687216
Abstract
A completely unsupervised mixture distribution network, namely the selforganising mixture network, is proposed for learning arbitrary density functions. The algorithm minimises the KullbackLeibler information by means of stochastic approximation methods. The density functions are modelled as mixtures of parametric distributions such as Gaussian and Cauchy. The first layer of the network is similar to the Kohonen's selforganising map (SOM), but with the parameters of the class conditional densities as the learning weights. The winning mechanism is based on maximum posterior probability, and the updating of weights can be limited to a small neighbourhood around the winner. The second layer accumulates the responses of these local nodes, weighted by the learning mixing parameters. The network possesses simple structure and computation, yet yields fast and robust convergence. Experimental results are also presented
Item Type:  Conference or Workshop Item (Paper) 

Additional Information:  A completely unsupervised mixture distribution network, namely the selforganising mixture network, is proposed for learning arbitrary density functions. The algorithm minimises the KullbackLeibler information by means of stochastic approximation methods. The density functions are modelled as mixtures of parametric distributions such as Gaussian and Cauchy. The first layer of the network is similar to the Kohonen's selforganising map (SOM), but with the parameters of the class conditional densities as the learning weights. The winning mechanism is based on maximum posterior probability, and the updating of weights can be limited to a small neighbourhood around the winner. The second layer accumulates the responses of these local nodes, weighted by the learning mixing parameters. The network possesses simple structure and computation, yet yields fast and robust convergence. Experimental results are also presented 
Keywords:  Gaussian distribution, convergence, probability, selforganising feature maps, unsupervised learning, algorithms, Bayesian methods, pattern classification 
Subjects:  G Mathematical and Computer Sciences > G730 Neural Computing 
Divisions:  College of Science > School of Computer Science 
ID Code:  5085 
Deposited By:  Tammie Farley 
Deposited On:  21 Apr 2012 07:35 
Last Modified:  13 Mar 2013 09:06 
Repository Staff Only: item control page