OR/AND neurons and the development of interpretable logic models

  • Authors:
  • W. Pedrycz;M. Reformat;K. Li

  • Affiliations:
  • Dept. of Electr. & Comput. Eng., Univ. of Alberta, Edmonton, Alta.;-;-

  • Venue:
  • IEEE Transactions on Neural Networks
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we are concerned with the concept of fuzzy logic networks and logic-based data analysis realized within this framework. The networks under discussion are homogeneous architectures comprising of OR/AND neurons originally introduced by Hirota and Pedrycz. Being treated here as generic processing units, OR/AND neurons are neurofuzzy constructs that exhibit well-defined logic characteristics and are endowed with a high level of parametric flexibility and come with significant interpretation abilities. The composite logic nature of the logic neurons becomes instrumental in covering a broad spectrum of logic dependencies whose character spread in-between between those being captured by plain and and or logic descriptors (connectives). From the functional standpoint, the developed network realizes a logic approximation of multidimensional mappings between unit hypercubes, that is transformations from [0,1]n to [0,1]m. The way in which the structure of the network has been formed is highly modular and becomes reflective of a general concept of decomposition of logic expressions and Boolean functions (as being commonly encountered in two-valued logic). In essence, given a collection of input variables, selected is their subset and transformed into new composite variable, which in turn is used in the consecutive module of the network. These intermediate synthetic variables are the result of the successive problem (mapping) decomposition. The development of the network is realized through genetic optimization. This helps address important issues of structural optimization (where we are concerned with a selection of a subset of variables and their allocation within the network) and reaching a global minimum when carrying out an extensive parametric optimization (adjustments of the connections of the neurons). The paper offers a comprehensive and user-interactive design procedure including a simple pruning mechanism whose intention is to enha- - nce the interpretability of the network while reducing its size. The experimental studies comprise of three parts. First, we demonstrate the performance of the network on Boolean data (that leads to some useful comparative observations considering a wealth of optimization tools available in two-valued logic and digital systems). Second, we discuss synthetic multivalued data that helps focus on the approximation abilities of the network. Finally, show the generation of logic expressions describing selected data sets coming from the machine learning repository