Discernibility-based variable granularity and kansei representations

  • Authors:
  • Yuji Muto;Mineichi Kudo

  • Affiliations:
  • Division of Computer Science, Graduate school of Information Science and Technology, Hokkaido University, Sapporo, Japan;Division of Computer Science, Graduate school of Information Science and Technology, Hokkaido University, Sapporo, Japan

  • Venue:
  • RSFDGrC'05 Proceedings of the 10th international conference on Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing - Volume Part I
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we discuss the most suitable “representation granularity”, keeping several types of discernibility including individually discernibility and class discernibility. In the traditional “reduction” sense, the goal is to find the smallest number of attributes such that they enable us to discern each tuple or each decision class. However, once we pay attention to the number of attribute values too, that is, the size of each attribute, another criterion is needed. Indeed, we should ask ourselves about which one is better in the following two situations: 1) we can discern them with a single attribute of size ten, and 2) we can do this with two attributes of size five. This study answers this question with some criteria. Especially, we deal with continuous attributes. If we evaluate this difference in the light of understandability, we may prefer the latter, because they give more simple descriptions. Such a combination of simple nominal description helps us as a language or as a Kansei representation. To do this, we propose some criteria and algorithms to find near-optimal solutions for those criteria. In addition, we show some results for some databases in UCI Machine Learning Repository.