# Feature selection for set-valued data based on D–S evidence theory

Original Source Here

### Abstract

Feature selection is one basic and critical technology for data mining, especially in current “big data era”. Rough set theory is sensitive to noise in feature selection due the stringent condition of an equivalence relation. However, D–S evidence theory is flexible to measure uncertainty of information. In this paper, we introduce robust feature evaluation metrics “belief function” and “plausibility function” into feature selection algorithm to avoid the defect that classification effect is affected by noise such as missing values, confusing data, etc. Firstly, similarity between information values in a set-valued information system (SVIS) is introduced and a variable parameter to control the similarity of samples is given. Secondly,
$$\theta$$
-lower and
$$\theta$$
-upper approximations in an SVIS are put forward. Then, the concepts of
$$\theta$$
-belief function,
$$\theta$$
-plausibility function,
$$\theta$$
-belief reduction and
$$\theta$$
-plausibility reduction are given. Moreover, several feature selection algorithms based on the D–S evidence theory in an SVIS are proposed. Experimental results and statistical test show that the proposed metric is insensitive to noise because it comprehensively considers the evidence at all levels, and the proposed algorithms are more robust than several state-of-the-art feature selection algorithms.

AI/ML

Trending AI/ML Article Identified & Digested via Granola by Ramsey Elbasheer; a Machine-Driven RSS Bot