Abstract
Hierarchical structured data are very common for data mining and other tasks in real-life world. How to select the optimal scale combination from a multi-scale decision table is critical for subsequent tasks. At present, the models for calculating the optimal scale combination mainly include lattice model, complement model and stepwise optimal scale selection model, which are mainly based on consistent multi-scale decision tables. The optimal scale selection model for inconsistent multi-scale decision tables has not been given. Based on this, firstly, this paper introduces the concept of complement and lattice model proposed by Li and Hu. Secondly, based on the concept of positive region consistency of inconsistent multi-scale decision tables, the paper proposes complement model and lattice model based on positive region consistent and gives the algorithm. Finally, some numerical experiments are employed to verify that the model has the same properties in processing inconsistent multi-scale decision tables as the complement model and lattice model in processing consistent multi-scale decision tables. And for the consistent multi-scale decision table, the same results can be obtained by using the model based on positive region consistent. However, the lattice model based on positive region consistent is more time-consuming and costly. The model proposed in this paper provides a new theoretical method for the optimal scale combination selection of the inconsistent multi-scale decision table.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
1.1 A brief review on multi-scale decision tables
Rough set theory was originally proposed by Professor Pawlak in 1982 (Pawlak 1992). Because of the mature mathematical foundation and unnecessary of prior knowledge, it is easy to use and become an effective tool for dealing with various incomplete information such as imprecise, inconsistent information. It is a powerful data analysis method. Rough set theory can, in the absence of prior knowledge, find out the classification of knowledge to determine the upper and lower approximation of the problem by describing the set of the given problem and then, analyze and process the uncertain data.
Based on the theory of Pawlak rough set, Wu and Leung introduced the concept of multi-scale decision table (MSDT) from the perspective of granular computing and analyzed the knowledge acquisition in it Wu and Leung (2011).
The concept of multi-scale is very common in our life. We can describe a thing from multiple angles, that is, from multiple scales. And it is also widely used in deep learning (Bharati et al. 2020; Li et al. 2020; Qian et al. 2017; Taverniers et al. 2021). Significantly, some applications of deep learning models to medical imaging and drug discovery for managing COVID-19 disease are studied by some literatures such as (Bharati et al. 2021a, b; Khamparia et al. 2021; Mondal et al. 2021a, b). Additionally, MSDT is one of the research objects in the field of knowledge discovery in database. So how to extract useful information and discover new knowledge from MSDT is worth research.
The general method of processing multi-scale decision tables is to limited multi-scale attributes on a certain scale, and then, we can obtain a series of single-scale decision tables (SSDT) whose each attribute only has one scale. At last, we can do data mining on a single-scale decision table we choose. In a multi-scale information table, if all the attributes are on the finest scales, then the most information of objects is included, but this process is of high cost. However, if all the attribute are on the coarsest scales, then some useful information may be lost. Therefore, one or several optimal scale combinations which can reduce the cost without losing useful information are existed.
However, Wu and Leung pointed out that their research is based on two assumptions (Wu and Leung 2013). One of them is that the number of scales of each attribute must be the same. Another one is that only the corresponding single attributes are able to combine into a subsystem in the process of decomposition for subsystems. Under the same assumptions, (Gu and Wu 2013) and (She et al. 2015) studied the knowledge acquisition and rule induction in multi-scale decision tables.
Later, Li and Hu extended their theory and broke these two assumptions (Li and Hu 2017; Li et al. 2017). They proposed lattice model and complement model to calculate the optimal scale combination. Based on the concept of multi-scale attribute significance they introduced, they proposed stepwise optimal scale selection model. For the attribute with different significance, the scale selection should be carried out step by step, which can effectively reduce the time of calculating and get the best results based on the attribute significance
Since then, the theoretical study on generalized multi-scale decision table has also been explored by some researchers. Xie et al. proposed three new types of rules and their extraction methods (Xie et al. 2018). In Hao et al. (2017), motivated by the fact that sequential three-way decisions are an effective mathematical tool in dealing with the data with information sequentially updated, Hao et al. used this methodology to investigate the optimal scale selection problem in a dynamic multi-scale decision table. And in Huang et al. (2019), Huang et al. addressed the issue of optimal scale selection and rule acquisition in dominance-based multi-scale intuitionistic fuzzy (IF) decision tables. What is more, on the basis of an IF inclusion measure, two novel multi-granulation decision-theoretic models have been developed in multi-scale IF information tables in Huang et al. (2020). In 2021, (Wang et al. 2021) firstly investigated the belief structure and the plausibility structure by defining belief and plausibility functions from the multi-granulation viewpoint and discuss how to construct multi-granulation rough set models in multi-scale information systems. Then, the optimal scale selection methods with various requirements are studied in two aspects of optimistic and pessimistic multi-granulation for a multi-scale decision information system.
Wu and Leung did a comparison study of optimal scale combination selection in multi-scale decision tables whose different attributes have different numbers of scales (Wu et al. 2017; Wu and Leung 2020). They formulate information granules with different scale combinations in multi-scale information systems and discuss their relationships. What is more, the definition and properties of lower and upper approximations of sets with different scale combination are proposed in their paper.
The relationship between rule extraction and feature matrix is further studied (Huang et al. 2020; Chen et al. 2019). Finally, the matrix is used to describe the scale combination, and the matrix method for optimal scale combination selection and the optimal scale combination keeping the positive region unchanged in the consistent and inconsistent generalized multi-scale decision information system are given, respectively.
In Bao et al. (2021), Bao et al. defined entropy optimal scale combination in multi-scale decision tables. They proved that the entropy optimal scale combination and classical optimal scale combination proposed previously are equivalent.
Recently, Zhan et al. (2021) establish group decision-making (GDM) idea on multi-scale information systems proposed by Wu and Leung from the perspective of multi-expert group decision-making (MEGDM). It can be applied to sorting problems on multi-scale information systems.
1.2 The motivation of our research
As a more general case, inconsistent decision tables are more common in daily life and knowledge discovery tasks. And consistent multi-scale tables can be used as special cases of inconsistent multi-scale tables.
Nevertheless, the current works on multi-scale decision tables and also optimal scale combination selection are mainly aimed to calculate the optimal scale combination in consistent multi-scale decision tables. They cannot be applied to more general scenarios. Before using, we must judge the type of the table. And for inconsistent multi-scale decision tables, we can only obtain an optimal scale combination. It is full of limitations if there is missing data in the table. Motivated by these, in this paper, we focus on how to get all the optimal scale combinations in inconsistent multi-scale decision tables. Complement model and lattice model based on positive region consistence are proposed, and the algorithms of them are given as well. Compared with the above models, our models are more generalized.
Our main contributions are summarized as follows:
-
We propose complement model and lattice model based on positive region consistent and give the algorithm for inconsistent multi-scale decision tables.
-
We conduct some numerical experiments to prove that the models based on positive region consistence can also deal with consistent multi-scale decision tables correctly.
The remainder parts of the paper are organized as follows. In Sect. 2, several basic notions of Pawlak rough set, information tables and decision tables, scale combination and attribute significance are reviewed. In Sect. 4, the concept of positive region consistent is introduced. And the optimal scale combination selection models for inconsistent multi-scale decision tables and their algorithm are proposed. Some numerical experiments are employed in Sect. 5. Finally, we conclude the paper with a summary and outlook the further in Sect. 6.
2 Preliminaries
In this section, we review several basic concepts and results of Pawlak rough set, information tables and decision tables, scale combination and attribute significance.
2.1 Pawlak rough set
Let U be a finite and nonempty set called universe of discourse. If \(R \subseteq U \times U\) is an equivalence relation on U, that is, R is a reflexive, symmetric and transitive binary relation on U, then the pair (U, R) is called a Pawlak approximation space (Pawlak 1992). The equivalence relation R partitions the universe of discourse U into disjoint subsets. Such partition is a quotient set of U and denoted by \(U/R=\{[x]_R | x \in U\}\), where \([x]_R=\{y \in U|(x, y) \in R\}\) is the R equivalence class containing x. The elements in U/R are called elementary sets. For any set \(X \in {\mathscr {P}}(U)\), lower and upper approximations are defined as follows:
Definition 1
Let U be a finite and nonempty set called universe of discourse. If \(X \in {\mathscr {P}}(U)\), lower and upper approximations of X are defined as:
where \({\mathscr {P}}(U)\) is the power set of U. Obviously, they can be defined by:
If and only if \({\underline{R}}(X) \ne {\overline{R}}(X)\), X cannot be precisely defined by R. \(({\underline{R}}(X),{\overline{R}}(X))\) is called the Pawlak rough set of X with respect to (w.r.t.) (U, R). The sets \(BN_R(X)={\overline{R}}(X)-{\underline{R}}(X)\), \(POS_R={\underline{R}}(X)\), \(NEG_R=U-{\overline{R}}(X)\) are, respectively, called the boundary, the positive region and the negation region of X w.r.t. (U,R).
The accuracy of rough set can be defined as (Pawlak 1992):
where \(|\cdot |\) is the cardinal number of set. For the empty set \(\emptyset \), we define \(\alpha _R(\emptyset )=1\). Obviously, \(0 \leqslant \alpha _R(X) \leqslant 1\).
Definition 2
(Wu and Leung 2013) Let U be a finite and nonempty universe of discourse. \({\mathbf {P}}_1\) and \({\mathbf {P}}_2\) are two partitions of U, For each \(A\in {\mathbf {P}}_1\), if there exists \(B\in {\mathbf {P}}_2\) such that \(A\subseteq B\), we say that \({\mathbf {P}}_1\) is finer than \({\mathbf {P}}_2\) or \({\mathbf {P}}_2\) is coarser than \({\mathbf {P}}_1\), denoted as \({\mathbf {P}}_1 \sqsubseteq {\mathbf {P}}_2\). If \(A\subset B\), we say \({\mathbf {P}}_1\) is strictly finer than \({\mathbf {P}}_2\), denoted as \({\mathbf {P}}_1 \sqsubset {\mathbf {P}}_2\).
2.2 Information table and decision table
Definition 3
(Wu and Leung 2011) An information table is a 2-tuple (U, A), where \(U=\{x_1,x_2,\ldots ,x_n\}\) is a finite and nonempty set called universe of discourse, \(A=\{a_1,a_2,\ldots ,a_m\}\) is a finite and nonempty set of attributes. For any \(a\in A\), there is \(a: U \rightarrow V_a\), that is, for any \(x\in U\), there is \(a(x)\in V_a\), where \(V_a=(a(x)|x\in U)\) called the domain of a.
For each attribute \(a \in A\), a is a surjective function from U to \(V_a\), and it determines an equivalence relation on U.
Definition 4
(Wu and Leung 2011) A decision table is a 2-tuple \((U,C\cup \{d\})\), where (U, C) is an information table, \(d \notin C\) is a special attribute called decision. In this case, C is called conditional attribute set. d is a map \(d: U\rightarrow V_d\) from U to \(V_d\).
Similarly, we can define the equivalence relation as:
Then, we obtain a partition \(U/R_d\) of U. For any \(B \subseteq C\), define equivalence \(R_B\) as:
If \(R_C \subseteq R_d\), then the decision table \((U,C\cup \{d\})\) is consistent; otherwise, it is inconsistent.
For the inconsistent decision table \((U,C\cup \{d\})\), the concept of generalized decision attribute is introduced by Wu and Leung (2020). For any \(B\subseteq C\), the generalized decision attribute of x w.r.t. B, denoted as \(\partial _B\), can be defined as (Komorowski et al. 1999):
According to Eq.7, we know that for any decision table \(S=(U,C\cup \{d\})\), although S may be inconsistent, \(S=(U,C\cup \partial _C)\) must be consistent.
Based on the single-scale decision table, Wu and Leung proposed the concept of multi-scale decision table (Wu and Leung 2011):
Definition 5
(Wu and Leung 2011) A multi-scale decision table can be denoted as \(S=(U,C\cup \{d\})\), where U is finite and nonempty object set called universe of discourse, A is finite and nonempty set of attribute, d is decision. Each attribute \(a_j\in C\) is a multi-scale attribute, that is, for the same object in U, attribute \(a_j\) can take on different values at different scales.
For each attribute \(a_j\in C\), we assume that the higher the level of scale is, the coarser the partition w.r.t. the scale becomes. If the attribute \(a_j\) has three levels of scale, its first level of scale \(a_j^1\) is finer than its second level of scale \(a_j^2\), and its second level of scale \(a_j^1\) is finer than its third level of scale \(a_j^2\).
2.3 Scales combination
The general method of processing multi-scale information tables is to limit multi-scale attributes on a certain scale, and then, we can obtain a series of single-scale information tables whose each attribute only has one scale. At last, we can do data mining on a single-scale information table we choose.
The concept of scales combination and scales collection and some properties was introduced by Li and Hu (2017).
Definition 6
(Li and Hu 2017) Let \(S=(U,A)\) be a multi-scale information table, where attribute \(a_i\) has \(I_i\) levels of scale, \(i=1,2,\ldots ,m\). If we restrict attribute \(a_1,a_2,\cdots ,a_m\) on their \(I_ith\) scale, respectively, we can obtain a single-scale information table \(S^K\), where \(K=(l_1;l_2;\ldots ;l_m)\). The combination \((l_1;l_2;\ldots ;l_m)\) is called the scales combination of S in \(S^K\). All the scales combination of S is called scales collection, denoted as \({\mathscr {L}}=\{(l_1;l_2;\cdots ;l_m)|1 \leqslant l_i \leqslant I_i,i=1,2,\ldots ,m\}\).
Definition 7
(Li and Hu 2017) Let \(S=(U,A)\) be a multi-scale information table, and \({\mathscr {L}}\) is the scales collection of S. For \(K_1, K_2\in {\mathscr {L}}\), if the elements of \(K_2\) are not less than the corresponding elements of \(K_1\), then we say that \(K_1\) is weaker than \(K_2\) or \(K_2\) is stronger than \(K_1\), denoted as \(K_1\preceq K_2\).
According to Definition 7, we know that \({\mathscr {L}}\) is an partial order relation. Thus, \(({\mathscr {L}},\preceq )\) is a partial order set, which is reflexive, antisymmetric and transitive. Furthermore, \(({\mathscr {L}},{\preceq })\) is a lattice in which every two elements have a unique supremum and a unique infimum. Obviously, we can get the proposition as follow.
Proposition 1
(Li and Hu 2017) Let \(K_1, K_2\in {\mathscr {L}}\) and \(K_1\preceq K_2\), if \(S^{K_2}=(U,C^{K_2}\cup \{d\})\) is consistent, then \(S^{K_1}\) is also consistent.
According to the Proposition 1, we can define the concept of optimal scale combination as follow (Li and Hu 2017).
Definition 8
(Li and Hu 2017) Let \({\mathscr {L}}\) be a scales collection of a consistent multi-scale decision table S, for \(K\in {\mathscr {L}}\), all the K meet the condition that if for all the \(K\in {\mathscr {L}}\) and \(K\preceq K\), \(S^K\) is consistent, but \(S^{K}\)(if there exists K) is inconsistent are the optimal scale combination of S.
Therefore, the consistency of multi-scale decision table can be defined by:
Definition 9
(Wu and Leung 2011) Let \(S=(U,C\cup \{d\})\) be a multi-scale decision table, and \({\mathbf {1}}_m=(1;1;\cdots ;1)\). If \(S^{{\mathbf {1}}_m}=(U,\{a_j^1|j=1,2,\ldots ,m\}\cup \{d\})\) whose all the attributes are on their finest level of scale is consistent, then the multi-scale decision table S is consistent.
Let \({\mathscr {L}}\) be the scales collection of \((U,C\cup \{d\})\). For an arbitrary \(K\in {\mathscr {L}}\), the corresponding equivalence relation \(R_{A^K}\) can be defined as
U can be partitioned by \(R_{A^K}\) into a family of equivalence classes as follows
where \([x]_{A^K}={y\in U|(x,y)\in R_{A^K}}\).
According to Eqs.8 and 9, we can know the relation between equivalence relation and subsets of attributes.
Proposition 2
(Li and Hu 2017) Let \(S=(U,C\cup \{d\})=(U,\{a_j^k|j=1,2,\cdots ,m,k=1,2,\ldots ,I_j\} \cup \{d\})\) be a multi-scale decision table. \({\mathscr {L}}\) is the scales collection of S. For \(K_0=(k_1;k_2;\cdots ;k_m)\in {\mathscr {L}}\) and an arbitrary subset \(C_1 \subseteq C\), there exists \(K_1\subseteq K_0\) such that the indexes of \(K_1\) in \(K_0\) are the same as those of \(C_1\) in C. Similarly, there exist a sequence \(C_m\subseteq \cdots \subseteq C_2 \subseteq C_1\subseteq C_0\) and the corresponding indexes sets \(K_m\subseteq \cdots \subseteq K_2 \subseteq K_1\subseteq K_0\). The following equations hold
3 Complement model and lattice model
In order to extend the application of multi-scale decision table, Li and Hu (2017) proposed complement model and lattice model.
3.1 Complement model
Let \(S=(U,C\cup \{d\})\) be a multi-scale decision table, and \(S^+\) be its complement system. Let \(I_i,i=1,2,\cdots ,m\) be the number of levels of scales of attribute \(a_i\), respectively, and they are not necessary to be the same. For some attributes with less number \(I_i\), we complement them with several known levels of scales to obtain a new multi-scale decision table, whose attributes have the same number of levels of scales.
Let \(I=\max \{I_1,I_2,\cdots ,I_m\}\), that is, the maximum of \(I_i\) and p be the index of attribute with the largest number of levels of scales. In case of multiple occurrences of the maximum values, the index corresponding to the first occurrence is returned. Firstly, the concept of scale vector is introduced.
Definition 10
(Li and Hu 2017) Let \(S=(U,C\cup \{d\})\) be a multi-scale decision table. Attribute \(a_i\in C\) has \(I_i\) levels of scales. \(C_i=(1,2,\ldots ,I_i)\) is called the original scale vector of \(a_i\), and \(C_i^+\) is the corresponding complement scale vector.
In order to ensure that the number of levels of scales of all the attribute are all the same in \(S^+\), other complement scale vector \(C_i^+(i\ne p)\) should be formed as \((l_{i1},l_{i2},\ldots ,I_{ij},l_{iI})\), where \(1\leqslant l_{ij} \leqslant I_i\) and \(l_{ij}\leqslant l_{ik}\) when \(j\leqslant k\). Moreover, to include more information about \(a_i\), the original scale vector \(C_i\) should be covered by \(C_i^+\). Thus, \(C_i^+\) should satisfy the following conditions:
(C1) \(l_{i1}=1,l_{iI}=I_i\)
(C2) \(dim(C_i^+)=I\)
(C3) \(0\leqslant l_{i,j+1}-l_{ij} \leqslant 1,\forall 1\leqslant j \leqslant I-1\)
Therefore, the number of possible choice for scale vector \(C_i^+\) of \(a_i\) is equivalent to choose \(I-I_i\) from \(I_i\) with replacement (Li and Hu 2017). According to Brualdi (2010), there are \(\left( {\begin{array}{c}I-1\\ I_i-1\end{array}}\right) \) choices of \(C_i^+\). Hence, we can get \(\begin{array}{c} \prod _{i=1}^m \left( {\begin{array}{c}I-1\\ I_i-1\end{array}}\right) \end{array}\) different new multi-scale decision tables.
These new multi-scale decision tables can be decomposed into I decision tables with the same decision attribute where I is the number of levels of scales. We can choose the table with optimal scales combination (Wu and Leung 2013).
3.2 Lattice model
Let \(S=(U,C\cup \{d\})\) be a multi-scale decision table, and \(I_i(i=1,2,\cdots ,m)\) be the number of levels of scales of \(a_i\) which are not necessary to be the same. \({\mathscr {L}}\) is scales collection of S, and \(|{\mathscr {L}}|=\begin{array}{c} \prod _{i=1}^m I_i \end{array}\). If S is consistent, then there exists the scales combination K which is the optimal scales combination of S. Lattice model is aimed to select all the optimal scales combination.
For a given multi-scale decision table, lattice model can be described via the following procedure:
-
1.
According to Definition 6, scales collection \({\mathscr {L}}\) of S can be calculated.
-
2.
Based on the consistence and the partial order relation between elements in \({\mathscr {L}}\) of S, the set of optimal scales combinations OSC can be obtained according to Definition 8.
-
3.
In the subsystem \(S^K\) confirmed by the optimal scale combination \(K\in OSC\), knowledge acquisition and other tasks can be done.
4 Optimal scale combination selection models for inconsistent multi-scale decision tables
4.1 Consistence of positive region
Let \(S=(U,C\cup \{d\})\) be a multi-scale decision table, where \(U=\{x_1,x_2,\ldots ,x_n\}\), \(C=\{a_1,a_2,\ldots , a_m\}\) and \(I_i\) is the number of levels of scales of \(a_i, (i=1,2,\cdots ,m)\). If \({\mathbf {1}}_m \preceq K_1\preceq K_2\preceq (I_1;I_2;\ldots ;I_m)\) and \(S^{K_1}\) is an inconsistent decision table, according to Proposition 1, \(S^{K_2}\) is also inconsistent. Hence, if S is inconsistent, that is \(S^{{\mathbf {1}}_m}\) is inconsistent, then \(S^K\) is also inconsistent for any \(K\in {\mathscr {L}}\).
For \(K\in {\mathscr {L}}\), there is an equivalence relation as follow:
For \(X\in U\), the upper and lower approximations of X are shown as:
where \([x]_{C^K}=\{y\in U|(x,y)\in R_{C^K}\}\).
The positive region under scale combination K in S is defined as (Wu and Leung 2013; Li et al. 2017):
where \(U/d=\{D_1,D_2,...,D_r\}\).
Definition 11
(Li et al. 2017) Let \(S=(U,C\cup \{d\})\) be a multi-scale decision table, where \(U=\{x_1,x_2,\ldots ,x_n\}\), \(C=\{a_1,a_2,\cdots ,a_m\}\) and \(I_i\) is the number of levels of scales of \(a_i, (i=1,2,\ldots ,m)\). For \(K\in {\mathscr {L}}\), if \(POS_{C^K}(d)=POS_{C^{{\mathbf {1}}_m}}(d)\), S is said to be positive region consistent. If \(S^K\) is positive region consistent and \(S^{K'}\)(if there exists \(K'\),\(K\preceq K'\) and \(K'\in {\mathscr {L}}\)) is not positive region consistent, then K is the positive region scale combination of S.
The algorithm of judging whether or not a given subsystem of a multi-scale decision table is positive region consistent has been given by Li et al. (2017).
4.2 Complement model and lattice model based on positive region consistence
We extend the application of complement model and lattice model in this subsection.
The complement model and lattice model proposed by Li and Hu are aimed to process the consistent multi-scale decision table. Therefore, we can combine positive region consistence with these two models to obtain the complement model and lattice model based on positive region consistence which can deal with inconsistent multi-scale decision tables.
In order to deal with multi-scale decision table by using complement model and lattice model based on positive region consistence, we only need to replace the judgement of consistence of subsystem with the judgement of positive region consistence in complement model and lattice model.
And for \(K\in {\mathscr {L}}\), “\(S^K\) is consistent” is included by “\(S^K\) is positive region consistent” (Li et al. 2017). Hence, complement model and lattice model are included by the complement model and lattice model based on positive region consistence. The new models can deal with all kinds of multi-scale decision tables.
Then, we propose the algorithm of complement model and lattice model based on positive region consistence.
5 Numerical experiments
In order to verify the feasibility of complement model(CM-PR) and lattice model(LM-PR) based on positive region consistence, some numerical experiments are employed in this section. And we compare the results of them with the result of stepwise optimal scale selection based on positive region consistence(SOSS-PR) proposed by Li et al. (2017).
Example 1
Table 1 is an inconsistent multi-decision table \(S=(U,C\cup \{d\})\), where \(U=\{x_1,x_2,\cdots ,x_{20}\}\), \(C=\{a_1,a_2,a_3,a_4\}\). We can notice that \(x_4\) and \(x_6\) are indistinguishable w.r.t. \(R_C\), but \(d(x_4)\ne d(x_6)\). The results obtained by using CM-PR, LM-PR and SOSS-PR, respectively, are shown in Table 2.
In order to evaluate the above algorithms more objectively, two data sets are collected from the University of California, Irvine (UCI) Machine Learning Repository (Lichman 2013). These two decision tables are single-scale decision tables. Thus, we use the method by Li and Hu (2017) to obtain their corresponding multi-scale decision tables. There are four steps in that method, but we only do the first three steps. The decision value of object x in multi-scale decision table is not change to \(\partial _{C^{(1;1;...;1)}}(x)\), that is, it keeps the original decision value. Then, the multi-scale decision tables we obtain are not inconsistent. And the details of them and the results are shown in Tables 3, 4, respectively.
Through the numerical experiments, it can be found that the optimal scale combination of CM-PR is weaker than that of LM-PR and the result of SOSS-PR is one of the results of LM-PR. Moreover, the running time of SOSS-PR is shorter than that of LM-PR. These conclusions are similar to the conclusions of the model deal with consistent multi-scale decision tables summarized by Li and Hu (2017) and Li et al. (2017).
In order to test the performance of CM-PR and LM-PR in consistent multi-scale decision tables, we use complement model, lattice model, stepwise optimal scale selection, CM-PR, LM-PR, SOSS-PR to deal with some consistent multi-scale decision tables, respectively. Tables 5, 6 and 7 are three consistent multi-scale decision tables collected from Li and Hu (2017). The results are shown in Tables 8, 9.
For Table 5, the set of optimal scales combination via complement model and CM-PR is {(3;2;3;2),(3;3;3;1)}, the set of optimal scales combination via lattice model and LM-PR is {(4;3;1;2),(4;2;2;2),(1;3;3;2),(4;3;2;1),(3;2;3;2),(3;3;3;1),(4;1;3;2)}, and the optimal scales combination via stepwise optimal scale selection and SOSS-PR is (4;3;1;2).
For Table 6, the set of optimal scales combination via complement model and CM-PR is {(2;1;3;3;3),(2;2;2;2;2),(1;2;3;3;3)}, the set of optimal scales combination via lattice model and LM-PR is {(2;2;4;1;3),(1;2;4;3;3),(2;1;4;3;3),(1;2;3;3;4), (2;2;4;3;2), (2;2;3;1;4),(2;2;2;3;3), (2;1;3;3;4)}, and the optimal scales combination via stepwise optimal scale selection and SOSS-PR is (2;2;4;3;2).
For Table 7, the set of optimal scales combination via complement model and CM-PR is {(2;2;2;2;4)}, the set of optimal scales combination via lattice model and LM-PR is {(2;2;2;2;4)}, and the optimal scales combination via stepwise optimal scale selection and SOSS-PR is (2;2;2;2;4).
The results are shown in Tables 8, 9
Moreover, for the data sets described in Table 3, we use the method proposed by Li and Hu (2017) to obtain their corresponding consistent multi-scale decision tables. And optimal scales combination on these two consistent multi-scale decision tables using three models based on consistence and three models based on positive region consistence are shown in Tables 10, 11, respectively.
Compared Table 8 with Table 9 and compared Table 10 with Table 11, some facts are verified. The running times of complement model and CM-PR have no static relationship and the running time of LM-PR is about two times longer than that of lattice model. The running time of SOSS-PR is slightly slower than that of stepwise optimal scale selection. When dealing with the consistent multi-scale decision table, the model based on the positive region consistence has the same results as the model based on consistence. Thus, CM-PR, LM-PR and SOSS-PR are also able to deal with consistent multi-scale decision tables efficiently.
In a word, for the general multi-scale decision tables, we can directly use the model based on positive region consistence to deal with them. For single-scale decision tables, we can only do the first three steps in the method proposed by Li and Hu (2017) to obtain their corresponding multi-scale decision tables. Generalized decision values are not need to calculate. Finally, the same results can be obtained by using the models based on positive region consistent. Moreover, the optimal scales combination obtained after converting single scale decision table to multi-scale decision table often has excellent performance in classification experiments (Li et al. 2017).
6 Conclusions
Based on the models and theories proposed by Li and Hu, this paper introduces some new methods which are more generalized to calculate all the optimal scale combinations in inconsistent multi-scale decision tables. It is also an expansion of the multi-scale decision tables studied by Wu and Leung, breaking their two strong assumptions.
Some numerical experiments are employed to verify that the new models have the same properties as the complement model and lattice model in dealing with the inconsistent multi-scale decision table. And for consistent multi-scale decision tables, the model based on positive region consistence can also get the same results.
In fact, “the consistence of the single-scale subsystem” is included by the “positive region consistence of it.” Therefore, the model based on positive region consistent can efficiently solve the problem of optimal scale selection in the consistent and inconsistent multi-scale decision tables.
However, the lattice model based on positive region consistence which can always calculate all the optimal scale combinations is more time-consuming and costly. So future research scopes will also include how to optimize the algorithm of lattice model and reduce the time complexity so that it can perform well in dealing with large data sets.
Data Availability
Enquiries about data availability should be directed to the authors.
References
Brualdi RA (2010) Introductory combinatorics, 5th Edition. Pearson Education
Bao H, Wu W-Z, Zheng J-W, Li T-J (2021) Entropy based optimal scale combination selection for generalized multi-scale information tables. Int J Mach Learn Cybern 12:1427–1437
Bharati S, Podder P, Mondal MRH (2020) Hybrid deep learning for detecting lung diseases from X-ray images. Inform Med Unlocked 20
Bharati S, Podder P, Mondal MRH, Prasath VBS (2021) Medical Imaging with Deep Learning for COVID-19 Diagnosis: A Comprehensive Review arXiv:2107.09602
Bharati S, Podder P, Mondal MRH, Prasath VBS (2021) CO-ResNet: optimized ResNet model for COVID-19 diagnosis from X-ray images. Int J Hybrid Intell Syst 17:71–85
Chen Y-S, Li J-J, Huang J-X (2019) Matrix method for the optimal scale selection of multi-scale information decision systems. Mathematics 7(3):290
Gu S-M, Wu W-Z (2013) On knowledge acquisition in multi-scale decision systems. Int J Mach Learn Cybern 4:477–486
Hao C, Li J, Fan M, Tsang ECC (2017) Optimal scale selection in dynamic multi-scale decision tables based on sequential three-way decisions. Inf Sci 415–416:213–232
Huang B, Li H-X, Feng G-F, Zhou X-Z (2019) Dominance-based rough sets in multi-scale intuitionistic fuzzy decision tables. Appl Math Comput 348:487–512
Huang B, Wu W-Z, Yan J-J, Li H-X, Zhou X-Z (2020) Inclusion measure-based multi-granulation decision-theoretic rough sets in multi-scale intuitionistic fuzzy information tables. Inf Sci 507:421–448
Huang J-X, Li W-K, Zhang X-P, Li J-J (2020) Knowledege acquisition and matrix method of generalized multi-scale information system. J Shanxi Univ 43(4):878–887
Khamparia A, Bharati S, Podder P, Gupta D, Khanna A, Phung TK, Thanh DNH (2021) Diagnosis of breast cancer based on modern mammography using hybrid transfer learning. Multidimension Syst Signal Process 32:747–765
Komorowski J, Pawlak Z, Polkowski L, Skowron A (1999) Rough sets: tutorial. Springer-Verlag, Berlin
Li F, Hu B-Q (2017) A new approach of optimal scale selection to multi-scale decision tables. Inf Sci 381:193–208
Li F, Hu B-Q, Wang J (2017) Stepwise optimal scale selection for multi-scale decision tables via attribute significance. Knowl-Based Syst 129:4–16
Li Y-C, Huang M-X, Zhang Y, Chen J, Xu H-X, Wang G, Feng W-L (2020) Automated gleason grading and gleason pattern region segmentation based on deep learning for pathological images of prostate cancer. Access IEEE 8:117714–117725
Lichman M (2013) UCI machine learning repository. http://archive.ics.uci.edu/ml
Mondal MRH, Bharati S, Podder P (2021) CO-IRv2: Optimized inceptionResNetV2 for COVID-19 detection from chest CT images. PLoS ONE 16:e0259179
Mondal MRH, Bharati S, Podder P (2021) Diagnosis of COVID-19 using machine learning and deep learning: a review. Current Med Imag 17:1403–1418
Pawlak Z (1992) Rough Sets:Theoretical Aspects of Reasoning about Data. Kluwer Academic Publisher
Qian X-L, Fu Y-W, Jiang Y-G, Xiang T, Xue X-Y (2017) Multi-scale deep learning architectures for person Re-identification. Proceedings of the IEEE International Conference on Computer Vision (ICCV) 5399-5408
She Y, Li J, Yang H (2015) A local approach to rule induction in multi-scale decision tables. Knowl-Based Syst 89:398–410
Taverniers Søren, Hall Eric J, Katsoulakis Markos A, Tartakovsky Daniel M (2021) Mutual information for explainable deep learning of multiscale systems. J Comput Phys 444
Wang H, Li W, Zhan T, Yuan K, Hu X (2021) Multi-granulation-based optimal scale selection in multi-scale information systems. Comput Electr Eng 92:107107
Wu W-Z, Leung Y (2011) Theory and applications of granular labelled partitions in multi-scale decision tables. Inf Sci 181:3878–3897
Wu W-Z, Leung Y (2013) Optimal scale selection for multi-scale decision tables. Int J Approximate Reason 54:1107–1129
Wu W-Z, Qian Y, Li TJ, Gu S-M (2017) On rule acquisition in incomplete multi-scale decision tables. Inf Sci 378:282–302
Wu W-Z, Leung Y (2020) A comparison study of optimal scale combination selection in generalized multi-scale decision tables. Int J Mach Learn Cybern 11:961–972
Xie J-P, Yang M-H, Li J-H, Zheng Z (2018) Rule acquisition and optimal scale selection in multi-scale formal decision contexts and their applications to smart city. Futur Gener Comput Syst 83:564–581
Zhan J-M, Zhang K, Wu W-Z (2021) An investigation on Wu-Leung multi-scale information systems and multi-expert group decision-making. Expert Syst Appl 170:114542
Acknowledgements
The authors are extremely grateful to the editor and anonymous referees for their valuable comments and helpful suggestions which helped to improve the presentation of this paper.
Funding
This research was supported by the National Natural Science Foundation of China (Grant no. 12101500), the Chinese Universities Scientific Fund (Grant no. 2452018054) and the College Students’ Innovation and Entrepreneurship Training Program (Grant No. X202010712184).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper
Human and animal rights
This article does not contain any studies with human participants or animals performed by the author.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Yingjie, Z., Bin, Y. Optimal scale combination selection for inconsistent multi-scale decision tables. Soft Comput 26, 6119–6129 (2022). https://doi.org/10.1007/s00500-022-07102-y
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00500-022-07102-y