International Journal of Database Theory and Application Vol.9, No.5 (2016), pp.77-90 http://dx.doi.org/10.14257/ijdta.2016.9.5.08 A Novel Soft Set Approach for Feature Selection Daoli Yanga, Zhi Xiaoa,b,1,2, Wei Xua, Xianning Wanga and Yuchen Pana a School of Economics and Business Administration, Chongqing University, Chongqing 400044, PR China b Chongqing Key Laboratory of Logistics, Chongqing University, Chongqing 400044, PR China xiaozhicqu@163.com Abstract Feature selection is an important preprocess for data mining. The soft set theory is a new mathematical tool to deal with uncertainties. Merging the soft set theory into the feature selection process based on rough sets facilitates the computation with equivalence classes and improves the efficiency. We propose a paired relation soft set model based on equivalence classes of the information system. Then we use it to present the lower approximate set in the form of soft sets and calculate the degree of dependency between relations. Furthermore, we give a new mapping to obtain equivalence classes of indiscernibility relations and propose a feature selection algorithm based on the paired relation soft set model. Compared to the algorithm based NSS, this algorithm shows 18.17% improvement on an average. Meantime, both of the algorithms show a good scalability. Keywords: soft set, feature selection, rough set, information system 1. Introduction Feature selection is an important issue in data mining. In the big data era, the scale of databases grows rapidly with an increasing number of features due to availability of more detailed information. The existence of irrelative features is inevitable. Feature selection can eliminate those features so that it reduces the dimensionality, improves the predictive accuracy and enhances comprehensibility of the induced concepts [1]. Besides, features may have approximate relationships with each other. Using a subset of features to represent other features and deduce the classification results is an imprecise process. The rough set theory is a powerful tool to process such kind of imprecision [2-4]. Thus, it is incorporated into feature selection algorithms. However, the computation with equivalence classes in these algorithms is inconvenient and hard to understand. The soft set theory can facilitate the computation and improve the efficiency. Recent literatures of rough set based feature selection algorithms mainly focus on three aspects: feature evaluation, search strategies and practical application. Feature evaluation provides the significance of the candidate features which is represented by dependency degree in the rough set theory. For different nature of data, calculation methods of the dependency degree are different. Several types of data have been studied, such as fuzzy data [5], data with noise [6], dynamic incomplete data [7], hybrid data [8], etc. As for search strategies, there are commonly used methods like sequential forward selection and sequential backward elimination [6], as well as many heuristic methods which are hybrid genetic algorithm [9], granular neural network [5], water drops algorithm [10], etc. Some literatures explore feature selection based on rough set for practical application in 1 2 Corresponding Author, email: xiaozhicqu@163.com Supported by the National Science Foundation of China (Grant No. 71171209) ISSN: 2005-4270 IJDTA Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) domains like computer security [11] and assessment of power systems [12]. But researches focusing on the improvement in efficiency when calculating the dependency degree of features are few. This kind of improvement is also crucial in improving performance of rough set based feature selection. The soft set theory was proposed by D. MOLODTSOV in 1999 [13]. It offers a rigid mathematical theory to deal with uncertainties. Unlike the probability theory and the fuzzy set theory suffering from the inadequacy of the parameterization tools, the soft set theory is free of those problems. This makes the theory easily applicable in practice [13] and very convenient to combine with other theories. In recent years, studies about the soft set theory are growing rapidly in many fields such as modern algebra [14], forecasting [15], decision making [16], data analysis [17], etc. It has been proved that every rough set is a soft set, which shows the necessity to combine two theories together [18]. As for the dependency degree in the rough set theory, Qin et al. [19] had found an efficient way to calculate it by constructing a soft set model (NSS) for an information system. Later, Mamat et al. [20] gave a new clustering method of attributes based on the same soft set model. Both of them suggest that the method based on the soft set theory is efficient and easy to apply. However, there are two remaining issues they don’t tackle. One is how to get equivalence classes of the indiscernibility relation of features based on the soft set theory and the other is how to apply the soft set model to feature selection which is different from feature cluster. Our main contributions are summarized as follows: (a) We propose a paired relation soft set model (PRSS) based on the information system. We use it to construct the lower approximate set, present the approximate set in the form of soft set and furthermore calculate the dependency degree between features. (b) We propose a method to acquire equivalence classes of indiscernibility relations based on the PRSS. (c) We present a feature selection algorithm based on the PRSS and apply it to seventeen UCI benchmark data sets. The result shows improvement in efficiency compared to the method based on NSS. The rest of our paper contains four sections. Section 2 presents the fundamental theory. Section 3 proposes the paired relation soft set model, a method to acquire equivalence classes of indiscernibility relations and the relative feature selection algorithm. Section 4 compares experimental results with the NSS method. The final section concludes our work. 2. Essential Rudiments 2.1. Rough Set Theory Definition 1 (Information system [4]): Suppose there is a pair =( , ) of non-empty, finite sets and , where is the universe set of objects, and is a set consisting of attributes, i.e. functions , where is the set of values of attribute , called the domain of . The pair =( , ) is called an information system. Sometimes, we distinguish in an information system =( , ) a partition of A into two classes , of attributes, called condition and decision (action) attributes, respectively. The tuple called a decision system [4]. Definition 2 (Indiscernibility relation [3]): In an information system , if and , then (intersection of all equivalence relations belonging to ), denoted by , will be called an indiscernibility relation over . The equivalence class of involving element is denoted by . The denotes the family of all equivalence classes of 78 Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) . Meantime we denote the family of all equivalence relations defined in as . Definition 3 (The lower and upper approximations [3]): With each subset and an equivalence relation , we associate two subsets: (1) Called the -lower and -upper approximations of respectively. Definition 4 (Degree of partial dependency [3]): Let =( , ) be an information system and , . We say that depends in a degree ( ) from , if and only if (2) Where is denotes the cardinality of the set and the -positive region of . Dependency degree shows the importance of for (3) . 2.2. Soft Set Theory Let be the initial universe set and be a set of parameters. Definition 5 (Soft set [13]): A pair of is called a soft set (over ) if and only if is a mapping of into the set of all subset of the set . In other words, the soft set is a parameterized family of subsets of the set . Every set , , from this family may be considered as the set of -elements of the soft set , or as the set of - approximate elements of the soft set [13]. Definition 6 (Soft subset [21]): For two soft sets and over a common universe set , the is a soft subset of if and , and are identical approximations. It is denoted as . Apparently, for a set of - approximate elements of , we can define a corresponding soft set which has only one parameter . . Let and be soft sets over , then operation * for soft sets [13] is , Where , , the sets and . Example 1: There are five houses evaluated by three criterions , and , and (4) is the Cartesian product of which are ready to be . Suppose . Then we get the soft set . If we want to know which houses are both expensive and beautiful, we get , and execute the intersection operation as follows (5) , (6) That is . 2.3. A Soft Set Model on Equivalence Class Computation in the rough set theory always involves intersection, union of the sets of equivalence classes and the dependency degree between attributes. It is inconvenient to execute these operations directly on the information system. Thus, Qin et al. [19] constructed a soft set model over equivalence classes to facilitate the computation. Copyright ⓒ 2016 SERSC 79 International Journal of Database Theory and Application Vol.9, No.5 (2016) Given an information system =( , ), denotes the set of all equivalence classes in the partitions , where . Let = be the initial universe set of objects and = be the set of parameters. denotes the power set of . Meantime, define mapping . The pair is a soft set model over equivalence classes (NSS) [19]. Table 1 shows the tabular representation of , where and is either 1 or 0, , . Table 1. The Tabular Representation of the Soft Set over Equivalence Classes Then they defined two soft set and where , : and (7) The lower and upper approximations of parameter defined as with respect to attribute , where . The lower and upper approximations of attribute defined as with respect to attribute are (8) are (9) where . In the following, we just present an example for soft set . is similar. Example 2: we consider the information system of house evaluation as shown in Table 2. Then, Then they The lower and upper approximations of parameter with respect to attribute are defined h The lower and upper approximations of attribute with respect to attribute are he In the following, we just present an example for soft set . is similar. Example 2: we consider the information system of house evaluation as shown in Table 2. partitions of each attribution are , and . The soft set is shown in Table 3. For , , and , , = and = . Thus, 80 = . Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) Table 2. An Information System of House Evaluation price appearance material expensive beautiful wood cheap beautiful metal cheap ugly brick Table 3. The Tabular Representation of Soft Set Based on NSS 1 0 1 0 1 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 1 0 0 1 1 0 1 0 1 0 0 0 1 1 0 0 1 0 0 1 0 1 0 0 1 3. Proposed Paired Relation Soft Set Technique 3.1. Paired Relation Soft Set Given an information system =( , ) and two relations , , . When examining the upper and lower approximation sets of with respect to , it is unnecessary to include all the equivalence classes of them into . As shown in Table 3, computation in cells with color is redundant. Thus, we change the initial universe set and parameter of NSS. Definition 7 (Paired relation soft set): Let and . Define the mapping . A pair of is called paired relation soft set for its universe set and parameters stemming from two relations and , respectively. Table 4. The Tabular Representation of PRSS \ Example 3: For , example 2, let and and . . , and . The tabular representation of PRSS for the appearance and price is shown in Table 4. Then we can construct the soft sets and , where mappings and are in the formula 7. The tabular representation of soft set and for appearance and price in Example 2 is shown in Table 5. Copyright ⓒ 2016 SERSC 81 International Journal of Database Theory and Application Vol.9, No.5 (2016) Table 5. The Tabular Representation of Soft Sets on PRSS and appearance Based appearance \ \ pric e 1 0 0 0 pric e 1 0 1 1 3.2. Computation of Dependency Degree Based on PRSS According to formula 2, the computation of the dependency degree requires the lower approximate set and the positive region. We firstly give a soft set with two parameter sets. Then, based on that soft set and , we present the soft set representation of the lower approximate set and the computation of dependency degree. A soft set with two parameter sets is shown as follows. Given an information system =( , ) and two relations , , . Let , . Define mapping . We get soft set . Meantime, let , , , and . The lower approximation set of parameter with respect to is represented by soft set where , is as shown in Table 5. Apparently, The lower approximation of parameter with respect to where (10) . is represented by soft set . Then, Compare (11) . with , we have = = (12) where . The lower approximation of relation with respect to is represented by soft set . Then, Compare (13) . and , we get, (14) where . Apparently, soft set is the -positive region of (see formula 2 and 3). To get the dependency degree, we also need the cardinality of . . (15) However, if we only want to get the dependency degree, it is unnecessary to calculate first. In the following, we provide a direct way to get the dependency degree. Based on soft set (see table 5), the cardinality of is 82 Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) . The cardinality of (16) is , The cardinality of (17) is . The degree of dependency of relation with respect to (18) is . Given a decision system represented by , (19) , the positive region of on is (20) The dependency degree of decision on is represented by . (21) Example 4: we present the low approximate set and the degree of dependency for (appearance) and (price) in Example 3 based on in Table 5. , , , . , , . The lower approximation of relation For the calculation of cardinality, with respect to , is . , , . , , . Then the dependency degree of relation with respect to is . 3.3. Computation of Equivalence Classes of Indiscernibility Relations We first introduce a new mapping as a supplement to mappings in formula 7. Then, represent equivalence classes of the indiscernibility relation formed by two relations using a soft set based on . Given a PRSS where the universal set , parameter set and , , , we propose Definition 8. Definition 8: Let be the power set of , and .We define mapping , for The soft set based on mapping . Copyright ⓒ 2016 SERSC is denoted by . . Let (22) and , 83 International Journal of Database Theory and Application Vol.9, No.5 (2016) Suppose and , , and . Define mapping . The equivalence classes of the indiscernibility relation formed by are represented by where . (23) Then . Example 5: For (appearance) and (price) in example 3, the tabular representation of is shown in Table 6. Then, . Table 6. The Tabular Representation of appearance \ price 3.4. Feature Selection Algorithm Based on PRSS Given a decision system , the feature selection algorithm based on PRSS is shown in Table 7. The algorithm selects features according to the dependency degree of decision provided by Algorithm 2 and uses the sequential forward selection [6] as the search strategy provided by Algorithm 1. The output is a feature ranking . Table 7. The Feature Selection Algorithm Based on PRSS Algorithm 1 Input X, Outpu t Algorithm 2 X is a sample set and , and are in step 2 of algorithm 1 , , is the degree F is a feature ranking Begin Step 1: Initialize While Step 2: Find Step 3: Step 4: End Return End Given the set with Begin If , then Construct of and based on PRSS Calculate Else Construct soft set based on PRSS Calculate Construct of and based on PRSS Calculate End End features selected, the th feature is determined by . 84 dependency (24) Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) Based on the ranking, we can get a feature ranking soft set and the is defined as , where . (25) Next, we use KNN classifier to cross-validate the classification accuracy of the data with these feature subsets. Those with the highest classification accuracy are the final feature subsets. Table 8. The Algorithm for the Degree of Dependency Based on NSS Algorithm 3 Input , , , Output and are in step 2 in table is the dependency degree Begin If , then Construct Calculate Else Construct soft set Calculate Construct Calculate End End of of and and based on NSS. according to formula 9. based on NSS. . based on NSS. . 3.5. Feature Selection Algorithm Based on NSS Feature selection algorithm based on NSS still uses Algorithm 1 as the search strategy. But the Algorithm 3 in Table 8 which provides the dependency degree of decision is different from Algorithm 2. Let and be the set of all equivalence classes in the partitions and . Example 6: Suppose and . , . Then is shown in Table 9. The construct of follows the same logic of Table 3. Table 9. The Tabular Representation of price appearance \ price appearance Copyright ⓒ 2016 SERSC 85 International Journal of Database Theory and Application Vol.9, No.5 (2016) 4. Experimental Results We compare the execute time of Algorithm 1 and 2 based on PRSS with that of Algorithm 1 and 3 basd on NSS using seventeen datasets abtained from the benchmark UCI Machine Learning Repository [22]. Those datasets are discribed in Table 10. The field named “Number of Classes” in this table shows the aggregation of the number of equvalence classes formed by each feature. We run the algorithms in MATLAB R2015a on a pc with a processor Intel(R) Core(TM) i5-2410M, a 8.0GB memory and Widows 7 operating system. Figure 1 shows datasets on which ranges of the executive time are (0 , 0.1 ) and (0.1, 1) in seconds. Figure 2 shows datasets on which ranges of the executive time are (1, 20) and (20, 120) in seconds. Table 11 summrizes the comparison of feature selection algorithm based on PRSS and NSS. It shows that PRSS improves efficiency to a certain degree on each dataset and achieves 18.17% improvement on an average. Figure 1. Excutive Time (Second) of Feature Selection Based on PRSS and NSS Table 10. Description of Datasets No Data sets Number of instances Number of features Number of Classes 1 2 3 4 Abalone Adult AI(Acute Inflammations ) CE(Car Evaluation) Chess(King-Rook vs. KingPawn) CMC(Contraceptive Method Choice) DER(Dermatology) Ecoli LC(Lung Cancer ) Lenses Mushroom (MR) Musk Nursery SFL(Solar Flare large) SFS(Solar Flare small) Statlog (statlog_satimage) 4177 20 120 1728 9 4 8 7 6077 10 58 25 3196 37 75 1473 10 74 366 336 32 24 8124 476 12960 1066 323 4435 34 9 55 6 23 169 9 13 13 37 135 707 154 36 119 25760 32 48 40 2752 5 6 7 8 9 10 11 12 13 14 15 16 86 Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) 17 Yeast 1484 10 1884 Figure 2. Excutive Time (Second) of Feature Selection Based on PRSS and NSS Figure 3. The Scalability of PRSS and NSS to the Number of Features Table 11. Comparison of Feature Selection Algorithm of PRSS and NSS No Data sets 1 2 3 4 Abalone Adult AI(Acute Inflammations ) CE(Car Evaluation) Chess(King-Rook vs. 5 King-Pawn) CMC(Contraceptive 6 Method Choice) 7 DER(Dermatology) 8 Ecoli 9 LC(Lung Cancer ) 10 Lenses 11 Mushroom 12 Musk 13 Nursery 14 SFL(Solar Flare large) 15 SFS(Solar Flare small) 16 Statlog(statlog_satimage) 17 Yeast Average improvement Copyright ⓒ 2016 SERSC Executing time (s) NSS PRSS 6.645 6.620825 0.0047 0.0032 0.0128 0.008 0.0282 0.0239 Improvement 10.3424 10.0919 2.42% 0.3662 0.3511 4.10% 0.5524 0.06 0.4183 0.0052 10.42 60.4035 1.5058 0.093 0.0283 106.3707 0.9246 0.4227 0.0506 0.1632 0.0027 10.2518 56.4015 1.4757 0.078 0.0173 105.8892 0.9113 23.49% 15.68% 60.99% 47.93% 1.61% 6.63% 2.00% 16.09% 38.95% 0.45% 1.44% 18.17% 0.37% 33.50% 37.80% 15.40% 87 International Journal of Database Theory and Application Vol.9, No.5 (2016) Figure 4. The Scalability of PRSS and NSS to the Number of Instances Figure 5. The Scalability of PRSS and NSS to the Number of Equvalence Classes We also analyze the scalability of both algorithms on three aspects, the number of features, the number of instances and the number of equvalence classes. Figure 3 shows the execute time of both algorithms with regard to the number of features in each dataset. Figure 4 shows the execute time of both algorithms with regard to the number of instances in each dataset. We also provide the change of the execute time with regard to the aggregation of the number of equvalence classes formed by each feature in Figure 5. In general, three figures show the execute time increase linearly as the number of features, instances and equvalence classes increase, respectively. Though the execute time varies at several points, it is at an acceptabe level. Thus, those three figures demonstrate a good scalibility of both feature selection algorithms. 5. Conclusion Merging the soft set theory into some core conceptions of the rough set theory facilitates the computation based on equvalence classes. We proposed a paired relation soft set model to construct the lower approximate set, presented the lower approximate set in the form of soft sets, and furthermore calculated the dependency degree of features. We also provided a new mapping to calculate equivalence classes of indiscernibility relations. Based on PRSS, the feature selection algorithm was given. Compared to the feature selection algorithm based on NSS, the PRSS algorithm shows 18.17% improvement on an average. Experimental results also show a good scalibility of both feature selection algorithms. Theoretically, the paper shows a way to combine the soft set theory and the rough set theory together to improve the feature selection algorithm. Practically, the novel method based on PRSS can accelerate data processing procedure and is helpful in dealing with big data. 88 Copyright ⓒ 2016 SERSC International Journal of Database Theory and Application Vol.9, No.5 (2016) In this paper, we only combined the soft set theory with the classic rough set model which is based on equivalence classes. To deal with data with fuzzy attributes, a fuzzy rough set model based on fuzzy similarity relations is more suitable. How to apply PRSS into a fuzzy rough set model and facilitate the computation will be a topic of furture reasearches. Acknowledgment We are very grateful to referees and editors. Their comments are very helpful in improving the paper. We also appreciate the research team members. References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] L. H. and M. H, “Feature Selection for Knowledge Discovery and Data Mining”, Kluwer, Boston, (1998). Z. Pawlak, “Rough Sets”, International Journal of Computer & Information Sciences, vol. 11, no. 5, (1982), pp. 341-356. Z. Pawlak, “Rough Sets: A Theoretical Aspect of Reasoning About Data”, Kluwer Academic Publisher, (1991). Z. Pawlak and A. Skowron, “Rudiments of Rough Sets”, Information Sciences, vol. 177, no. 1, (2007), pp. 3-27. A. Ganivada, S. S. Ray and S. K. Pal, “Fuzzy Rough Sets, and a Granular Neural Network for Unsupervised Feature Selection”, Neural Networks, vol. 48, (2013), pp. 91-108. Q. H. Hu, S. A. An and D. R. Yu, “Soft Fuzzy Rough Sets for Robust Feature Evaluation and Selection”, Information Sciences, vol. 180, no. 22, (2010), pp. 4384-4400. W. H. Shu and H. Shen, “Incremental Feature Selection Based on Rough Set in Dynamic Incomplete Data”, Pattern Recognition, vol. 47, no. 12, (2014), pp. 3890-3906. A. P. Zeng, T. R. Li, D. Liu, J. B. Zhang and H. M. Chen, “A Fuzzy Rough Set Approach for Incremental Feature Selection on Hybrid Information Systems”, Fuzzy Sets and Systems, vol. 258, (2015), pp. 39-60. S. Y. Jing, “A Hybrid Genetic Algorithm for Feature Subset Selection in Rough Set Theory”, Soft Computing, vol. 18, no. 7, (2014), pp. 1373-1382. B. O. Alijla, L. C. Peng, A. T. Khader and M. A. Al-Betar, “Intelligent Water Drops Algorithm for Rough Set Feature Selection”, Intelligent Information and Database Systems (Aciids 2013), Pt Ii, vol. 7803, (2013), pp. 356-365. V. Rampure and A. Tiwari, “A Rough Set Based Feature Selection on Kdd Cup 99 Data Set”, International Journal of Database Theory and Application, vol. 8, no. 1, (2015), pp. 149-156. X. P. Gu, Y. Li and J. H. Jia, “Feature Selection for Transient Stability Assessment Based on Kernelized Fuzzy Rough Sets and Memetic Algorithm”, International Journal of Electrical Power & Energy Systems, vol. 64, (2015), pp. 664-670. D. Molodtsov, “Soft Set Theory - First Results”, Computers & Mathematics with Applications, vol. 37, no. 4-5, (1999), pp. 19-31. G. Muhiuddin, F. Feng and Y. B. Jun, “Subalgebras of Bck/Bci-Algebras Based on Cubic Soft Sets”, The Scientific World Journal, vol. 2014, (2014), pp. 9. W. Xu, Z. Xiao, X. Dang, D. L. Yang and X. L. Yang, “Financial Ratio Selection for Business Failure Prediction Using Soft Set Theory”, Knowledge-Based Systems, vol. 63, (2014), pp. 59-67. Z. Xiao, W. J. Chen, and L. L. Li, “A Method Based on Interval-Valued Fuzzy Soft Set for MultiAttribute Group Decision-Making Problems under Uncertain Environment”, Knowledge and Information Systems, vol. 34, no. 3, (2013), pp. 653-669. Y. Zou and Z. Xiao, “Data Analysis Approaches of Soft Sets under Incomplete Information”, Knowledge-Based Systems, vol. 21, no. 8, (2008), pp. 941-945. T. Herawan and M. M. Deris, “A Direct Proof of Every Rough Set Is a Soft Set”, in 2009 3rd Asia International Conference on Modelling and Simulation, Bandung, Bali, Indonesia, (2009), pp. 119-124. H. W. Qin, X. Q. Ma, J. M. Zain, and T. Herawan, “A Novel Soft Set Approach in Selecting Clustering Attribute”, Knowledge-Based Systems, vol. 36, (2012), pp. 139-145. R. Mamat, T. Herawan and M. M. Denis, “Mar: Maximum Attribute Relative of Soft Set for Clustering Attribute Selection”, Knowledge-Based Systems, vol. 52, (2013), pp. 11-20. P. K. Maji, R. Biswas, and A. R. Roy, “Soft Set Theory”, Computers & Mathematics with Applications, vol. 45, no. 4-5, (2003), pp. 555-562. Uci Machine Learning Repository. Available: http://archive.ics.uci.edu/ml/datasets.html Copyright ⓒ 2016 SERSC 89 International Journal of Database Theory and Application Vol.9, No.5 (2016) Authors Daoli Yang, He received the B.A. degree in Electronic Business from Shandong University in 2006. Currently, he is a Ph.D. candidate in the School of Economics and Business Administration at Chongqing University. His research interests include supply chain management, information intelligent analysis and soft set theory. Zhi Xiao, He received the B.S. degree in Department of Mathematics from Southwest Normal University in 1982. And he received his Ph.D. degree in Technology Economy and Management from Chongqing University in 2003. He is currently a Professor in the School of Economics and Business Administration at Chongqing University. His research interests include information intelligent analysis, data mining, and soft set theory. Wei Xu, He received the B.A. degree in Logistic Management from Jiangxi University of Finance and Economic in 2008. He received his Ph.D. degree in Quantitative Economics from Chongqing University in 2014. His research interests include intelligent decision-making, information intelligent analysis and business Intelligence. Xianning Wang, He is a Ph.D. candidate in the School of Economics and Business Administration at Chongqing University from 2012. His research interests mainly include dynamic economics modeling and regional economy development. Yuchen Pan, She is a Ph.D. candidate in the School of Economics and Business Administration at Chongqing University. Her research interests include information intelligent analysis and prediction, and financial information application. 90 Copyright ⓒ 2016 SERSC