On a Hybridization of Deep Learning and Rough Set Based Granular Computing
<p>The diagram shows a scheme of our experimental part. The exact design of the neural network is in <a href="#algorithms-13-00063-f003" class="html-fig">Figure 3</a>. The data that is fed into the neural network is normalized after granulation to a range of <math display="inline"><semantics> <mrow> <mo><</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>></mo> </mrow> </semantics></math>.</p> "> Figure 2
<p>A diagram of the neural network used to learn the Australian credit system. Neural networks for the other two systems differ only in the number of inputs determined by the number of conditional attributes.</p> "> Figure 3
<p>Results for 10 learning cycles, using 10 splits; for Australian credit data set; In ‘percentage of objects’ ax, we have the percentage size of granulated data vs accuracy of classification in ‘Accuracy’ ax; in. The results are not perfectly evenly matched or at the same points on the x-axis, due to the fact that the size reduction levels of the training systems varied.</p> "> Figure 4
<p>Mean result for 10 learning cycles, using 10 splits; for Australian credit data set; The only way to show the average values from the experiments was to calculate the average accuracy for specific granulation radii. Hence, on the x-axis we have the granulation radii (approximation levels). The figure shows the result from <a href="#algorithms-13-00063-t005" class="html-table">Table 5</a>.</p> "> Figure 5
<p>Results for 10 learning cycles, using 10 splits; for Heart Disease data set; In ‘percentage of objects’ ax, we have the percentage size of granulated data vs accuracy of classification in ‘Accuracy’ ax; in. The results are not perfectly evenly matched or at the same points on the x-axis, due to the fact that the size reduction levels of the training systems varied.</p> "> Figure 6
<p>Mean result for 10 learning cycles, using 10 splits; for Heart disease data set; Mean result for 10 learning cycles, using 10 splits; for Australian credit data set; The only way to show the average values from the experiments was to calculate the average accuracy for specific granulation radii. Hence, on the x-axis we have the granulation radii (approximation levels). The figure shows the result from <a href="#algorithms-13-00063-t006" class="html-table">Table 6</a>.</p> "> Figure 7
<p>Results for 10 learning cycles, using 10 splits; for Pima Indians Diabetes data set; In ‘percentage of objects’ ax, we have the percentege size of granulated data vs accuracy of classification in ‘Accuracy’ ax; in. The results are not perfectly evenly matched or at the same points on the x-axis, due to the fact that the size reduction levels of the training systems varied.</p> "> Figure 8
<p>Mean results for 10 learning cycles, using 10 splits; for Pima Indians Diabetes data set; Mean result for 10 learning cycles, using 10 splits; for Australian credit data set; The only way to show the average values from the experiments was to calculate the average accuracy for specific granulation radii. Hence, on the x-axis we have the granulation radii (approximation levels). The figure shows the result from <a href="#algorithms-13-00063-t007" class="html-table">Table 7</a>.</p> "> Figure 9
<p>Results for 10 learning cycles, using 10 splits; for Australian credit data set converted to dummy variables (after conversion to dummy variables its 35 attributes); In ‘percentage of objects’ ax, we have the percentage size of granulated data vs accuracy of classification in ‘Accuracy’ ax; in. The results are not perfectly evenly matched or at the same points on the x-axis, due to the fact that the size reduction levels of the training systems varied.</p> "> Figure 10
<p>Dummy variables - mean result for 10 learning cycles, using 10 splits; for Australian credit data set; after conversion to Dummy variables its 35 attributes.</p> ">
Abstract
:1. Introduction
2. Reducing the Size of Decision-Making Systems Based on Their Granular Reflections
- First step: granulation. We begin with computation of granules around each training object using selected method. In the method used in this article, by surrounding the objects of the training system class with objects indiscernible to the degree determined by the granulation radius.
- Second step: the process of covering. The training decision system is covered by selected granules. After the calculation of granules in point 1, a group of granules that cover the entire training system with their objects is searched for.
- Third step: building the granular reflections. The granular reflection of original training decision system is derived from the granules selected in step 2. We form new objects by converting granules using majority voting.
2.1. Standard Granulation
2.2. Concept Dependent Granulation
2.2.1. Toy Example of Concept Dependent Granulation
3. Design of the Experimental Part
4. Procedure for Performed Experiments
- 1.
- Data input (original decision system),
- 2.
- Data random split in the ratio 70-30 per cent TRN-TST,
- 3.
- Granulation step, covering step, new objects generation—see Section 2.2.1,
- 4.
- Neural network learning step for each set of objects (for each granulation radius) see Figure 1,
- 5.
- Classification step for each test set, based on approximated data see Figure 2,
- 6.
- Compute accuracy, time, number of objects and compare vs first set. The whole procedure is repeated 10 times.
The Results of Experiments
- gran_rad—granulation radius as a percentage value,
- no_of_gran_objects—number of new objects in tested decision system after the granulation process,
- percentage_of_objects—percentage of objects in tested decision system comparing to the primary decision system size,
- time_to_learn—time that was needed to complete the learning process using given data,
- accuracy—classification accuracy for given neural network.
5. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Haykin, S.S. Neural Networks: A Comprehensive Foundation; Prentice Hall: Upper Saddle River, NJ, USA, 1999; ISBN 978-0-13-273350-2. [Google Scholar]
- Połap, D.; Woźniak, M.; Wei, W.; Damaševičius, R. Multi-threaded learning control mechanism for neural networks. Future Gener. Comput. Syst. 2018, 87, 16–34. [Google Scholar] [CrossRef]
- Woźniak, M.; Połap, D. Intelligent Home Systems for Ubiquitous User Support by Using Neural Networks and Rule-Based Approach. IEEE Trans. Ind. Inform. 2020, 16, 2651–2658. [Google Scholar] [CrossRef]
- Novikoff, A.B. On convergence proofs on perceptrons. In Proceedings of the Symposium on the Mathematical Theory of Automata, New York, NY, USA, 24–26 April 1962; Polytechnic Institute of Brooklyn: Brooklyn, NY, USA, 1962; Volume 12, pp. 615–622. [Google Scholar]
- Bryson, A.E.; Ho, Y.-C. Applied Optimal Control: Optimization, Estimation, and Control; Blaisdell Publishing Company: Waltham, MA, USA; Xerox College Publishing: Lexington, MA, USA, 1969; p. 481. [Google Scholar]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Nielsen, M. Neural Networks and Deep Learning; Determination Press: San Francisco, CA, USA, 2015. [Google Scholar]
- Dick, S.; Kandel, A. Granular Computing in Neural Networks. In Granular Computing. Studies in Fuzziness and Soft Computing; Pedrycz, W., Ed.; Physica: Heidelberg, Germany, 2001; Volume 70. [Google Scholar]
- Leng, J.; Chen, Q.; Mao, N.; Jiang, P. Combining granular computing technique with deep learning for service planning under social manufacturing contexts. Knowl.-Based Syst. 2018, 143, 295–306, ISSN 0950-7051. [Google Scholar] [CrossRef]
- Ghiasi, B.; Sheikhian, H.; Zeynolabedin, A.; Niksokhan, M.H. Granular computing-neural network model for prediction of longitudinal dispersion coefficients in rivers. Water Sci. Technol. 2020. [Google Scholar] [CrossRef] [PubMed]
- Pawlak, Z. Rough Sets: Theoretical Aspects of Reasoning about Data; Kluwer: Alphen aan den Rijn, The Netherlands, 1991. [Google Scholar]
- Skowron, A.; Rauszer, C. The discernibility matrices and functions in information systems. In Intelligent Decision Support. Handbook of Applications and Advances of Rough Set Theory; Słowiński, R., Ed.; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1992; pp. 331–362. [Google Scholar]
- Pawlak, Z.; Skowron, A. A rough set approach for decision rules generation. In Proceedings of the IJCAI’93 Workshop W12: The Management of Uncertainty in AI, Chambery Savoie, France, 30 August 1993; ICSResearch Report 23/93. Warsaw University of Technology: Warsaw, Poland, 1993. [Google Scholar]
- Polkowski, L.; Skowron, A. Rough mereology. In Proceedings of the ISMIS’94, Charlotte, NC, USA, 16–19 October 1994. LNCS 867. [Google Scholar]
- Polkowski, L. Approximate Reasoning by Parts. An Introduction to Rough Mereology; Springer: Berlin, Germany, 2011. [Google Scholar]
- Polkowski, L. A model of granular computing with applications. In Proceedings of the 2006 IEEE International Conference on Granular Computing, Atlanta, GA, USA, 10–12 May 2006. [Google Scholar]
- Polkowski, L. A unified approach to granulation of knowledge and granular computing based on rough mereology: A survey. In Handbook of Granular Computing; John Wiley and Sons: New York, NY, USA, 2008; pp. 375–401. [Google Scholar]
- Polkowski, L. Formal granular calculi based on rough inclusions. In Proceedings of the IEEE 2005 Conference on Granular Computing GrC05, Beijing, China, 25–27 July 2005; IEEE Press: New York, NY, USA; pp. 57–62. [Google Scholar]
- Polkowski, L.; Artiemjew, P. Granular Computing in Decision Approximation—An Application of Rough Mereology. In Intelligent Systems Reference Library 77; Springer: Berlin/Heidelberg, Germany, 2015; pp. 1–422. ISBN 978-3-319-12879-5. [Google Scholar]
- University of California. Irvine Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 5 March 2020).
- Szypulski, J.; Artiemjew, P. The Rough Granular Approach to Classifier Synthesis by Means of SVM. In Proceedings of the International Joint Conference on Rough Sets, IJCRS’15, Tianjin, China, 20–23 November 2015; Lecture Notes in Computer Science (LNCS). Springer: Heidelberg, Germany, 2015; pp. 256–263. [Google Scholar]
d | |||||
---|---|---|---|---|---|
2 | 1 | 2 | 1 | 1 | |
3 | 2 | 3 | 3 | 1 | |
1 | 5 | 1 | 2 | 1 | |
6 | 2 | 3 | 8 | 2 | |
4 | 5 | 8 | 6 | 2 | |
5 | 1 | 8 | 1 | 2 |
1 | 0 | 0 | x | x | x | |
0 | 1 | 0 | x | x | x | |
0 | 0 | 1 | x | x | x | |
x | x | x | 1 | 0 | 0 | |
x | x | x | 0 | 1 | 1 | |
x | x | x | 0 | 1 | 1 |
2 | 1 | 2 | 1 | 1 | |
3 | 2 | 3 | 3 | 1 | |
1 | 5 | 1 | 2 | 1 | |
6 | 2 | 3 | 8 | 2 | |
5 | 5 | 8 | 6 | 2 |
15 | 690 | |
9 | 768 | |
14 | 270 |
no_of_gran_objects | percentage_of_objects | time_to_learn | accuracy | |
---|---|---|---|---|
Mean | Mean | Mean | Mean | |
gran_rad | ||||
0.0667 | 2.0 | 0.4149 | 0.3666 | 0.5646 |
0.1333 | 2.0 | 0.4149 | 0.3607 | 0.5337 |
0.2000 | 3.4 | 0.7054 | 0.3691 | 0.5423 |
0.2667 | 5.1 | 1.0581 | 0.3685 | 0.5154 |
0.3333 | 8.2 | 1.7012 | 0.3696 | 0.5192 |
0.4000 | 16.0 | 3.3195 | 0.3778 | 0.5577 |
0.4667 | 31.6 | 6.5560 | 0.3777 | 0.6236 |
0.5333 | 65.3 | 13.5477 | 0.3916 | 0.7764 |
0.6000 | 145.3 | 30.1452 | 0.4287 | 0.8125 |
0.6667 | 283.8 | 58.8797 | 0.7464 | 0.8399 |
0.7333 | 412.9 | 85.6639 | 0.8210 | 0.8534 |
0.8000 | 468.8 | 97.2614 | 0.8585 | 0.8587 |
0.8667 | 477.9 | 99.1494 | 0.8532 | 0.8553 |
0.9333 | 479.3 | 99.4398 | 0.8817 | 0.8553 |
1.0000 | 482.0 | 100.0000 | 0.8995 | 0.8562 |
no_of_gran_objects | percentage_of_objects | time_to_learn | accuracy | |
---|---|---|---|---|
Mean | Mean | Mean | Mean | |
gran_rad | ||||
0.0714 | 2.0 | 0.9434 | 0.3702 | 0.6801 |
0.1429 | 2.3 | 1.0849 | 0.3786 | 0.6505 |
0.2143 | 2.8 | 1.3208 | 0.3708 | 0.6231 |
0.2857 | 4.5 | 2.1226 | 0.3758 | 0.7132 |
0.3571 | 8.8 | 4.1509 | 0.3928 | 0.7110 |
0.4286 | 17.0 | 8.0189 | 0.4064 | 0.7187 |
0.5000 | 35.7 | 16.8396 | 0.4209 | 0.7198 |
0.5714 | 70.4 | 33.2075 | 0.4299 | 0.7560 |
0.6429 | 122.7 | 57.8774 | 0.4571 | 0.7813 |
0.7143 | 177.9 | 83.9151 | 0.4872 | 0.8231 |
0.7857 | 204.6 | 96.5094 | 0.4942 | 0.8297 |
0.8571 | 211.4 | 99.7170 | 0.4995 | 0.8220 |
0.9286 | 211.4 | 99.7170 | 0.5023 | 0.8209 |
1.0000 | 211.4 | 99.7170 | 0.5010 | 0.8253 |
no_of_gran_objects | percentage_of_objects | time_to_learn | accuracy | |
---|---|---|---|---|
Mean | Mean | Mean | Mean | |
gran_rad | ||||
0.1111 | 2.0 | 0.3724 | 0.3679 | 0.5392 |
0.2222 | 32.6 | 6.0708 | 0.4544 | 0.5584 |
0.3333 | 145.3 | 27.0577 | 0.4899 | 0.7009 |
0.4444 | 331.0 | 61.6387 | 0.7895 | 0.7563 |
0.5556 | 477.8 | 88.9758 | 0.8457 | 0.7723 |
0.6667 | 533.0 | 99.2551 | 0.8643 | 0.7714 |
0.7778 | 537.0 | 100.0000 | 0.8882 | 0.7684 |
0.8889 | 537.0 | 100.0000 | 0.9313 | 0.7671 |
1.0000 | 537.0 | 100.0000 | 0.9417 | 0.7680 |
no_of_gran_objects | percentage_of_objects | time_to_learn | accuracy | |
---|---|---|---|---|
Mean | Mean | Mean | Mean | |
gran_rad | ||||
0.025 | 2.0 | 0.4149 | 0.3598 | 0.5534 |
0.050 | 2.0 | 0.4149 | 0.4274 | 0.5647 |
0.075 | 2.0 | 0.4149 | 0.4314 | 0.4836 |
0.100 | 2.0 | 0.4149 | 0.4373 | 0.4744 |
0.125 | 2.0 | 0.4149 | 0.4391 | 0.5536 |
0.150 | 2.0 | 0.4149 | 0.4327 | 0.5841 |
0.175 | 2.0 | 0.4149 | 0.4305 | 0.5778 |
0.200 | 2.0 | 0.4149 | 0.4349 | 0.4928 |
0.225 | 2.0 | 0.4149 | 0.4404 | 0.4826 |
0.250 | 2.0 | 0.4149 | 0.4348 | 0.5048 |
0.275 | 2.0 | 0.4149 | 0.4342 | 0.5082 |
0.300 | 2.0 | 0.4149 | 0.4586 | 0.5261 |
0.325 | 2.0 | 0.4149 | 0.4494 | 0.5652 |
0.350 | 2.0 | 0.4149 | 0.4476 | 0.5130 |
0.375 | 2.0 | 0.4149 | 0.4369 | 0.4797 |
0.400 | 2.0 | 0.4149 | 0.4608 | 0.5329 |
0.425 | 2.0 | 0.4149 | 0.4444 | 0.5256 |
0.450 | 2.0 | 0.4149 | 0.4443 | 0.5179 |
0.475 | 2.0 | 0.4149 | 0.4516 | 0.5135 |
0.500 | 2.0 | 0.4149 | 0.4436 | 0.5855 |
0.525 | 2.0 | 0.4149 | 0.4417 | 0.5034 |
0.550 | 2.2 | 0.4564 | 0.4517 | 0.4638 |
0.575 | 2.9 | 0.6017 | 0.4431 | 0.5063 |
0.600 | 3.7 | 0.7676 | 0.4476 | 0.5546 |
0.625 | 5.2 | 1.0788 | 0.4571 | 0.4947 |
0.650 | 8.7 | 1.8050 | 0.4506 | 0.5594 |
0.675 | 12.1 | 2.5104 | 0.4754 | 0.5237 |
0.700 | 18.0 | 3.7344 | 0.4782 | 0.6222 |
0.725 | 29.4 | 6.0996 | 0.5029 | 0.6961 |
0.750 | 49.9 | 10.3527 | 0.5087 | 0.7140 |
0.775 | 80.4 | 16.6805 | 0.5228 | 0.7889 |
0.800 | 134.8 | 27.9668 | 0.7577 | 0.7903 |
0.825 | 214.3 | 44.4606 | 0.8252 | 0.8169 |
0.850 | 331.3 | 68.7344 | 0.8719 | 0.8319 |
0.875 | 427.3 | 88.6515 | 0.9257 | 0.8386 |
0.900 | 470.7 | 97.6556 | 0.9505 | 0.8319 |
0.925 | 478.2 | 99.2116 | 0.9665 | 0.8309 |
0.950 | 479.5 | 99.4813 | 0.9729 | 0.8377 |
0.975 | 482.0 | 100.0000 | 0.9688 | 0.8329 |
1.000 | 482.0 | 100.0000 | 0.9697 | 0.8386 |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ropiak, K.; Artiemjew, P. On a Hybridization of Deep Learning and Rough Set Based Granular Computing. Algorithms 2020, 13, 63. https://doi.org/10.3390/a13030063
Ropiak K, Artiemjew P. On a Hybridization of Deep Learning and Rough Set Based Granular Computing. Algorithms. 2020; 13(3):63. https://doi.org/10.3390/a13030063
Chicago/Turabian StyleRopiak, Krzysztof, and Piotr Artiemjew. 2020. "On a Hybridization of Deep Learning and Rough Set Based Granular Computing" Algorithms 13, no. 3: 63. https://doi.org/10.3390/a13030063
APA StyleRopiak, K., & Artiemjew, P. (2020). On a Hybridization of Deep Learning and Rough Set Based Granular Computing. Algorithms, 13(3), 63. https://doi.org/10.3390/a13030063