Differential Replication for Credit Scoring in Regulated Environments
<p>(<b>a</b>) Binary classification dataset, (<b>b</b>) decision function learned by an artificial neural network, (<b>c</b>) synthetic data points sampled uniformly at random from the original feature space and labeled according to the predictions of the neural net, and (<b>d</b>) decision function of a decision tree-based copy.</p> "> Figure 2
<p>Diagram for Scenario 1, where we copy the whole predictive system, composed by both the preprocessing step and the logistic regression classifier, and obtain a new classifier applied directly on the decomposable raw data attributes.</p> "> Figure 3
<p>Diagram for Scenario 2, where we remove the preprocessing step and instead train a higher performance tree model on the original data attributes. We then copy this classifier by means of a self-explanatory model to extract global explanations.</p> "> Figure 4
<p>Distribution of copy accuracies for scenarios (<b>a</b>) 1 and (<b>b</b>) 2.</p> "> Figure 5
<p>Feature importances for the original gradient-boosted tree (blue) and the copy decision tree (red).</p> "> Figure 6
<p>Comparison of normalized first order coefficients for original (blue) and copy models (red).</p> "> Figure 7
<p>Accuracy for different copy depths.</p> ">
Abstract
:1. Introduction
2. Related Work
3. Adapting Models to the Demands of Their Environment
3.1. Environmental Adaptation and Differential Replication
3.2. Methodologies for Differential Replication
3.3. Background on the Copying Process
3.4. Approximating the Copying Process with the Simplified Single-Pass Copy
- Step 1: Synthetic sample generation. This step of the process accounts for finding the optimal set of data points for the copy. Because we approximate this step with a single iteration of an alternating scheme, it suffices to draw samples from a distribution that represents the coverage properties of the space where we want to build the copy with confidence (For high-dimensional synthetic samples, we can simply use ).
- Step 2: Label the synthetic set of samples. The optimal set is labeled according to the class prediction outputs of . This defines the synthetic dataset,
- Step 3: Finding the optimized model. Given a synthetic set , this step finds the value of the parameters of the copy model using regularized empirical risk keeping constant, as follows:
Algorithm 1: Simplified single-pass copy. |
|
4. Use Case
4.1. Scenario 1: De-Obfuscation of the Attribute Preprocessing
4.2. Scenario 2: Regulatory Compliant High Performance Copies
5. Experiments
5.1. Dataset
5.2. Experimental Settings
5.3. Validation and Discussion of Results
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Smith, J. Machine Learning Systems: Designs that Scale; Manning: New York, NY, USA, 2018. [Google Scholar]
- Polyzotis, N.; Roy, S.; Whang, S.E.; Zinkevich, M. Data Management Challenges in Production Machine Learning. In Proceedings of the 2017 ACM International Conference on Management of Data, Chicago, IL, USA, 14–19 May 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 1723–1726. [Google Scholar]
- Sculley, D.; Holt, G.; Golovin, D.; Davydov, E.; Phillips, T.; Ebner, D.; Chaudhary, V.; Young, M.; Crespo, J.F.; Dennison, D. Hidden Technical Debt in Machine Systems. Adv. Neural Inf. Process. Syst. 2015, 28, 2503–2511. [Google Scholar]
- Unceta, I.; Nin, J.; Pujol, O. Risk Mitigation in Algorithmic Accountability: The Role of Machine Learning Copies. PLoS ONE 2020, 15, e0241286. [Google Scholar] [CrossRef] [PubMed]
- Fredrikson, M.; Jha, S.; Ristenpart, T. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, 12–16 October 2015. [Google Scholar]
- Shokri, R.; Tech, C.; Stronati, M.; Shmatikov, V. Membership Inference Attacks Against Machine Learning Models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 3–18. [Google Scholar]
- Song, C.; Ristenpart, T.; Shmatikov, V. Machine Learning Models that Remember Too Much. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 587–601. [Google Scholar]
- European Union Commission. Legislation; European Union Commission: Brussels, Belgium, 2016. [Google Scholar]
- Goodman, B.; Flaxman, S. European Union Regulations on Algorithmic Decision-Making and a Right to Explanation. AI Mag. 2017, 38, 50–57. [Google Scholar] [CrossRef] [Green Version]
- Selbst, A.D.; Powles, J. Meaningful Information and the Right to Explanation. Int. Data Priv. Law 2017, 7, 233–242. [Google Scholar] [CrossRef]
- Barocas, S.; Selbst, A.D. Big Data’s Disparate Impact. Calif. Law Rev. 2016, 104, 671–732. [Google Scholar] [CrossRef]
- Friedman, B.; Nissenbaum, H. Bias in Computer Systems. ACM Trans. Inf. Sys. 1996, 14, 330–347. [Google Scholar] [CrossRef]
- Hardt, M. How Big Data Is Unfair. Available online: https://medium.com/@mrtz/how-big-data-is-unfair-9aa544d739de (accessed on 26 March 2021).
- Unceta, I.; Nin, J.; Pujol, O. Environmental Adaptation and Differential Replication in Machine Learning. Entropy 2020, 22, 1122. [Google Scholar] [CrossRef] [PubMed]
- Unceta, I.; Nin, J.; Pujol, O. Copying Machine Learning Classifiers. IEEE Access 2020, 8, 160268–160284. [Google Scholar] [CrossRef]
- Executive Office of the President. The National Artificial Intelligence Research and Development Strategic Plan; Technical Report; National Science and Technology Council: Washington, DC, USA, 2016. [Google Scholar]
- Executive Office of the President. Preparing for the Future of Artificial Intelligence; Technical Report; National Science and Technology Council: Washington, DC, USA, 2016. [Google Scholar]
- European Parliament. Civil Law Rules on Robotics-European Parliament Resolution of 16 February 2017 with Recommendations to the Commission on Civil Law Rules on Robotics (2015/2103(INL)); Technical Report P8TA-PROV(2017)00 51; European Parliament: Brussels, Belgium, 2017. [Google Scholar]
- News, X. China Rolls Out Three-Year Program for AI Growth. Available online: http://english.www.gov.cn/state_council/ministries/2016/05/23/content_281475355720632.htm (accessed on 26 March 2021).
- Financial Stability Board. Financial Stability Implications from FinTech Supervisory and Regulatory Issues that Merit Authorities’ Attention; Technical Report; Financial Stability Board: Basel, Switzerland, 2017. [Google Scholar]
- Alonso, A.; Carbó, J. Machine Learning in Credit Risk: Measuring the Dilemma between Prediction and Supervisory Cost. Banco Esp. Work. Pap. 2020. [Google Scholar] [CrossRef]
- Bussmann, N.; Giudici, P.; Marinelli, D.; Papenbrock, J. Explainable Machine Learning in Credit Risk Management. Comput. Econ. 2020, 57, 203–216. [Google Scholar] [CrossRef]
- Rudin, C. Please Stop Explaining Black Box Models for High-Stakes Decisions. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NIPS 2018), Workshop on Critiquing and Correcting Trends in Machine Learning, Montreal, QC, Canada, 8 December 2018. [Google Scholar]
- Izonin, I.; Tkachenko, R.; Kryvinska, N.; Tkachenko, P.; Greguš, M. Multiple Linear Regression Based on Coefficients Identification Using Non-iterative SGTM Neural-like Structure. In Advances in Computational Intelligence; IWANN 2019. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2019; Volume 11506, pp. 467–479. [Google Scholar] [CrossRef]
- Bücker, M.; Szepannek, G.; Gosiewska, A.; Biecek, P. Transparency, Auditability and eXplainability of Machine Learning Models in Credit Scoring. arXiv 2020, arXiv:2009.13384. [Google Scholar]
- Lessmann, S.; Baesens, B.; Seow, H.V.; Thomas, L.C. Benchmarking state-of-the-art classification algorithms for credit scoring: An update of research. Eur. J. Oper. Res. 2015, 247, 124–136. [Google Scholar] [CrossRef] [Green Version]
- Louzada, F.; Ara, A.; Fernandes, G.B. Classification methods applied to credit scoring: Systematic review and overall comparison. Surv. Oper. Res. Manag. Sci. 2016, 21, 117–134. [Google Scholar] [CrossRef] [Green Version]
- Dastile, X.; Celik, T.; Potsane, M. Statistical and machine learning models in credit scoring: A systematic literature survey. Appl. Soft Comput. 2020, 91, 106263. [Google Scholar] [CrossRef]
- Fair Isaac Corporation (FICO). Introduction to Scorecard for FICO Model Builder; Technical Report; FICO: San Jose, CA, USA, 2011. [Google Scholar]
- Lou, Y.; Caruana, R.; Gehrke, J. Intelligible Models for Classification and Regression. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Beijing, China, 12–16 August 2012. [Google Scholar]
- Lipton, Z.C. The Mythos of Model Interpretability. Workshop Human Interpret in Mach Learn. ACM QEUE 2016. pp. 96–100. Available online: https://queue.acm.org/detail.cfm?id=3241340 (accessed on 26 March 2021).
- Ribeiro, M.T.; Singh, S.; Guestrin, C. Why Should I Trust You? Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. Anchors: High-Precision Model-Agnostic Explanations. 2018. Available online: https://dblp.org/rec/conf/aaai/Ribeiro0G18.html (accessed on 26 March 2021).
- Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. In Proceedings of the Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4765–4774. [Google Scholar]
- Shapley, L.S. A value for N-person games. Contrib. Theory Games 1953, 2, 307–317. [Google Scholar]
- Fisher, A.; Rudin, C.; Dominici, F. All Models are Wrong, but Many are Useful: Learning a Variable’s Importance by Studying an Entire Class of Prediction Models Simultaneously. J. Mach. Learn. Res. 2019, 20, 1–81. [Google Scholar]
- Mancisidor, R.A.; Kampffmeyer, M.; Aas, K.; Jenssen, R. Deep generative models for reject inference in credit scoring. Knowl. Based Syst. 2020, 196, 105758. [Google Scholar] [CrossRef] [Green Version]
- Ariza-Garzón, M.J.; Arroyo, J.; Caparrini, A.; Segovia-Vargas, M. Explainability of a Machine Learning Granting Scoring Model in Peer-to-Peer Lending. IEEE Access 2020, 8, 64873–64890. [Google Scholar] [CrossRef]
- Brown, I.; Mues, C. An experimental comparison of classification algorithms for imbalanced credit scoring data sets. Expert Syst. Appl. 2012, 39, 3446–3453. [Google Scholar] [CrossRef] [Green Version]
- Tkachenko, R.; Doroshenko, A.; Izonin, I.; Tsymbal, Y.; Havrysh, B. Imbalance Data Classification via Neural-Like Structures of Geometric Transformations Model: Local and Global Approaches. In Advances in Computer Science for Engineering and Education; Springer: Berlin/Heidelberg, Germany, 2019; pp. 112–122. [Google Scholar] [CrossRef]
- Zhang, L.; Ray, H.; Priestley, J.; Tan, S. A descriptive study of variable discretization and cost-sensitive logistic regression on imbalanced credit data. J. Appl. Stat. 2020, 47, 568–581. [Google Scholar] [CrossRef] [Green Version]
- Menagie, M. A Comparison of Machine Learning Algorithms Using an Insufficient Number of Labeled Observations. Ph.D. Thesis, Vrije Universiteit, Amsterdam, The Netherlands, 2018. [Google Scholar]
- Barocas, S.; Boyd, D. Engaging the Ethics of Data Science in Practice. Commun. ACM 2017, 60, 23–25. [Google Scholar] [CrossRef]
- Veale, M.; Binns, R. Fairer Machine Learning in the Real World: Mitigating Discrimination Without Collecting Sensitive Data. Bid Data Soc. 2017, 4, 1–17. [Google Scholar] [CrossRef]
- Kroll, J. The Fallacy of Inscrutability. Philos. Trans. Roya Soc. 2018, 376. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Lu, J.; Liu, A.; Dong, F.; Gu, F.; Gama, J.; Zhang, G. Learning under Concept Drift: A Review. IEEE Trans. Knowl. Data Eng. 2019, 31, 2346–2363. [Google Scholar] [CrossRef] [Green Version]
- Loeffel, P. Adaptive machine learning algorithms for data streams subject to concept drifts. In Machine Learning [cs.LG]; Université Pierre et Marie Curie: Paris, France, 2017. [Google Scholar]
- Bottou, L.; Le Cun, Y. Large Scale Online Learning. Adv. Neural Inf. Process. Syst. 2004, 16, 217–234. [Google Scholar]
- Pan, S.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
- Torrey, L.; Shavlik, J. Transfer Learning. In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques; Soria Olivas, E., Martín Guerrero, J., Martinez-Sober, M., Magdalena-Benedito, J., Serrano López, A., Eds.; IGI Global: Pennsylvania, PA, USA, 2010; pp. 242–264. [Google Scholar] [CrossRef]
- Weiss, K.; Khoshgoftaar, T.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3. [Google Scholar] [CrossRef] [Green Version]
- Csurka, G. A Comprehensive Survey on Domain Adaptation for Visual Applications. In Domain Adaptation in Computer Vision Applications; Springer International Publishing: Cham, Switzerland, 2017; pp. 1–35. [Google Scholar] [CrossRef]
- Li, D.; Yang, Y.; Song, Y.; Hospedales, T. Deeper, Broader and Artier Domain Generalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5542–5550. [Google Scholar]
- Barque, M.; Martin, S.; Vianin, J.; Genoud, D.; Wannier, D. Improving wind power prediction with retraining machine learning algorithms. In International Workshop on Big Data and Information Security; IEEE Computer Society: Washington, DC, USA, 2018; pp. 43–48. [Google Scholar] [CrossRef]
- Mena, J.; Brando, A.; Pujol, O.; Vitrià, J. Uncertainty estimation for black-box classification models: A use case for sentiment analysis. In Pattern Recognition and Image Analysis; IbPRIA 2019. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2019; Volume 11867, pp. 29–40. [Google Scholar] [CrossRef]
- Mena, J.; Pujol, O.; Vitrià, J. Uncertainty-Based Rejection Wrappers for Black-Box Classifiers. IEEE Access 2020, 8, 101721–101746. [Google Scholar] [CrossRef]
- Bhattacharya, B.; Poulsen, R.; Toussaint, G. Application of Proximity Graphs to Editing Nearest Neighbor Decision Rule. In International Symposium on Information Theory; IEEE Computer Society: Washington, DC, USA; Santa Monica, CA, USA, 1981. [Google Scholar]
- Bhattacharya, B.; Mukherjee, K.; Toussaint, G.T. Geometric decision rules for instance-based learning algorithms. In Proceedings of the First International Conference on Pattern Recognition and Machine Intelligence, Kolkata, India, 20–22 December 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 60–69. [Google Scholar] [CrossRef] [Green Version]
- Mukherjee, K. Application of the Gabriel Graph to Instance Based Learning. Master’s Thesis, Simon Fraser University, Burnaby, BC, Canada, 2004. [Google Scholar]
- Buciluă, C.; Caruana, R.; Niculescu-Mizil, A. Model Compression. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Association for Computing Machinery, Philadelphia, PA, USA, 20–23 August 2006; pp. 535–541. [Google Scholar] [CrossRef]
- Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. NIPS Deep Learning and Representation Learning Workshop. 2015. Available online: https://www.cs.toronto.edu/~hinton/absps/distillation.pdf (accessed on 26 March 2021).
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: Washington, DC, USA, 2016; pp. 2818–2826. [Google Scholar] [CrossRef] [Green Version]
- Yang, C.; Xie, L.; Qiao, S.; Yuille, A.L. Training Deep Neural Networks in Generations: A More Tolerant Teacher Educates Better Students. Proc. AAAI Conf. Artif. Intell. 2019, 33, 5628–5635. [Google Scholar] [CrossRef] [Green Version]
- Müller, R.; Kornblith, S.; Hinton, G. When Does Label Smoothing Help? In Proceedings of the 33rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; MIT Press: Cambridge, MA, USA, 2019. [Google Scholar]
- Yuan, L.; Tay, F.E.; Li, G.; Wang, T.; Feng, J. Revisiting Knowledge Distillation via Label Smoothing Regularization. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; IEEE Computer Society: Washington, DC, USA, 2020. [Google Scholar]
- Bagherinezhad, H.; Horton, M.; Rastegari, M.; Farhadi, A. Label Refinery: Improving ImageNet Classification through Label Progression. arXiv 2018, arXiv:1805.02641. [Google Scholar]
- Vapnik, V.N. The Nature of Statistical Learning Theory; Springer: Berlin/Heidelberg, Germany, 2000; p. 226. [Google Scholar]
- Unceta, I.; Nin, J.; Pujol, O. From Batch to Online Learning Using Copies. Front. Artif. Intell. Appl. 2019, 319, 125–134. [Google Scholar]
- Escalera, S.; Masip, D.; Puertas, E.; Radeva, P.; Pujol, O. Online Error-Correcting Output Codes. Pattern Recognit. Lett. 2009, 32, 458–467. [Google Scholar] [CrossRef]
- Unceta, I.; Nin, J.; Pujol, O. Using Copies to Remove Sensitive Data: A Case Study on Fair Superhero Alignment Prediction. In Pattern Recognition and Image Analysis; IbPRIA 2019. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany; Madrid, Spain, 2019; Volume 11867, pp. 182–193. [Google Scholar] [CrossRef]
- Docherty, A.; Viort, F. Better Banking: Understanding and Addressing the Failures in Risk Management, Governance and Regulation; Wiley: Hoboken, NJ, USA, 2013; pp. 1–363. [Google Scholar]
- BBVA Research. Situación Inmobiliaria México. Primer Semestre 2018; BBVA Research: Madrid, Spain, 2018. (In Spanish) [Google Scholar]
- Unceta, I.; Palacios, D.; Nin, J.; Pujol, O. Sampling Unknown Decision Functions to Build Classifier Copies. In Proceedings of the 17th International Conference on Modeling Decisions for Artificial Intelligence, Sant Cugat, Spain, 2–4 September 2020; pp. 192–204. [Google Scholar] [CrossRef]
Attribute | Description |
---|---|
indebtedness | Level of indebtedness |
credit_amount | Amount of credit |
property_value | Property value |
loan_to_value | Loan to value |
duration | Duration of the loan |
studies | Level of studies |
poverty_index | Marginalization/poverty index |
age | Age |
est_soc_income | Estimated socio-demographic income |
value_m2 | Value per square meter |
est_income | Estimated income |
installment | Monthly installment |
n_family_unit | Members of the family unit |
est_mila_income | Estimated income based on MILA model |
p_default | Percentage of defaulted contracts in the last 4 months from those signed during the previous 12 to 24 months |
zip_code | ZIP code |
municipality | Municipality |
economy_level | Level of economy |
Attribute | Description |
---|---|
zip_code_municipality | Bivariate attribute resulting from the concatenation of features zip_code and municipality |
est_soc_income/est_mila_income | Univariate attribute resulting from the ratio between features est_soc_income and est_mila_income |
property_value/installment | Univariate attribute resulting from the ratio between features property_value and installment |
indebtedness/loan_to_value | Univariate attribute resulting from the ratio between features indebtedness and loan_to_value |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Unceta, I.; Nin, J.; Pujol, O. Differential Replication for Credit Scoring in Regulated Environments. Entropy 2021, 23, 407. https://doi.org/10.3390/e23040407
Unceta I, Nin J, Pujol O. Differential Replication for Credit Scoring in Regulated Environments. Entropy. 2021; 23(4):407. https://doi.org/10.3390/e23040407
Chicago/Turabian StyleUnceta, Irene, Jordi Nin, and Oriol Pujol. 2021. "Differential Replication for Credit Scoring in Regulated Environments" Entropy 23, no. 4: 407. https://doi.org/10.3390/e23040407
APA StyleUnceta, I., Nin, J., & Pujol, O. (2021). Differential Replication for Credit Scoring in Regulated Environments. Entropy, 23(4), 407. https://doi.org/10.3390/e23040407