[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3456172.3456193acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccdeConference Proceedingsconference-collections
research-article

Hybrid feature tweaking: Combining random forest similarity tweaking with CLPFD

Published: 06 August 2021 Publication History

Abstract

When using prediction models created from data, it is in certain cases not sufficient for the users to only get a prediction, sometimes accompanied with a probability of the predictive outcome. Instead, a more elaborate answer is required, like given the predictive outcome, how can this outcome be changed to a wished outcome, i.e., feature tweaking. In this paper we introduce a novel hybrid method for performing feature tweaking that builds upon Random Forest Similarity Tweaking and utilizing a Constraint Logic Programming solver for the Finite Domain (CLPFD). This hybrid method is compared to only using a CLPFD solver and to using a previously known feature tweaking algorithm, Actionable Feature Tweaking. The results show that the hybrid method provides a good balance between the distances, comparing the original example and the tweaked example, and completeness, the number of successfully tweaked examples, compared to the other methods. Another benefit with the novel method, is that the user can specify a prediction threshold for feature tweaking and adjust weights of features to mimic the real-world cost of changing feature values.

References

[1]
Jonas Biteus, and Tony Lindgren. 2017. Planning Flexible Maintenance for Heavy Trucks using Machine Learning Models, Constraint Programming, and Route Optimization. SAE International Journal of Materials and Manufacturing 10, 3, 306-315.
[2]
Vasiliki Kougia, John Pavlopoulos, and Ion Androutsopoulos. 2019. AUEB NLP Group at ImageCLEFmed Caption 2019. Working Notes of the Conference and Labs of the Evaluation Forum (CLEF), Lugano, Switzerland, 9–12.
[3]
Päivi Parviainen, Maarit Tihinen, Jukka Kääriäinen, and Susanna Teppola. 2017. Tackling the Digitalisation Challenge: How to Benefit from Digitalisation in Practice. International Journal of Information Systems and Project Management, 5, 1, 63-77.
[4]
Tony Lindgren. 2019. On Data Driven Organizations and the Necessity of Interpretable Models. In proceedings of the Second EAI International Conference (SGIoT), Niagara Falls, ON, Canada, 121-130.
[5]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. "Why Should I Trust You?": Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '16). Association for Computing Machinery, New York, NY, USA, 1135–1144.
[6]
Zachary C. Lipton. 2018. The mythos of model interpretability. Commun. ACM 61, 10 (October 2018), 36–43.
[7]
Christoph Molnar. 2019. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable. https://christophm.github.io/interpretable-ml-book/.
[8]
Gabriele Tolomei, Fabrizio Silvestri, Andrew Haines, and Mounia Lalmas. 2017. Interpretable Predictions of Tree-based Ensembles via Actionable Feature Tweaking. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '17). Association for Computing Machinery, New York, NY, USA, 465–474.
[9]
Tony Lindgren, Panagiotis Papapetrou, Isak Samsten and Lars Asker. 2019. Example-Based Feature Tweaking Using Random Forests. IEEE 20th International Conference on Information Reuse and Integration for Data Science (IRI), Los Angeles, CA, USA, 2019, pp. 53-60
[10]
Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel 2011. Scikit-learn: Machine learning in Python." the Journal of machine Learning research 12, 2825-2830.
[11]
Mats Carlsson. 2020. SICStus Prolog user's Manual.
[12]
MiniZinc Challenge 2019 The Challenge. https://www.minizinc.org/challenge2019/challenge.html
[13]
Blake, Catherine. 1998. UCI repository of machine learning databases. http://www.ics.uci.edu/∼mlearn/MLRepository.html

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICCDE '21: Proceedings of the 2021 7th International Conference on Computing and Data Engineering
January 2021
110 pages
ISBN:9781450388450
DOI:10.1145/3456172
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 06 August 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Feature tweaking
  2. Model interpretability
  3. Random forest

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICCDE 2021

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 23
    Total Downloads
  • Downloads (Last 12 months)5
  • Downloads (Last 6 weeks)0
Reflects downloads up to 17 Jan 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media