[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3534678.3539126acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Open access

User Behavior Pre-training for Online Fraud Detection

Published: 14 August 2022 Publication History

Abstract

The outbreak of COVID-19 burgeons newborn services on online platforms and simultaneously buoys multifarious online fraud activities. Due to the rapid technological and commercial innovation that opens up an ever-expanding set of products, the insufficient labeling data renders existing supervised or semi-supervised fraud detection models ineffective in these emerging services. However, the ever accumulated user behavioral data on online platforms might be helpful in improving the performance of fraud detection on newborn services. To this end, in this paper, we propose to pre-train user behavior sequences, which consist of orderly arranged actions, from the large-scale unlabeled data sources for online fraud detection. Recent studies illustrate accurate extraction of user intentions~(formed by consecutive actions) in behavioral sequences can propel improvements in the performance of online fraud detection. By anatomizing the characteristic of online fraud activities, we devise a model named UB-PTM that learns knowledge of fraud activities by three agent tasks at different granularities, i.e., action, intention, and sequence levels, from large-scale unlabeled data. Extensive experiments on three downstream transaction and user-level online fraud detection tasks demonstrate that our UB-PTM is able to outperform the state-of-the-art designing for specific tasks.

Supplemental Material

MP4 File
Presentation Video

References

[1]
Alex Beutel, Wanhong Xu, Venkatesan Guruswami, Christopher Palow, and Christos Faloutsos. 2013. Copycatch: stopping group attacks by spotting lockstep behavior in social networks. In WWW. 119--130.
[2]
Richard J Bolton and David J Hand. 2002. Statistical fraud detection: A review. Statistical science, Vol. 17, 3 (2002), 235--255.
[3]
Bernardo Branco, Pedro Abreu, Ana Sofia Gomes, Mariana SC Almeida, Jo ao Tiago Ascens ao, and Pedro Bizarro. 2020. Interleaved sequence RNNs for fraud detection. In KDD. 3101--3109.
[4]
Shaosheng Cao, XinXing Yang, Cen Chen, Jun Zhou, Xiaolong Li, and Yuan Qi. 2019. Titant: Online real-time transaction fraud detection in ant financial. arXiv preprint arXiv:1906.07407 (2019).
[5]
Joy Iong-Zong Chen and Kong-Long Lai. 2021. Deep convolution neural network model for credit-card fraud detection and alert. Journal of Artificial Intelligence, Vol. 3, 02 (2021), 101--112.
[6]
Mark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, and Ilya Sutskever. 2020. Generative Pretraining From Pixels. ICML, Vol. 119 (2020), 1691--1703.
[7]
Dawei Cheng, Xiaoyang Wang, Ying Zhang, and Liqing Zhang. 2020. Graph neural network for fraud detection via spatial-temporal attention. IEEE Transactions on Knowledge and Data Engineering (2020).
[8]
Ronan Collobert and Jason Weston. 2008. A unified architecture for natural language processing: Deep neural networks with multitask learning. In ICML. 160--167.
[9]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT.
[10]
Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. Advances in NIPS, Vol. 32 (2019).
[11]
Jerome H Friedman. 2001. Greedy function approximation: a gradient boosting machine. Annals of statistics (2001), 1189--1232.
[12]
David J Hand. 2005. Good practice in retail credit scorecard assessment. Journal of the Operational Research Society, Vol. 56, 9 (2005), 1109--1117.
[13]
Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. 2021. Masked Autoencoders Are Scalable Vision Learners. arXiv e-prints (Nov. 2021), arXiv:2111.06377.
[14]
Binbin Hu, Zhiqiang Zhang, Chuan Shi, Jun Zhou, Xiaolong Li, and Yuan Qi. 2019. Cash-out user detection based on attributed heterogeneous information network with a hierarchical attention mechanism. In AAAI, Vol. 33. 946--953.
[15]
Xue Jiang, Zhuoran Zheng, Chen Lyu, Liang Li, and Lei Lyu. 2021. TreeBERT: A tree-based pre-trained model for programming language. In UAI. PMLR, 54--63.
[16]
Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In ICLR.
[17]
Jie Lei, Linjie Li, Luowei Zhou, Zhe Gan, Tamara L. Berg, Mohit Bansal, and Jingjing Liu. 2021. Less Is More: ClipBERT for Video-and-Language Learning via Sparse Sampling. In CVPR. 7331--7341.
[18]
Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019).
[19]
Yuan Li, Yiheng Sun, and Noshir Contractor. 2017. Graph mining assisted semi-supervised learning for fraudulent cash-out detection. In ASONAM. 546--553.
[20]
Wangli Lin, Li Sun, Qiwei Zhong, Can Liu, Jinghua Feng, Xiang Ao, and Hao Yang. 2021. Online Credit Payment Fraud Detection via Structure-Aware Hierarchical Recurrent Neural Network. In IJCAI.
[21]
Can Liu, Li Sun, Xiang Ao, Jinghua Feng, Qing He, and Hao Yang. 2021. Intention-aware heterogeneous graph attention networks for fraud transactions detection. In KDD. 3280--3288.
[22]
Can Liu, Qiwei Zhong, Xiang Ao, Li Sun, Wangli Lin, Jinghua Feng, Qing He, and Jiayu Tang. 2020. Fraud Transactions Detection via Behavior Tree with Local Intention Calibration. In KDD. 3035--3043.
[23]
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019).
[24]
Ziqi Liu, Chaochao Chen, Xinxing Yang, Jun Zhou, Xiaolong Li, and Le Song. 2018. Heterogeneous graph neural networks for malicious account detection. In CIKM. 2077--2085.
[25]
Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. In NIPS, Vol. 32.
[26]
Mohit Iyyer Matt Gardner Christopher Clark Kenton Lee Matthew E.Peters, Mark Neumann and Luke Zettlemoyer. 2018. Deep contextualized word representations. In NAACL-HLT.
[27]
Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In NIPS.
[28]
Feiyang Pan, Shuokai Li, Xiang Ao, Pingzhong Tang, and Qing He. 2019. Warm up cold-start advertisements: Improving ctr predictions via learning to learn id embeddings. In SIGIR. 695--704.
[29]
Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et almbox. 2019. PyTorch: An imperative style, high-performance deep learning library. In NIPS.
[30]
Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. (2018).
[31]
Mike Schuster and Kuldip K Paliwal. 1997. Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, Vol. 45, 11 (1997), 2673--2681.
[32]
KR Seeja and Masoumeh Zareapoor. 2014. Fraudminer: A novel credit card fraud detection model based on frequent itemset mining. The Scientific World Journal, Vol. 2014 (2014).
[33]
Chen Sun, Austin Myers, Carl Vondrick, Kevin Murphy, and Cordelia Schmid. 2019. VideoBERT: A Joint Model for Video and Language Representation Learning. In ICCV. 7463--7472.
[34]
Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Hao Tian, Hua Wu, and Haifeng Wang. 2020. Ernie 2.0: A continual pre-training framework for language understanding. In AAAI, Vol. 34. 8968--8975.
[35]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In NIPS.
[36]
Stijn Viaene, Richard A Derrig, and Guido Dedene. 2004. A case study of applying boosting Naive Bayes to claim fraud diagnosis. IEEE Trans Knowl Data Eng, Vol. 16, 5 (2004), 612--620.
[37]
Cheng Wang, Changqi Wang, Hangyu Zhu, and Jipeng Cui. 2020. LAW: learning automatic windows for online payment fraud detection. IEEE Transactions on Dependable and Secure Computing, Vol. 18, 5 (2020), 2122--2135.
[38]
Daixin Wang, Jianbin Lin, Peng Cui, Quanhui Jia, Zhen Wang, Yanming Fang, Quan Yu, Jun Zhou, Shuang Yang, and Yuan Qi. 2019. A semi-supervised graph attentive network for financial fraud detection. In ICDM. IEEE, 598--607.
[39]
Wenhan Xiong, Jingfei Du, William Yang Wang, and Veselin Stoyanov. 2020. Pretrained encyclopedia: Weakly supervised knowledge-pretrained language model. In ICLR.
[40]
Zhengyan Zhang, Xu Han, Zhiyuan Liu, Xin Jiang, Maosong Sun, and Qun Liu. 2019. ERNIE: Enhanced language representation with informative entities. In ACL.
[41]
Panpan Zheng, Shuhan Yuan, and Xintao Wu. 2019. Safe: A neural survival analysis model for fraud early detection. In AAAI, Vol. 33. 1278--1285.
[42]
Qiwei Zhong, Yang Liu, Xiang Ao, Binbin Hu, Jinghua Feng, Jiayu Tang, and Qing He. 2020. Financial Defaulter Detection on Online Credit Payment via Multi-view Attributed Heterogeneous Information Network. In WWW. 785--795.
[43]
Xiaoqian Zhu, Xiang Ao, Zidi Qin, Yanpeng Chang, Yang Liu, Qing He, and Jianping Li. 2021. Intelligent financial fraud detection practices in post-pandemic era. The Innovation, Vol. 2, 4 (2021), 100176.

Cited By

View all
  • (2024)Credit Card Fraud Detection via Intelligent Sampling and Self-supervised LearningACM Transactions on Intelligent Systems and Technology10.1145/364128315:2(1-29)Online publication date: 28-Mar-2024
  • (2024)Robust Sequence-Based Self-Supervised Representation Learning for Anti-Money LaunderingProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3680078(4571-4578)Online publication date: 21-Oct-2024
  • (2024)A Payment Transaction Pre-training Model for Fraud Transaction DetectionProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3679670(932-941)Online publication date: 21-Oct-2024
  • Show More Cited By

Index Terms

  1. User Behavior Pre-training for Online Fraud Detection

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    KDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
    August 2022
    5033 pages
    ISBN:9781450393850
    DOI:10.1145/3534678
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 14 August 2022

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. fraud detection
    2. pre-training model
    3. user behavior modeling

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    KDD '22
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)954
    • Downloads (Last 6 weeks)124
    Reflects downloads up to 13 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Credit Card Fraud Detection via Intelligent Sampling and Self-supervised LearningACM Transactions on Intelligent Systems and Technology10.1145/364128315:2(1-29)Online publication date: 28-Mar-2024
    • (2024)Robust Sequence-Based Self-Supervised Representation Learning for Anti-Money LaunderingProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3680078(4571-4578)Online publication date: 21-Oct-2024
    • (2024)A Payment Transaction Pre-training Model for Fraud Transaction DetectionProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3679670(932-941)Online publication date: 21-Oct-2024
    • (2024)Burstiness-aware Bipartite Graph Neural Networks for Fraudulent User Detection on Rating PlatformsCompanion Proceedings of the ACM Web Conference 202410.1145/3589335.3651475(834-837)Online publication date: 13-May-2024
    • (2024)Detecting Evolving Fraudulent Behavior in Online Payment Services: Open-Category and Concept-DriftIEEE Transactions on Services Computing10.1109/TSC.2024.342288017:5(2180-2193)Online publication date: Sep-2024
    • (2023)Knowledge-inspired Subdomain Adaptation for Cross-Domain Knowledge TransferProceedings of the 32nd ACM International Conference on Information and Knowledge Management10.1145/3583780.3614946(234-244)Online publication date: 21-Oct-2023
    • (2023)Sequence As Genes: An User Behavior Modeling Framework for Fraud Transaction Detection in E-commerceProceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3580305.3599905(5194-5203)Online publication date: 6-Aug-2023
    • (2023)When to Pre-Train Graph Neural Networks? From Data Generation Perspective!Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3580305.3599548(142-153)Online publication date: 6-Aug-2023

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media