[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3534678.3542603acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
abstract

Accelerated GNN Training with DGL and RAPIDS cuGraph in a Fraud Detection Workflow

Published: 14 August 2022 Publication History

Abstract

Graph Neural Networks (GNNs) have gained the interest of industry with Relational Graph Convolutional Networks (R-GCNs) showing promise for fraud detection. Taking existing workflows that leverage graph features to train a gradient boosted decision tree (GBDT) and replacing the graph features with GNN produced embedding achieves an increase in accuracy. However, recent work has shown that the combination of graph attributes with GNN embeddings provides the biggest lift in accuracy.
Whether to use a GNN is half of the picture. Data loading, data cleaning and prep (ETL), and graph processing are critical first steps before graph features or GNN training can be performed. Moreover, the entire process is interactive, optimizing training and validation, for shorter model delivery cycles. Quicker model updates are the key to staying ahead of evolving fraud techniques. McDonald and Deotte [1] published a BLOG on the importance of being able to iterate quickly in finding a solution.
The RAPIDS [2] suite of open-source software libraries gives the data scientist the freedom to execute end-to-end analytics work-flows on GPUs. The ETL and data loading portion is handled by RAPIDS cuDF, which utilizes a familiar DataFrame API. The GBDT process is handled by RAPIDS cuML that has an implementation of XGBoost and RandomForest. The graph analytic portion is handled by RAPIDS cuGraph. Recently cuGraph announced integration into Deep Graph Library (DGL) [3]. For GNN training, graph sampling can consume up to 80% of the training time. RAPIDS cuGraph sampling algorithms execute 10x to 100x faster than similar CPU versions and scale to support massive size graphs. Join us as we dive into GNNs for fraud detection and as we demonstrate how RAPIDS + DGL drastically reduces training time. We will cover everything from accelerating data load and data prep to accelerated GNN training with cuGraph + DGL.

References

[1]
Carol McDonald, Chris Deotte, NVIDIA Developer Technical BLOG, Jan 26, 2021
[2]
RAPIDS, https://rapids.ai/
[3]
Deep Graph Library, https://www.dgl.ai/
[4]
https://developer.nvidia.com/blog/leveraging-machine-learning-to-detect-fraud-tips-to-developing-a-winning-kaggle-solution/

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
August 2022
5033 pages
ISBN:9781450393850
DOI:10.1145/3534678
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 August 2022

Check for updates

Author Tags

  1. etl
  2. gpus
  3. gradient boosted decision tree
  4. graph analytics
  5. graph neural networks

Qualifiers

  • Abstract

Conference

KDD '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 271
    Total Downloads
  • Downloads (Last 12 months)54
  • Downloads (Last 6 weeks)3
Reflects downloads up to 10 Dec 2024

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media