Zhou et al., 2022 - Google Patents
Model-architecture co-design for high performance temporal gnn inference on fpgaZhou et al., 2022
View PDF- Document ID
- 14259240996651151967
- Author
- Zhou H
- Zhang B
- Kannan R
- Prasanna V
- Busart C
- Publication year
- Publication venue
- 2022 IEEE International Parallel and Distributed Processing Symposium (IPDPS)
External Links
Snippet
Temporal Graph Neural Networks (TGNNs) are powerful models to capture temporal, structural, and contextual information on temporal graphs. The generated temporal node embeddings outperform other methods in many downstream tasks. Real-world applications …
- 230000002123 temporal effect 0 title abstract description 62
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/30286—Information retrieval; Database structures therefor; File system structures therefor in structured data stores
- G06F17/30386—Retrieval requests
- G06F17/30424—Query processing
- G06F17/30533—Other types of queries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/30286—Information retrieval; Database structures therefor; File system structures therefor in structured data stores
- G06F17/30386—Retrieval requests
- G06F17/30424—Query processing
- G06F17/30442—Query optimisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/30286—Information retrieval; Database structures therefor; File system structures therefor in structured data stores
- G06F17/30587—Details of specialised database models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
- G06F17/5009—Computer-aided design using simulation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/11—Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/76—Architectures of general purpose stored programme computers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/18—Digital computers in general; Data processing equipment in general in which a programme is changed according to experience gained by the computer itself during a complete run; Learning machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for programme control, e.g. control unit
- G06F9/06—Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F19/00—Digital computing or data processing equipment or methods, specially adapted for specific applications
- G06F19/10—Bioinformatics, i.e. methods or systems for genetic or protein-related data processing in computational molecular biology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F2217/00—Indexing scheme relating to computer aided design [CAD]
- G06F2217/78—Power analysis and optimization
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computer systems based on specific mathematical models
- G06N7/005—Probabilistic networks
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Zhou et al. | Model-architecture co-design for high performance temporal gnn inference on fpga | |
Geng et al. | O3BNN-R: An out-of-order architecture for high-performance and regularized BNN inference | |
Schorn et al. | Automated design of error-resilient and hardware-efficient deep neural networks | |
Zhou et al. | BlockGNN: Towards efficient GNN acceleration using block-circulant weight matrices | |
Yudanov et al. | GPU-based simulation of spiking neural networks with real-time performance & high accuracy | |
Mohaidat et al. | A survey on neural network hardware accelerators | |
Liu et al. | An unbiased mcmc fpga-based accelerator in the land of custom precision arithmetic | |
Yan et al. | FPGAN: an FPGA accelerator for graph attention networks with software and hardware co-optimization | |
Lee et al. | Neuroengine: a hardware-based event-driven simulation system for advanced brain-inspired computing | |
AU2020326407B2 (en) | Extending finite rank deep kernel learning to forecasting over long time horizons | |
Lopes et al. | A machine learning approach to accelerating DSE of reconfigurable accelerator systems | |
Chen et al. | Exploiting on-chip heterogeneity of versal architecture for GNN inference acceleration | |
Jamma et al. | Hardware accelerators for the K-nearest neighbor algorithm using high level synthesis | |
Sakellariou et al. | An FPGA accelerator for spiking neural network simulation and training | |
Zhao et al. | RF-RISA: A novel flexible random forest accelerator based on FPGA | |
Oluwasakin et al. | Minimization of high computational cost in data preprocessing and modeling using MPI4Py | |
Metz et al. | Towards neural hardware search: Power estimation of cnns for gpgpus with dynamic frequency scaling | |
Wang et al. | A parallel digital VLSI architecture for integrated support vector machine training and classification | |
De Maio et al. | A roadmap to post-moore era for distributed systems | |
Peroni et al. | ARGA: Approximate reuse for GPGPU acceleration | |
Han et al. | ADE-HGNN: Accelerating HGNNs through Attention Disparity Exploitation | |
Singha et al. | LEAPER: Fast and Accurate FPGA-based System Performance Prediction via Transfer Learning | |
Kotlar et al. | Energy efficient implementation of tensor operations using dataflow paradigm for machine learning | |
Guo et al. | Algorithms and architecture support of degree-based quantization for graph neural networks | |
Ramadurgam | Optimized Embedded Architectures and Techniques for Machine Learning Algorithms for On-Chip AI Acceleration |