OMEGA: A Low-Latency GNN Serving System for Large Graphs

Type: Preprint

Publication Date: 2025-01-14

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2501.08547

Abstract

Graph Neural Networks (GNNs) have been widely adopted for their ability to compute expressive node representations in graph datasets. However, serving GNNs on large graphs is challenging due to the high communication, computation, and memory overheads of constructing and executing computation graphs, which represent information flow across large neighborhoods. Existing approximation techniques in training can mitigate the overheads but, in serving, still lead to high latency and/or accuracy loss. To this end, we propose OMEGA, a system that enables low-latency GNN serving for large graphs with minimal accuracy loss through two key ideas. First, OMEGA employs selective recomputation of precomputed embeddings, which allows for reusing precomputed computation subgraphs while selectively recomputing a small fraction to minimize accuracy loss. Second, we develop computation graph parallelism, which reduces communication overhead by parallelizing the creation and execution of computation graphs across machines. Our evaluation with large graph datasets and GNN models shows that OMEGA significantly outperforms state-of-the-art techniques.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ PDF Chat Acceleration Algorithms in GNNs: A Survey 2024 Lu Ma
Zeang Sheng
Xunkai Li
Xin-Yi Gao
Zhezheng Hao
Ling Yang
Wentao Zhang
Bin Cui
+ Communication-Efficient Graph Neural Networks with Probabilistic Neighborhood Expansion Analysis and Caching 2023 Tim Kaler
Alexandros-Stavros Iliopoulos
Philip Murzynowski
Tao B. Schardl
Charles E. Leiserson
Jie Chen
+ Redundancy-Free Computation Graphs for Graph Neural Networks 2019 Zhihao Jia
Sina Lin
Rex Ying
Jiaxuan You
Jure Leskovec
Alex Aiken
+ A Survey on Graph Neural Network Acceleration: Algorithms, Systems, and Customized Hardware 2023 Shichang Zhang
Atefeh Sohrabizadeh
Cheng Wan
Zijie Huang
Ziniu Hu
Yewen Wang
Ying-yan
Lin
Jason Cong
Yizhou Sun
+ Dorylus: Affordable, Scalable, and Accurate GNN Training over Billion-Edge Graphs 2021 John Thorpe
Yifan Qiao
Jonathan Eyolfson
Shen Teng
Guanzhou Hu
Zhihao Jia
Jinliang Wei
Keval Vora
Ravi Netravali
Miryung Kim
+ PDF Chat FeatGraph: A Flexible and Efficient Backend for Graph Neural Network Systems 2020 Yuwei Hu
Zihao Ye
Minjie Wang
Jiali Yu
Da Zheng
Mu Li
Zheng Zhang
Zhiru Zhang
Yida Wang
+ FeatGraph: A Flexible and Efficient Backend for Graph Neural Network Systems 2020 Yuwei Hu
Zihao Ye
Minjie Wang
Jiali Yu
Da Zheng
Mu Li
Zheng Zhang
Zhiru Zhang
Yida Wang
+ Distributed Graph Neural Network Training with Periodic Stale Representation Synchronization 2022 Zheng Chai
Guangji Bai
Liang Zhao
Yue Cheng
+ PDF Chat Can Graph Reordering Speed Up Graph Neural Network Training? An Experimental Study 2024 Nikolai Merkel
Pierre Toussing
Ruben Mayer
Hans‐Arno Jacobsen
+ BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing 2021 Tianfeng Liu
Yangrui Chen
Dan Li
Chuan Wu
Yibo Zhu
Jun He
Yanghua Peng
Hongzheng Chen
Hongzhi Chen
Chuanxiong Guo
+ PDF Chat CATGNN: Cost-Efficient and Scalable Distributed Training for Graph Neural Networks 2024 Xin Huang
Weipeng Zhuo
Minh Vuong
Shiju Li
Jongryool Kim
Bradley Rees
Chul‐Ho Lee
+ SCARA: Scalable Graph Neural Networks with Feature-Oriented Optimization 2022 Ningyi Liao
Dingheng Mo
Siqiang Luo
Xiang Li
Pengcheng Yin
+ PDF Chat Exact Acceleration of Subgraph Graph Neural Networks by Eliminating Computation Redundancy 2024 Qian Tao
Xiyuan Wang
Muhan Zhang
Shu‐Xian Hu
Wenyuan Yu
Jingren Zhou
+ PDF Chat DistDGL: Distributed Graph Neural Network Training for Billion-Scale Graphs 2020 Da Zheng
Chao Ma
Minjie Wang
Jinjing Zhou
Qidong Su
Xiang Song
Quan Gan
Zheng Zhang
George Karypis
+ Dorylus: Affordable, Scalable, and Accurate GNN Training with Distributed CPU Servers and Serverless Threads 2021 John Thorpe
Yifan Qiao
Jonathan Eyolfson
Shen Teng
Guanzhou Hu
Zhihao Jia
Jinliang Wei
Keval Vora
Ravi Netravali
Miryung Kim
+ GNNear: Accelerating Full-Batch Training of Graph Neural Networks with Near-Memory Processing 2021 Zhe Zhou
Cong Li
Xuechao Wei
Xiaoyang Wang
Guangyu Sun
+ MariusGNN: Resource-Efficient Out-of-Core Training of Graph Neural Networks 2022 Roger Waleffe
Jason Mohoney
Theodoros Rekatsinas
Shivaram Venkataraman
+ PDF Chat Heta: Distributed Training of Heterogeneous Graph Neural Networks 2024 Yuchen Zhong
Junwei Su
Chuan Wu
Minjie Wang
+ PDF Chat Scalable Graph Neural Network Training 2021 Marco Serafini
Hui Guan
+ PDF Chat Scalable Graph Neural Network Training 2021 Marco Serafini
Hui Guan

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors