GraalNN: Context-Sensitive Static Profiling with Graph Neural Networks
Accurate static profile prediction is crucial for achieving optimal program performance in the absence of dynamic profiles. However, existing static profiling methods struggle to fully exploit the complex structure of the compiler’s intermediate representation (IR) and fail to effectively utilize the calling context information needed for accurate profile inference. To address these limitations, we introduce \textit{GraalNN}, a Graph Neural Network (GNN)-based static profiling framework that directly learns structural information from control-flow graphs (CFGs). This reduces the reliance on handcrafted features and minimizes the effort required for feature engineering while improving the model’s ability to predict profiles based on captured control-flow dependencies.
GraalNN follows a two-stage approach: it predicts context-insensitive profiles during the parsing phase and refines them with context-sensitive profiles during the inlining phase. This approach achieves a runtime speedup of 10.13% on a wide range of industry-standard benchmarks, surpassing existing static profiling techniques by over 2.5%. Additionally, our method limits the binary size increase to just 2.94%, maintaining compilation efficiency and outperforming all existing static profiling solutions by at least 0.9%.
Mon 3 MarDisplayed time zone: Pacific Time (US & Canada) change
14:00 - 15:20 | ML Tools & OptimizationMain Conference at Casuarina Ballroom (Level 2) Chair(s): Jeronimo Castrillon TU Dresden, Germany | ||
14:00 20mTalk | VEGA: Automatically Generating Compiler Backends Using a Pre-Trained Transformer Model Main Conference Ming Zhong SKLP, Institute of Computing Technology, CAS, Fang Lv Institute of Computing Technology, Chinese Academy of Sciences, Lulin Wang SKLP, ICT, CAS Beijing, China, Lei Qiu SKLP, Institute of Computing Technology, CAS; University of Chinese Academy of Sciences, Yingying Wang SKLP, ICT, CAS Beijing, China, Ying Liu Institute of Computing Technology, Chinese Academy of Sciences, Huimin Cui Institute of Computing Technology, Chinese Academy of Sciences, Xiaobing Feng ICT CAS, Jingling Xue UNSW Sydney | ||
14:20 20mTalk | IntelliGen: Instruction-Level Auto-Tuning for Tensor Program with Monotonic Memory Optimization Main Conference Zixuan Ma Tsinghua University, Haojie Wang Tsinghua University, Jingze Xing Tsinghua University, Shuhong Huang Tsinghua University, Liyan Zheng Tsinghua University, Chen Zhang Tsinghua University, Huanqi Cao Tsinghua University, Kezhao Huang Tsinghua University, Mingshu Zhai Tsinghua University, Shizhi Tang Tsinghua University, Penghan Wang Tsinghua University, Jidong Zhai Tsinghua University | ||
14:40 20mTalk | GraalNN: Context-Sensitive Static Profiling with Graph Neural Networks Main Conference | ||
15:00 20mTalk | LLM-Vectorizer: LLM-based Verified Loop Vectorizer Main Conference Jubi Taneja Microsoft Research, Avery Laird University of Toronto, Cong Yan Microsoft Research, Madan Musuvathi Microsoft Research, Shuvendu K. Lahiri Microsoft Research |