Accurate static profile prediction is crucial for achieving optimal program performance in the absence of dynamic profiles. However, existing static profiling methods struggle to fully exploit the complex structure of the compiler’s intermediate representation (IR) and fail to effectively utilize the calling context information needed for accurate profile inference. To address these limitations, we introduce \textit{GraalNN}, a Graph Neural Network (GNN)-based static profiling framework that directly learns structural information from control-flow graphs (CFGs). This reduces the reliance on handcrafted features and minimizes the effort required for feature engineering while improving the model’s ability to predict profiles based on captured control-flow dependencies.

GraalNN follows a two-stage approach: it predicts context-insensitive profiles during the parsing phase and refines them with context-sensitive profiles during the inlining phase. This approach achieves a runtime speedup of 10.13% on a wide range of industry-standard benchmarks, surpassing existing static profiling techniques by over 2.5%. Additionally, our method limits the binary size increase to just 2.94%, maintaining compilation efficiency and outperforming all existing static profiling solutions by at least 0.9%.

Mon 3 Mar

Displayed time zone: Pacific Time (US & Canada) change

14:00 - 15:20
ML Tools & OptimizationMain Conference at Casuarina Ballroom (Level 2)
Chair(s): Jeronimo Castrillon TU Dresden, Germany
14:00
20m
Talk
VEGA: Automatically Generating Compiler Backends Using a Pre-Trained Transformer Model
Main Conference
Ming Zhong SKLP, Institute of Computing Technology, CAS, Fang Lv Institute of Computing Technology, Chinese Academy of Sciences, Lulin Wang SKLP, ICT, CAS Beijing, China, Lei Qiu SKLP, Institute of Computing Technology, CAS; University of Chinese Academy of Sciences, Yingying Wang SKLP, ICT, CAS Beijing, China, Ying Liu Institute of Computing Technology, Chinese Academy of Sciences, Huimin Cui Institute of Computing Technology, Chinese Academy of Sciences, Xiaobing Feng ICT CAS, Jingling Xue UNSW Sydney
14:20
20m
Talk
IntelliGen: Instruction-Level Auto-Tuning for Tensor Program with Monotonic Memory Optimization
Main Conference
Zixuan Ma Tsinghua University, Haojie Wang Tsinghua University, Jingze Xing Tsinghua University, Shuhong Huang Tsinghua University, Liyan Zheng Tsinghua University, Chen Zhang Tsinghua University, Huanqi Cao Tsinghua University, Kezhao Huang Tsinghua University, Mingshu Zhai Tsinghua University, Shizhi Tang Tsinghua University, Penghan Wang Tsinghua University, Jidong Zhai Tsinghua University
14:40
20m
Talk
GraalNN: Context-Sensitive Static Profiling with Graph Neural Networks
Main Conference
Lazar Milikic Oracle Labs, Milan Cugurovic Oracle Labs, Vojin Jovanovic Oracle Labs
15:00
20m
Talk
LLM-Vectorizer: LLM-based Verified Loop Vectorizer
Main Conference
Jubi Taneja Microsoft Research, Avery Laird University of Toronto, Cong Yan Microsoft Research, Madan Musuvathi Microsoft Research, Shuvendu K. Lahiri Microsoft Research