Close

Presentation

Accelerating Scientific Workflows with LLM-Driven Compiler Optimizations for Generated High-Performance Hardware
DescriptionThe optimization of computation kernels is central to high performance computing, directly impacting applications from scientific computing to artificial intelligence (AI). In experimental workflows with high-throughput or streaming data, software-only execution often becomes a bottleneck, motivating custom hardware accelerators. Field-programmable gate arrays and application-specific integrated circuits excel at these workloads by exploiting parallelism, pipelining, and low latency. Yet, mapping optimized kernels to hardware with high-level synthesis (HLS) requires significant manual effort. To address this, we propose a large language model (LLM)-driven optimization approach. Our method leverages the MLIR compiler infrastructure and modern LLMs’ capability to synthesize code to create tailored optimization strategies for hardware targets through HLS. This approach achieved 2.7x speedup for an electron energy loss spectroscopy autoencoder model targeting the Virtex UltraScale+. These results show that LLM-driven optimization offers a low-effort, high-performance alternative to manual workflows, paving the way for agentic AI in compilers and high performance computing.