XLA: Accelerated Linear Algebra, Google Developers, 2024 (Google) - Provides an overview of XLA, the domain-specific compiler for linear algebra, detailing its optimization passes including algebraic simplifications for machine learning workloads.
TVM: An End-to-End Deep Learning Compiler Stack for CPUs, GPUs, and Specialized Accelerators, Tianqi Chen, Thierry Moreau, Ziheng Jiang, Lianmin Zheng, Eddie Yan, Haichen Shen, Meghan Cowan, Leyuan Wang, Yuwei Hu, Luis Ceze, Carlos Guestrin, and Arvind Krishnamurthy, 201813th USENIX Symposium on Operating Systems Design and Implementation (OSDI '18) (USENIX Association)DOI: 10.5555/3342356.3342465 - Describes TVM's graph-level optimization passes, including techniques for computation graph simplification and transformation, which underpins advanced algebraic optimizations in deep learning compilers.
Halide: A Language and Compiler for Optimizing Parallel Programs on GPUs and CPUs, Jonathan Ragan-Kelley, Connelly Barnes, Andrew Adams, Sylvain Paris, Frédo Durand, Saman Amarasinghe, 2013ACM SIGPLAN Notices, Vol. 48 (Association for Computing Machinery (ACM))DOI: 10.1145/2499370.2462176 - Presents Halide, a programming language and compiler that introduces a declarative approach to image processing pipelines, providing foundational concepts for graph IRs and optimization passes used in modern ML compilers.