The LLM Compiler is Meta's most recent hit
Meta recently unveiled the Meta Large Language Model (LLM) Compiler, a suite of open-source models building on Code Llama's achievements by delivering improved code optimization and compiler reasoning performance. The LLM Compiler is a groundbreaking technology that addresses a significant gap in applying large language models to compiler optimization.
The LLM Compiler is trained on a massive corpus of 546 billion tokens of LLVM-IR and is available in 7 billion and 13 billion parameter sizes. In the research paper, Meta also reports on fine-tuned versions of the LLM Compiler which demonstrate impressive capabilities in code size optimization and disassembly from x86_64 and ARM back into LLVM-IR. In tests, it achieved 77% of the optimizing potential of an autotuning search and showed a 45% success rate in round-trip disassembly.
The implications of the LLM Compiler are far-reaching, promising faster compile times, more efficient code, and new tools for understanding complex systems. Meta has released the LLM Compiler under a permissive commercial license, opening the doors for academic researchers and industry practitioners to leverage the technology based on their individual needs, while advancing the research further.