TY - GEN
T1 - LLMatic
T2 - 2024 Genetic and Evolutionary Computation Conference, GECCO 2024
AU - Nasir, Muhammad Umair
AU - Earle, Sam
AU - Togelius, Julian
AU - James, Steven
AU - Cleghorn, Christopher
N1 - Publisher Copyright:
© 2024 Copyright is held by the owner/author(s). Publication rights licensed to ACM.
PY - 2024/7/14
Y1 - 2024/7/14
N2 - Large language models (LLMs) have emerged as powerful tools capable of accomplishing a broad spectrum of tasks. Their abilities span numerous areas, and one area where they have made a significant impact is in the domain of code generation. Here, we propose using the coding abilities of LLMs to introduce meaningful variations to code defining neural networks. Meanwhile, Quality-Diversity (QD) algorithms are known to discover diverse and robust solutions. By merging the code-generating abilities of LLMs with the diversity and robustness of QD solutions, we introduce LLMatic, a Neural Architecture Search (NAS) algorithm. While LLMs struggle to conduct NAS directly through prompts, LLMatic uses a procedural approach, leveraging QD for prompts and network architecture to create diverse and high-performing networks. We test LLMatic on the CIFAR-10 and NAS-bench-201 benchmarks, demonstrating that it can produce competitive networks while evaluating just 2, 000 candidates, even without prior knowledge of the benchmark domain or exposure to any previous top-performing models for the benchmark. The open-sourced code is available at https://github.com/umair-nasir14/LLMatic.
AB - Large language models (LLMs) have emerged as powerful tools capable of accomplishing a broad spectrum of tasks. Their abilities span numerous areas, and one area where they have made a significant impact is in the domain of code generation. Here, we propose using the coding abilities of LLMs to introduce meaningful variations to code defining neural networks. Meanwhile, Quality-Diversity (QD) algorithms are known to discover diverse and robust solutions. By merging the code-generating abilities of LLMs with the diversity and robustness of QD solutions, we introduce LLMatic, a Neural Architecture Search (NAS) algorithm. While LLMs struggle to conduct NAS directly through prompts, LLMatic uses a procedural approach, leveraging QD for prompts and network architecture to create diverse and high-performing networks. We test LLMatic on the CIFAR-10 and NAS-bench-201 benchmarks, demonstrating that it can produce competitive networks while evaluating just 2, 000 candidates, even without prior knowledge of the benchmark domain or exposure to any previous top-performing models for the benchmark. The open-sourced code is available at https://github.com/umair-nasir14/LLMatic.
KW - large language models
KW - neural architecture search
KW - neural networks
KW - quality-diversity optimization
UR - http://www.scopus.com/inward/record.url?scp=85197054494&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85197054494&partnerID=8YFLogxK
U2 - 10.1145/3638529.3654017
DO - 10.1145/3638529.3654017
M3 - Conference contribution
AN - SCOPUS:85197054494
T3 - GECCO 2024 - Proceedings of the 2024 Genetic and Evolutionary Computation Conference
SP - 1110
EP - 1118
BT - GECCO 2024 - Proceedings of the 2024 Genetic and Evolutionary Computation Conference
PB - Association for Computing Machinery, Inc
Y2 - 14 July 2024 through 18 July 2024
ER -