Loading...
 

Large Language Models for and with Evolutionary Computation Workshop

Webpage: TBA (in case of acceptance)

Description

Large language models (LLMs), along with other foundational models in generative AI, have significantly changed the traditional expectations of artificial intelligence and machine learning systems. An LLM takes natural language text prompts as input and generates responses by matching patterns and completing sequences, providing output in natural language. In contrast, evolutionary computation (EC) is inspired by Neo-Darwinian evolution and focuses on black-box search and optimization. But what connects these two approaches?

One answer is evolutionary search heuristics (LLM with EC), with operators that use LLMs to fulfill their function. This hybridization turns the conventional paradigm that ECs use on its head, and in turn, sometimes yields high-performing and novel EC systems.

Another answer is using LLM for EC. LLMs may help researchers select feasible candidates from the pool of algorithms based on user-specified goals and provide a basic description of the methods or propose novel hybrid methods. Further, the models can help identify and describe distinct components suitable for adaptive enhancement or hybridization and provide a pseudo-code, implementation, and reasoning for the proposed methodology. Finally, LLMs have the potential to transform automated metaheuristic design and configuration by generating codes, iteratively improving the initially designed solutions or algorithm templates (with or without performance or other data-driven feedback), and even guiding implementations.

This workshop aims to encourage innovative approaches that leverage the strengths of LLMs and EC techniques, thus enabling the creation of more adaptive, efficient, and scalable algorithms by integrating evolutionary mechanisms with advanced LLM capabilities. Thanks to the collaborative platform for researchers and practitioners, the workshop may Inspire novel research directions that could reshape AI, specifically LLMs, and optimization fields through this hybridization and achieve a better understanding and explanation of how these two seemingly disparate fields are related and how knowledge of their functions and operations can be leveraged.

It includes (but is not restricted to the following topics):
- Evolutionary Prompt Engineering
- Optimisation of LLM Architectures
- LLM-Guided Evolutionary Algorithms
- How can an EA using an LLM evolve different of units of evolution, e.g. code, strings, images, multi-modal candidates?
- How can an EA using an LLM solve prompt composition or other LLM development and use challenges?
- How can an EA using an LLM integrate design explorations related to cooperation, modularity, reuse, or competition?
- How can an EA using an LLM model biology?
- How can an EA using an LLM intrinsically, or with guidance, support open-ended evolution?
- What new variants hybridizing EC and/or another search heuristic are possible and in what respects are they advantageous?
- What are new ways of using LLMs for evolutionary operators, e.g. new ways of generating variation through LLMs, as with LMX or ELM, or new ways of using LLMs for selection, as with e.g. Quality-Diversity through AI Feedback)
- How well does an EA using an LLM scale with population size and problem complexity?
- What is the most accurate computational complexity of an EA using an LLM?
- What makes a good EA plus LLM benchmark?
- LLMs for (automated) generation of EC.
- Understanding, fine-tuning, and adaptation of Large Language Models for EC. How large do LLMs need to be? Are there benefits for using larger/smaller ones? Ones trained on different datasets or in different ways?
- Implementing/generating methodology for population dynamics analysis, population diversity measures, control, and analysis and visualization.
- Generating rules for EC (boundary and constraints handling strategies).
- The performance improvement, testing, and efficiency of the improved algorithms.
- Reasoning for component-wise analysis of algorithms.
- Connection of LLM and other ML techniques for EC (Reinforcement learning, AutoML)
- Generation and reasoning for parallel approaches for EC algorithms.
- Benchmarking and Comparative Studies of LLM-generated algorithms.
- Applications of LLM and EC (not limited to):

+ constrained optimization + multi-objective optimization + expensive and surrogate assisted optimization + dynamic and uncertain optimization + large-scale optimization + combinatorial/discrete optimization

Organizers

Erik Hemberg
Erik Hemberg is a Research Scientist in the AnyScale Learning For All (ALFA) group at Massachusetts Institute of Technology Computer Science and Artificial Intelligence Lab, USA. He has a PhD in Computer Science from University College Dublin, Ireland and an MSc in Industrial Engineering and Applied Mathematics from Chalmers University of Technology, Sweden. His research interests involves coevolutionary algorithms, grammatical representations and generative models. His work focuses on developing autonomous, pro-active cyber defenses that are anticipatory and adapt to counter attacks. He is also interested in automated semantic parsing of law, and data science for education and healthcare.


 
Roman Senkerik

Roman Senkerik was born in Zlin, the Czech Republic, in 1981. He received an MSc degree in technical cybernetics from the Tomas Bata University in Zlin, Faculty of applied informatics in 2004, the Ph.D. degree also in technical Cybernetics, in 2008, from the same university, and Assoc. prof. Degree in Informatics from VSB – Technical University of Ostrava, in 2013.

From 2008 to 2013 he was a Research Assistant and Lecturer with the Tomas Bata University in Zlin, Faculty of applied informatics. Since 2014 he is an Associate Professor and since 2017 Head of the A.I.Lab https://ailab.fai.utb.cz/ with the Department of Informatics and Artificial Intelligence, Tomas Bata University in Zlin. He is the author of more than 40 journal papers, 250 conference papers, and several book chapters as well as editorial notes. His research interests are the development of evolutionary algorithms, their modifications and benchmarking, soft computing methods, and their interdisciplinary applications in optimization and cyber-security, machine learning, neuro-evolution, data science, the theory of chaos, and complex systems. He is a recognized reviewer for many leading journals in computer science/computational intelligence. He was a part of the organizing teams for special sessions/workshops/symposiums at GECCO, IEEE WCCI, CEC, or SSCI events.


 
Joel Lehman
Joel Lehman is a machine learning researcher interested in algorithmic creativity, evolutionary algorithms, artificial life, and AI for wellbeing. Most recently he was a research scientist at OpenAI co-leading the Open-Endedness team (studying algorithms that can innovate endlessly). Previously he was a founding member of Uber AI Labs, first employee of Geometric Intelligence (acquired by Uber), and a tenure track professor at the IT University of Copenhagen. He co-wrote with Kenneth Stanley a popular science book called "Why Greatness Cannot Be Planned" on what AI search algorithms imply for individual and societal accomplishment.


Una-May O’Reilly
Dr. Una-May O'Reilly is the leader of ALFA Group at Massachusetts Institute of Technology's Computer Science and Artificial Intelligence Lab. An evolutionary computation researcher for 20+ years, she is broadly interested in adversarial intelligence — the intelligence that emerges and is recruited while learning and adapting in competitive settings. Her interest has led her to study settings where security is under threat, for which she has developed machine learning algorithms that variously model the arms races of tax compliance and auditing, malware and its detection, cyber network attacks and defenses, and adversarial paradigms in deep learning. She is passionately interested in programming and genetic programming. She is a recipient of the EvoStar Award for Outstanding Achievements in Evolutionary Computation in Europe and the ACM SIGEVO Award Recognizing Outstanding Achievements in Evolutionary Computation. Devoted to the field and committed to its growth, she served on the ACM SIGEVO executive board from SIGEVO's inception and held different officer positions before retiring from it in 2023. She co-founded the annual workshops for Women@GECCO and has proudly watched their evolution to Women+@GECCO. She was on the founding editorial boards and continues to serve on the editorial boards of Genetic Programming and Evolvable Machines, and ACM Transactions on Evolutionary Learning and Optimization. She has received a GECCO best paper award and an GECCO test of time award. She is honored to be a member of SPECIES, a member of the Julian Miller Award committee, and to chair the 2023 and 2024 committees selecting SIGEVO Awards Recognizing Outstanding Achievements in Evolutionary Computation.


 
Michal Pluhacek
Assoc. prof Michal Pluhacek received his Ph.D. degree in Information Technologies in 2016 with the dissertation topic: Modern method of development and modifications of evolutionary computational techniques. He became an assoc. prof. in 2023 after successfully defending his habilitation thesis on the topic „Inner Dynamics of Evolutionary Computation Techniques: Meaning for Practice.“ He currently works as a senior researcher at the Regional Research Centre CEBIA-Tech of Tomas Bata University in Zlin, Czech Republic. He is the author of many journal and conference papers on Particle Swarm Optimization and related topics. His research focus includes swarm intelligence theory and applications and artificial intelligence in general. In 2019, he finished six-months long research stay at New Jersey Institute of Technology, USA, focusing on swarm intelligence and swarm robotics. Recently, he is focusing his research on the interconnection of evolutionary computing and the large language models. More info: https://ailab.fai.utb.cz/our-team/


 
Niki van Stein


 
Pier Luca Lanzi
Pier Luca Lanzi received the Laurea degree in computer science from the Université degli Studi di Udine and the Ph.D. degree in Computer and Automation Engineering from the Politecnico di Milano. He is an associate professor at the Politecnico di Milano, Dept. of Electronics and Information. His research areas include genetic and evolutionary computation, reinforcement learning, and machine learning. He is interested in applications to data mining and autonomous agents. He is member of the editorial board of the "Evolutionary Computation Journal" and Editor in chief of SIGEVOlution, the ACM Newsletter of SIGEVO, the Special Interest Group on Genetic and Evolutionary Computation.


Tome Eftimov