YouTube каталог
Self Improving Agents in 5 Minutes
🛠 How-to
en

Auto Agent: самовдосконалювані AI-агенти

Developers Digest10 днів тому4 квіт. 2026Impact 6/10
AI Аналіз

Новий проєкт Auto Agent розширює ідеї auto research Андрія Карпатого, оптимізуючи структуру самого AI-агента, що дозволяє автономно вдосконалювати агентів. Такий підхід дає змогу експериментувати з різними завданнями, підказками, інструментами та оркестрацією, потенційно приводячи до ефективніших та спеціалізованих AI-структур.

Ключові тези

  • Auto Agent оптимізує структуру агента, а не код навчання.
  • Проєкт використовує мета-агента для експериментів з різними завданнями, підказками та інструментами.
  • Підхід дозволяє автономно виявляти специфічні для домену інструменти та логіку оркестрації.
Можливості

Зниження витрат на інжиніринг AI-агентів на 30-50% за рахунок автоматизації • Можливість швидкої адаптації AI-агентів до нових завдань та доменів • Пошук неочевидних конфігурацій AI-агентів, які можуть бути ефективнішими за ручні налаштування

Нюанси

Ключова ідея полягає в автоматизації процесу інжинірингу AI-агентів, що може значно знизити витрати та час на розробку, але потребує ретельного моніторингу та валідації результатів.

Опис відео

In this video, I'm going to be taking a look at self-improving agents. Specifically, how a new project extends ideas from Andre Carpathy's auto research into something that is arguably even much bigger. Now, if you've been following the channel, I covered things like self-improving skills, continual learning, as well as how to run clog code autonomously. And one of the interesting things and areas that more and more people are starting to look at is this idea of actually autonomously improving the agent harness themselves. Now, I saw this post from Kevin Goo on X. And Kevin put out a really interesting project called Auto Agent, which I'll link within the description of the video. And if you're not familiar with Auto Research, the idea with this was basically give an AI agent a small but real LLM training setup, and let it experiment overnight. It modifies the training code, trains for 5 minutes, check if the results have improved, and then it decides whether to keep or discards the results, and it will repeat that process. And effectively overnight what you can do is you can determine based on these experiments is this actually of value. Auto agent instead of optimizing the training code what it optimizes for is the agent harness itself. So effectively it's sort of like a meta agent that experiments on different tasks of the agents prompts tools as well as orchestration. Now the results of this are really impressive which I'll touch on in just a moment. But all in all the pattern is effectively the same loop but just on different scopes. Auto research will optimize the training code whereas auto aagent will optimize the agent harness itself. The looping mechanism is effectively identical but the implications are very different. Now one thing that people really liked with Carpath setup within auto research it was very simple. It was effectively a few different files, one GPU as well as a 5minute training budget. There was the prepared pi which was fixed. This was the data prep, the tokenizer, the data loader. Nobody effectively touched this. There was the train pi and this was what the agent actually edits. The full GPT model optimizes the training loop architecture, hyperparameters, batch size, everything was effectively fair game within that file. And then the program MD is what the human edits. And this file has the instructions what to tell the agent how to try, what to avoid, how to evaluate the different results. And the key insight here was you're not writing Python anymore. You're effectively just writing the markdown file. You're programming in natural language. The human programs the agent and the agent programs the code and then the agent will just run all night. And Carpathy calls this the story of how it all began because the implication is this is what research looks like from here on out. Now with auto agent it takes that same research loop but it makes it work for any domain not just ML training. Now the architecture with this it's a bit different. Effectively you have a meta agent and a task agent. The task agent is one for doing the actual work on whatever domain you care about. It starts with almost nothing. It's just a bash tool. It reads the programm for research direction. Then it experiments with the agent.py, which is the task agents code. And then within auto agent, there's an adapter where it connects to whatever benchmark you're evaluating against. And the meta agent will spin up thousands of parallel sandboxes, run the task agent on evaluation tasks, read the results and the reasoning traces, and decide what to keep and what to revert. Very similar same type of idea to auto research. Effectively, again, overnight, the agent has domain specific tooling, verification loops, orchestration logic, things that nobody programmed that it all discovered autonomously by the meta agent just by running in a loop. And now, like they show within the repo, they have two different examples. They show spreadsheet bench as well as terminal bench. And you can see when it goes through these different iterations, how it's actually improving the harness itself as it verifies this is actually a better result than the previous one. It would go and continue building off of that. Here are some of the results with a very similar idea to what Carpathy had within Auto Research. Now within Auto agent, the loop is effectively the same. It's going to edit the agent harness similar to how Auto Research would edit the train pie. It would run on those tasks and then it would check the benchmark and then repeat. Effectively the same loop, it's just a different target. What's really interesting with this and why it matters is arguably every domain needs a bit of a different harness. And the harness engineering in of itself requires someone who understands both the domain as well as how the models behave. And the thing with companies is oftent times they don't have one workflow. Being able to optimize different harnesses that might live at different parts of the stack or different parts of the process. This allows you to potentially explore areas where you might have an optimized harness where you can run with cheaper models that are geared towards specific tasks instead of just having one monolithic harness that tries to do everything. I think this is just one of probably many projects of what we're going to see. But the domain experts are going to be really valuable with these types of projects cuz they're going to be able to define what are good instructions for the outcomes that you want from these different meta agents. And the interesting thing with this is it's arguably just another level of abstraction similar to code. We used to write the actual syntax of all of the different code. And now increasingly more and more of code is just written by AI models. A similar thing potentially could happen with harnesses where instead of actually engineering the harness manually, we could just let agents engineer the harnesses themselves. Effectively define what success looks like, point the meta agent at it, and then come back in 24 hours and see the results. That's it for this video. If you found this video useful, please like, comment, share, and subscribe. Otherwise, until the next one.