-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
"L2 Agent" #37
Comments
Seems that L4 requires the bot to write a custom plugin at runtime to handle a novel task. This could be really interesting (and feasible) with automated CI checking. I feel like this might be quite slow to run CI on every commit, but would be incredible to see it self build, save, and install a new plugin for future runs, which the L2 described in this specification would be able to invoke in the future. If we could pull off L4 I'm sure we could go viral/trend in programmer news. Most of the infrastructure is in place, but making robust CI end-to-end tests seems like a month long project. |
I previously experimented with building a L2 agent using V1
I agree both v interesting and definitely feasible. A "simple" V1 is possible if we map safe commands and safe direct actions that'll fire the intended plugins. Concerns and Questions for V2: Implementation Strategy:
Operational Flow:
Challenges:
Potential Development Paths:
|
Intuitively I believe that providing all the context and doing direct invocations (not writing the slash command) seems like the best approach. However, this can get expensive because it would require the larger model, and we would be using a lot of context. We probably would need to rely on tagging the bot if this is the case which is not as interesting. I was under the impression that we have standardized payload interfaces for all of the plugins, and that we just need to understand the help menu of each plugin. |
/start |
! This task does not reflect a business priority at the moment. You may start tasks with one of the following labels: Priority: 3 (High), Priority: 4 (Urgent), Priority: 5 (Emergency) |
/start |
Tip
|
/stop |
! Adding a label to issue failed! |
I was reading my friends blog post and was inspired to think about AI systems in a more structured way. They have these AI "level" designations.
It would be interesting to make an L2 agent according to the definition in the blog post:
This is a stepping stone to L3 according to the blog because L3 coordinates L2 and below.
We can make this a command interface where we can tag the bot and ask for requests in plain language:
@ubiquity-os give me the wallet address of @0x4007
In the above example, we should pass the entire help menu to ChatGPT and it can invoke the correct plugin based on the command description.
I think this should be quite straightforward to implement, and is a useful stepping stone towards a more advanced AI powered system.
We can use ChatGPT 4o mini because this seems pretty simple to just look at the help menu.
Advanced Version
As a more advanced version of this plugin, we can listen for every comment (no bot tag required) and the bot can jump in to help if it thinks it can based on any comment. For example, if somebody asks to be assigned to a task, perhaps the bot can somehow invoke /start on behalf of that user (which inherits all of the checks, like if they are already assigned to too many other open tasks etc)
This makes the bot's presence much more pronounced, and it will truly feel like a helpful, and proactive member of the team instead of "a tool" that must be specifically called upon for help.
Remark
I suppose if it calls other plugins with LLMs (like conversation rewards, somehow) then technically this would be considered an L3 class system.
The text was updated successfully, but these errors were encountered: