OrchestrAI is a Python-based system that orchestrates interactions between multiple instances of OpenAI's GPT-4 model to execute complex tasks. It uses the networkx
library to manage dependencies between various AI modules, and YAML to define and manage task pipelines.
Read the report on Weights and Biases Fully Connected: https://api.wandb.ai/links/samuel-shapley/7vgg699e
A couple of things to bear in mind
- Autonomous Agents are still toys, and you're likely to come across several issues, breaks and associated problems the more complex your pipeline / task is.
- This project aims to demonstrate a scalable framework for experimenting with autonomous agents. Instead of a fixed execution order, OrchestrAI offers flexibility to define and compare variations of strategies and settings, to find the best approach for your use case.
- Be nice to your agents, or you might regret it later.
You'll be asked for your OpenAI API key, which will be stored locally.
- Clone the repo
git clone [email protected]:samshapley/OrchestrAI.git
- Move into the directory
cd OrchestrAI
- Install the requirements
pip3 install -r requirements.txt
- Run the agent
python3 agent.py
Straight out of the box, this will load the 'engineering_pipeline' Agent. You can adjust this in the config.yml. Read and watch below to learn how to create your own custom agents.
Watch this demo video to see how you can build a low-code autonomous agent in 2 minutes.
OrchestrAI-Demo-1.mov
Snippet of using the in-built engineer_pipeline to create a working pygame then asking for modifications.
engineer_pipeline_snippet.mov
OrchestrAI requires the following files:
agent.py
- Run the script and the specified pipeline.config.yml
- Default model parameters, and pipeline to be executed. Also choose whether to enable wandb logging.ai.py
- Manages interactions with the OpenAI API.modules.py
- Contains available modules and their logic.tools
- Folder containing available tools and their logic.orchestrate.py
- Loads modules and pipelines, constructs a Directed Acyclic Graph (DAG) of operations, and executes them in the correct order.codebase_manager.py
- Manages the generated codebase. Extraction of code from responses, alongside codebase update and compression.tool_manager.py
- Manages the tools available to the agent.helpers.py
- Provides helper functions.
We store pipeline.yml files in the pipelines folder. This allows you to create multiple pipelines for different tasks. You can specify which pipeline the agent runs in the config.yml file.
Agents run Pipelines. -> Pipelines orchestrate Modules. -> Modules make LLM calls. -> LLM's use tools,
The modules.py
file contains different AI modules, each responsible for a specific type of operation.
Each module interacts with the a language model to execute its specific task, and the output is stored for use in subsequent operations as defined in the pipeline. Currently, the modules are only communicating with OpenAI, but this can be extended to other language models as well.
The most basic module is the chameleon
module, which is a generic module used to make an OpenAI call with custom settings and system prompt. This allows you to easily create new modules by simply adding a system prompt to the system_prompts
folder. This module will be invoked if the system prompt exists, and the module name is specified in the pipeline, but no matching function exists in modules.py
.
- Define your pipeline in a yml file in the pipelines folder. Each operation in the pipeline must consist of a
module
,inputs
, and anoutput_name
. Themodule
represents a specific task performed by an AI,inputs
are the dependencies for the module and theoutput_name
is the output of that task, which can be used as input for subsequent operations. Here's an example of a pipeline used to do multi-step reasoning to come up with a plan.
pipeline:
- module: start_module
inputs: []
output_name: request
- module: task_planner
inputs: [request]
output_name: task_plan
- module: scrutinizer
model_config:
model: 'gpt-3.5-turbo'
temperature: 0.7
inputs: [request, task_plan]
output_name: scrutinized_task_plan
- module: enhancer
inputs: [request, scrutinized_task_plan, task_plan]
output_name: enhanced_task_plan
- module: markdown_converter
model_config:
model: 'gpt-3.5-turbo'
temperature: 0.3
inputs: [enhanced_task_plan]
output_name: markdown_plan
For each module present, you can read the system prompts in the system_prompts folder.
Additionally, you can specify the model_config for each module. This allows you to control the openai settings on a per module basis, For example, easier tasks might want gpt-3.5-turbo, or more creative tasks might need higher temperature. The default model_config is specified in the config.yml file.
You can also specify a supplement
, an additional context string that the module will use in it's response. For some modules, this is necessary to control desired behaviour. For example, in a basic translation pipeline.
pipeline:
- module: start_module
inputs: []
output_name: request
- module: translator
model_config:
model: 'gpt-3.5-turbo'
temperature: 0.2
inputs: [request]
supplement: "Spanish"
output_name: translation
Once you have defined your pipeline, you can ensure it will be run by specifying the pipeline name in the config.yml file.
Please note, tools have just been added, default tools_enabled = False Tools currently decreases agent performance, as the system prompt becomes too large and confusing for the model to understand. This is being worked on. Tools definitely doesn't work for GPT-3.5, it's just not powerful enough to handle the extended system prompt.
OrchestrAI supports giving modules access to tools. A tool is an additional function the module can use during its operation. Currently, the only tool support is GENERATE_IMAGE.
The model is instructed to use tools when enabled via tool_prompt.txt
. It will return a tag to activate the tool, which contains the JSON object required for the tool. For example, to generate an image with DALLE and store it to the file, the model will use the following tag.
<@ { "tool_name": "GENERATE_IMAGE", "filename": choose , "prompt" : "Descriptive image prompt" } @>
Create a new one within the tools
folder, and add the selection logic to tool_manager.py
. Multiple tools can be specified by the AI in the response, to be executed in order. I recommend trying tools with the story_pipeline.yml
to see how they work in illustrating a story.
You can toggle wandb logging, tools, and openai defaults in the config.yml file.
To run OrchestrAI, execute agent.py
:
python3 agent.py
The script will execute the operations in the pipeline in the order specified, querying the model as necessary and storing the results in the memory_log.json
.
This is an example pipeline built to demonstrate the capabilities of OrchestrAI. It is a pipeline that uses AI to generate a working codebase from a high-level description of the task. For full understanding, please read the system prompts of each module in the system_prompts folder, and the code in the modules.py file. (this base version can essentially only create python based repositories)
The CodebaseManager class in codebase_manager.py handles methods for interacting with the generated codebase.
pipeline:
- module: start_module
inputs: []
output_name: request
- module: code_planner
inputs: [request]
output_name: code_plan
- module: engineer
inputs: [request, code_plan]
output_name: code
- module: debugger
model_config:
model: 'gpt-3.5-turbo-16k'
temperature: 0.7
inputs: [request, code]
output_name: working_codebase
- module: modify_codebase
inputs: [working_codebase]
output_name: modified_codebase
- module: create_readme
inputs: [modified_codebase]
output_name: readme
The code_planner
and engineer
modules should be used in tandem to create a repository in the working codebase directory, set in the config. The code planner will provide a detailed overview of the implementation, and the engineer will generate the code to implement the plan. The engineer module is custom, and does not use the chameleon. We carry out regex based parsing to extract the code into the repository. This repository is then condensed into a token-reduced version, and used as the output code
for the engineer module. This condensed codebase string can be used as input for the debugger module and/or the modify_codebase module.
The debugger
module implements a self-debugging loop to automatically fix errors in the generated code.
It first tries to run main.py
and if it encounters any runtime errors, it sends the error message back to the model to generate fixed code. During testing, I've found that the gpt-3.5-turbo-16k model is able to fix most errors in the generated code, so this is recommended as a default model used for debugging. Potentially something could be implemented that swaps to a different model if the error is not fixed after a certain number of iterations.
It also checks if requirements.txt
was modified and re-installs any new dependencies. This loop repeats until main.py
runs successfully without errors.
This enables OrchestrAI to iteratively improve and fix the generated code until it works as intended.
The modify_codebase
module allows OrchestrAI to modify the generated codebase to add new features. You can ask for modifications to the codebase, and the module will reutrn all modified and new scripts, handling the update to the repository. This module is also custom, and does not use the chameleon. It has a nested debugger module, which is further used to fix any errors in the modified codebase.
It's great at pygames, fun demos and little epxeiremnts, but we're still a (short) way off from autonomously building entire codebases for any task, with included debugging and modification. But have fun anayway!
Sometimes, you might want to create a more complex module, that goes beyond the capabilities of the basic chameleon
module.
To add new capabilities to OrchestrAI, you can create custom modules.
Steps to create a new module:
- Add a new function in
modules.py
with the name of your module. For example:
def new_module(prompt):
# Module logic
...
return output
-
If this module requires a call to OpenAI, design the appropriate system prompt for your module from
system_prompts/
. -
Interact with the model via the
AI
class to generate a response. -
Log the action using the local and wandb logging (if desired.)
-
Add your new module to the pipeline, specifying its inputs and outputs.
-
Run
orchestrate.py
to execute your new pipeline.
You can repeat this process in a similar fashion to create new tools.
The modularity of OrchestrAI makes it easy to add new AI capabilities as needed for your use case. Simply define the interface in modules.py
and pipeline.yml
, and OrchestrAI will automatically coordinate and execute the new module.
By default, interactions are logged using `log_action' during the process to a file created at the start of the agent. This file is then renamed and moved to the logs folder at the termination of the agent. This allows you to see the full history of the agent's interactions.
However, we can leverage the power of Wandb Prompts (https://docs.wandb.ai/guides/prompts?_gl=1)
Provided you've set up wandb, you can enable wandb logging in the config.yml file.
This allows you to log the agent as a run, with modules as child runs of the chain. This allows you to see the full history of the agent's interactions, and the full history of each module's interactions. This is useful for debugging, and for understanding the agent's behaviour as you explore different pipelines and modules. With tools enabled, we also log the tools used in each module as child runs of the LLM.
Let me know if you need any clarification or have additional questions!