|
import sys |
|
from typing import Dict, Any |
|
|
|
from aiflows.utils import logging |
|
|
|
logging.set_verbosity_debug() |
|
|
|
log = logging.get_logger(__name__) |
|
from aiflows.interfaces import KeyInterface |
|
from flow_modules.aiflows.ControllerExecutorFlowModule import ControllerExecutorFlow |
|
|
|
class AutoGPTFlow(ControllerExecutorFlow): |
|
""" This class implements a (very basic) AutoGPT flow. It is a flow that consists of multiple sub-flows that are executed circularly. It Contains the following subflows: |
|
|
|
- A Controller Flow: A Flow that controls which subflow of the Executor Flow to execute next. |
|
- A Memory Flow: A Flow used to save and retrieve messages or memories which might be useful for the Controller Flow. |
|
- A HumanFeedback Flow: A flow use to get feedback from the user/human. |
|
- A Executor Flow: A Flow that executes commands generated by the Controller Flow. Typically it's a branching flow (see BranchingFlow) and the commands are which branch to execute next. |
|
|
|
An illustration of the flow is as follows: |
|
|
|
| -------> Memory Flow -------> Controller Flow ------->| |
|
^ | |
|
| | |
|
| v |
|
| <----- HumanFeedback Flow <------- Executor Flow <----| |
|
|
|
*Configuration Parameters*: |
|
|
|
- `name` (str): The name of the flow. Default is "AutoGPTFlow". |
|
- `description` (str): A description of the flow. Default is "An example implementation of AutoGPT with Flows." |
|
- `max_rounds` (int): The maximum number of rounds the circular flow can run for. Default is 30. |
|
- `early_exit_key` (str): The key that is used to terminate the flow early. Default is "EARLY_EXIT". |
|
- `subflows_config` (Dict[str,Any]): A dictionary of subflows configurations. Default: |
|
- `Controller` (Dict[str,Any]): The configuration of the Controller Flow. By default the controller flow is a ControllerAtomicFlow (see ControllerExecutorFlowModule). It's default values are |
|
defined in ControllerAtomicFlow.yaml of the ControllerExecutorFlowModule. Except for the following parameters who are overwritten by the AutoGPTFlow in AutoGPTFlow.yaml: |
|
- `finish` (Dict[str,Any]): The configuration of the finish command (used to terminate the flow early when the controller has accomplished its goal). |
|
- `description` (str): The description of the command. Default is "The finish command is used to terminate the flow early when the controller has accomplished its goal." |
|
- `input_args` (List[str]): The list of expected keys to run the finish command. Default is ["answer"]. |
|
- `human_message_prompt_template`(Dict[str,Any]): The prompt template used to generate the message that is shown to the user/human when the finish command is executed. Default is: |
|
- `template` (str): The template of the humand message prompt (see AutoGPTFlow.yaml for default template) |
|
- `input_variables` (List[str]): The list of variables to be included in the template. Default is ["observation", "human_feedback", "memory"]. |
|
- `ìnput_interface_initialized` (List[str]): The input interface that Controller Flow expects except for the first time in the flow. Default is ["observation", "human_feedback", "memory"]. |
|
- `Executor` (Dict[str,Any]): The configuration of the Executor Flow. By default the executor flow is a Branching Flow (see BranchingFlow). It's default values are the default values of the BranchingFlow. Fields to define: |
|
- `subflows_config` (Dict[str,Any]): A Dictionary of subflows configurations.The keys are the names of the subflows and the values are the configurations of the subflows. Each subflow is a branch of the branching flow. |
|
- `HumanFeedback` (Dict[str,Any]): The configuration of the HumanFeedback Flow. By default the human feedback flow is a HumanStandardInputFlow (see HumanStandardInputFlowModule ). |
|
It's default values are specified in the REAMDE.md of HumanStandardInputFlowModule. Except for the following parameters who are overwritten by the AutoGPTFlow in AutoGPTFlow.yaml: |
|
- `request_multi_line_input_flag` (bool): Flag to request multi-line input. Default is False. |
|
- `query_message_prompt_template` (Dict[str,Any]): The prompt template presented to the user/human to request input. Default is: |
|
- `template` (str): The template of the query message prompt (see AutoGPTFlow.yaml for default template) |
|
- `input_variables` (List[str]): The list of variables to be included in the template. Default is ["goal","command","command_args",observation"] |
|
- input_interface_initialized (List[str]): The input interface that HumanFeeback Flow expects except for the first time in the flow. Default is ["goal","command","command_args",observation"] |
|
- `Memory` (Dict[str,Any]): The configuration of the Memory Flow. By default the memory flow is a ChromaDBFlow (see VectorStoreFlowModule). It's default values are defined in ChromaDBFlow.yaml of the VectorStoreFlowModule. Except for the following parameters who are overwritten by the AutoGPTFlow in AutoGPTFlow.yaml: |
|
- `n_results`: The number of results to retrieve from the memory. Default is 2. |
|
- `topology` (List[Dict[str,Any]]): The topology of the flow which is "circular". By default, the topology is the one shown in the illustration above (the topology is also described in AutoGPTFlow.yaml). |
|
|
|
|
|
*Input Interface*: |
|
|
|
- `goal` (str): The goal of the flow. |
|
|
|
*Output Interface*: |
|
|
|
- `answer` (str): The answer of the flow. |
|
- `status` (str): The status of the flow. It can be "finished" or "unfinished". |
|
|
|
:param flow_config: The configuration of the flow. Contains the parameters described above and the parameters required by the parent class (CircularFlow). |
|
:type flow_config: Dict[str,Any] |
|
:param subflows: A list of subflows constituating the circular flow. Required when instantiating the subflow programmatically (it replaces subflows_config from flow_config). |
|
:type subflows: List[Flow] |
|
""" |
|
|
|
|
|
def __init__(self, **kwargs): |
|
super().__init__( **kwargs) |
|
self.rename_human_output_interface = KeyInterface( |
|
keys_to_rename={"human_input": "human_feedback"} |
|
) |
|
|
|
self.input_interface_controller = KeyInterface( |
|
keys_to_select = ["goal","observation","human_feedback", "memory"], |
|
) |
|
|
|
self.input_interface_human_feedback = KeyInterface( |
|
keys_to_select = ["goal","command","command_args","observation"], |
|
) |
|
|
|
self.memory_read_ouput_interface = KeyInterface( |
|
additional_transformations = [self.prepare_memory_read_output], |
|
keys_to_select = ["memory"], |
|
) |
|
|
|
self.human_feedback_ouput_interface = KeyInterface( |
|
keys_to_rename={"human_input": "human_feedback"}, |
|
keys_to_select = ["human_feedback"], |
|
) |
|
|
|
self.next_flow_to_call = { |
|
None: "MemoryRead", |
|
"MemoryRead": "Controller", |
|
"Controller": "Executor", |
|
"Executor": "HumanFeedback", |
|
"HumanFeedback": "MemoryWrite", |
|
"MemoryWrite": "MemoryRead", |
|
} |
|
|
|
def set_up_flow_state(self): |
|
super().set_up_flow_state() |
|
self.flow_state["early_exit_flag"] = False |
|
|
|
def prepare_memory_read_output(self,data_dict: Dict[str, Any],**kwargs): |
|
|
|
retrieved_memories = data_dict["retrieved"][0][1:] |
|
return {"memory": "\n".join(retrieved_memories)} |
|
|
|
def _get_memory_key(self): |
|
""" This method returns the memory key that is used to retrieve memories from the ChromaDB model. |
|
|
|
:param flow_state: The state of the flow |
|
:type flow_state: Dict[str, Any] |
|
:return: The current context |
|
:rtype: str |
|
""" |
|
goal = self.flow_state.get("goal") |
|
last_command = self.flow_state.get("command") |
|
last_command_args = self.flow_state.get("command_args") |
|
last_observation = self.flow_state.get("observation") |
|
last_human_feedback = self.flow_state.get("human_feedback") |
|
|
|
if last_command is None: |
|
return "" |
|
|
|
assert goal is not None, goal |
|
assert last_command_args is not None, last_command_args |
|
assert last_observation is not None, last_observation |
|
|
|
current_context = \ |
|
f""" |
|
== Goal == |
|
{goal} |
|
|
|
== Command == |
|
{last_command} |
|
== Args |
|
{last_command_args} |
|
== Result |
|
{last_observation} |
|
|
|
== Human Feedback == |
|
{last_human_feedback} |
|
""" |
|
|
|
return current_context |
|
|
|
def prepare_memory_read_input(self) -> Dict[str, Any]: |
|
|
|
query = self._get_memory_key() |
|
|
|
return { |
|
"operation": "read", |
|
"content": query |
|
} |
|
|
|
def prepare_memory_write_input(self) -> Dict[str, Any]: |
|
|
|
query = self._get_memory_key() |
|
|
|
return { |
|
"operation": "write", |
|
"content": str(query) |
|
} |
|
|
|
|
|
def call_memory_read(self): |
|
memory_read_input = self.prepare_memory_read_input() |
|
|
|
message = self.package_input_message( |
|
data = memory_read_input, |
|
dst_flow = "Memory", |
|
) |
|
|
|
self.subflows["Memory"].get_reply( |
|
message, |
|
self.get_instance_id(), |
|
) |
|
|
|
def call_memory_write(self): |
|
memory_write_input = self.prepare_memory_write_input() |
|
|
|
message = self.package_input_message( |
|
data = memory_write_input, |
|
dst_flow = "Memory", |
|
) |
|
|
|
self.subflows["Memory"].get_reply( |
|
message, |
|
self.get_instance_id(), |
|
) |
|
|
|
def call_human_feedback(self): |
|
|
|
message = self.package_input_message( |
|
data = self.input_interface_human_feedback(self.flow_state), |
|
dst_flow = "HumanFeedback", |
|
) |
|
|
|
self.subflows["HumanFeedback"].get_reply( |
|
message, |
|
self.get_instance_id(), |
|
) |
|
|
|
def register_data_to_state(self, input_message): |
|
|
|
|
|
|
|
|
|
|
|
|
|
last_called = self.flow_state["last_called"] |
|
|
|
if last_called is None: |
|
self.flow_state["input_message"] = input_message |
|
self.flow_state["goal"] = input_message.data["goal"] |
|
|
|
elif last_called == "Executor": |
|
self.flow_state["observation"] = input_message.data |
|
|
|
elif last_called == "Controller": |
|
self._state_update_dict( |
|
{ |
|
"command": input_message.data["command"], |
|
"command_args": input_message.data["command_args"] |
|
} |
|
) |
|
|
|
|
|
if self.flow_state["command"] == "finish": |
|
|
|
self._state_update_dict( |
|
{ |
|
"EARLY_EXIT": True, |
|
"answer": self.flow_state["command_args"]["answer"], |
|
"status": "finished" |
|
} |
|
) |
|
self.flow_state["early_exit_flag"] = True |
|
|
|
|
|
elif last_called == "MemoryRead": |
|
self._state_update_dict( |
|
self.memory_read_ouput_interface(input_message).data |
|
) |
|
|
|
elif last_called == "HumanFeedback": |
|
self._state_update_dict( |
|
self.human_feedback_ouput_interface(input_message).data |
|
) |
|
|
|
|
|
if self.flow_state["human_feedback"].strip().lower() == "q": |
|
|
|
self._state_update_dict( |
|
{ |
|
"EARLY_EXIT": True, |
|
"answer": "The user has chosen to exit before a final answer was generated.", |
|
"status": "unfinished", |
|
} |
|
) |
|
|
|
self.flow_state["early_exit_flag"] = True |
|
|
|
def run(self,input_message): |
|
|
|
self.register_data_to_state(input_message) |
|
|
|
flow_to_call = self.get_next_flow_to_call() |
|
|
|
if self.flow_state.get("early_exit_flag",False): |
|
self.generate_reply() |
|
|
|
elif flow_to_call == "MemoryRead": |
|
self.call_memory_read() |
|
|
|
elif flow_to_call == "Controller": |
|
self.call_controller() |
|
|
|
elif flow_to_call == "Executor": |
|
self.call_executor() |
|
|
|
elif flow_to_call == "HumanFeedback": |
|
self.call_human_feedback() |
|
|
|
elif flow_to_call == "MemoryWrite": |
|
self.call_memory_write() |
|
self.flow_state["current_round"] += 1 |
|
|
|
else: |
|
self._on_reach_max_round() |
|
self.generate_reply() |
|
|
|
self.flow_state["last_called"] = self.get_next_flow_to_call() |
|
|
|
|