diff --git a/scripts/agent_manager.py b/scripts/agent_manager.py index 285b9f1d7b..9939332b80 100644 --- a/scripts/agent_manager.py +++ b/scripts/agent_manager.py @@ -1,14 +1,16 @@ import openai next_key = 0 -agents = {} # key, (task, full_message_history, model) +agents = {} # key, (task, full_message_history, model) # Create new GPT agent + + def create_agent(task, prompt, model): global next_key global agents - messages = [{"role": "user", "content": prompt},] + messages = [{"role": "user", "content": prompt}, ] # Start GTP3 instance response = openai.ChatCompletion.create( @@ -22,12 +24,15 @@ def create_agent(task, prompt, model): messages.append({"role": "assistant", "content": agent_reply}) key = next_key - next_key += 1 # This is done instead of len(agents) to make keys unique even if agents are deleted + # This is done instead of len(agents) to make keys unique even if agents + # are deleted + next_key += 1 agents[key] = (task, messages, model) return key, agent_reply + def message_agent(key, message): global agents @@ -50,12 +55,14 @@ def message_agent(key, message): return agent_reply + def list_agents(): global agents # Return a list of agent keys and their tasks return [(key, task) for key, (task, _, _) in agents.items()] + def delete_agent(key): global agents @@ -64,5 +71,3 @@ def delete_agent(key): return True except KeyError: return False - - diff --git a/scripts/ai_functions.py b/scripts/ai_functions.py index 0f7780935d..0bd6c139a9 100644 --- a/scripts/ai_functions.py +++ b/scripts/ai_functions.py @@ -2,10 +2,14 @@ from typing import List, Optional import json import openai -def call_ai_function(function, args, description, model = "gpt-4"): + +def call_ai_function(function, args, description, model="gpt-4"): # parse args to comma seperated string args = ", ".join(args) - messages = [{"role": "system", "content": f"You are now the following python function: ```# {description}\n{function}```\n\nOnly respond with your `return` value."},{"role": "user", "content": args}] + messages = [{"role": "system", + "content": f"You are now the following python function: ```# {description}\n{function}```\n\nOnly respond with your `return` value."}, + {"role": "user", + "content": args}] response = openai.ChatCompletion.create( model=model, @@ -15,7 +19,8 @@ def call_ai_function(function, args, description, model = "gpt-4"): return response.choices[0].message["content"] -### Evaluating code +# Evaluating code + def evaluate_code(code: str) -> List[str]: function_string = "def analyze_code(code: str) -> List[str]:" @@ -26,7 +31,7 @@ def evaluate_code(code: str) -> List[str]: return json.loads(result_string) -### Improving code +# Improving code def improve_code(suggestions: List[str], code: str) -> str: function_string = "def generate_improved_code(suggestions: List[str], code: str) -> str:" @@ -37,7 +42,7 @@ def improve_code(suggestions: List[str], code: str) -> str: return result_string -### Writing tests +# Writing tests def write_tests(code: str, focus: List[str]) -> str: function_string = "def create_test_cases(code: str, focus: Optional[str] = None) -> str:" @@ -45,4 +50,4 @@ def write_tests(code: str, focus: List[str]) -> str: description_string = """Generates test cases for the existing code, focusing on specific areas if required.""" result_string = call_ai_function(function_string, args, description_string) - return result_string \ No newline at end of file + return result_string diff --git a/scripts/browse.py b/scripts/browse.py index 6e38e9e598..af7ce57633 100644 --- a/scripts/browse.py +++ b/scripts/browse.py @@ -1,7 +1,7 @@ from googlesearch import search import requests from bs4 import BeautifulSoup -from readability import Document# +from readability import Document import openai @@ -24,18 +24,21 @@ def scrape_text(url): return text + def extract_hyperlinks(soup): hyperlinks = [] for link in soup.find_all('a', href=True): hyperlinks.append((link.text, link['href'])) return hyperlinks + def format_hyperlinks(hyperlinks): formatted_links = [] for link_text, link_url in hyperlinks: formatted_links.append(f"{link_text} ({link_url})") return formatted_links + def scrape_links(url): response = requests.get(url) @@ -49,9 +52,10 @@ def scrape_links(url): script.extract() hyperlinks = extract_hyperlinks(soup) - + return format_hyperlinks(hyperlinks) + def split_text(text, max_length=8192): paragraphs = text.split("\n") current_length = 0 @@ -69,22 +73,33 @@ def split_text(text, max_length=8192): if current_chunk: yield "\n".join(current_chunk) -def summarize_text(text, is_website = True): + +def summarize_text(text, is_website=True): if text == "": return "Error: No text to summarize" - + print("Text length: " + str(len(text)) + " characters") summaries = [] chunks = list(split_text(text)) for i, chunk in enumerate(chunks): - print("Summarizing chunk " + str(i+1) + " / " + str(len(chunks))) + print("Summarizing chunk " + str(i + 1) + " / " + str(len(chunks))) if is_website: - messages = [{"role": "user", "content": "Please summarize the following website text, do not describe the general website, but instead concisely extract the specifc information this subpage contains.: " + chunk},] + messages = [ + { + "role": "user", + "content": "Please summarize the following website text, do not describe the general website, but instead concisely extract the specifc information this subpage contains.: " + + chunk}, + ] else: - messages = [{"role": "user", "content": "Please summarize the following text, focusing on extracting concise and specific information: " + chunk},] + messages = [ + { + "role": "user", + "content": "Please summarize the following text, focusing on extracting concise and specific information: " + + chunk}, + ] - response= openai.ChatCompletion.create( + response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=messages, max_tokens=300, @@ -98,9 +113,19 @@ def summarize_text(text, is_website = True): # Summarize the combined summary if is_website: - messages = [{"role": "user", "content": "Please summarize the following website text, do not describe the general website, but instead concisely extract the specifc information this subpage contains.: " + combined_summary},] + messages = [ + { + "role": "user", + "content": "Please summarize the following website text, do not describe the general website, but instead concisely extract the specifc information this subpage contains.: " + + combined_summary}, + ] else: - messages = [{"role": "user", "content": "Please summarize the following text, focusing on extracting concise and specific infomation: " + combined_summary},] + messages = [ + { + "role": "user", + "content": "Please summarize the following text, focusing on extracting concise and specific infomation: " + + combined_summary}, + ] response = openai.ChatCompletion.create( model="gpt-3.5-turbo", @@ -109,4 +134,4 @@ def summarize_text(text, is_website = True): ) final_summary = response.choices[0].message.content - return final_summary \ No newline at end of file + return final_summary diff --git a/scripts/chat.py b/scripts/chat.py index 9dc14b7ce3..99d99fb073 100644 --- a/scripts/chat.py +++ b/scripts/chat.py @@ -5,6 +5,7 @@ import keys # Initialize the OpenAI API client openai.api_key = keys.OPENAI_API_KEY + def create_chat_message(role, content): """ Create a chat message with the given role and content. @@ -18,7 +19,14 @@ def create_chat_message(role, content): """ return {"role": role, "content": content} -def chat_with_ai(prompt, user_input, full_message_history, permanent_memory, token_limit, debug = False): + +def chat_with_ai( + prompt, + user_input, + full_message_history, + permanent_memory, + token_limit, + debug=False): while True: try: """ @@ -34,8 +42,12 @@ def chat_with_ai(prompt, user_input, full_message_history, permanent_memory, tok Returns: str: The AI's response. """ - current_context = [create_chat_message("system", prompt), create_chat_message("system", f"Permanent memory: {permanent_memory}")] - current_context.extend(full_message_history[-(token_limit - len(prompt) - len(permanent_memory) - 10):]) + current_context = [ + create_chat_message( + "system", prompt), create_chat_message( + "system", f"Permanent memory: {permanent_memory}")] + current_context.extend( + full_message_history[-(token_limit - len(prompt) - len(permanent_memory) - 10):]) current_context.extend([create_chat_message("user", user_input)]) # Debug print the current context @@ -45,7 +57,8 @@ def chat_with_ai(prompt, user_input, full_message_history, permanent_memory, tok # Skip printing the prompt if message["role"] == "system" and message["content"] == prompt: continue - print(f"{message['role'].capitalize()}: {message['content']}") + print( + f"{message['role'].capitalize()}: {message['content']}") print("----------- END OF CONTEXT ----------------") response = openai.ChatCompletion.create( @@ -56,8 +69,12 @@ def chat_with_ai(prompt, user_input, full_message_history, permanent_memory, tok assistant_reply = response.choices[0].message["content"] # Update full message history - full_message_history.append(create_chat_message("user", user_input)) - full_message_history.append(create_chat_message("assistant", assistant_reply)) + full_message_history.append( + create_chat_message( + "user", user_input)) + full_message_history.append( + create_chat_message( + "assistant", assistant_reply)) return assistant_reply except openai.RateLimitError: diff --git a/scripts/commands.py b/scripts/commands.py index 4316fd67a4..610856321b 100644 --- a/scripts/commands.py +++ b/scripts/commands.py @@ -28,6 +28,7 @@ def get_command(response): except Exception as e: return "Error:", str(e) + def execute_command(command_name, arguments): try: if command_name == "google": @@ -41,7 +42,10 @@ def execute_command(command_name, arguments): elif command_name == "memory_ovr": return overwrite_memory(arguments["key"], arguments["string"]) elif command_name == "start_agent": - return start_agent(arguments["name"], arguments["task"], arguments["prompt"]) + return start_agent( + arguments["name"], + arguments["task"], + arguments["prompt"]) elif command_name == "message_agent": return message_agent(arguments["key"], arguments["message"]) elif command_name == "list_agents": @@ -51,7 +55,9 @@ def execute_command(command_name, arguments): elif command_name == "navigate_website": return navigate_website(arguments["action"], arguments["username"]) elif command_name == "register_account": - return register_account(arguments["username"], arguments["website"]) + return register_account( + arguments["username"], + arguments["website"]) elif command_name == "get_text_summary": return get_text_summary(arguments["url"]) elif command_name == "get_hyperlinks": @@ -66,6 +72,9 @@ def execute_command(command_name, arguments): return delete_file(arguments["file"]) elif command_name == "browse_website": return browse_website(arguments["url"]) + # TODO: Change these to take in a file rather than pasted code, if + # non-file is given, return instructions "Input should be a python + # filepath, write your code to file and try again" elif command_name == "evaluate_code": return ai.evaluate_code(arguments["code"]) elif command_name == "improve_code": @@ -82,8 +91,11 @@ def execute_command(command_name, arguments): except Exception as e: return "Error: " + str(e) + def get_datetime(): - return "Current date and time: " + datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S") + return "Current date and time: " + \ + datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S") + def google_search(query, num_results=8): search_results = [] @@ -92,6 +104,7 @@ def google_search(query, num_results=8): return json.dumps(search_results, ensure_ascii=False, indent=4) + def browse_website(url): summary = get_text_summary(url) links = get_hyperlinks(url) @@ -104,20 +117,24 @@ def browse_website(url): return result + def get_text_summary(url): text = browse.scrape_text(url) summary = browse.summarize_text(text) return """ "Result" : """ + summary + def get_hyperlinks(url): link_list = browse.scrape_links(url) return link_list + def commit_memory(string): _text = f"""Committing memory with string "{string}" """ mem.permanent_memory.append(string) return _text + def delete_memory(key): if key >= 0 and key < len(mem.permanent_memory): _text = "Deleting memory with key " + str(key) @@ -127,6 +144,8 @@ def delete_memory(key): else: print("Invalid key, cannot delete memory.") return None + + def overwrite_memory(key, string): if key >= 0 and key < len(mem.permanent_memory): _text = "Overwriting memory with key " + \ @@ -138,10 +157,12 @@ def overwrite_memory(key, string): print("Invalid key, cannot overwrite memory.") return None + def shutdown(): print("Shutting down...") quit() + def start_agent(name, task, prompt, model="gpt-3.5-turbo"): global cfg @@ -164,6 +185,7 @@ def start_agent(name, task, prompt, model="gpt-3.5-turbo"): return f"Agent {name} created with key {key}. First response: {agent_response}" + def message_agent(key, message): global cfg agent_response = agents.message_agent(key, message) @@ -174,12 +196,14 @@ def message_agent(key, message): return f"Agent {key} responded: {agent_response}" + def list_agents(): return agents.list_agents() + def delete_agent(key): result = agents.delete_agent(key) - if result == False: + if not result: return f"Agent {key} does not exist." return f"Agent {key} deleted." @@ -189,12 +213,15 @@ def navigate_website(action, username): print(_text) return "Command not implemented yet." + def register_account(username, website): - _text = "Registering account with username " + username + " and website " + website + _text = "Registering account with username " + \ + username + " and website " + website print(_text) return "Command not implemented yet." + def check_notifications(website): _text = "Checking notifications from " + website print(_text) - return "Command not implemented yet." \ No newline at end of file + return "Command not implemented yet." diff --git a/scripts/config.py b/scripts/config.py index f7e73aad05..bc7ebf7180 100644 --- a/scripts/config.py +++ b/scripts/config.py @@ -7,7 +7,9 @@ class Singleton(type): def __call__(cls, *args, **kwargs): if cls not in cls._instances: - cls._instances[cls] = super(Singleton, cls).__call__(*args, **kwargs) + cls._instances[cls] = super( + Singleton, cls).__call__( + *args, **kwargs) return cls._instances[cls] @@ -24,4 +26,4 @@ class Config(metaclass=Singleton): self.continuous_mode = value def set_speak_mode(self, value: bool): - self.speak_mode = value \ No newline at end of file + self.speak_mode = value diff --git a/scripts/execute_code.py b/scripts/execute_code.py index 74f871a7b2..cfd818d49b 100644 --- a/scripts/execute_code.py +++ b/scripts/execute_code.py @@ -1,6 +1,7 @@ import docker import os + def execute_python_file(file): workspace_folder = "auto_gpt_workspace" @@ -16,11 +17,15 @@ def execute_python_file(file): client = docker.from_env() # You can replace 'python:3.8' with the desired Python image/version - # You can find available Python images on Docker Hub: https://hub.docker.com/_/python + # You can find available Python images on Docker Hub: + # https://hub.docker.com/_/python container = client.containers.run( 'python:3.8', f'python {file}', - volumes={os.path.abspath(workspace_folder): {'bind': '/workspace', 'mode': 'ro'}}, + volumes={ + os.path.abspath(workspace_folder): { + 'bind': '/workspace', + 'mode': 'ro'}}, working_dir='/workspace', stderr=True, stdout=True, @@ -34,4 +39,4 @@ def execute_python_file(file): return logs except Exception as e: - return f"Error: {str(e)}" \ No newline at end of file + return f"Error: {str(e)}" diff --git a/scripts/file_operations.py b/scripts/file_operations.py index 998c5635e9..62b3dc4b7d 100644 --- a/scripts/file_operations.py +++ b/scripts/file_operations.py @@ -6,7 +6,8 @@ working_directory = "auto_gpt_workspace" if not os.path.exists(working_directory): os.makedirs(working_directory) - + + def safe_join(base, *paths): new_path = os.path.join(base, *paths) norm_new_path = os.path.normpath(new_path) @@ -16,6 +17,7 @@ def safe_join(base, *paths): return norm_new_path + def read_file(filename): try: filepath = safe_join(working_directory, filename) @@ -25,6 +27,7 @@ def read_file(filename): except Exception as e: return "Error: " + str(e) + def write_to_file(filename, text): try: filepath = safe_join(working_directory, filename) @@ -34,6 +37,7 @@ def write_to_file(filename, text): except Exception as e: return "Error: " + str(e) + def append_to_file(filename, text): try: filepath = safe_join(working_directory, filename) @@ -43,10 +47,11 @@ def append_to_file(filename, text): except Exception as e: return "Error: " + str(e) + def delete_file(filename): try: filepath = safe_join(working_directory, filename) os.remove(filepath) return "File deleted successfully." except Exception as e: - return "Error: " + str(e) \ No newline at end of file + return "Error: " + str(e) diff --git a/scripts/main.py b/scripts/main.py index de0f846657..d51ad0fc91 100644 --- a/scripts/main.py +++ b/scripts/main.py @@ -12,11 +12,19 @@ from enum import Enum, auto import sys from config import Config + class Argument(Enum): CONTINUOUS_MODE = "continuous-mode" SPEAK_MODE = "speak-mode" -def print_to_console(title, title_color, content, speak_text = False, min_typing_speed=0.05, max_typing_speed=0.01): + +def print_to_console( + title, + title_color, + content, + speak_text=False, + min_typing_speed=0.05, + max_typing_speed=0.01): global cfg if speak_text and cfg.speak_mode: speak.say_text(f"{title}. {content}") @@ -34,6 +42,7 @@ def print_to_console(title, title_color, content, speak_text = False, min_typing max_typing_speed = max_typing_speed * 0.95 print() + def print_assistant_thoughts(assistant_reply): global ai_name global cfg @@ -54,9 +63,15 @@ def print_assistant_thoughts(assistant_reply): assistant_thoughts_plan = None assistant_thoughts_criticism = None assistant_thoughts_speak = None - - print_to_console(f"{ai_name.upper()} THOUGHTS:", Fore.YELLOW, assistant_thoughts_text) - print_to_console("REASONING:", Fore.YELLOW, assistant_thoughts_reasoning) + + print_to_console( + f"{ai_name.upper()} THOUGHTS:", + Fore.YELLOW, + assistant_thoughts_text) + print_to_console( + "REASONING:", + Fore.YELLOW, + assistant_thoughts_reasoning) if assistant_thoughts_plan: print_to_console("PLAN:", Fore.YELLOW, "") if assistant_thoughts_plan: @@ -64,12 +79,16 @@ def print_assistant_thoughts(assistant_reply): # Split the input_string using the newline character and dash lines = assistant_thoughts_plan.split('\n') - # Iterate through the lines and print each one with a bullet point + # Iterate through the lines and print each one with a bullet + # point for line in lines: # Remove any "-" characters from the start of the line line = line.lstrip("- ") print_to_console("- ", Fore.GREEN, line.strip()) - print_to_console("CRITICISM:", Fore.YELLOW, assistant_thoughts_criticism) + print_to_console( + "CRITICISM:", + Fore.YELLOW, + assistant_thoughts_criticism) # Speak the assistant's thoughts if cfg.speak_mode and assistant_thoughts_speak: @@ -81,27 +100,45 @@ def print_assistant_thoughts(assistant_reply): except Exception as e: print_to_console("Error: \n", Fore.RED, str(e)) + def construct_prompt(): global ai_name # Construct the prompt - print_to_console("Welcome to Auto-GPT! ", Fore.GREEN, "Enter the name of your AI and its role below. Entering nothing will load defaults.", speak_text=True) + print_to_console( + "Welcome to Auto-GPT! ", + Fore.GREEN, + "Enter the name of your AI and its role below. Entering nothing will load defaults.", + speak_text=True) # Get AI Name from User - print_to_console("Name your AI: ", Fore.GREEN, "For example, 'Entrepreneur-GPT'") + print_to_console( + "Name your AI: ", + Fore.GREEN, + "For example, 'Entrepreneur-GPT'") ai_name = input("AI Name: ") if ai_name == "": ai_name = "Entrepreneur-GPT" - print_to_console(f"{ai_name} here!", Fore.LIGHTBLUE_EX, "I am at your service.", speak_text=True) + print_to_console( + f"{ai_name} here!", + Fore.LIGHTBLUE_EX, + "I am at your service.", + speak_text=True) # Get AI Role from User - print_to_console("Describe your AI's role: ", Fore.GREEN, "For example, 'an AI designed to autonomously develop and run businesses with the sole goal of increasing your net worth.'") + print_to_console( + "Describe your AI's role: ", + Fore.GREEN, + "For example, 'an AI designed to autonomously develop and run businesses with the sole goal of increasing your net worth.'") ai_role = input(f"{ai_name} is: ") if ai_role == "": ai_role = "an AI designed to autonomously develop and run businesses with the sole goal of increasing your net worth." # Enter up to 5 goals for the AI - print_to_console("Enter up to 5 goals for your AI: ", Fore.GREEN, "For example: \nIncrease net worth \nGrow Twitter Account \nDevelop and manage multiple businesses autonomously'") + print_to_console( + "Enter up to 5 goals for your AI: ", + Fore.GREEN, + "For example: \nIncrease net worth \nGrow Twitter Account \nDevelop and manage multiple businesses autonomously'") print("Enter nothing to load defaults, enter nothing when finished.", flush=True) ai_goals = [] for i in range(5): @@ -110,7 +147,8 @@ def construct_prompt(): break ai_goals.append(ai_goal) if len(ai_goals) == 0: - ai_goals = ["Increase net worth", "Grow Twitter Account", "Develop and manage multiple businesses autonomously"] + ai_goals = ["Increase net worth", "Grow Twitter Account", + "Develop and manage multiple businesses autonomously"] prompt = data.load_prompt() prompt_start = """Your decisions must always be made independently without seeking user assistance. Play to your strengths as an LLM and pursue simple strategies with no legal complications.""" @@ -124,6 +162,8 @@ def construct_prompt(): return full_prompt # Check if the python script was executed with arguments, get those arguments + + def parse_arguments(): global cfg cfg.set_continuous_mode(False) @@ -131,12 +171,16 @@ def parse_arguments(): for arg in sys.argv[1:]: if arg == Argument.CONTINUOUS_MODE.value: print_to_console("Continuous Mode: ", Fore.RED, "ENABLED") - print_to_console("WARNING: ", Fore.RED, "Continuous mode is not recommended. It is potentially dangerous and may cause your AI to run forever or carry out actions you would not usually authorise. Use at your own risk.") + print_to_console( + "WARNING: ", + Fore.RED, + "Continuous mode is not recommended. It is potentially dangerous and may cause your AI to run forever or carry out actions you would not usually authorise. Use at your own risk.") cfg.set_continuous_mode(True) elif arg == Argument.SPEAK_MODE.value: print_to_console("Speak Mode: ", Fore.GREEN, "ENABLED") cfg.set_speak_mode(True) + cfg = Config() parse_arguments() @@ -152,7 +196,12 @@ user_input = "NEXT COMMAND" while True: # Send message to AI, get response with Spinner("Thinking... "): - assistant_reply = chat.chat_with_ai(prompt, user_input, full_message_history, mem.permanent_memory, token_limit) + assistant_reply = chat.chat_with_ai( + prompt, + user_input, + full_message_history, + mem.permanent_memory, + token_limit) # Print Assistant thoughts print_assistant_thoughts(assistant_reply) @@ -162,14 +211,19 @@ while True: command_name, arguments = cmd.get_command(assistant_reply) except Exception as e: print_to_console("Error: \n", Fore.RED, str(e)) - if not cfg.continuous_mode: ### GET USER AUTHORIZATION TO EXECUTE COMMAND ### - # Get key press: Prompt the user to press enter to continue or escape to exit + # Get key press: Prompt the user to press enter to continue or escape + # to exit user_input = "" - print_to_console("NEXT ACTION: ", Fore.CYAN, f"COMMAND = {Fore.CYAN}{command_name}{Style.RESET_ALL} ARGUMENTS = {Fore.CYAN}{arguments}{Style.RESET_ALL}") - print("Enter 'y' to authorise command or 'n' to exit program...", flush=True) + print_to_console( + "NEXT ACTION: ", + Fore.CYAN, + f"COMMAND = {Fore.CYAN}{command_name}{Style.RESET_ALL} ARGUMENTS = {Fore.CYAN}{arguments}{Style.RESET_ALL}") + print( + "Enter 'y' to authorise command or 'n' to exit program...", + flush=True) while True: console_input = input(Fore.MAGENTA + "Input:" + Style.RESET_ALL) if console_input.lower() == "y": @@ -184,22 +238,31 @@ while True: if user_input != "NEXT COMMAND": print("Exiting...", flush=True) break - - print_to_console("-=-=-=-=-=-=-= COMMAND AUTHORISED BY USER -=-=-=-=-=-=-=", Fore.MAGENTA, "") + + print_to_console( + "-=-=-=-=-=-=-= COMMAND AUTHORISED BY USER -=-=-=-=-=-=-=", + Fore.MAGENTA, + "") else: # Print command - print_to_console("NEXT ACTION: ", Fore.CYAN, f"COMMAND = {Fore.CYAN}{command_name}{Style.RESET_ALL} ARGUMENTS = {Fore.CYAN}{arguments}{Style.RESET_ALL}") + print_to_console( + "NEXT ACTION: ", + Fore.CYAN, + f"COMMAND = {Fore.CYAN}{command_name}{Style.RESET_ALL} ARGUMENTS = {Fore.CYAN}{arguments}{Style.RESET_ALL}") # Exectute command if command_name.lower() != "error": result = f"Command {command_name} returned: {cmd.execute_command(command_name, arguments)}" else: - result =f"Command {command_name} threw the following error: " + arguments + result = f"Command {command_name} threw the following error: " + arguments - # Check if there's a result from the command append it to the message history - if result != None: + # Check if there's a result from the command append it to the message + # history + if result is not None: full_message_history.append(chat.create_chat_message("system", result)) print_to_console("SYSTEM: ", Fore.YELLOW, result) else: - full_message_history.append(chat.create_chat_message("system", "Unable to execute command")) + full_message_history.append( + chat.create_chat_message( + "system", "Unable to execute command")) print_to_console("SYSTEM: ", Fore.YELLOW, "Unable to execute command") diff --git a/scripts/speak.py b/scripts/speak.py index 91b4dea8b9..13ceb8d904 100644 --- a/scripts/speak.py +++ b/scripts/speak.py @@ -11,8 +11,9 @@ tts_headers = { } -def say_text(text, voice_index = 0): - tts_url = "https://api.elevenlabs.io/v1/text-to-speech/{voice_id}".format(voice_id=voices[voice_index]) +def say_text(text, voice_index=0): + tts_url = "https://api.elevenlabs.io/v1/text-to-speech/{voice_id}".format( + voice_id=voices[voice_index]) formatted_message = {"text": text} response = requests.post( @@ -27,4 +28,3 @@ def say_text(text, voice_index = 0): else: print("Request failed with status code:", response.status_code) print("Response content:", response.content) - diff --git a/scripts/spinner.py b/scripts/spinner.py index ebb5fb9862..2a48dfecfd 100644 --- a/scripts/spinner.py +++ b/scripts/spinner.py @@ -3,6 +3,7 @@ import threading import itertools import time + class Spinner: def __init__(self, message="Loading...", delay=0.1): self.spinner = itertools.cycle(['-', '/', '|', '\\'])