diff --git a/autogpt_platform/backend/backend/blocks/llm.py b/autogpt_platform/backend/backend/blocks/llm.py index bafbaba0fb..f510d680f5 100644 --- a/autogpt_platform/backend/backend/blocks/llm.py +++ b/autogpt_platform/backend/backend/blocks/llm.py @@ -362,6 +362,13 @@ class AITextGeneratorBlock(Block): yield "error", str(e) +class SummaryStyle(Enum): + CONCISE = "concise" + DETAILED = "detailed" + BULLET_POINTS = "bullet points" + NUMBERED_LIST = "numbered list" + + class AITextSummarizerBlock(Block): class Input(BlockSchema): text: str @@ -370,6 +377,8 @@ class AITextSummarizerBlock(Block): default=LlmModel.GPT4_TURBO, description="The language model to use for summarizing the text.", ) + focus: str = "general information" + style: SummaryStyle = SummaryStyle.CONCISE api_key: BlockSecret = SecretField(value="") # TODO: Make this dynamic max_tokens: int = 4000 # Adjust based on the model's context window @@ -440,7 +449,7 @@ class AITextSummarizerBlock(Block): raise ValueError("Failed to get a response from the LLM.") def _summarize_chunk(self, chunk: str, input_data: Input) -> str: - prompt = f"Summarize the following text concisely:\n\n{chunk}" + prompt = f"Summarize the following text in a {input_data.style} form. Focus your summary on the topic of `{input_data.focus}` if present, otherwise just provide a general summary:\n\n```{chunk}```" llm_response = self.llm_call( AIStructuredResponseGeneratorBlock.Input( @@ -454,13 +463,10 @@ class AITextSummarizerBlock(Block): return llm_response["summary"] def _combine_summaries(self, summaries: list[str], input_data: Input) -> str: - combined_text = " ".join(summaries) + combined_text = "\n\n".join(summaries) if len(combined_text.split()) <= input_data.max_tokens: - prompt = ( - "Provide a final, concise summary of the following summaries:\n\n" - + combined_text - ) + prompt = f"Provide a final summary of the following section summaries in a {input_data.style} form, focus your summary on the topic of `{input_data.focus}` if present:\n\n ```{combined_text}```\n\n Just respond with the final_summary in the format specified." llm_response = self.llm_call( AIStructuredResponseGeneratorBlock.Input( diff --git a/autogpt_platform/docker-compose.yml b/autogpt_platform/docker-compose.yml index be6f1f49ed..dcde6567f1 100644 --- a/autogpt_platform/docker-compose.yml +++ b/autogpt_platform/docker-compose.yml @@ -96,36 +96,6 @@ services: file: ./supabase/docker/docker-compose.yml service: rest - realtime: - <<: *supabase-services - extends: - file: ./supabase/docker/docker-compose.yml - service: realtime - - storage: - <<: *supabase-services - extends: - file: ./supabase/docker/docker-compose.yml - service: storage - - imgproxy: - <<: *supabase-services - extends: - file: ./supabase/docker/docker-compose.yml - service: imgproxy - - meta: - <<: *supabase-services - extends: - file: ./supabase/docker/docker-compose.yml - service: meta - - functions: - <<: *supabase-services - extends: - file: ./supabase/docker/docker-compose.yml - service: functions - analytics: <<: *supabase-services extends: diff --git a/autogpt_platform/frontend/src/components/edit/control/BlocksControl.tsx b/autogpt_platform/frontend/src/components/edit/control/BlocksControl.tsx index e09017fefd..cfea589964 100644 --- a/autogpt_platform/frontend/src/components/edit/control/BlocksControl.tsx +++ b/autogpt_platform/frontend/src/components/edit/control/BlocksControl.tsx @@ -77,11 +77,7 @@ export const BlocksControl: React.FC = ({ return ( { - if (!open) { - resetFilters(); - } - }} + onOpenChange={(open) => open || resetFilters()} > diff --git a/autogpt_platform/infra/helm/autogpt-server/values.dev.yaml b/autogpt_platform/infra/helm/autogpt-server/values.dev.yaml index 8d80932ae7..5d17aa4078 100644 --- a/autogpt_platform/infra/helm/autogpt-server/values.dev.yaml +++ b/autogpt_platform/infra/helm/autogpt-server/values.dev.yaml @@ -85,7 +85,7 @@ env: NUM_NODE_WORKERS: 5 REDIS_HOST: "redis-dev-master.redis-dev.svc.cluster.local" REDIS_PORT: "6379" - BACKEND_CORS_ALLOW_ORIGINS: ["https://dev-builder.agpt.co"] + BACKEND_CORS_ALLOW_ORIGINS: '["https://dev-builder.agpt.co"]' SUPABASE_SERVICE_ROLE_KEY: "" GITHUB_CLIENT_ID: "" GITHUB_CLIENT_SECRET: "" diff --git a/docs/content/server/setup.md b/docs/content/server/setup.md index 3a33fc781f..d6ffa78422 100644 --- a/docs/content/server/setup.md +++ b/docs/content/server/setup.md @@ -9,7 +9,7 @@ This guide will help you setup the server and builder for the project. -We also offer this in video format. You can check it out [here](https://github.com/Significant-Gravitas/AutoGPT#how-to-get-started). +We also offer this in video format. You can check it out [here](https://github.com/Significant-Gravitas/AutoGPT?tab=readme-ov-file#how-to-setup-for-self-hosting). !!! warning **DO NOT FOLLOW ANY OUTSIDE TUTORIALS AS THEY WILL LIKELY BE OUT OF DATE** @@ -20,6 +20,7 @@ To setup the server, you need to have the following installed: - [Node.js](https://nodejs.org/en/) - [Docker](https://docs.docker.com/get-docker/) +- [Git](https://git-scm.com/downloads) ### Checking if you have Node.js & NPM installed @@ -59,6 +60,16 @@ docker-compose -v Once you have Docker and Docker Compose installed, you can proceed to the next step. +## Cloning the Repository +The first step is cloning the AutoGPT repository to your computer. +To do this, open a terminal window in a folder on your computer and run: +``` +git clone https://github.com/Significant-Gravitas/AutoGPT.git +``` +If you get stuck, follow [this guide](https://docs.github.com/en/repositories/creating-and-managing-repositories/cloning-a-repository). + +Once that's complete you can close this terminal window. + ## Running the backend services To run the backend services, follow these steps: