Compare commits

..

54 Commits

Author SHA1 Message Date
SwiftyOS
5e0dad5a96 passing logger context 2024-09-13 20:04:54 +02:00
SwiftyOS
f598ba27b6 renamed folder 2024-09-13 19:57:39 +02:00
SwiftyOS
286202cc66 get agents working 2024-09-13 19:55:02 +02:00
SwiftyOS
55eb917162 formatting 2024-09-13 19:04:36 +02:00
SwiftyOS
c2843eecfa moved categories to api level 2024-09-13 19:03:11 +02:00
SwiftyOS
3971fbd800 Added test data 2024-09-13 19:02:12 +02:00
SwiftyOS
e04838feb5 added swagger instructions to the docs 2024-09-13 18:38:46 +02:00
SwiftyOS
3ef1f6e380 Added swagger docs 2024-09-13 18:25:38 +02:00
SwiftyOS
705c63b801 Merge branch 'master' into go 2024-09-13 17:52:32 +02:00
SwiftyOS
da007e3a22 add config from env 2024-09-12 19:02:50 +02:00
SwiftyOS
9e18c26e65 fixing db queries 2024-09-12 19:02:16 +02:00
SwiftyOS
30e3d65711 fixed cors added prometheus 2024-09-12 18:35:11 +02:00
SwiftyOS
b9c26b1a6b Added readme 2024-09-12 18:19:26 +02:00
SwiftyOS
220a127e51 Added cors 2024-09-12 18:19:18 +02:00
SwiftyOS
004e49edb1 added metrics 2024-09-12 18:07:24 +02:00
SwiftyOS
bf21bb1fa5 allow changing of port 2024-09-12 18:05:28 +02:00
SwiftyOS
530ddf2c34 Added migrations and docker-compose 2024-09-12 17:55:02 +02:00
SwiftyOS
33ee2f2ee5 adding tests 2024-09-12 17:11:42 +02:00
SwiftyOS
ae1d410b65 adding tests 2024-09-12 16:50:17 +02:00
SwiftyOS
196a5d6b59 formatting 2024-09-12 16:30:53 +02:00
SwiftyOS
82c1249d33 added rest of routes 2024-09-12 16:30:45 +02:00
SwiftyOS
9389a30298 fixes 2024-09-12 15:50:49 +02:00
SwiftyOS
9b58faeeb6 added analytics route 2024-09-12 15:48:11 +02:00
SwiftyOS
9cb55c5ac0 added all agent routes 2024-09-12 15:38:46 +02:00
SwiftyOS
eb1df12ce8 adding all the routes 2024-09-12 15:12:23 +02:00
SwiftyOS
12f40596b3 added handler stubs 2024-09-12 13:59:52 +02:00
SwiftyOS
3af26a9379 added models 2024-09-12 13:49:28 +02:00
SwiftyOS
4041ed3e33 deleted setup script 2024-09-12 13:45:27 +02:00
SwiftyOS
ee78653425 updated auth 2024-09-12 13:45:11 +02:00
SwiftyOS
76c5a27044 remove dead layer 2024-09-12 12:55:19 +02:00
SwiftyOS
c6aba70dd4 add a go server 2024-09-12 12:54:58 +02:00
SwiftyOS
470e7036b9 remove try 2024-09-12 12:45:14 +02:00
SwiftyOS
f6608754aa remove dead layer 2024-09-12 12:05:41 +02:00
SwiftyOS
bd70ab00e0 fixed missing prisma binaries 2024-09-12 12:02:00 +02:00
SwiftyOS
2c940b381a updated depends checks 2024-09-12 11:48:19 +02:00
SwiftyOS
03b30ebf5b updated default config 2024-09-12 11:43:29 +02:00
SwiftyOS
0d8c2a820e Sorting alphabetical 2024-09-12 11:39:21 +02:00
SwiftyOS
5ce562e11f updated config based on review comments 2024-09-12 11:20:05 +02:00
Swifty
fcf2247c20 Merge branch 'master' into remove-forge-and-autogpt 2024-09-12 11:00:46 +02:00
SwiftyOS
7326ee1221 add exp backoff on connection issues 2024-09-12 10:59:40 +02:00
SwiftyOS
70c7a3b1f3 sleep for prisma issues 2024-09-12 10:58:27 +02:00
SwiftyOS
6bdab5b777 updated version 2024-09-12 10:20:23 +02:00
SwiftyOS
5f5e31ac19 trying to get around github action linting issue 2024-09-12 10:20:09 +02:00
SwiftyOS
9e71b658d6 linting not working in github actions.. 2024-09-12 10:14:52 +02:00
SwiftyOS
b7b23d68b4 linting issue 2024-09-12 10:10:45 +02:00
SwiftyOS
630f401cee formatting 2024-09-12 10:05:44 +02:00
SwiftyOS
94fbcfb501 updated env.example 2024-09-12 09:47:56 +02:00
SwiftyOS
8102f78030 Updated logging 2024-09-12 09:43:11 +02:00
SwiftyOS
b1eb259bb3 fixed docker compose watch 2024-09-11 16:37:23 +02:00
SwiftyOS
c738eb3bc6 Reduced built image size 2024-09-11 16:27:57 +02:00
SwiftyOS
60fca5c5f0 Update build process to reduce image size 2024-09-11 16:10:06 +02:00
SwiftyOS
bba9836735 Merge remote-tracking branch 'origin/master' into remove-forge-and-autogpt 2024-09-11 15:47:40 +02:00
SwiftyOS
ad76bd1300 update lock files 2024-09-11 15:46:22 +02:00
SwiftyOS
fa16c207e0 Remove forge and autogpt 2024-09-10 14:10:59 +02:00
97 changed files with 11086 additions and 2914 deletions

View File

@@ -34,6 +34,3 @@ yarn-error.log*
# typescript
*.tsbuildinfo
next-env.d.ts
# Sentry Config File
.env.sentry-build-plugin

View File

@@ -1,4 +1,3 @@
import { withSentryConfig } from "@sentry/nextjs";
import dotenv from "dotenv";
// Load environment variables
@@ -29,56 +28,4 @@ const nextConfig = {
},
};
export default withSentryConfig(nextConfig, {
// For all available options, see:
// https://github.com/getsentry/sentry-webpack-plugin#options
org: "significant-gravitas",
project: "builder",
// Only print logs for uploading source maps in CI
silent: !process.env.CI,
// For all available options, see:
// https://docs.sentry.io/platforms/javascript/guides/nextjs/manual-setup/
// Upload a larger set of source maps for prettier stack traces (increases build time)
widenClientFileUpload: true,
// Automatically annotate React components to show their full name in breadcrumbs and session replay
reactComponentAnnotation: {
enabled: true,
},
// Route browser requests to Sentry through a Next.js rewrite to circumvent ad-blockers.
// This can increase your server load as well as your hosting bill.
// Note: Check that the configured route will not match with your Next.js middleware, otherwise reporting of client-
// side errors will fail.
tunnelRoute: "/monitoring",
// Hides source maps from generated client bundles
hideSourceMaps: true,
// Automatically tree-shake Sentry logger statements to reduce bundle size
disableLogger: true,
// Enables automatic instrumentation of Vercel Cron Monitors. (Does not yet work with App Router route handlers.)
// See the following for more information:
// https://docs.sentry.io/product/crons/
// https://vercel.com/docs/cron-jobs
automaticVercelMonitors: true,
async headers() {
return [
{
source: "/:path*",
headers: [
{
key: "Document-Policy",
value: "js-profiling",
},
],
},
];
},
});
export default nextConfig;

View File

@@ -27,7 +27,6 @@
"@radix-ui/react-switch": "^1.1.0",
"@radix-ui/react-toast": "^1.2.1",
"@radix-ui/react-tooltip": "^1.1.2",
"@sentry/nextjs": "^8",
"@supabase/ssr": "^0.4.0",
"@supabase/supabase-js": "^2.45.0",
"@tanstack/react-table": "^8.20.5",

View File

@@ -1,57 +0,0 @@
// This file configures the initialization of Sentry on the client.
// The config you add here will be used whenever a users loads a page in their browser.
// https://docs.sentry.io/platforms/javascript/guides/nextjs/
import * as Sentry from "@sentry/nextjs";
Sentry.init({
dsn: "https://fe4e4aa4a283391808a5da396da20159@o4505260022104064.ingest.us.sentry.io/4507946746380288",
// Add optional integrations for additional features
integrations: [
Sentry.replayIntegration(),
Sentry.httpClientIntegration(),
Sentry.replayCanvasIntegration(),
Sentry.reportingObserverIntegration(),
Sentry.browserProfilingIntegration(),
// Sentry.feedbackIntegration({
// // Additional SDK configuration goes in here, for example:
// colorScheme: "system",
// }),
],
// Define how likely traces are sampled. Adjust this value in production, or use tracesSampler for greater control.
tracesSampleRate: 1,
// Set `tracePropagationTargets` to control for which URLs trace propagation should be enabled
tracePropagationTargets: [
"localhost",
/^https:\/\/dev\-builder\.agpt\.co\/api/,
],
beforeSend(event, hint) {
// Check if it is an exception, and if so, show the report dialog
if (event.exception && event.event_id) {
Sentry.showReportDialog({ eventId: event.event_id });
}
return event;
},
// Define how likely Replay events are sampled.
// This sets the sample rate to be 10%. You may want this to be 100% while
// in development and sample at a lower rate in production
replaysSessionSampleRate: 0.1,
// Define how likely Replay events are sampled when an error occurs.
replaysOnErrorSampleRate: 1.0,
// Setting this option to true will print useful information to the console while you're setting up Sentry.
debug: false,
// Set profilesSampleRate to 1.0 to profile every transaction.
// Since profilesSampleRate is relative to tracesSampleRate,
// the final profiling rate can be computed as tracesSampleRate * profilesSampleRate
// For example, a tracesSampleRate of 0.5 and profilesSampleRate of 0.5 would
// result in 25% of transactions being profiled (0.5*0.5=0.25)
profilesSampleRate: 1.0,
});

View File

@@ -1,16 +0,0 @@
// This file configures the initialization of Sentry for edge features (middleware, edge routes, and so on).
// The config you add here will be used whenever one of the edge features is loaded.
// Note that this config is unrelated to the Vercel Edge Runtime and is also required when running locally.
// https://docs.sentry.io/platforms/javascript/guides/nextjs/
import * as Sentry from "@sentry/nextjs";
Sentry.init({
dsn: "https://fe4e4aa4a283391808a5da396da20159@o4505260022104064.ingest.us.sentry.io/4507946746380288",
// Define how likely traces are sampled. Adjust this value in production, or use tracesSampler for greater control.
tracesSampleRate: 1,
// Setting this option to true will print useful information to the console while you're setting up Sentry.
debug: false,
});

View File

@@ -1,23 +0,0 @@
// This file configures the initialization of Sentry on the server.
// The config you add here will be used whenever the server handles a request.
// https://docs.sentry.io/platforms/javascript/guides/nextjs/
import * as Sentry from "@sentry/nextjs";
// import { NodeProfilingIntegration } from "@sentry/profiling-node";
Sentry.init({
dsn: "https://fe4e4aa4a283391808a5da396da20159@o4505260022104064.ingest.us.sentry.io/4507946746380288",
// Define how likely traces are sampled. Adjust this value in production, or use tracesSampler for greater control.
tracesSampleRate: 1,
// Setting this option to true will print useful information to the console while you're setting up Sentry.
debug: false,
// Integrations
integrations: [
Sentry.anrIntegration(),
// NodeProfilingIntegration,
// Sentry.fsIntegration(),
],
});

View File

@@ -1,27 +0,0 @@
"use client";
import * as Sentry from "@sentry/nextjs";
import NextError from "next/error";
import { useEffect } from "react";
export default function GlobalError({
error,
}: {
error: Error & { digest?: string };
}) {
useEffect(() => {
Sentry.captureException(error);
}, [error]);
return (
<html>
<body>
{/* `NextError` is the default Next.js error page component. Its type
definition requires a `statusCode` prop. However, since the App Router
does not expose status codes for errors, we simply pass 0 to render a
generic error message. */}
<NextError statusCode={0} />
</body>
</html>
);
}

View File

@@ -3,7 +3,6 @@ import { revalidatePath } from "next/cache";
import { redirect } from "next/navigation";
import { createServerClient } from "@/lib/supabase/server";
import { z } from "zod";
import * as Sentry from "@sentry/nextjs";
const loginFormSchema = z.object({
email: z.string().email().min(2).max(64),
@@ -11,53 +10,45 @@ const loginFormSchema = z.object({
});
export async function login(values: z.infer<typeof loginFormSchema>) {
return await Sentry.withServerActionInstrumentation("login", {}, async () => {
const supabase = createServerClient();
const supabase = createServerClient();
if (!supabase) {
redirect("/error");
}
if (!supabase) {
redirect("/error");
}
// We are sure that the values are of the correct type because zod validates the form
const { data, error } = await supabase.auth.signInWithPassword(values);
// We are sure that the values are of the correct type because zod validates the form
const { data, error } = await supabase.auth.signInWithPassword(values);
if (error) {
return error.message;
}
if (error) {
return error.message;
}
if (data.session) {
await supabase.auth.setSession(data.session);
}
if (data.session) {
await supabase.auth.setSession(data.session);
}
revalidatePath("/", "layout");
redirect("/profile");
});
revalidatePath("/", "layout");
redirect("/profile");
}
export async function signup(values: z.infer<typeof loginFormSchema>) {
return await Sentry.withServerActionInstrumentation(
"signup",
{},
async () => {
const supabase = createServerClient();
const supabase = createServerClient();
if (!supabase) {
redirect("/error");
}
if (!supabase) {
redirect("/error");
}
// We are sure that the values are of the correct type because zod validates the form
const { data, error } = await supabase.auth.signUp(values);
// We are sure that the values are of the correct type because zod validates the form
const { data, error } = await supabase.auth.signUp(values);
if (error) {
return error.message;
}
if (error) {
return error.message;
}
if (data.session) {
await supabase.auth.setSession(data.session);
}
if (data.session) {
await supabase.auth.setSession(data.session);
}
revalidatePath("/", "layout");
redirect("/profile");
},
);
revalidatePath("/", "layout");
redirect("/profile");
}

View File

@@ -27,7 +27,7 @@ import "@xyflow/react/dist/style.css";
import { CustomNode } from "./CustomNode";
import "./flow.css";
import { Link } from "@/lib/autogpt-server-api";
import { getTypeColor, filterBlocksByType } from "@/lib/utils";
import { getTypeColor } from "@/lib/utils";
import { history } from "./history";
import { CustomEdge } from "./CustomEdge";
import ConnectionLine from "./ConnectionLine";
@@ -36,19 +36,14 @@ import { SaveControl } from "@/components/edit/control/SaveControl";
import { BlocksControl } from "@/components/edit/control/BlocksControl";
import {
IconPlay,
IconUndo2,
IconRedo2,
IconSquare,
IconOutput,
IconUndo2,
} from "@/components/ui/icons";
import { startTutorial } from "./tutorial";
import useAgentGraph from "@/hooks/useAgentGraph";
import { v4 as uuidv4 } from "uuid";
import { useRouter, usePathname, useSearchParams } from "next/navigation";
import { LogOut } from "lucide-react";
import RunnerUIWrapper, {
RunnerUIWrapperRef,
} from "@/components/RunnerUIWrapper";
// This is for the history, this is the minimum distance a block must move before it is logged
// It helps to prevent spamming the history with small movements especially when pressing on a input in a block
@@ -106,8 +101,6 @@ const FlowEditor: React.FC<{
// State to control if blocks menu should be pinned open
const [pinBlocksPopover, setPinBlocksPopover] = useState(false);
const runnerUIRef = useRef<RunnerUIWrapperRef>(null);
useEffect(() => {
const params = new URLSearchParams(window.location.search);
@@ -557,21 +550,9 @@ const FlowEditor: React.FC<{
onClick: handleRedo,
},
{
label: !savedAgent
? "Please save the agent to run"
: !isRunning
? "Run"
: "Stop",
label: !isRunning ? "Run" : "Stop",
icon: !isRunning ? <IconPlay /> : <IconSquare />,
onClick: !isRunning
? () => runnerUIRef.current?.runOrOpenInput()
: requestStopRun,
disabled: !savedAgent,
},
{
label: "Runner Output",
icon: <LogOut size={18} strokeWidth={1.8} />,
onClick: () => runnerUIRef.current?.openRunnerOutput(),
onClick: !isRunning ? requestSaveAndRun : requestStopRun,
},
];
@@ -607,21 +588,12 @@ const FlowEditor: React.FC<{
<SaveControl
agentMeta={savedAgent}
onSave={(isTemplate) => requestSave(isTemplate ?? false)}
agentDescription={agentDescription}
onDescriptionChange={setAgentDescription}
agentName={agentName}
onNameChange={setAgentName}
/>
</ControlPanel>
</ReactFlow>
</div>
<RunnerUIWrapper
ref={runnerUIRef}
nodes={nodes}
setNodes={setNodes}
isRunning={isRunning}
requestSaveAndRun={requestSaveAndRun}
/>
</FlowContext.Provider>
);
};

View File

@@ -1,141 +0,0 @@
import React, {
useState,
useCallback,
forwardRef,
useImperativeHandle,
} from "react";
import RunnerInputUI from "./runner-ui/RunnerInputUI";
import RunnerOutputUI from "./runner-ui/RunnerOutputUI";
import { Node } from "@xyflow/react";
import { filterBlocksByType } from "@/lib/utils";
import { BlockIORootSchema } from "@/lib/autogpt-server-api/types";
interface RunnerUIWrapperProps {
nodes: Node[];
setNodes: React.Dispatch<React.SetStateAction<Node[]>>;
isRunning: boolean;
requestSaveAndRun: () => void;
}
export interface RunnerUIWrapperRef {
openRunnerInput: () => void;
openRunnerOutput: () => void;
runOrOpenInput: () => void;
}
const RunnerUIWrapper = forwardRef<RunnerUIWrapperRef, RunnerUIWrapperProps>(
({ nodes, setNodes, isRunning, requestSaveAndRun }, ref) => {
const [isRunnerInputOpen, setIsRunnerInputOpen] = useState(false);
const [isRunnerOutputOpen, setIsRunnerOutputOpen] = useState(false);
const getBlockInputsAndOutputs = useCallback(() => {
const inputBlocks = filterBlocksByType(
nodes,
(node) => node.data.block_id === "c0a8e994-ebf1-4a9c-a4d8-89d09c86741b",
);
const outputBlocks = filterBlocksByType(
nodes,
(node) => node.data.block_id === "363ae599-353e-4804-937e-b2ee3cef3da4",
);
const inputs = inputBlocks.map((node) => ({
id: node.id,
type: "input" as const,
inputSchema: node.data.inputSchema as BlockIORootSchema,
hardcodedValues: {
name: (node.data.hardcodedValues as any).name || "",
description: (node.data.hardcodedValues as any).description || "",
value: (node.data.hardcodedValues as any).value,
placeholder_values:
(node.data.hardcodedValues as any).placeholder_values || [],
limit_to_placeholder_values:
(node.data.hardcodedValues as any).limit_to_placeholder_values ||
false,
},
}));
const outputs = outputBlocks.map((node) => ({
id: node.id,
type: "output" as const,
outputSchema: node.data.outputSchema as BlockIORootSchema,
hardcodedValues: {
name: (node.data.hardcodedValues as any).name || "Output",
description:
(node.data.hardcodedValues as any).description ||
"Output from the agent",
value: (node.data.hardcodedValues as any).value,
},
result: (node.data.executionResults as any)?.at(-1)?.data?.output,
}));
return { inputs, outputs };
}, [nodes]);
const handleInputChange = useCallback(
(nodeId: string, field: string, value: string) => {
setNodes((nds) =>
nds.map((node) => {
if (node.id === nodeId) {
return {
...node,
data: {
...node.data,
hardcodedValues: {
...(node.data.hardcodedValues as any),
[field]: value,
},
},
};
}
return node;
}),
);
},
[setNodes],
);
const openRunnerInput = () => setIsRunnerInputOpen(true);
const openRunnerOutput = () => setIsRunnerOutputOpen(true);
const runOrOpenInput = () => {
const { inputs } = getBlockInputsAndOutputs();
if (inputs.length > 0) {
openRunnerInput();
} else {
requestSaveAndRun();
}
};
useImperativeHandle(ref, () => ({
openRunnerInput,
openRunnerOutput,
runOrOpenInput,
}));
return (
<>
<RunnerInputUI
isOpen={isRunnerInputOpen}
onClose={() => setIsRunnerInputOpen(false)}
blockInputs={getBlockInputsAndOutputs().inputs}
onInputChange={handleInputChange}
onRun={() => {
setIsRunnerInputOpen(false);
requestSaveAndRun();
}}
isRunning={isRunning}
/>
<RunnerOutputUI
isOpen={isRunnerOutputOpen}
onClose={() => setIsRunnerOutputOpen(false)}
blockOutputs={getBlockInputsAndOutputs().outputs}
/>
</>
);
},
);
RunnerUIWrapper.displayName = "RunnerUIWrapper";
export default RunnerUIWrapper;

View File

@@ -9,7 +9,6 @@ import {
import FeaturedAgentsTable from "./FeaturedAgentsTable";
import { AdminAddFeaturedAgentDialog } from "./AdminAddFeaturedAgentDialog";
import { revalidatePath } from "next/cache";
import * as Sentry from "@sentry/nextjs";
export default async function AdminFeaturedAgentsControl({
className,
@@ -56,15 +55,9 @@ export default async function AdminFeaturedAgentsControl({
component: <Button>Remove</Button>,
action: async (rows) => {
"use server";
return await Sentry.withServerActionInstrumentation(
"removeFeaturedAgent",
{},
async () => {
const all = rows.map((row) => removeFeaturedAgent(row.id));
await Promise.all(all);
revalidatePath("/marketplace");
},
);
const all = rows.map((row) => removeFeaturedAgent(row.id));
await Promise.all(all);
revalidatePath("/marketplace");
},
},
]}

View File

@@ -2,23 +2,16 @@
import AutoGPTServerAPI from "@/lib/autogpt-server-api";
import MarketplaceAPI from "@/lib/marketplace-api";
import { revalidatePath } from "next/cache";
import * as Sentry from "@sentry/nextjs";
export async function approveAgent(
agentId: string,
version: number,
comment: string,
) {
return await Sentry.withServerActionInstrumentation(
"approveAgent",
{},
async () => {
const api = new MarketplaceAPI();
await api.approveAgentSubmission(agentId, version, comment);
console.debug(`Approving agent ${agentId}`);
revalidatePath("/marketplace");
},
);
const api = new MarketplaceAPI();
await api.approveAgentSubmission(agentId, version, comment);
console.debug(`Approving agent ${agentId}`);
revalidatePath("/marketplace");
}
export async function rejectAgent(
@@ -26,117 +19,67 @@ export async function rejectAgent(
version: number,
comment: string,
) {
return await Sentry.withServerActionInstrumentation(
"rejectAgent",
{},
async () => {
const api = new MarketplaceAPI();
await api.rejectAgentSubmission(agentId, version, comment);
console.debug(`Rejecting agent ${agentId}`);
revalidatePath("/marketplace");
},
);
const api = new MarketplaceAPI();
await api.rejectAgentSubmission(agentId, version, comment);
console.debug(`Rejecting agent ${agentId}`);
revalidatePath("/marketplace");
}
export async function getReviewableAgents() {
return await Sentry.withServerActionInstrumentation(
"getReviewableAgents",
{},
async () => {
const api = new MarketplaceAPI();
return api.getAgentSubmissions();
},
);
const api = new MarketplaceAPI();
return api.getAgentSubmissions();
}
export async function getFeaturedAgents(
page: number = 1,
pageSize: number = 10,
) {
return await Sentry.withServerActionInstrumentation(
"getFeaturedAgents",
{},
async () => {
const api = new MarketplaceAPI();
const featured = await api.getFeaturedAgents(page, pageSize);
console.debug(`Getting featured agents ${featured.agents.length}`);
return featured;
},
);
const api = new MarketplaceAPI();
const featured = await api.getFeaturedAgents(page, pageSize);
console.debug(`Getting featured agents ${featured.agents.length}`);
return featured;
}
export async function getFeaturedAgent(agentId: string) {
return await Sentry.withServerActionInstrumentation(
"getFeaturedAgent",
{},
async () => {
const api = new MarketplaceAPI();
const featured = await api.getFeaturedAgent(agentId);
console.debug(`Getting featured agent ${featured.agentId}`);
return featured;
},
);
const api = new MarketplaceAPI();
const featured = await api.getFeaturedAgent(agentId);
console.debug(`Getting featured agent ${featured.agentId}`);
return featured;
}
export async function addFeaturedAgent(
agentId: string,
categories: string[] = ["featured"],
) {
return await Sentry.withServerActionInstrumentation(
"addFeaturedAgent",
{},
async () => {
const api = new MarketplaceAPI();
await api.addFeaturedAgent(agentId, categories);
console.debug(`Adding featured agent ${agentId}`);
revalidatePath("/marketplace");
},
);
const api = new MarketplaceAPI();
await api.addFeaturedAgent(agentId, categories);
console.debug(`Adding featured agent ${agentId}`);
revalidatePath("/marketplace");
}
export async function removeFeaturedAgent(
agentId: string,
categories: string[] = ["featured"],
) {
return await Sentry.withServerActionInstrumentation(
"removeFeaturedAgent",
{},
async () => {
const api = new MarketplaceAPI();
await api.removeFeaturedAgent(agentId, categories);
console.debug(`Removing featured agent ${agentId}`);
revalidatePath("/marketplace");
},
);
const api = new MarketplaceAPI();
await api.removeFeaturedAgent(agentId, categories);
console.debug(`Removing featured agent ${agentId}`);
revalidatePath("/marketplace");
}
export async function getCategories() {
return await Sentry.withServerActionInstrumentation(
"getCategories",
{},
async () => {
const api = new MarketplaceAPI();
const categories = await api.getCategories();
console.debug(
`Getting categories ${categories.unique_categories.length}`,
);
return categories;
},
);
const api = new MarketplaceAPI();
const categories = await api.getCategories();
console.debug(`Getting categories ${categories.unique_categories.length}`);
return categories;
}
export async function getNotFeaturedAgents(
page: number = 1,
pageSize: number = 100,
) {
return await Sentry.withServerActionInstrumentation(
"getNotFeaturedAgents",
{},
async () => {
const api = new MarketplaceAPI();
const agents = await api.getNotFeaturedAgents(page, pageSize);
console.debug(`Getting not featured agents ${agents.agents.length}`);
return agents;
},
);
const api = new MarketplaceAPI();
const agents = await api.getNotFeaturedAgents(page, pageSize);
console.debug(`Getting not featured agents ${agents.agents.length}`);
return agents;
}

View File

@@ -19,7 +19,6 @@ import React from "react";
export type Control = {
icon: React.ReactNode;
label: string;
disabled?: boolean;
onClick: () => void;
};
@@ -51,18 +50,15 @@ export const ControlPanel = ({
{controls.map((control, index) => (
<Tooltip key={index} delayDuration={500}>
<TooltipTrigger asChild>
<div>
<Button
variant="ghost"
size="icon"
onClick={() => control.onClick()}
data-id={`control-button-${index}`}
disabled={control.disabled || false}
>
{control.icon}
<span className="sr-only">{control.label}</span>
</Button>
</div>
<Button
variant="ghost"
size="icon"
onClick={() => control.onClick()}
data-id={`control-button-${index}`}
>
{control.icon}
<span className="sr-only">{control.label}</span>
</Button>
</TooltipTrigger>
<TooltipContent side="right">{control.label}</TooltipContent>
</Tooltip>

View File

@@ -18,8 +18,6 @@ import {
interface SaveControlProps {
agentMeta: GraphMeta | null;
agentName: string;
agentDescription: string;
onSave: (isTemplate: boolean | undefined) => void;
onNameChange: (name: string) => void;
onDescriptionChange: (description: string) => void;
@@ -37,9 +35,7 @@ interface SaveControlProps {
export const SaveControl = ({
agentMeta,
onSave,
agentName,
onNameChange,
agentDescription,
onDescriptionChange,
}: SaveControlProps) => {
/**
@@ -79,7 +75,7 @@ export const SaveControl = ({
id="name"
placeholder="Enter your agent name"
className="col-span-3"
value={agentName}
defaultValue={agentMeta?.name || ""}
onChange={(e) => onNameChange(e.target.value)}
/>
<Label htmlFor="description">Description</Label>
@@ -87,21 +83,9 @@ export const SaveControl = ({
id="description"
placeholder="Your agent description"
className="col-span-3"
value={agentDescription}
defaultValue={agentMeta?.description || ""}
onChange={(e) => onDescriptionChange(e.target.value)}
/>
{agentMeta?.version && (
<>
<Label htmlFor="version">Version</Label>
<Input
id="version"
placeholder="Version"
className="col-span-3"
value={agentMeta?.version || "-"}
disabled
/>
</>
)}
</div>
</CardContent>
<CardFooter className="flex flex-col items-stretch gap-2">

View File

@@ -1,16 +1,9 @@
"use server";
import * as Sentry from "@sentry/nextjs";
import MarketplaceAPI, { AnalyticsEvent } from "@/lib/marketplace-api";
export async function makeAnalyticsEvent(event: AnalyticsEvent) {
return await Sentry.withServerActionInstrumentation(
"makeAnalyticsEvent",
{},
async () => {
const apiUrl = process.env.AGPT_SERVER_API_URL;
const api = new MarketplaceAPI();
await api.makeAnalyticsEvent(event);
},
);
const apiUrl = process.env.AGPT_SERVER_API_URL;
const api = new MarketplaceAPI();
await api.makeAnalyticsEvent(event);
}

View File

@@ -380,7 +380,7 @@ const NodeKeyValueInput: FC<{
<Input
type="text"
placeholder="Value"
defaultValue={value ?? ""}
value={value ?? ""}
onBlur={(e) =>
updateKeyValuePairs(
keyValuePairs.toSpliced(index, 1, {
@@ -563,7 +563,7 @@ const NodeStringInput: FC<{
<Input
type="text"
id={selfKey}
defaultValue={schema.secret && value ? "********" : value}
value={schema.secret && value ? "********" : value}
readOnly={schema.secret}
placeholder={
schema?.placeholder || `Enter ${beautifyString(displayName)}`
@@ -658,7 +658,7 @@ const NodeNumberInput: FC<{
<Input
type="number"
id={selfKey}
defaultValue={value}
value={value}
onBlur={(e) => handleInputChange(selfKey, parseFloat(e.target.value))}
placeholder={
schema.placeholder || `Enter ${beautifyString(displayName)}`

View File

@@ -1,61 +0,0 @@
import React from "react";
import { Input } from "@/components/ui/input";
import {
Select,
SelectContent,
SelectItem,
SelectTrigger,
SelectValue,
} from "@/components/ui/select";
interface InputBlockProps {
id: string;
name: string;
description?: string;
value: string;
placeholder_values?: any[];
onInputChange: (id: string, field: string, value: string) => void;
}
export function InputBlock({
id,
name,
description,
value,
placeholder_values,
onInputChange,
}: InputBlockProps) {
return (
<div className="space-y-1">
<h3 className="text-base font-semibold">{name || "Unnamed Input"}</h3>
{description && <p className="text-sm text-gray-600">{description}</p>}
<div>
{placeholder_values && placeholder_values.length > 1 ? (
<Select
onValueChange={(value) => onInputChange(id, "value", value)}
value={value}
>
<SelectTrigger className="w-full">
<SelectValue placeholder="Select a value" />
</SelectTrigger>
<SelectContent>
{placeholder_values.map((placeholder, index) => (
<SelectItem key={index} value={placeholder.toString()}>
{placeholder.toString()}
</SelectItem>
))}
</SelectContent>
</Select>
) : (
<Input
id={`${id}-Value`}
value={value}
onChange={(e) => onInputChange(id, "value", e.target.value)}
placeholder={placeholder_values?.[0]?.toString() || "Enter value"}
className="w-full"
/>
)}
</div>
</div>
);
}

View File

@@ -1,33 +0,0 @@
import React from "react";
import { ScrollArea } from "@/components/ui/scroll-area";
import { InputBlock } from "./RunnerInputBlock";
import { BlockInput } from "./RunnerInputUI";
interface InputListProps {
blockInputs: BlockInput[];
onInputChange: (nodeId: string, field: string, value: string) => void;
}
export function InputList({ blockInputs, onInputChange }: InputListProps) {
return (
<ScrollArea className="h-[20vh] overflow-auto pr-4 sm:h-[30vh] md:h-[40vh] lg:h-[50vh]">
<div className="space-y-4">
{blockInputs && blockInputs.length > 0 ? (
blockInputs.map((block) => (
<InputBlock
key={block.id}
id={block.id}
name={block.hardcodedValues.name}
description={block.hardcodedValues.description}
value={block.hardcodedValues.value?.toString() || ""}
placeholder_values={block.hardcodedValues.placeholder_values}
onInputChange={onInputChange}
/>
))
) : (
<p>No input blocks available.</p>
)}
</div>
</ScrollArea>
);
}

View File

@@ -1,74 +0,0 @@
import React from "react";
import {
Dialog,
DialogContent,
DialogHeader,
DialogTitle,
DialogDescription,
DialogFooter,
} from "@/components/ui/dialog";
import { Button } from "@/components/ui/button";
import { BlockIORootSchema } from "@/lib/autogpt-server-api/types";
import { InputList } from "./RunnerInputList";
export interface BlockInput {
id: string;
inputSchema: BlockIORootSchema;
hardcodedValues: {
name: string;
description: string;
value: any;
placeholder_values?: any[];
limit_to_placeholder_values?: boolean;
};
}
interface RunSettingsUiProps {
isOpen: boolean;
onClose: () => void;
blockInputs: BlockInput[];
onInputChange: (nodeId: string, field: string, value: string) => void;
onRun: () => void;
isRunning: boolean;
}
export function RunnerInputUI({
isOpen,
onClose,
blockInputs,
onInputChange,
onRun,
isRunning,
}: RunSettingsUiProps) {
const handleRun = () => {
onRun();
onClose();
};
return (
<Dialog open={isOpen} onOpenChange={onClose}>
<DialogContent className="flex max-h-[80vh] flex-col overflow-hidden sm:max-w-[400px] md:max-w-[500px] lg:max-w-[600px]">
<DialogHeader className="px-4 py-4">
<DialogTitle className="text-2xl">Run Settings</DialogTitle>
<DialogDescription className="mt-2 text-sm">
Configure settings for running your agent.
</DialogDescription>
</DialogHeader>
<div className="flex-grow overflow-y-auto px-4 py-4">
<InputList blockInputs={blockInputs} onInputChange={onInputChange} />
</div>
<DialogFooter className="px-6 py-4">
<Button
onClick={handleRun}
className="px-8 py-2 text-lg"
disabled={isRunning}
>
{isRunning ? "Running..." : "Run"}
</Button>
</DialogFooter>
</DialogContent>
</Dialog>
);
}
export default RunnerInputUI;

View File

@@ -1,94 +0,0 @@
import React from "react";
import {
Sheet,
SheetContent,
SheetHeader,
SheetTitle,
SheetDescription,
} from "@/components/ui/sheet";
import { ScrollArea } from "@/components/ui/scroll-area";
import { BlockIORootSchema } from "@/lib/autogpt-server-api/types";
import { Label } from "@/components/ui/label";
import { Textarea } from "@/components/ui/textarea";
interface BlockOutput {
id: string;
outputSchema: BlockIORootSchema;
hardcodedValues: {
name: string;
description: string;
};
result?: any;
}
interface OutputModalProps {
isOpen: boolean;
onClose: () => void;
blockOutputs: BlockOutput[];
}
const formatOutput = (output: any): string => {
if (typeof output === "object") {
try {
return JSON.stringify(output, null, 2);
} catch (error) {
return `Error formatting output: ${(error as Error).message}`;
}
}
return String(output);
};
export function RunnerOutputUI({
isOpen,
onClose,
blockOutputs,
}: OutputModalProps) {
return (
<Sheet open={isOpen} onOpenChange={onClose}>
<SheetContent
side="right"
className="flex h-full w-full flex-col overflow-hidden sm:max-w-[500px]"
>
<SheetHeader className="px-2 py-2">
<SheetTitle className="text-xl">Run Outputs</SheetTitle>
<SheetDescription className="mt-1 text-sm">
View the outputs from your agent run.
</SheetDescription>
</SheetHeader>
<div className="flex-grow overflow-y-auto px-2 py-2">
<ScrollArea className="h-full overflow-auto pr-4">
<div className="space-y-4">
{blockOutputs && blockOutputs.length > 0 ? (
blockOutputs.map((block) => (
<div key={block.id} className="space-y-1">
<Label className="text-base font-semibold">
{block.hardcodedValues.name || "Unnamed Output"}
</Label>
{block.hardcodedValues.description && (
<Label className="block text-sm text-gray-600">
{block.hardcodedValues.description}
</Label>
)}
<div className="rounded-md bg-gray-100 p-2">
<Textarea
readOnly
value={formatOutput(block.result ?? "No output yet")}
className="resize-none whitespace-pre-wrap break-words border-none bg-transparent text-sm"
/>
</div>
</div>
))
) : (
<p>No output blocks available.</p>
)}
</div>
</ScrollArea>
</div>
</SheetContent>
</Sheet>
);
}
export default RunnerOutputUI;

View File

@@ -6,7 +6,20 @@ export interface InputProps
extends React.InputHTMLAttributes<HTMLInputElement> {}
const Input = React.forwardRef<HTMLInputElement, InputProps>(
({ className, type, ...props }, ref) => {
({ className, type, value, ...props }, ref) => {
// This ref allows the `Input` component to be both controlled and uncontrolled.
// The HTMLvalue will only be updated if the value prop changes, but the user can still type in the input.
ref = ref || React.createRef<HTMLInputElement>();
React.useEffect(() => {
if (
ref &&
ref.current &&
ref.current.value !== value &&
type !== "file"
) {
ref.current.value = value;
}
}, [value, type, ref]);
return (
<input
type={type}
@@ -16,6 +29,7 @@ const Input = React.forwardRef<HTMLInputElement, InputProps>(
className,
)}
ref={ref}
defaultValue={type !== "file" ? value : undefined}
{...props}
/>
);

View File

@@ -16,7 +16,6 @@ import {
import { Connection, MarkerType } from "@xyflow/react";
import Ajv from "ajv";
import { useCallback, useEffect, useMemo, useRef, useState } from "react";
import { useRouter, useSearchParams, usePathname } from "next/navigation";
const ajv = new Ajv({ strict: false, allErrors: true });
@@ -25,11 +24,6 @@ export default function useAgentGraph(
template?: boolean,
passDataToBeads?: boolean,
) {
const [router, searchParams, pathname] = [
useRouter(),
useSearchParams(),
usePathname(),
];
const [savedAgent, setSavedAgent] = useState<Graph | null>(null);
const [agentDescription, setAgentDescription] = useState<string>("");
const [agentName, setAgentName] = useState<string>("");
@@ -139,8 +133,8 @@ export default function useAgentGraph(
id: node.id,
type: "custom",
position: {
x: node?.metadata?.position?.x || 0,
y: node?.metadata?.position?.y || 0,
x: node.metadata.position.x,
y: node.metadata.position.y,
},
data: {
block_id: block.id,
@@ -313,7 +307,7 @@ export default function useAgentGraph(
(template ? api.getTemplate(flowID) : api.getGraph(flowID)).then(
(graph) => {
console.debug("Loading graph");
console.log("Loading graph");
loadGraph(graph);
},
);
@@ -644,59 +638,31 @@ export default function useAgentGraph(
links: links,
};
// To avoid saving the same graph, we compare the payload with the saved agent.
// Differences in IDs are ignored.
const comparedPayload = {
...(({ id, ...rest }) => rest)(payload),
nodes: payload.nodes.map(
({ id, data, input_nodes, output_nodes, ...rest }) => rest,
),
links: payload.links.map(({ source_id, sink_id, ...rest }) => rest),
};
const comparedSavedAgent = {
name: savedAgent?.name,
description: savedAgent?.description,
nodes: savedAgent?.nodes.map((v) => ({
block_id: v.block_id,
input_default: v.input_default,
metadata: v.metadata,
})),
links: savedAgent?.links.map((v) => ({
sink_name: v.sink_name,
source_name: v.source_name,
})),
};
let newSavedAgent = null;
if (savedAgent && deepEquals(comparedPayload, comparedSavedAgent)) {
console.warn("No need to save: Graph is the same as version on server");
newSavedAgent = savedAgent;
if (savedAgent && deepEquals(payload, savedAgent)) {
console.debug(
"No need to save: Graph is the same as version on server",
);
// Trigger state change
setSavedAgent(savedAgent);
return;
} else {
console.debug(
"Saving new Graph version; old vs new:",
comparedPayload,
savedAgent,
payload,
);
setNodesSyncedWithSavedAgent(false);
newSavedAgent = savedAgent
? await (savedAgent.is_template
? api.updateTemplate(savedAgent.id, payload)
: api.updateGraph(savedAgent.id, payload))
: await (asTemplate
? api.createTemplate(payload)
: api.createGraph(payload));
console.debug("Response from the API:", newSavedAgent);
}
// Route the URL to the new flow ID if it's a new agent.
if (!savedAgent) {
const path = new URLSearchParams(searchParams);
path.set("flowID", newSavedAgent.id);
router.push(`${pathname}?${path.toString()}`);
return;
}
setNodesSyncedWithSavedAgent(false);
const newSavedAgent = savedAgent
? await (savedAgent.is_template
? api.updateTemplate(savedAgent.id, payload)
: api.updateGraph(savedAgent.id, payload))
: await (asTemplate
? api.createTemplate(payload)
: api.createGraph(payload));
console.debug("Response from the API:", newSavedAgent);
// Update the node IDs on the frontend
setSavedAgent(newSavedAgent);

View File

@@ -1,13 +0,0 @@
import * as Sentry from "@sentry/nextjs";
export async function register() {
if (process.env.NEXT_RUNTIME === "nodejs") {
await import("../sentry.server.config");
}
if (process.env.NEXT_RUNTIME === "edge") {
await import("../sentry.edge.config");
}
}
export const onRequestError = Sentry.captureRequestError;

View File

@@ -46,7 +46,7 @@ export default class MarketplaceAPI {
pageSize: number = 10,
): Promise<AgentListResponse> {
return this._get(
`/top-downloads/agents?page=${page}&page_size=${pageSize}`,
`agents/top-downloads?page=${page}&page_size=${pageSize}`,
);
}
@@ -54,7 +54,7 @@ export default class MarketplaceAPI {
page: number = 1,
pageSize: number = 10,
): Promise<AgentListResponse> {
return this._get(`/featured/agents?page=${page}&page_size=${pageSize}`);
return this._get(`/agents/featured?page=${page}&page_size=${pageSize}`);
}
async searchAgents(

View File

@@ -7,7 +7,6 @@ export function createClient() {
process.env.NEXT_PUBLIC_SUPABASE_ANON_KEY!,
);
} catch (error) {
console.error("error creating client", error);
return null;
}
}

View File

@@ -24,16 +24,15 @@ export function deepEquals(x: any, y: any): boolean {
const ok = Object.keys,
tx = typeof x,
ty = typeof y;
const res =
return (
x &&
y &&
tx === ty &&
(tx === "object"
? ok(x).length === ok(y).length &&
ok(x).every((key) => deepEquals(x[key], y[key]))
: x === y);
return res;
: x === y)
);
}
/** Get tailwind text color class from type name */
@@ -185,7 +184,7 @@ export const categoryColorMap: Record<string, string> = {
SEARCH: "bg-blue-300/[.7]",
BASIC: "bg-purple-300/[.7]",
INPUT: "bg-cyan-300/[.7]",
OUTPUT: "bg-red-300/[.7]",
OUTPUT: "bg-brown-300/[.7]",
LOGIC: "bg-teal-300/[.7]",
};
@@ -195,10 +194,3 @@ export function getPrimaryCategoryColor(categories: Category[]): string {
}
return categoryColorMap[categories[0].category] || "bg-gray-300/[.7]";
}
export function filterBlocksByType<T>(
blocks: T[],
predicate: (block: T) => boolean,
): T[] {
return blocks.filter(predicate);
}

View File

@@ -1,23 +1,16 @@
import { redirect } from "next/navigation";
import getServerUser from "@/hooks/getServerUser";
import React from "react";
import * as Sentry from "@sentry/nextjs";
export async function withRoleAccess(allowedRoles: string[]) {
"use server";
return await Sentry.withServerActionInstrumentation(
"withRoleAccess",
{},
async () => {
return async function <T extends React.ComponentType<any>>(Component: T) {
const { user, role, error } = await getServerUser();
return async function <T extends React.ComponentType<any>>(Component: T) {
const { user, role, error } = await getServerUser();
if (error || !user || !role || !allowedRoles.includes(role)) {
redirect("/unauthorized");
}
if (error || !user || !role || !allowedRoles.includes(role)) {
redirect("/unauthorized");
}
return Component;
};
},
);
return Component;
};
}

File diff suppressed because it is too large Load Diff

View File

@@ -7,7 +7,7 @@ from pydantic import BaseModel, Field, SecretStr, field_serializer
class _BaseCredentials(BaseModel):
id: str = Field(default_factory=lambda: str(uuid4()))
provider: str
title: Optional[str]
title: str
@field_serializer("*")
def dump_secret_strings(value: Any, _info):
@@ -18,8 +18,6 @@ class _BaseCredentials(BaseModel):
class OAuth2Credentials(_BaseCredentials):
type: Literal["oauth2"] = "oauth2"
username: Optional[str]
"""Username of the third-party service user that these credentials belong to"""
access_token: SecretStr
access_token_expires_at: Optional[int]
"""Unix timestamp (seconds) indicating when the access token expires (if at all)"""

View File

@@ -9,8 +9,7 @@ REDIS_HOST=localhost
REDIS_PORT=6379
REDIS_PASSWORD=password
ENABLE_AUTH=false
ENABLE_CREDIT=false
AUTH_ENABLED=false
APP_ENV="local"
PYRO_HOST=localhost
SENTRY_DSN=

View File

@@ -17,10 +17,6 @@ ENV POETRY_VERSION=1.8.3 \
POETRY_NO_INTERACTION=1 \
POETRY_VIRTUALENVS_CREATE=false \
PATH="$POETRY_HOME/bin:$PATH"
# Upgrade pip and setuptools to fix security vulnerabilities
RUN pip3 install --upgrade pip setuptools
RUN pip3 install poetry
# Copy and install dependencies
@@ -45,10 +41,6 @@ ENV POETRY_VERSION=1.8.3 \
POETRY_VIRTUALENVS_CREATE=false \
PATH="$POETRY_HOME/bin:$PATH"
# Upgrade pip and setuptools to fix security vulnerabilities
RUN pip3 install --upgrade pip setuptools
# Copy only necessary files from builder
COPY --from=builder /app /app
COPY --from=builder /usr/local/lib/python3.11 /usr/local/lib/python3.11

View File

@@ -14,8 +14,7 @@ class ReadCsvBlock(Block):
skip_columns: list[str] = []
class Output(BlockSchema):
row: dict[str, str]
all_data: list[dict[str, str]]
data: dict[str, str]
def __init__(self):
super().__init__(
@@ -28,15 +27,8 @@ class ReadCsvBlock(Block):
"contents": "a, b, c\n1,2,3\n4,5,6",
},
test_output=[
("row", {"a": "1", "b": "2", "c": "3"}),
("row", {"a": "4", "b": "5", "c": "6"}),
(
"all_data",
[
{"a": "1", "b": "2", "c": "3"},
{"a": "4", "b": "5", "c": "6"},
],
),
("data", {"a": "1", "b": "2", "c": "3"}),
("data", {"a": "4", "b": "5", "c": "6"}),
],
)
@@ -61,7 +53,8 @@ class ReadCsvBlock(Block):
for _ in range(input_data.skip_rows):
next(reader)
def process_row(row):
# join the data with the header
for row in reader:
data = {}
for i, value in enumerate(row):
if i not in input_data.skip_columns:
@@ -69,12 +62,4 @@ class ReadCsvBlock(Block):
data[header[i]] = value.strip() if input_data.strip else value
else:
data[str(i)] = value.strip() if input_data.strip else value
return data
all_data = []
for row in reader:
processed_row = process_row(row)
all_data.append(processed_row)
yield "row", processed_row
yield "all_data", all_data
yield "data", data

View File

@@ -1,7 +1,6 @@
import logging
from enum import Enum
from json import JSONDecodeError
from typing import Any, List, NamedTuple
from typing import List, NamedTuple
import anthropic
import ollama
@@ -25,7 +24,6 @@ LlmApiKeys = {
class ModelMetadata(NamedTuple):
provider: str
context_window: int
cost_factor: int
class LlmModel(str, Enum):
@@ -57,29 +55,26 @@ class LlmModel(str, Enum):
MODEL_METADATA = {
LlmModel.GPT4O_MINI: ModelMetadata("openai", 128000, cost_factor=10),
LlmModel.GPT4O: ModelMetadata("openai", 128000, cost_factor=12),
LlmModel.GPT4_TURBO: ModelMetadata("openai", 128000, cost_factor=11),
LlmModel.GPT3_5_TURBO: ModelMetadata("openai", 16385, cost_factor=8),
LlmModel.CLAUDE_3_5_SONNET: ModelMetadata("anthropic", 200000, cost_factor=14),
LlmModel.CLAUDE_3_HAIKU: ModelMetadata("anthropic", 200000, cost_factor=13),
LlmModel.LLAMA3_8B: ModelMetadata("groq", 8192, cost_factor=6),
LlmModel.LLAMA3_70B: ModelMetadata("groq", 8192, cost_factor=9),
LlmModel.MIXTRAL_8X7B: ModelMetadata("groq", 32768, cost_factor=7),
LlmModel.GEMMA_7B: ModelMetadata("groq", 8192, cost_factor=6),
LlmModel.GEMMA2_9B: ModelMetadata("groq", 8192, cost_factor=7),
LlmModel.LLAMA3_1_405B: ModelMetadata("groq", 8192, cost_factor=10),
# Limited to 16k during preview
LlmModel.LLAMA3_1_70B: ModelMetadata("groq", 131072, cost_factor=15),
LlmModel.LLAMA3_1_8B: ModelMetadata("groq", 131072, cost_factor=13),
LlmModel.OLLAMA_LLAMA3_8B: ModelMetadata("ollama", 8192, cost_factor=7),
LlmModel.OLLAMA_LLAMA3_405B: ModelMetadata("ollama", 8192, cost_factor=11),
LlmModel.GPT4O_MINI: ModelMetadata("openai", 128000),
LlmModel.GPT4O: ModelMetadata("openai", 128000),
LlmModel.GPT4_TURBO: ModelMetadata("openai", 128000),
LlmModel.GPT3_5_TURBO: ModelMetadata("openai", 16385),
LlmModel.CLAUDE_3_5_SONNET: ModelMetadata("anthropic", 200000),
LlmModel.CLAUDE_3_HAIKU: ModelMetadata("anthropic", 200000),
LlmModel.LLAMA3_8B: ModelMetadata("groq", 8192),
LlmModel.LLAMA3_70B: ModelMetadata("groq", 8192),
LlmModel.MIXTRAL_8X7B: ModelMetadata("groq", 32768),
LlmModel.GEMMA_7B: ModelMetadata("groq", 8192),
LlmModel.GEMMA2_9B: ModelMetadata("groq", 8192),
LlmModel.LLAMA3_1_405B: ModelMetadata(
"groq", 8192
), # Limited to 16k during preview
LlmModel.LLAMA3_1_70B: ModelMetadata("groq", 131072),
LlmModel.LLAMA3_1_8B: ModelMetadata("groq", 131072),
LlmModel.OLLAMA_LLAMA3_8B: ModelMetadata("ollama", 8192),
LlmModel.OLLAMA_LLAMA3_405B: ModelMetadata("ollama", 8192),
}
for model in LlmModel:
if model not in MODEL_METADATA:
raise ValueError(f"Missing MODEL_METADATA metadata for model: {model}")
class AIStructuredResponseGeneratorBlock(Block):
class Input(BlockSchema):
@@ -94,7 +89,7 @@ class AIStructuredResponseGeneratorBlock(Block):
)
class Output(BlockSchema):
response: dict[str, Any]
response: dict[str, str]
error: str
def __init__(self):
@@ -140,33 +135,16 @@ class AIStructuredResponseGeneratorBlock(Block):
)
return response.choices[0].message.content or ""
elif provider == "anthropic":
system_messages = [p["content"] for p in prompt if p["role"] == "system"]
sysprompt = " ".join(system_messages)
messages = []
last_role = None
for p in prompt:
if p["role"] in ["user", "assistant"]:
if p["role"] != last_role:
messages.append({"role": p["role"], "content": p["content"]})
last_role = p["role"]
else:
# If the role is the same as the last one, combine the content
messages[-1]["content"] += "\n" + p["content"]
sysprompt = "".join([p["content"] for p in prompt if p["role"] == "system"])
usrprompt = [p for p in prompt if p["role"] == "user"]
client = anthropic.Anthropic(api_key=api_key)
try:
response = client.messages.create(
model=model.value,
max_tokens=4096,
system=sysprompt,
messages=messages,
)
return response.content[0].text if response.content else ""
except anthropic.APIError as e:
error_message = f"Anthropic API error: {str(e)}"
logger.error(error_message)
raise ValueError(error_message)
response = client.messages.create(
model=model.value,
max_tokens=4096,
system=sysprompt,
messages=usrprompt, # type: ignore
)
return response.content[0].text if response.content else ""
elif provider == "groq":
client = Groq(api_key=api_key)
response_format = {"type": "json_object"} if json_format else None
@@ -217,16 +195,14 @@ class AIStructuredResponseGeneratorBlock(Block):
prompt.append({"role": "user", "content": input_data.prompt})
def parse_response(resp: str) -> tuple[dict[str, Any], str | None]:
def parse_response(resp: str) -> tuple[dict[str, str], str | None]:
try:
parsed = json.loads(resp)
if not isinstance(parsed, dict):
return {}, f"Expected a dictionary, but got {type(parsed)}"
miss_keys = set(input_data.expected_format.keys()) - set(parsed.keys())
if miss_keys:
return parsed, f"Missing keys: {miss_keys}"
return parsed, None
except JSONDecodeError as e:
except Exception as e:
return {}, f"JSON decode error: {e}"
logger.info(f"LLM request: {prompt}")
@@ -250,16 +226,7 @@ class AIStructuredResponseGeneratorBlock(Block):
if input_data.expected_format:
parsed_dict, parsed_error = parse_response(response_text)
if not parsed_error:
yield "response", {
k: (
json.loads(v)
if isinstance(v, str)
and v.startswith("[")
and v.endswith("]")
else (", ".join(v) if isinstance(v, list) else v)
)
for k, v in parsed_dict.items()
}
yield "response", {k: str(v) for k, v in parsed_dict.items()}
return
else:
yield "response", {"response": response_text}
@@ -334,7 +301,7 @@ class AITextGeneratorBlock(Block):
yield "error", str(e)
class AITextSummarizerBlock(Block):
class TextSummarizerBlock(Block):
class Input(BlockSchema):
text: str
model: LlmModel = LlmModel.GPT4_TURBO
@@ -352,8 +319,8 @@ class AITextSummarizerBlock(Block):
id="c3d4e5f6-7g8h-9i0j-1k2l-m3n4o5p6q7r8",
description="Utilize a Large Language Model (LLM) to summarize a long text.",
categories={BlockCategory.AI, BlockCategory.TEXT},
input_schema=AITextSummarizerBlock.Input,
output_schema=AITextSummarizerBlock.Output,
input_schema=TextSummarizerBlock.Input,
output_schema=TextSummarizerBlock.Output,
test_input={"text": "Lorem ipsum..." * 100},
test_output=("summary", "Final summary of a long text"),
test_mock={
@@ -445,7 +412,7 @@ class AITextSummarizerBlock(Block):
else:
# If combined summaries are still too long, recursively summarize
return self._run(
AITextSummarizerBlock.Input(
TextSummarizerBlock.Input(
text=combined_text,
api_key=input_data.api_key,
model=input_data.model,

View File

@@ -1,264 +0,0 @@
import random
from collections import defaultdict
from enum import Enum
from typing import Any, Dict, List, Optional, Union
from autogpt_server.data.block import Block, BlockCategory, BlockOutput, BlockSchema
from autogpt_server.data.model import SchemaField
class SamplingMethod(str, Enum):
RANDOM = "random"
SYSTEMATIC = "systematic"
TOP = "top"
BOTTOM = "bottom"
STRATIFIED = "stratified"
WEIGHTED = "weighted"
RESERVOIR = "reservoir"
CLUSTER = "cluster"
class DataSamplingBlock(Block):
class Input(BlockSchema):
data: Union[Dict[str, Any], List[Union[dict, List[Any]]]] = SchemaField(
description="The dataset to sample from. Can be a single dictionary, a list of dictionaries, or a list of lists.",
placeholder="{'id': 1, 'value': 'a'} or [{'id': 1, 'value': 'a'}, {'id': 2, 'value': 'b'}, ...]",
)
sample_size: int = SchemaField(
description="The number of samples to take from the dataset.",
placeholder="10",
default=10,
)
sampling_method: SamplingMethod = SchemaField(
description="The method to use for sampling.",
default=SamplingMethod.RANDOM,
)
accumulate: bool = SchemaField(
description="Whether to accumulate data before sampling.",
default=False,
)
random_seed: Optional[int] = SchemaField(
description="Seed for random number generator (optional).",
default=None,
)
stratify_key: Optional[str] = SchemaField(
description="Key to use for stratified sampling (required for stratified sampling).",
default=None,
)
weight_key: Optional[str] = SchemaField(
description="Key to use for weighted sampling (required for weighted sampling).",
default=None,
)
cluster_key: Optional[str] = SchemaField(
description="Key to use for cluster sampling (required for cluster sampling).",
default=None,
)
class Output(BlockSchema):
sampled_data: List[Union[dict, List[Any]]] = SchemaField(
description="The sampled subset of the input data."
)
sample_indices: List[int] = SchemaField(
description="The indices of the sampled data in the original dataset."
)
def __init__(self):
super().__init__(
id="4a448883-71fa-49cf-91cf-70d793bd7d87",
description="This block samples data from a given dataset using various sampling methods.",
categories={BlockCategory.LOGIC},
input_schema=DataSamplingBlock.Input,
output_schema=DataSamplingBlock.Output,
test_input={
"data": [
{"id": i, "value": chr(97 + i), "group": i % 3} for i in range(10)
],
"sample_size": 3,
"sampling_method": SamplingMethod.STRATIFIED,
"accumulate": False,
"random_seed": 42,
"stratify_key": "group",
},
test_output=[
(
"sampled_data",
[
{"id": 0, "value": "a", "group": 0},
{"id": 1, "value": "b", "group": 1},
{"id": 8, "value": "i", "group": 2},
],
),
("sample_indices", [0, 1, 8]),
],
)
self.accumulated_data = []
def run(self, input_data: Input) -> BlockOutput:
if input_data.accumulate:
if isinstance(input_data.data, dict):
self.accumulated_data.append(input_data.data)
elif isinstance(input_data.data, list):
self.accumulated_data.extend(input_data.data)
else:
raise ValueError(f"Unsupported data type: {type(input_data.data)}")
# If we don't have enough data yet, return without sampling
if len(self.accumulated_data) < input_data.sample_size:
return
data_to_sample = self.accumulated_data
else:
# If not accumulating, use the input data directly
data_to_sample = (
input_data.data
if isinstance(input_data.data, list)
else [input_data.data]
)
if input_data.random_seed is not None:
random.seed(input_data.random_seed)
data_size = len(data_to_sample)
if input_data.sample_size > data_size:
raise ValueError(
f"Sample size ({input_data.sample_size}) cannot be larger than the dataset size ({data_size})."
)
indices = []
if input_data.sampling_method == SamplingMethod.RANDOM:
indices = random.sample(range(data_size), input_data.sample_size)
elif input_data.sampling_method == SamplingMethod.SYSTEMATIC:
step = data_size // input_data.sample_size
start = random.randint(0, step - 1)
indices = list(range(start, data_size, step))[: input_data.sample_size]
elif input_data.sampling_method == SamplingMethod.TOP:
indices = list(range(input_data.sample_size))
elif input_data.sampling_method == SamplingMethod.BOTTOM:
indices = list(range(data_size - input_data.sample_size, data_size))
elif input_data.sampling_method == SamplingMethod.STRATIFIED:
if not input_data.stratify_key:
raise ValueError(
"Stratify key must be provided for stratified sampling."
)
strata = defaultdict(list)
for i, item in enumerate(data_to_sample):
if isinstance(item, dict):
strata_value = item.get(input_data.stratify_key)
elif hasattr(item, input_data.stratify_key):
strata_value = getattr(item, input_data.stratify_key)
else:
raise ValueError(
f"Stratify key '{input_data.stratify_key}' not found in item {item}"
)
if strata_value is None:
raise ValueError(
f"Stratify value for key '{input_data.stratify_key}' is None"
)
strata[str(strata_value)].append(i)
# Calculate the number of samples to take from each stratum
stratum_sizes = {
k: max(1, int(len(v) / data_size * input_data.sample_size))
for k, v in strata.items()
}
# Adjust sizes to ensure we get exactly sample_size samples
while sum(stratum_sizes.values()) != input_data.sample_size:
if sum(stratum_sizes.values()) < input_data.sample_size:
stratum_sizes[
max(stratum_sizes, key=lambda k: stratum_sizes[k])
] += 1
else:
stratum_sizes[
max(stratum_sizes, key=lambda k: stratum_sizes[k])
] -= 1
for stratum, size in stratum_sizes.items():
indices.extend(random.sample(strata[stratum], size))
elif input_data.sampling_method == SamplingMethod.WEIGHTED:
if not input_data.weight_key:
raise ValueError("Weight key must be provided for weighted sampling.")
weights = []
for item in data_to_sample:
if isinstance(item, dict):
weight = item.get(input_data.weight_key)
elif hasattr(item, input_data.weight_key):
weight = getattr(item, input_data.weight_key)
else:
raise ValueError(
f"Weight key '{input_data.weight_key}' not found in item {item}"
)
if weight is None:
raise ValueError(
f"Weight value for key '{input_data.weight_key}' is None"
)
try:
weights.append(float(weight))
except ValueError:
raise ValueError(
f"Weight value '{weight}' cannot be converted to a number"
)
if not weights:
raise ValueError(
f"No valid weights found using key '{input_data.weight_key}'"
)
indices = random.choices(
range(data_size), weights=weights, k=input_data.sample_size
)
elif input_data.sampling_method == SamplingMethod.RESERVOIR:
indices = list(range(input_data.sample_size))
for i in range(input_data.sample_size, data_size):
j = random.randint(0, i)
if j < input_data.sample_size:
indices[j] = i
elif input_data.sampling_method == SamplingMethod.CLUSTER:
if not input_data.cluster_key:
raise ValueError("Cluster key must be provided for cluster sampling.")
clusters = defaultdict(list)
for i, item in enumerate(data_to_sample):
if isinstance(item, dict):
cluster_value = item.get(input_data.cluster_key)
elif hasattr(item, input_data.cluster_key):
cluster_value = getattr(item, input_data.cluster_key)
else:
raise TypeError(
f"Item {item} does not have the cluster key '{input_data.cluster_key}'"
)
clusters[str(cluster_value)].append(i)
# Randomly select clusters until we have enough samples
selected_clusters = []
while (
sum(len(clusters[c]) for c in selected_clusters)
< input_data.sample_size
):
available_clusters = [c for c in clusters if c not in selected_clusters]
if not available_clusters:
break
selected_clusters.append(random.choice(available_clusters))
for cluster in selected_clusters:
indices.extend(clusters[cluster])
# If we have more samples than needed, randomly remove some
if len(indices) > input_data.sample_size:
indices = random.sample(indices, input_data.sample_size)
else:
raise ValueError(f"Unknown sampling method: {input_data.sampling_method}")
sampled_data = [data_to_sample[i] for i in indices]
# Clear accumulated data after sampling if accumulation is enabled
if input_data.accumulate:
self.accumulated_data = []
yield "sampled_data", sampled_data
yield "sample_indices", indices

View File

@@ -1,263 +0,0 @@
from abc import ABC, abstractmethod
from datetime import datetime, timezone
from enum import Enum
from typing import Any, Optional, Type
import prisma.errors
from prisma import Json
from prisma.enums import UserBlockCreditType
from prisma.models import UserBlockCredit
from pydantic import BaseModel
from autogpt_server.blocks.llm import (
MODEL_METADATA,
AIConversationBlock,
AIStructuredResponseGeneratorBlock,
AITextGeneratorBlock,
AITextSummarizerBlock,
)
from autogpt_server.blocks.talking_head import CreateTalkingAvatarVideoBlock
from autogpt_server.data.block import Block, BlockInput
from autogpt_server.util.settings import Config
class BlockCostType(str, Enum):
RUN = "run" # cost X credits per run
BYTE = "byte" # cost X credits per byte
SECOND = "second" # cost X credits per second
class BlockCost(BaseModel):
cost_amount: int
cost_filter: BlockInput
cost_type: BlockCostType
def __init__(
self,
cost_amount: int,
cost_type: BlockCostType = BlockCostType.RUN,
cost_filter: Optional[BlockInput] = None,
**data: Any,
) -> None:
super().__init__(
cost_amount=cost_amount,
cost_filter=cost_filter or {},
cost_type=cost_type,
**data,
)
llm_cost = [
BlockCost(
cost_type=BlockCostType.RUN,
cost_filter={
"model": model,
"api_key": None, # Running LLM with user own API key is free.
},
cost_amount=metadata.cost_factor,
)
for model, metadata in MODEL_METADATA.items()
]
BLOCK_COSTS: dict[Type[Block], list[BlockCost]] = {
AIConversationBlock: llm_cost,
AITextGeneratorBlock: llm_cost,
AIStructuredResponseGeneratorBlock: llm_cost,
AITextSummarizerBlock: llm_cost,
CreateTalkingAvatarVideoBlock: [
BlockCost(cost_amount=15, cost_filter={"api_key": None})
],
}
class UserCreditBase(ABC):
def __init__(self, num_user_credits_refill: int):
self.num_user_credits_refill = num_user_credits_refill
@abstractmethod
async def get_or_refill_credit(self, user_id: str) -> int:
"""
Get the current credit for the user and refill if no transaction has been made in the current cycle.
Returns:
int: The current credit for the user.
"""
pass
@abstractmethod
async def spend_credits(
self,
user_id: str,
user_credit: int,
block: Block,
input_data: BlockInput,
data_size: float,
run_time: float,
) -> int:
"""
Spend the credits for the user based on the block usage.
Args:
user_id (str): The user ID.
user_credit (int): The current credit for the user.
block (Block): The block that is being used.
input_data (BlockInput): The input data for the block.
data_size (float): The size of the data being processed.
run_time (float): The time taken to run the block.
Returns:
int: amount of credit spent
"""
pass
@abstractmethod
async def top_up_credits(self, user_id: str, amount: int):
"""
Top up the credits for the user.
Args:
user_id (str): The user ID.
amount (int): The amount to top up.
"""
pass
class UserCredit(UserCreditBase):
async def get_or_refill_credit(self, user_id: str) -> int:
cur_time = self.time_now()
cur_month = cur_time.replace(day=1, hour=0, minute=0, second=0, microsecond=0)
nxt_month = cur_month.replace(month=cur_month.month + 1)
user_credit = await UserBlockCredit.prisma().group_by(
by=["userId"],
sum={"amount": True},
where={
"userId": user_id,
"createdAt": {"gte": cur_month, "lt": nxt_month},
"isActive": True,
},
)
if user_credit:
credit_sum = user_credit[0].get("_sum") or {}
return credit_sum.get("amount", 0)
key = f"MONTHLY-CREDIT-TOP-UP-{cur_month}"
try:
await UserBlockCredit.prisma().create(
data={
"amount": self.num_user_credits_refill,
"type": UserBlockCreditType.TOP_UP,
"userId": user_id,
"transactionKey": key,
"createdAt": self.time_now(),
}
)
except prisma.errors.UniqueViolationError:
pass # Already refilled this month
return self.num_user_credits_refill
@staticmethod
def time_now():
return datetime.now(timezone.utc)
@staticmethod
def _block_usage_cost(
block: Block,
input_data: BlockInput,
data_size: float,
run_time: float,
) -> tuple[int, BlockInput]:
block_costs = BLOCK_COSTS.get(type(block))
if not block_costs:
return 0, {}
for block_cost in block_costs:
if all(input_data.get(k) == b for k, b in block_cost.cost_filter.items()):
if block_cost.cost_type == BlockCostType.RUN:
return block_cost.cost_amount, block_cost.cost_filter
if block_cost.cost_type == BlockCostType.SECOND:
return (
int(run_time * block_cost.cost_amount),
block_cost.cost_filter,
)
if block_cost.cost_type == BlockCostType.BYTE:
return (
int(data_size * block_cost.cost_amount),
block_cost.cost_filter,
)
return 0, {}
async def spend_credits(
self,
user_id: str,
user_credit: int,
block: Block,
input_data: BlockInput,
data_size: float,
run_time: float,
validate_balance: bool = True,
) -> int:
cost, matching_filter = self._block_usage_cost(
block=block, input_data=input_data, data_size=data_size, run_time=run_time
)
if cost <= 0:
return 0
if validate_balance and user_credit < cost:
raise ValueError(f"Insufficient credit: {user_credit} < {cost}")
await UserBlockCredit.prisma().create(
data={
"userId": user_id,
"amount": -cost,
"type": UserBlockCreditType.USAGE,
"blockId": block.id,
"metadata": Json(
{
"block": block.name,
"input": matching_filter,
}
),
"createdAt": self.time_now(),
}
)
return cost
async def top_up_credits(self, user_id: str, amount: int):
await UserBlockCredit.prisma().create(
data={
"userId": user_id,
"amount": amount,
"type": UserBlockCreditType.TOP_UP,
"createdAt": self.time_now(),
}
)
class DisabledUserCredit(UserCreditBase):
async def get_or_refill_credit(self, *args, **kwargs) -> int:
return 0
async def spend_credits(self, *args, **kwargs) -> int:
return 0
async def top_up_credits(self, *args, **kwargs):
pass
def get_user_credit_model() -> UserCreditBase:
config = Config()
if config.enable_credit.lower() == "true":
return UserCredit(config.num_user_credits_refill)
else:
return DisabledUserCredit(0)
def get_block_costs() -> dict[str, list[BlockCost]]:
return {block().id: costs for block, costs in BLOCK_COSTS.items()}

View File

@@ -1,9 +1,9 @@
from collections import defaultdict
from datetime import datetime, timezone
from enum import Enum
from multiprocessing import Manager
from typing import Any, Generic, TypeVar
from prisma.enums import AgentExecutionStatus
from prisma.models import (
AgentGraphExecution,
AgentNodeExecution,
@@ -21,14 +21,12 @@ from autogpt_server.util import json, mock
class GraphExecution(BaseModel):
user_id: str
graph_exec_id: str
graph_id: str
start_node_execs: list["NodeExecution"]
graph_id: str
class NodeExecution(BaseModel):
user_id: str
graph_exec_id: str
graph_id: str
node_exec_id: str
@@ -36,7 +34,13 @@ class NodeExecution(BaseModel):
data: BlockInput
ExecutionStatus = AgentExecutionStatus
class ExecutionStatus(str, Enum):
INCOMPLETE = "INCOMPLETE"
QUEUED = "QUEUED"
RUNNING = "RUNNING"
COMPLETED = "COMPLETED"
FAILED = "FAILED"
T = TypeVar("T")
@@ -144,7 +148,6 @@ async def create_graph_execution(
data={
"agentGraphId": graph_id,
"agentGraphVersion": graph_version,
"executionStatus": ExecutionStatus.QUEUED,
"AgentNodeExecutions": {
"create": [ # type: ignore
{
@@ -256,20 +259,10 @@ async def upsert_execution_output(
)
async def update_graph_execution_start_time(graph_exec_id: str):
await AgentGraphExecution.prisma().update(
where={"id": graph_exec_id},
data={
"executionStatus": ExecutionStatus.RUNNING,
"startedAt": datetime.now(tz=timezone.utc),
},
)
async def update_graph_execution_stats(graph_exec_id: str, stats: dict[str, Any]):
await AgentGraphExecution.prisma().update(
where={"id": graph_exec_id},
data={"executionStatus": ExecutionStatus.COMPLETED, "stats": json.dumps(stats)},
data={"stats": json.dumps(stats)},
)

View File

@@ -17,10 +17,8 @@ if TYPE_CHECKING:
from autogpt_server.blocks.basic import AgentInputBlock
from autogpt_server.data import db
from autogpt_server.data.block import Block, BlockData, BlockInput, get_block
from autogpt_server.data.credit import get_user_credit_model
from autogpt_server.data.execution import (
ExecutionQueue,
ExecutionResult,
ExecutionStatus,
GraphExecution,
NodeExecution,
@@ -47,41 +45,25 @@ from autogpt_server.util.type import convert
logger = logging.getLogger(__name__)
class LogMetadata:
def __init__(
self,
user_id: str,
graph_eid: str,
graph_id: str,
node_eid: str,
node_id: str,
block_name: str,
):
self.metadata = {
"component": "ExecutionManager",
"user_id": user_id,
"graph_eid": graph_eid,
"graph_id": graph_id,
"node_eid": node_eid,
"node_id": node_id,
"block_name": block_name,
}
self.prefix = f"[ExecutionManager|uid:{user_id}|gid:{graph_id}|nid:{node_id}]|geid:{graph_eid}|nid:{node_eid}|{block_name}]"
def get_log_metadata(
graph_eid: str,
graph_id: str,
node_eid: str,
node_id: str,
block_name: str,
) -> dict:
return {
"component": "ExecutionManager",
"graph_eid": graph_eid,
"graph_id": graph_id,
"node_eid": node_eid,
"node_id": node_id,
"block_name": block_name,
}
def info(self, msg: str, **extra):
logger.info(msg, extra={"json_fields": {**self.metadata, **extra}})
def warning(self, msg: str, **extra):
logger.warning(msg, extra={"json_fields": {**self.metadata, **extra}})
def error(self, msg: str, **extra):
logger.error(msg, extra={"json_fields": {**self.metadata, **extra}})
def debug(self, msg: str, **extra):
logger.debug(msg, extra={"json_fields": {**self.metadata, **extra}})
def exception(self, msg: str, **extra):
logger.exception(msg, extra={"json_fields": {**self.metadata, **extra}})
def get_log_prefix(graph_eid: str, node_eid: str, block_name: str = "-"):
return f"[ExecutionManager][graph-eid-{graph_eid}|node-eid-{node_eid}|{block_name}]"
T = TypeVar("T")
@@ -107,7 +89,6 @@ def execute_node(
Returns:
The subsequent node to be enqueued, or None if there is no subsequent node.
"""
user_id = data.user_id
graph_exec_id = data.graph_exec_id
graph_id = data.graph_id
node_exec_id = data.node_exec_id
@@ -118,10 +99,9 @@ def execute_node(
def wait(f: Coroutine[Any, Any, T]) -> T:
return loop.run_until_complete(f)
def update_execution(status: ExecutionStatus) -> ExecutionResult:
def update_execution(status: ExecutionStatus):
exec_update = wait(update_execution_status(node_exec_id, status))
api_client.send_execution_update(exec_update.model_dump())
return exec_update
node = wait(get_node(node_id))
@@ -131,35 +111,43 @@ def execute_node(
return
# Sanity check: validate the execution input.
log_metadata = LogMetadata(
user_id=user_id,
log_metadata = get_log_metadata(
graph_eid=graph_exec_id,
graph_id=graph_id,
node_eid=node_exec_id,
node_id=node_id,
block_name=node_block.name,
)
prefix = get_log_prefix(
graph_eid=graph_exec_id,
node_eid=node_exec_id,
block_name=node_block.name,
)
input_data, error = validate_exec(node, data.data, resolve_input=False)
if input_data is None:
log_metadata.error(f"Skip execution, input validation error: {error}")
logger.error(
"{prefix} Skip execution, input validation error",
extra={"json_fields": {**log_metadata, "error": error}},
)
return
# Execute the node
input_data_str = json.dumps(input_data)
input_size = len(input_data_str)
log_metadata.info("Executed node with input", input=input_data_str)
logger.info(
f"{prefix} Executed node with input",
extra={"json_fields": {**log_metadata, "input": input_data_str}},
)
update_execution(ExecutionStatus.RUNNING)
user_credit = get_user_credit_model()
output_size = 0
try:
credit = wait(user_credit.get_or_refill_credit(user_id))
if credit < 0:
raise ValueError(f"Insufficient credit: {credit}")
for output_name, output_data in node_block.execute(input_data):
output_size += len(json.dumps(output_data))
log_metadata.info("Node produced output", output_name=output_data)
logger.info(
f"{prefix} Node produced output",
extra={"json_fields": {**log_metadata, output_name: output_data}},
)
wait(upsert_execution_output(node_exec_id, output_name, output_data))
for execution in _enqueue_next_nodes(
@@ -167,25 +155,20 @@ def execute_node(
loop=loop,
node=node,
output=(output_name, output_data),
user_id=user_id,
graph_exec_id=graph_exec_id,
graph_id=graph_id,
log_metadata=log_metadata,
):
yield execution
r = update_execution(ExecutionStatus.COMPLETED)
s = input_size + output_size
t = (
(r.end_time - r.start_time).total_seconds()
if r.end_time and r.start_time
else 0
)
wait(user_credit.spend_credits(user_id, credit, node_block, input_data, s, t))
update_execution(ExecutionStatus.COMPLETED)
except Exception as e:
error_msg = str(e)
log_metadata.exception(f"Node execution failed with error {error_msg}")
error_msg = f"{e.__class__.__name__}: {e}"
logger.exception(
f"{prefix} Node execution failed with error",
extra={"json_fields": {**log_metadata, error: error_msg}},
)
wait(upsert_execution_output(node_exec_id, "error", error_msg))
update_execution(ExecutionStatus.FAILED)
@@ -211,10 +194,9 @@ def _enqueue_next_nodes(
loop: asyncio.AbstractEventLoop,
node: Node,
output: BlockData,
user_id: str,
graph_exec_id: str,
graph_id: str,
log_metadata: LogMetadata,
log_metadata: dict,
) -> list[NodeExecution]:
def wait(f: Coroutine[Any, Any, T]) -> T:
return loop.run_until_complete(f)
@@ -227,7 +209,6 @@ def _enqueue_next_nodes(
)
api_client.send_execution_update(exec_update.model_dump())
return NodeExecution(
user_id=user_id,
graph_exec_id=graph_exec_id,
graph_id=graph_id,
node_exec_id=node_exec_id,
@@ -281,11 +262,17 @@ def _enqueue_next_nodes(
# Incomplete input data, skip queueing the execution.
if not next_node_input:
log_metadata.warning(f"Skipped queueing {suffix}")
logger.warning(
f"Skipped queueing {suffix}",
extra={"json_fields": {**log_metadata}},
)
return enqueued_executions
# Input is complete, enqueue the execution.
log_metadata.info(f"Enqueued {suffix}")
logger.info(
f"Enqueued {suffix}",
extra={"json_fields": {**log_metadata}},
)
enqueued_executions.append(
add_enqueued_execution(next_node_exec_id, next_node_id, next_node_input)
)
@@ -311,9 +298,11 @@ def _enqueue_next_nodes(
idata, msg = validate_exec(next_node, idata)
suffix = f"{next_output_name}>{next_input_name}~{ineid}:{msg}"
if not idata:
log_metadata.info(f"Enqueueing static-link skipped: {suffix}")
logger.info(
f"{log_metadata} Enqueueing static-link skipped: {suffix}"
)
continue
log_metadata.info(f"Enqueueing static-link execution {suffix}")
logger.info(f"{log_metadata} Enqueueing static-link execution {suffix}")
enqueued_executions.append(
add_enqueued_execution(iexec.node_exec_id, next_node_id, idata)
)
@@ -454,18 +443,22 @@ class Executor:
def on_node_execution(
cls, q: ExecutionQueue[NodeExecution], node_exec: NodeExecution
):
log_metadata = LogMetadata(
user_id=node_exec.user_id,
log_metadata = get_log_metadata(
graph_eid=node_exec.graph_exec_id,
graph_id=node_exec.graph_id,
node_eid=node_exec.node_exec_id,
node_id=node_exec.node_id,
block_name="-",
)
prefix = get_log_prefix(
graph_eid=node_exec.graph_exec_id,
node_eid=node_exec.node_exec_id,
block_name="-",
)
execution_stats = {}
timing_info, _ = cls._on_node_execution(
q, node_exec, log_metadata, execution_stats
q, node_exec, log_metadata, prefix, execution_stats
)
execution_stats["walltime"] = timing_info.wall_time
execution_stats["cputime"] = timing_info.cpu_time
@@ -480,19 +473,29 @@ class Executor:
cls,
q: ExecutionQueue[NodeExecution],
node_exec: NodeExecution,
log_metadata: LogMetadata,
log_metadata: dict,
prefix: str,
stats: dict[str, Any] | None = None,
):
try:
log_metadata.info(f"Start node execution {node_exec.node_exec_id}")
logger.info(
f"{prefix} Start node execution {node_exec.node_exec_id}",
extra={"json_fields": {**log_metadata}},
)
for execution in execute_node(
cls.loop, cls.agent_server_client, node_exec, stats
):
q.add(execution)
log_metadata.info(f"Finished node execution {node_exec.node_exec_id}")
logger.info(
f"{prefix} Finished node execution {node_exec.node_exec_id}",
extra={"json_fields": {**log_metadata}},
)
except Exception as e:
log_metadata.exception(
f"Failed node execution {node_exec.node_exec_id}: {e}"
logger.exception(
f"Failed node execution {node_exec.node_exec_id}: {e}",
extra={
**log_metadata,
},
)
@classmethod
@@ -514,12 +517,10 @@ class Executor:
@classmethod
def on_graph_executor_stop(cls):
prefix = f"[on_graph_executor_stop {cls.pid}]"
logger.info(f"{prefix}Disconnecting DB...")
cls.loop.run_until_complete(db.disconnect())
logger.info(f"{prefix} ⏳ Terminating node executor pool...")
logger.info(
f"[on_graph_executor_stop {cls.pid}]Terminating node executor pool..."
)
cls.executor.terminate()
logger.info(f"{prefix} ✅ Finished cleanup")
@classmethod
def _init_node_executor_pool(cls):
@@ -531,16 +532,20 @@ class Executor:
@classmethod
@error_logged
def on_graph_execution(cls, graph_exec: GraphExecution, cancel: threading.Event):
log_metadata = LogMetadata(
user_id=graph_exec.user_id,
log_metadata = get_log_metadata(
graph_eid=graph_exec.graph_exec_id,
graph_id=graph_exec.graph_id,
node_id="*",
node_eid="*",
block_name="-",
)
prefix = get_log_prefix(
graph_eid=graph_exec.graph_exec_id,
node_eid="*",
block_name="-",
)
timing_info, node_count = cls._on_graph_execution(
graph_exec, cancel, log_metadata
graph_exec, cancel, log_metadata, prefix
)
cls.loop.run_until_complete(
@@ -560,9 +565,13 @@ class Executor:
cls,
graph_exec: GraphExecution,
cancel: threading.Event,
log_metadata: LogMetadata,
log_metadata: dict,
prefix: str,
) -> int:
log_metadata.info(f"Start graph execution {graph_exec.graph_exec_id}")
logger.info(
f"{prefix} Start graph execution {graph_exec.graph_exec_id}",
extra={"json_fields": {**log_metadata}},
)
n_node_executions = 0
finished = False
@@ -572,7 +581,10 @@ class Executor:
if finished:
return
cls.executor.terminate()
log_metadata.info(f"Terminated graph execution {graph_exec.graph_exec_id}")
logger.info(
f"{prefix} Terminated graph execution {graph_exec.graph_exec_id}",
extra={"json_fields": {**log_metadata}},
)
cls._init_node_executor_pool()
cancel_thread = threading.Thread(target=cancel_handler)
@@ -610,9 +622,10 @@ class Executor:
# Re-enqueueing the data back to the queue will disrupt the order.
execution.wait()
log_metadata.debug(
f"Dispatching node execution {exec_data.node_exec_id} "
logger.debug(
f"{prefix} Dispatching node execution {exec_data.node_exec_id} "
f"for node {exec_data.node_id}",
extra={**log_metadata},
)
running_executions[exec_data.node_id] = cls.executor.apply_async(
cls.on_node_execution,
@@ -622,8 +635,10 @@ class Executor:
# Avoid terminating graph execution when some nodes are still running.
while queue.empty() and running_executions:
log_metadata.debug(
f"Queue empty; running nodes: {list(running_executions.keys())}"
logger.debug(
"Queue empty; running nodes: "
f"{list(running_executions.keys())}",
extra={"json_fields": {**log_metadata}},
)
for node_id, execution in list(running_executions.items()):
if cancel.is_set():
@@ -632,13 +647,20 @@ class Executor:
if not queue.empty():
break # yield to parent loop to execute new queue items
log_metadata.debug(f"Waiting on execution of node {node_id}")
logger.debug(
f"Waiting on execution of node {node_id}",
extra={"json_fields": {**log_metadata}},
)
execution.wait(3)
log_metadata.info(f"Finished graph execution {graph_exec.graph_exec_id}")
logger.info(
f"{prefix} Finished graph execution {graph_exec.graph_exec_id}",
extra={"json_fields": {**log_metadata}},
)
except Exception as e:
log_metadata.exception(
f"Failed graph execution {graph_exec.graph_exec_id}: {e}"
logger.exception(
f"{prefix} Failed graph execution {graph_exec.graph_exec_id}: {e}",
extra={"json_fields": {**log_metadata}},
)
finally:
if not cancel.is_set():
@@ -725,7 +747,6 @@ class ExecutionManager(AppService):
for node_exec in node_execs:
starting_node_execs.append(
NodeExecution(
user_id=user_id,
graph_exec_id=node_exec.graph_exec_id,
graph_id=node_exec.graph_id,
node_exec_id=node_exec.node_exec_id,
@@ -741,7 +762,6 @@ class ExecutionManager(AppService):
self.agent_server_client.send_execution_update(exec_update.model_dump())
graph_exec = GraphExecution(
user_id=user_id,
graph_id=graph_id,
graph_exec_id=graph_exec_id,
start_node_execs=starting_node_execs,

View File

@@ -0,0 +1,99 @@
import time
from typing import Optional
from urllib.parse import urlencode
import requests
from autogpt_libs.supabase_integration_credentials_store import OAuth2Credentials
from autogpt_server.integrations.oauth import BaseOAuthHandler
class GitHubOAuthHandler(BaseOAuthHandler):
"""
Based on the documentation at:
- [Authorizing OAuth apps - GitHub Docs](https://docs.github.com/en/apps/oauth-apps/building-oauth-apps/authorizing-oauth-apps)
- [Refreshing user access tokens - GitHub Docs](https://docs.github.com/en/apps/creating-github-apps/authenticating-with-a-github-app/refreshing-user-access-tokens)
Notes:
- By default, token expiration is disabled on GitHub Apps. This means the access
token doesn't expire and no refresh token is returned by the authorization flow.
- When token expiration gets enabled, any existing tokens will remain non-expiring.
- When token expiration gets disabled, token refreshes will return a non-expiring
access token *with no refresh token*.
""" # noqa
PROVIDER_NAME = "github"
def __init__(self, client_id: str, client_secret: str, redirect_uri: str):
self.client_id = client_id
self.client_secret = client_secret
self.redirect_uri = redirect_uri
self.auth_base_url = "https://github.com/login/oauth/authorize"
self.token_url = "https://github.com/login/oauth/access_token"
def get_login_url(self, scopes: list[str], state: str) -> str:
params = {
"client_id": self.client_id,
"redirect_uri": self.redirect_uri,
"scope": " ".join(scopes),
"state": state,
}
return f"{self.auth_base_url}?{urlencode(params)}"
def exchange_code_for_tokens(self, code: str) -> OAuth2Credentials:
return self._request_tokens({"code": code, "redirect_uri": self.redirect_uri})
def _refresh_tokens(self, credentials: OAuth2Credentials) -> OAuth2Credentials:
if not credentials.refresh_token:
return credentials
return self._request_tokens(
{
"refresh_token": credentials.refresh_token.get_secret_value(),
"grant_type": "refresh_token",
}
)
def _request_tokens(
self,
params: dict[str, str],
current_credentials: Optional[OAuth2Credentials] = None,
) -> OAuth2Credentials:
request_body = {
"client_id": self.client_id,
"client_secret": self.client_secret,
**params,
}
headers = {"Accept": "application/json"}
response = requests.post(self.token_url, data=request_body, headers=headers)
response.raise_for_status()
token_data: dict = response.json()
now = int(time.time())
new_credentials = OAuth2Credentials(
provider=self.PROVIDER_NAME,
title=current_credentials.title if current_credentials else "GitHub",
access_token=token_data["access_token"],
# Token refresh responses have an empty `scope` property (see docs),
# so we have to get the scope from the existing credentials object.
scopes=(
token_data.get("scope", "").split(",")
or (current_credentials.scopes if current_credentials else [])
),
# Refresh token and expiration intervals are only given if token expiration
# is enabled in the GitHub App's settings.
refresh_token=token_data.get("refresh_token"),
access_token_expires_at=(
now + expires_in
if (expires_in := token_data.get("expires_in", None))
else None
),
refresh_token_expires_at=(
now + expires_in
if (expires_in := token_data.get("refresh_token_expires_in", None))
else None
),
)
if current_credentials:
new_credentials.id = current_credentials.id
return new_credentials

View File

@@ -0,0 +1,96 @@
from autogpt_libs.supabase_integration_credentials_store import OAuth2Credentials
from google.auth.transport.requests import Request
from google.oauth2.credentials import Credentials
from google_auth_oauthlib.flow import Flow
from pydantic import SecretStr
from .oauth import BaseOAuthHandler
class GoogleOAuthHandler(BaseOAuthHandler):
"""
Based on the documentation at https://developers.google.com/identity/protocols/oauth2/web-server
""" # noqa
PROVIDER_NAME = "google"
def __init__(self, client_id: str, client_secret: str, redirect_uri: str):
self.client_id = client_id
self.client_secret = client_secret
self.redirect_uri = redirect_uri
self.token_uri = "https://oauth2.googleapis.com/token"
def get_login_url(self, scopes: list[str], state: str) -> str:
flow = self._setup_oauth_flow(scopes)
flow.redirect_uri = self.redirect_uri
authorization_url, _ = flow.authorization_url(
access_type="offline",
include_granted_scopes="true",
state=state,
prompt="consent",
)
return authorization_url
def exchange_code_for_tokens(self, code: str) -> OAuth2Credentials:
flow = self._setup_oauth_flow(None)
flow.redirect_uri = self.redirect_uri
flow.fetch_token(code=code)
google_creds = flow.credentials
# Google's OAuth library is poorly typed so we need some of these:
assert google_creds.token
assert google_creds.refresh_token
assert google_creds.expiry
assert google_creds.scopes
return OAuth2Credentials(
provider=self.PROVIDER_NAME,
title="Google",
access_token=SecretStr(google_creds.token),
refresh_token=SecretStr(google_creds.refresh_token),
access_token_expires_at=int(google_creds.expiry.timestamp()),
refresh_token_expires_at=None,
scopes=google_creds.scopes,
)
def _refresh_tokens(self, credentials: OAuth2Credentials) -> OAuth2Credentials:
# Google credentials should ALWAYS have a refresh token
assert credentials.refresh_token
google_creds = Credentials(
token=credentials.access_token.get_secret_value(),
refresh_token=credentials.refresh_token.get_secret_value(),
token_uri=self.token_uri,
client_id=self.client_id,
client_secret=self.client_secret,
scopes=credentials.scopes,
)
# Google's OAuth library is poorly typed so we need some of these:
assert google_creds.refresh_token
assert google_creds.scopes
google_creds.refresh(Request())
assert google_creds.expiry
return OAuth2Credentials(
id=credentials.id,
provider=self.PROVIDER_NAME,
title=credentials.title,
access_token=SecretStr(google_creds.token),
refresh_token=SecretStr(google_creds.refresh_token),
access_token_expires_at=int(google_creds.expiry.timestamp()),
refresh_token_expires_at=None,
scopes=google_creds.scopes,
)
def _setup_oauth_flow(self, scopes: list[str] | None) -> Flow:
return Flow.from_client_config(
{
"web": {
"client_id": self.client_id,
"client_secret": self.client_secret,
"auth_uri": "https://accounts.google.com/o/oauth2/auth",
"token_uri": self.token_uri,
}
},
scopes=scopes,
)

View File

@@ -0,0 +1,76 @@
from base64 import b64encode
from urllib.parse import urlencode
import requests
from autogpt_libs.supabase_integration_credentials_store import OAuth2Credentials
from autogpt_server.integrations.oauth import BaseOAuthHandler
class NotionOAuthHandler(BaseOAuthHandler):
"""
Based on the documentation at https://developers.notion.com/docs/authorization
Notes:
- Notion uses non-expiring access tokens and therefore doesn't have a refresh flow
- Notion doesn't use scopes
"""
PROVIDER_NAME = "notion"
def __init__(self, client_id: str, client_secret: str, redirect_uri: str):
self.client_id = client_id
self.client_secret = client_secret
self.redirect_uri = redirect_uri
self.auth_base_url = "https://api.notion.com/v1/oauth/authorize"
self.token_url = "https://api.notion.com/v1/oauth/token"
def get_login_url(self, scopes: list[str], state: str) -> str:
params = {
"client_id": self.client_id,
"redirect_uri": self.redirect_uri,
"response_type": "code",
"owner": "user",
"state": state,
}
return f"{self.auth_base_url}?{urlencode(params)}"
def exchange_code_for_tokens(self, code: str) -> OAuth2Credentials:
request_body = {
"grant_type": "authorization_code",
"code": code,
"redirect_uri": self.redirect_uri,
}
auth_str = b64encode(f"{self.client_id}:{self.client_secret}".encode()).decode()
headers = {
"Authorization": f"Basic {auth_str}",
"Accept": "application/json",
}
response = requests.post(self.token_url, json=request_body, headers=headers)
response.raise_for_status()
token_data = response.json()
return OAuth2Credentials(
provider=self.PROVIDER_NAME,
title=token_data.get("workspace_name", "Notion"),
access_token=token_data["access_token"],
refresh_token=None,
access_token_expires_at=None, # Notion tokens don't expire
refresh_token_expires_at=None,
scopes=[],
metadata={
"owner": token_data["owner"],
"bot_id": token_data["bot_id"],
"workspace_id": token_data["workspace_id"],
"workspace_name": token_data.get("workspace_name"),
"workspace_icon": token_data.get("workspace_icon"),
},
)
def _refresh_tokens(self, credentials: OAuth2Credentials) -> OAuth2Credentials:
# Notion doesn't support token refresh
return credentials
def needs_refresh(self, credentials: OAuth2Credentials) -> bool:
# Notion access tokens don't expire
return False

View File

@@ -0,0 +1,48 @@
import time
from abc import ABC, abstractmethod
from typing import ClassVar
from autogpt_libs.supabase_integration_credentials_store import OAuth2Credentials
class BaseOAuthHandler(ABC):
PROVIDER_NAME: ClassVar[str]
@abstractmethod
def __init__(self, client_id: str, client_secret: str, redirect_uri: str): ...
@abstractmethod
def get_login_url(self, scopes: list[str], state: str) -> str:
"""Constructs a login URL that the user can be redirected to"""
...
@abstractmethod
def exchange_code_for_tokens(self, code: str) -> OAuth2Credentials:
"""Exchanges the acquired authorization code from login for a set of tokens"""
...
@abstractmethod
def _refresh_tokens(self, credentials: OAuth2Credentials) -> OAuth2Credentials:
"""Implements the token refresh mechanism"""
...
def refresh_tokens(self, credentials: OAuth2Credentials) -> OAuth2Credentials:
if credentials.provider != self.PROVIDER_NAME:
raise ValueError(
f"{self.__class__.__name__} can not refresh tokens "
f"for other provider '{credentials.provider}'"
)
return self._refresh_tokens(credentials)
def get_access_token(self, credentials: OAuth2Credentials) -> str:
"""Returns a valid access token, refreshing it first if needed"""
if self.needs_refresh(credentials):
credentials = self.refresh_tokens(credentials)
return credentials.access_token.get_secret_value()
def needs_refresh(self, credentials: OAuth2Credentials) -> bool:
"""Indicates whether the given tokens need to be refreshed"""
return (
credentials.access_token_expires_at is not None
and credentials.access_token_expires_at < int(time.time()) + 300
)

View File

@@ -23,7 +23,6 @@ class GitHubOAuthHandler(BaseOAuthHandler):
""" # noqa
PROVIDER_NAME = "github"
EMAIL_ENDPOINT = "https://api.github.com/user/emails"
def __init__(self, client_id: str, client_secret: str, redirect_uri: str):
self.client_id = client_id
@@ -70,13 +69,10 @@ class GitHubOAuthHandler(BaseOAuthHandler):
response.raise_for_status()
token_data: dict = response.json()
username = self._request_username(token_data["access_token"])
now = int(time.time())
new_credentials = OAuth2Credentials(
provider=self.PROVIDER_NAME,
title=current_credentials.title if current_credentials else None,
username=username,
title=current_credentials.title if current_credentials else "GitHub",
access_token=token_data["access_token"],
# Token refresh responses have an empty `scope` property (see docs),
# so we have to get the scope from the existing credentials object.
@@ -101,19 +97,3 @@ class GitHubOAuthHandler(BaseOAuthHandler):
if current_credentials:
new_credentials.id = current_credentials.id
return new_credentials
def _request_username(self, access_token: str) -> str | None:
url = "https://api.github.com/user"
headers = {
"Accept": "application/vnd.github+json",
"Authorization": f"Bearer {access_token}",
"X-GitHub-Api-Version": "2022-11-28",
}
response = requests.get(url, headers=headers)
if not response.ok:
return None
# Get the login (username)
return response.json().get("login")

View File

@@ -1,8 +1,5 @@
from autogpt_libs.supabase_integration_credentials_store import OAuth2Credentials
from google.auth.external_account_authorized_user import (
Credentials as ExternalAccountCredentials,
)
from google.auth.transport.requests import AuthorizedSession, Request
from google.auth.transport.requests import Request
from google.oauth2.credentials import Credentials
from google_auth_oauthlib.flow import Flow
from pydantic import SecretStr
@@ -16,7 +13,6 @@ class GoogleOAuthHandler(BaseOAuthHandler):
""" # noqa
PROVIDER_NAME = "google"
EMAIL_ENDPOINT = "https://www.googleapis.com/oauth2/v2/userinfo"
def __init__(self, client_id: str, client_secret: str, redirect_uri: str):
self.client_id = client_id
@@ -41,8 +37,6 @@ class GoogleOAuthHandler(BaseOAuthHandler):
flow.fetch_token(code=code)
google_creds = flow.credentials
username = self._request_email(google_creds)
# Google's OAuth library is poorly typed so we need some of these:
assert google_creds.token
assert google_creds.refresh_token
@@ -50,8 +44,7 @@ class GoogleOAuthHandler(BaseOAuthHandler):
assert google_creds.scopes
return OAuth2Credentials(
provider=self.PROVIDER_NAME,
title=None,
username=username,
title="Google",
access_token=SecretStr(google_creds.token),
refresh_token=SecretStr(google_creds.refresh_token),
access_token_expires_at=int(google_creds.expiry.timestamp()),
@@ -59,15 +52,6 @@ class GoogleOAuthHandler(BaseOAuthHandler):
scopes=google_creds.scopes,
)
def _request_email(
self, creds: Credentials | ExternalAccountCredentials
) -> str | None:
session = AuthorizedSession(creds)
response = session.get(self.EMAIL_ENDPOINT)
if not response.ok:
return None
return response.json()["email"]
def _refresh_tokens(self, credentials: OAuth2Credentials) -> OAuth2Credentials:
# Google credentials should ALWAYS have a refresh token
assert credentials.refresh_token
@@ -88,10 +72,9 @@ class GoogleOAuthHandler(BaseOAuthHandler):
assert google_creds.expiry
return OAuth2Credentials(
provider=self.PROVIDER_NAME,
id=credentials.id,
provider=self.PROVIDER_NAME,
title=credentials.title,
username=credentials.username,
access_token=SecretStr(google_creds.token),
refresh_token=SecretStr(google_creds.refresh_token),
access_token_expires_at=int(google_creds.expiry.timestamp()),

View File

@@ -49,18 +49,10 @@ class NotionOAuthHandler(BaseOAuthHandler):
response = requests.post(self.token_url, json=request_body, headers=headers)
response.raise_for_status()
token_data = response.json()
# Email is only available for non-bot users
email = (
token_data["owner"]["person"]["email"]
if "person" in token_data["owner"]
and "email" in token_data["owner"]["person"]
else None
)
return OAuth2Credentials(
provider=self.PROVIDER_NAME,
title=token_data.get("workspace_name"),
username=email,
title=token_data.get("workspace_name", "Notion"),
access_token=token_data["access_token"],
refresh_token=None,
access_token_expires_at=None, # Notion tokens don't expire

View File

@@ -4,10 +4,6 @@ from typing import Annotated, Literal
from autogpt_libs.supabase_integration_credentials_store import (
SupabaseIntegrationCredentialsStore,
)
from autogpt_libs.supabase_integration_credentials_store.types import (
Credentials,
OAuth2Credentials,
)
from fastapi import APIRouter, Body, Depends, HTTPException, Path, Query, Request
from pydantic import BaseModel
from supabase import Client
@@ -52,11 +48,8 @@ async def login(
class CredentialsMetaResponse(BaseModel):
id: str
type: Literal["oauth2", "api_key"]
title: str | None
scopes: list[str] | None
username: str | None
credentials_id: str
credentials_type: Literal["oauth2", "api_key"]
@integrations_api_router.post("/{provider}/callback")
@@ -80,53 +73,13 @@ async def callback(
logger.warning(f"Code->Token exchange failed for provider {provider}: {e}")
raise HTTPException(status_code=400, detail=str(e))
# TODO: Allow specifying `title` to set on `credentials`
store.add_creds(user_id, credentials)
return CredentialsMetaResponse(
id=credentials.id,
type=credentials.type,
title=credentials.title,
scopes=credentials.scopes,
username=credentials.username,
credentials_id=credentials.id,
credentials_type=credentials.type,
)
@integrations_api_router.get("/{provider}/credentials")
async def list_credentials(
provider: Annotated[str, Path(title="The provider to list credentials for")],
user_id: Annotated[str, Depends(get_user_id)],
store: Annotated[SupabaseIntegrationCredentialsStore, Depends(get_store)],
) -> list[CredentialsMetaResponse]:
credentials = store.get_creds_by_provider(user_id, provider)
return [
CredentialsMetaResponse(
id=cred.id,
type=cred.type,
title=cred.title,
scopes=cred.scopes if isinstance(cred, OAuth2Credentials) else None,
username=cred.username if isinstance(cred, OAuth2Credentials) else None,
)
for cred in credentials
]
@integrations_api_router.get("/{provider}/credentials/{cred_id}")
async def get_credential(
provider: Annotated[str, Path(title="The provider to retrieve credentials for")],
cred_id: Annotated[str, Path(title="The ID of the credentials to retrieve")],
user_id: Annotated[str, Depends(get_user_id)],
store: Annotated[SupabaseIntegrationCredentialsStore, Depends(get_store)],
) -> Credentials:
credential = store.get_creds_by_id(user_id, cred_id)
if not credential:
raise HTTPException(status_code=404, detail="Credentials not found")
if credential.provider != provider:
raise HTTPException(
status_code=404, detail="Credentials do not match the specified provider"
)
return credential
# -------- UTILITIES --------- #

View File

@@ -15,7 +15,6 @@ from autogpt_server.data import execution as execution_db
from autogpt_server.data import graph as graph_db
from autogpt_server.data import user as user_db
from autogpt_server.data.block import BlockInput, CompletedBlockOutput
from autogpt_server.data.credit import get_block_costs, get_user_credit_model
from autogpt_server.data.queue import AsyncEventQueue, AsyncRedisEventQueue
from autogpt_server.data.user import get_or_create_user
from autogpt_server.executor import ExecutionManager, ExecutionScheduler
@@ -33,7 +32,6 @@ class AgentServer(AppService):
mutex = KeyedMutex()
use_redis = True
_test_dependency_overrides = {}
_user_credit_model = get_user_credit_model()
def __init__(self, event_queue: AsyncEventQueue | None = None):
super().__init__(port=Config().agent_server_port)
@@ -93,11 +91,6 @@ class AgentServer(AppService):
endpoint=self.get_graph_blocks,
methods=["GET"],
)
api_router.add_api_route(
path="/blocks/costs",
endpoint=self.get_graph_block_costs,
methods=["GET"],
)
api_router.add_api_route(
path="/blocks/{block_id}/execute",
endpoint=self.execute_graph_block,
@@ -203,11 +196,6 @@ class AgentServer(AppService):
endpoint=self.update_schedule,
methods=["PUT"],
)
api_router.add_api_route(
path="/credits",
endpoint=self.get_user_credits,
methods=["GET"],
)
api_router.add_api_route(
path="/settings",
@@ -277,10 +265,6 @@ class AgentServer(AppService):
def get_graph_blocks(cls) -> list[dict[Any, Any]]:
return [v.to_dict() for v in block.get_blocks().values()]
@classmethod
def get_graph_block_costs(cls) -> dict[Any, Any]:
return get_block_costs()
@classmethod
def execute_graph_block(
cls, block_id: str, data: BlockInput
@@ -497,25 +481,6 @@ class AgentServer(AppService):
return await execution_db.list_executions(graph_id, graph_version)
@classmethod
async def get_graph_run_status(
cls,
graph_id: str,
graph_exec_id: str,
user_id: Annotated[str, Depends(get_user_id)],
) -> execution_db.ExecutionStatus:
graph = await graph_db.get_graph(graph_id, user_id=user_id)
if not graph:
raise HTTPException(status_code=404, detail=f"Graph #{graph_id} not found.")
execution = await execution_db.get_graph_execution(graph_exec_id, user_id)
if not execution:
raise HTTPException(
status_code=404, detail=f"Execution #{graph_exec_id} not found."
)
return execution.executionStatus
@classmethod
async def get_graph_run_node_execution_results(
cls,
@@ -557,11 +522,6 @@ class AgentServer(AppService):
execution_scheduler.update_schedule(schedule_id, is_enabled, user_id=user_id)
return {"id": schedule_id}
async def get_user_credits(
self, user_id: Annotated[str, Depends(get_user_id)]
) -> dict[str, int]:
return {"credits": await self._user_credit_model.get_or_refill_credit(user_id)}
def get_execution_schedules(
self, graph_id: str, user_id: Annotated[str, Depends(get_user_id)]
) -> dict[str, str]:

View File

@@ -252,6 +252,7 @@ Here are a couple of sample of the Block class implementation:
async def block_autogen_agent():
async with SpinTestServer() as server:
test_manager = server.exec_manager
test_user = await create_test_user()
test_graph = await create_graph(create_test_graph(), user_id=test_user.id)
input_data = {"input": "Write me a block that writes a string into a file."}
@@ -260,8 +261,10 @@ async def block_autogen_agent():
)
print(response)
result = await wait_execution(
exec_manager=test_manager,
graph_id=test_graph.id,
graph_exec_id=response["id"],
num_execs=10,
timeout=1200,
user_id=test_user.id,
)

View File

@@ -153,6 +153,7 @@ async def create_test_user() -> User:
async def reddit_marketing_agent():
async with SpinTestServer() as server:
exec_man = server.exec_manager
test_user = await create_test_user()
test_graph = await create_graph(create_test_graph(), user_id=test_user.id)
input_data = {"subreddit": "AutoGPT"}
@@ -160,7 +161,9 @@ async def reddit_marketing_agent():
test_graph.id, input_data, test_user.id
)
print(response)
result = await wait_execution(test_user.id, test_graph.id, response["id"], 120)
result = await wait_execution(
exec_man, test_user.id, test_graph.id, response["id"], 13, 120
)
print(result)

View File

@@ -75,6 +75,7 @@ def create_test_graph() -> graph.Graph:
async def sample_agent():
async with SpinTestServer() as server:
exec_man = server.exec_manager
test_user = await create_test_user()
test_graph = await create_graph(create_test_graph(), test_user.id)
input_data = {"input_1": "Hello", "input_2": "World"}
@@ -82,7 +83,9 @@ async def sample_agent():
test_graph.id, input_data, test_user.id
)
print(response)
result = await wait_execution(test_user.id, test_graph.id, response["id"], 10)
result = await wait_execution(
exec_man, test_user.id, test_graph.id, response["id"], 4, 10
)
print(result)

View File

@@ -42,15 +42,15 @@ class Config(UpdateTrackingModel["Config"], BaseSettings):
"""Config for the server."""
num_graph_workers: int = Field(
default=10,
default=1,
ge=1,
le=1000,
le=100,
description="Maximum number of workers to use for graph execution.",
)
num_node_workers: int = Field(
default=5,
default=1,
ge=1,
le=1000,
le=100,
description="Maximum number of workers to use for node execution within a single graph.",
)
pyro_host: str = Field(
@@ -61,14 +61,6 @@ class Config(UpdateTrackingModel["Config"], BaseSettings):
default="false",
description="If authentication is enabled or not",
)
enable_credit: str = Field(
default="false",
description="If user credit system is enabled or not",
)
num_user_credits_refill: int = Field(
default=1500,
description="Number of credits to refill for each user",
)
# Add more configuration fields as needed
model_config = SettingsConfigDict(

View File

@@ -5,7 +5,6 @@ from autogpt_server.data import db
from autogpt_server.data.block import Block, initialize_blocks
from autogpt_server.data.execution import ExecutionResult, ExecutionStatus
from autogpt_server.data.queue import AsyncEventQueue
from autogpt_server.data.user import create_default_user
from autogpt_server.executor import ExecutionManager, ExecutionScheduler
from autogpt_server.server import AgentServer
from autogpt_server.server.rest_api import get_user_id
@@ -65,7 +64,6 @@ class SpinTestServer:
await db.connect()
await initialize_blocks()
await create_default_user("false")
return self
@@ -84,18 +82,25 @@ class SpinTestServer:
async def wait_execution(
exec_manager: ExecutionManager,
user_id: str,
graph_id: str,
graph_exec_id: str,
num_execs: int,
timeout: int = 20,
) -> list:
async def is_execution_completed():
status = await AgentServer().get_graph_run_status(
execs = await AgentServer().get_graph_run_node_execution_results(
graph_id, graph_exec_id, user_id
)
if status == ExecutionStatus.FAILED:
raise Exception("Execution failed")
return status == ExecutionStatus.COMPLETED
return (
exec_manager.queue.empty()
and len(execs) == num_execs
and all(
v.status in [ExecutionStatus.COMPLETED, ExecutionStatus.FAILED]
for v in execs
)
)
# Wait for the executions to complete
for i in range(timeout):

View File

@@ -1,5 +1,4 @@
{
"num_graph_workers": 10,
"num_node_workers": 5,
"num_user_credits_refill": 1500
"num_node_workers": 5
}

View File

@@ -1,39 +0,0 @@
/*
Warnings:
- The `executionStatus` column on the `AgentNodeExecution` table would be dropped and recreated. This will lead to data loss if there is data in the column.
*/
-- CreateEnum
CREATE TYPE "AgentExecutionStatus" AS ENUM ('INCOMPLETE', 'QUEUED', 'RUNNING', 'COMPLETED', 'FAILED');
-- CreateEnum
CREATE TYPE "UserBlockCreditType" AS ENUM ('TOP_UP', 'USAGE');
-- AlterTable
ALTER TABLE "AgentGraphExecution" ADD COLUMN "executionStatus" "AgentExecutionStatus" NOT NULL DEFAULT 'COMPLETED',
ADD COLUMN "startedAt" TIMESTAMP(3);
-- AlterTable
ALTER TABLE "AgentNodeExecution" DROP COLUMN "executionStatus",
ADD COLUMN "executionStatus" "AgentExecutionStatus" NOT NULL DEFAULT 'COMPLETED';
-- CreateTable
CREATE TABLE "UserBlockCredit" (
"transactionKey" TEXT NOT NULL,
"createdAt" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
"userId" TEXT NOT NULL,
"blockId" TEXT,
"amount" INTEGER NOT NULL,
"type" "UserBlockCreditType" NOT NULL,
"isActive" BOOLEAN NOT NULL DEFAULT true,
"metadata" JSONB,
CONSTRAINT "UserBlockCredit_pkey" PRIMARY KEY ("transactionKey","userId")
);
-- AddForeignKey
ALTER TABLE "UserBlockCredit" ADD CONSTRAINT "UserBlockCredit_userId_fkey" FOREIGN KEY ("userId") REFERENCES "User"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
-- AddForeignKey
ALTER TABLE "UserBlockCredit" ADD CONSTRAINT "UserBlockCredit_blockId_fkey" FOREIGN KEY ("blockId") REFERENCES "AgentBlock"("id") ON DELETE SET NULL ON UPDATE CASCADE;

View File

@@ -289,7 +289,7 @@ description = "Shared libraries across NextGen AutoGPT"
optional = false
python-versions = ">=3.10,<4.0"
files = []
develop = true
develop = false
[package.dependencies]
colorama = "^0.4.6"
@@ -2022,7 +2022,6 @@ description = "Pure-Python implementation of ASN.1 types and DER/BER/CER codecs
optional = false
python-versions = ">=3.8"
files = [
{file = "pyasn1-0.6.1-py3-none-any.whl", hash = "sha256:0d632f46f2ba09143da3a8afe9e33fb6f92fa2320ab7e886e2d0f7672af84629"},
{file = "pyasn1-0.6.1.tar.gz", hash = "sha256:6f580d2bdd84365380830acf45550f2511469f673cb4a5ae3857a3170128b034"},
]
@@ -2033,7 +2032,6 @@ description = "A collection of ASN.1-based protocols modules"
optional = false
python-versions = ">=3.8"
files = [
{file = "pyasn1_modules-0.4.1-py3-none-any.whl", hash = "sha256:49bfa96b45a292b711e986f222502c1c9a5e1f4e568fc30e2574a6c7d07838fd"},
{file = "pyasn1_modules-0.4.1.tar.gz", hash = "sha256:c28e2dbf9c06ad61c71a075c7e0f9fd0f1b0bb2d2ad4377f240d33ac2ab60a7c"},
]
@@ -3623,4 +3621,4 @@ type = ["pytest-mypy"]
[metadata]
lock-version = "2.0"
python-versions = "^3.10"
content-hash = "311c527a1d1947af049dac27c7a2b2f49d7fa4cdede52ef436422a528b0ad866"
content-hash = "fbc928c40dc95041f7750ab34677fa3eebacd06a84944de900dedd639f847a9c"

View File

@@ -13,7 +13,7 @@ python = "^3.10"
aio-pika = "^9.4.3"
anthropic = "^0.25.1"
apscheduler = "^3.10.4"
autogpt-libs = { path = "../autogpt_libs", develop = true }
autogpt-libs = { path = "../autogpt_libs" }
click = "^8.1.7"
croniter = "^2.0.5"
discord-py = "^2.4.0"

View File

@@ -22,7 +22,6 @@ model User {
AgentGraphs AgentGraph[]
AgentGraphExecutions AgentGraphExecution[]
AgentGraphExecutionSchedules AgentGraphExecutionSchedule[]
UserBlockCredit UserBlockCredit[]
@@index([id])
@@index([email])
@@ -30,9 +29,9 @@ model User {
// This model describes the Agent Graph/Flow (Multi Agent System).
model AgentGraph {
id String @default(uuid())
version Int @default(1)
createdAt DateTime @default(now())
id String @default(uuid())
version Int @default(1)
createdAt DateTime @default(now())
updatedAt DateTime? @updatedAt
name String?
@@ -112,26 +111,13 @@ model AgentBlock {
// Prisma requires explicit back-references.
ReferencedByAgentNode AgentNode[]
UserBlockCredit UserBlockCredit[]
}
// This model describes the status of an AgentGraphExecution or AgentNodeExecution.
enum AgentExecutionStatus {
INCOMPLETE
QUEUED
RUNNING
COMPLETED
FAILED
}
// This model describes the execution of an AgentGraph.
model AgentGraphExecution {
id String @id @default(uuid())
createdAt DateTime @default(now())
id String @id @default(uuid())
createdAt DateTime @default(now())
updatedAt DateTime? @updatedAt
startedAt DateTime?
executionStatus AgentExecutionStatus @default(COMPLETED)
agentGraphId String
agentGraphVersion Int @default(1)
@@ -159,10 +145,12 @@ model AgentNodeExecution {
Input AgentNodeExecutionInputOutput[] @relation("AgentNodeExecutionInput")
Output AgentNodeExecutionInputOutput[] @relation("AgentNodeExecutionOutput")
executionStatus AgentExecutionStatus @default(COMPLETED)
// sqlite does not support enum
// enum Status { INCOMPLETE, QUEUED, RUNNING, SUCCESS, FAILED }
executionStatus String
// Final JSON serialized input data for the node execution.
executionData String?
addedTime DateTime @default(now())
addedTime DateTime @default(now())
queuedTime DateTime?
startedTime DateTime?
endedTime DateTime?
@@ -190,8 +178,8 @@ model AgentNodeExecutionInputOutput {
// This model describes the recurring execution schedule of an Agent.
model AgentGraphExecutionSchedule {
id String @id
createdAt DateTime @default(now())
id String @id
createdAt DateTime @default(now())
updatedAt DateTime? @updatedAt
agentGraphId String
@@ -211,27 +199,3 @@ model AgentGraphExecutionSchedule {
@@index([isEnabled])
}
enum UserBlockCreditType {
TOP_UP
USAGE
}
model UserBlockCredit {
transactionKey String @default(uuid())
createdAt DateTime @default(now())
userId String
user User @relation(fields: [userId], references: [id])
blockId String?
block AgentBlock? @relation(fields: [blockId], references: [id])
amount Int
type UserBlockCreditType
isActive Boolean @default(true)
metadata Json?
@@id(name: "creditTransactionIdentifier", [transactionKey, userId])
}

View File

@@ -1,90 +0,0 @@
from datetime import datetime
import pytest
from prisma.models import UserBlockCredit
from autogpt_server.blocks.llm import AITextGeneratorBlock
from autogpt_server.data.credit import UserCredit
from autogpt_server.data.user import DEFAULT_USER_ID
from autogpt_server.util.test import SpinTestServer
REFILL_VALUE = 1000
user_credit = UserCredit(REFILL_VALUE)
@pytest.mark.asyncio(scope="session")
async def test_block_credit_usage(server: SpinTestServer):
current_credit = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
spending_amount_1 = await user_credit.spend_credits(
DEFAULT_USER_ID,
current_credit,
AITextGeneratorBlock(),
{"model": "gpt-4-turbo"},
0.0,
0.0,
validate_balance=False,
)
assert spending_amount_1 > 0
spending_amount_2 = await user_credit.spend_credits(
DEFAULT_USER_ID,
current_credit,
AITextGeneratorBlock(),
{"model": "gpt-4-turbo", "api_key": "owned_api_key"},
0.0,
0.0,
validate_balance=False,
)
assert spending_amount_2 == 0
new_credit = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
assert new_credit == current_credit - spending_amount_1 - spending_amount_2
@pytest.mark.asyncio(scope="session")
async def test_block_credit_top_up(server: SpinTestServer):
current_credit = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
await user_credit.top_up_credits(DEFAULT_USER_ID, 100)
new_credit = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
assert new_credit == current_credit + 100
@pytest.mark.asyncio(scope="session")
async def test_block_credit_reset(server: SpinTestServer):
month1 = datetime(2022, 1, 15)
month2 = datetime(2022, 2, 15)
user_credit.time_now = lambda: month2
month2credit = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
# Month 1 result should only affect month 1
user_credit.time_now = lambda: month1
month1credit = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
await user_credit.top_up_credits(DEFAULT_USER_ID, 100)
assert await user_credit.get_or_refill_credit(DEFAULT_USER_ID) == month1credit + 100
# Month 2 balance is unaffected
user_credit.time_now = lambda: month2
assert await user_credit.get_or_refill_credit(DEFAULT_USER_ID) == month2credit
@pytest.mark.asyncio(scope="session")
async def test_credit_refill(server: SpinTestServer):
# Clear all transactions within the month
await UserBlockCredit.prisma().update_many(
where={
"userId": DEFAULT_USER_ID,
"createdAt": {
"gte": datetime(2022, 2, 1),
"lt": datetime(2022, 3, 1),
},
},
data={"isActive": False},
)
user_credit.time_now = lambda: datetime(2022, 2, 15)
balance = await user_credit.get_or_refill_credit(DEFAULT_USER_ID)
assert balance == REFILL_VALUE

View File

@@ -4,7 +4,7 @@ import pytest
from autogpt_server.blocks.basic import AgentInputBlock, StoreValueBlock
from autogpt_server.data.graph import Graph, Link, Node
from autogpt_server.data.user import DEFAULT_USER_ID
from autogpt_server.data.user import DEFAULT_USER_ID, create_default_user
from autogpt_server.server.model import CreateGraph
from autogpt_server.util.test import SpinTestServer
@@ -22,6 +22,8 @@ async def test_graph_creation(server: SpinTestServer):
Args:
server (SpinTestServer): The test server instance.
"""
await create_default_user("false")
value_block = StoreValueBlock().id
input_block = AgentInputBlock().id

View File

@@ -4,6 +4,7 @@ from prisma.models import User
from autogpt_server.blocks.basic import FindInDictionaryBlock, StoreValueBlock
from autogpt_server.blocks.maths import CalculatorBlock, Operation
from autogpt_server.data import execution, graph
from autogpt_server.executor import ExecutionManager
from autogpt_server.server import AgentServer
from autogpt_server.usecases.sample import create_test_graph, create_test_user
from autogpt_server.util.test import SpinTestServer, wait_execution
@@ -11,6 +12,7 @@ from autogpt_server.util.test import SpinTestServer, wait_execution
async def execute_graph(
agent_server: AgentServer,
test_manager: ExecutionManager,
test_graph: graph.Graph,
test_user: User,
input_data: dict,
@@ -21,8 +23,9 @@ async def execute_graph(
graph_exec_id = response["id"]
# Execution queue should be empty
result = await wait_execution(test_user.id, test_graph.id, graph_exec_id)
assert result and len(result) == num_execs
assert await wait_execution(
test_manager, test_user.id, test_graph.id, graph_exec_id, num_execs
)
return graph_exec_id
@@ -105,6 +108,7 @@ async def test_agent_execution(server: SpinTestServer):
data = {"input_1": "Hello", "input_2": "World"}
graph_exec_id = await execute_graph(
server.agent_server,
server.exec_manager,
test_graph,
test_user,
data,
@@ -165,7 +169,7 @@ async def test_input_pin_always_waited(server: SpinTestServer):
test_user = await create_test_user()
test_graph = await graph.create_graph(test_graph, user_id=test_user.id)
graph_exec_id = await execute_graph(
server.agent_server, test_graph, test_user, {}, 3
server.agent_server, server.exec_manager, test_graph, test_user, {}, 3
)
executions = await server.agent_server.get_graph_run_node_execution_results(
@@ -246,7 +250,7 @@ async def test_static_input_link_on_graph(server: SpinTestServer):
test_user = await create_test_user()
test_graph = await graph.create_graph(test_graph, user_id=test_user.id)
graph_exec_id = await execute_graph(
server.agent_server, test_graph, test_user, {}, 8
server.agent_server, server.exec_manager, test_graph, test_user, {}, 8
)
executions = await server.agent_server.get_graph_run_node_execution_results(
test_graph.id, graph_exec_id, test_user.id

View File

@@ -163,7 +163,7 @@ services:
environment:
- DATABASE_URL=postgresql://agpt_user:pass123@postgres:5432/agpt_local?connect_timeout=60&schema=market
ports:
- "8015:8015"
- "8015:8000"
networks:
- app-network

View File

@@ -17,10 +17,6 @@ ENV POETRY_VERSION=1.8.3 \
POETRY_NO_INTERACTION=1 \
POETRY_VIRTUALENVS_CREATE=false \
PATH="$POETRY_HOME/bin:$PATH"
# Upgrade pip and setuptools to fix security vulnerabilities
RUN pip3 install --upgrade pip setuptools
RUN pip3 install poetry
# Copy and install dependencies
@@ -39,9 +35,6 @@ FROM python:3.11-slim-buster AS server_dependencies
WORKDIR /app
# Upgrade pip and setuptools to fix security vulnerabilities
RUN pip3 install --upgrade pip setuptools
# Copy only necessary files from builder
COPY --from=builder /app /app
COPY --from=builder /usr/local/lib/python3.11 /usr/local/lib/python3.11
@@ -65,4 +58,4 @@ FROM server_dependencies AS server
ENV DATABASE_URL=""
ENV PORT=8015
CMD ["poetry", "run", "app"]
CMD ["uvicorn", "market.app:app", "--reload"]

View File

@@ -353,9 +353,7 @@ async def search_db(
async def get_top_agents_by_downloads(
page: int = 1,
page_size: int = 10,
submission_status: prisma.enums.SubmissionStatus = prisma.enums.SubmissionStatus.APPROVED,
page: int = 1, page_size: int = 10
) -> TopAgentsDBResponse:
"""Retrieve the top agents by download count.
@@ -376,7 +374,6 @@ async def get_top_agents_by_downloads(
analytics = await prisma.models.AnalyticsTracker.prisma().find_many(
include={"agent": True},
order={"downloads": "desc"},
where={"agent": {"is": {"submissionStatus": submission_status}}},
skip=skip,
take=page_size,
)
@@ -444,10 +441,7 @@ async def set_agent_featured(
async def get_featured_agents(
category: str = "featured",
page: int = 1,
page_size: int = 10,
submission_status: prisma.enums.SubmissionStatus = prisma.enums.SubmissionStatus.APPROVED,
category: str = "featured", page: int = 1, page_size: int = 10
) -> FeaturedAgentResponse:
"""Retrieve a list of featured agents from the database based on the provided category.
@@ -469,7 +463,6 @@ async def get_featured_agents(
where={
"featuredCategories": {"has": category},
"isActive": True,
"agent": {"is": {"submissionStatus": submission_status}},
},
include={"agent": {"include": {"AnalyticsTracker": True}}},
skip=skip,

View File

@@ -5,7 +5,6 @@ import typing
import fastapi
import fastapi.responses
import prisma
import prisma.enums
import market.db
import market.model
@@ -39,10 +38,6 @@ async def list_agents(
sort_order: typing.Literal["asc", "desc"] = fastapi.Query(
"desc", description="Sort order (asc or desc)"
),
submission_status: prisma.enums.SubmissionStatus = fastapi.Query(
default=prisma.enums.SubmissionStatus.APPROVED,
description="Filter by submission status",
),
):
"""
Retrieve a list of agents based on the provided filters.
@@ -57,7 +52,6 @@ async def list_agents(
description_threshold (int): Fuzzy search threshold (default: 60, min: 0, max: 100).
sort_by (str): Field to sort by (default: "createdAt").
sort_order (str): Sort order (asc or desc) (default: "desc").
submission_status (str): Filter by submission status (default: "APPROVED").
Returns:
market.model.AgentListResponse: A response containing the list of agents and pagination information.
@@ -76,7 +70,6 @@ async def list_agents(
description_threshold=description_threshold,
sort_by=sort_by,
sort_order=sort_order,
submission_status=submission_status,
)
agents = [
@@ -217,10 +210,6 @@ async def top_agents_by_downloads(
page_size: int = fastapi.Query(
10, ge=1, le=100, description="Number of items per page"
),
submission_status: prisma.enums.SubmissionStatus = fastapi.Query(
default=prisma.enums.SubmissionStatus.APPROVED,
description="Filter by submission status",
),
):
"""
Retrieve a list of top agents based on the number of downloads.
@@ -228,7 +217,6 @@ async def top_agents_by_downloads(
Args:
page (int): Page number (default: 1).
page_size (int): Number of items per page (default: 10, min: 1, max: 100).
submission_status (str): Filter by submission status (default: "APPROVED").
Returns:
market.model.AgentListResponse: A response containing the list of top agents and pagination information.
@@ -240,7 +228,6 @@ async def top_agents_by_downloads(
result = await market.db.get_top_agents_by_downloads(
page=page,
page_size=page_size,
submission_status=submission_status,
)
ret = market.model.AgentListResponse(
@@ -287,10 +274,6 @@ async def get_featured_agents(
page_size: int = fastapi.Query(
10, ge=1, le=100, description="Number of items per page"
),
submission_status: prisma.enums.SubmissionStatus = fastapi.Query(
default=prisma.enums.SubmissionStatus.APPROVED,
description="Filter by submission status",
),
):
"""
Retrieve a list of featured agents based on the provided category.
@@ -299,7 +282,6 @@ async def get_featured_agents(
category (str): Category of featured agents (default: "featured").
page (int): Page number (default: 1).
page_size (int): Number of items per page (default: 10, min: 1, max: 100).
submission_status (str): Filter by submission status (default: "APPROVED").
Returns:
market.model.AgentListResponse: A response containing the list of featured agents and pagination information.
@@ -312,7 +294,6 @@ async def get_featured_agents(
category=category,
page=page,
page_size=page_size,
submission_status=submission_status,
)
ret = market.model.AgentListResponse(

48
rnd/rest-api-go/.gitignore vendored Normal file
View File

@@ -0,0 +1,48 @@
# Binaries for programs and plugins
*.exe
*.exe~
*.dll
*.so
*.dylib
# Test binary, built with `go test -c`
*.test
# Output of the go coverage tool, specifically when used with LiteIDE
*.out
# Dependency directories (remove the comment below to include it)
# vendor/
# Go workspace file
go.work
# IDE-specific files
.idea/
.vscode/
# OS-specific files
.DS_Store
Thumbs.db
# Binary output directory
/bin/
# Log files
*.log
# Environment variables file
.env
# Air temporary files (if using Air for live reloading)
tmp/
# Compiled Object files, Static and Dynamic libs (Shared Objects)
*.o
*.a
# Debug files
debug
# Project-specific build outputs
/gosrv

View File

@@ -0,0 +1,34 @@
# Build stage
FROM golang:1.23.1-alpine AS builder
WORKDIR /app
# Copy go mod and sum files
COPY go.mod go.sum ./
# Download all dependencies
RUN go mod download
# Copy the source code
COPY . .
# Build the application
RUN CGO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -ldflags="-w -s" -o main .
# Run stage
FROM alpine:latest
RUN apk --no-cache add ca-certificates
ENV GIN_MODE=release
WORKDIR /root/
# Copy the pre-built binary file from the previous stage
COPY --from=builder /app/main .
COPY --from=builder /app/config.yaml .
# Expose port 8080 to the outside world
EXPOSE 8080
# Command to run the executable
CMD ["./main"]

122
rnd/rest-api-go/README.md Normal file
View File

@@ -0,0 +1,122 @@
# Market API
This project is a Go-based API for a marketplace application. It provides endpoints for managing agents, handling user authentication, and performing administrative tasks.
## Project Structure
The project is organized into several packages:
- `config`: Handles configuration loading and management
- `docs`: Contains the Swagger documentation
- `database`: Contains database migrations and interaction logic
- `handlers`: Implements HTTP request handlers
- `middleware`: Contains middleware functions for the API
- `models`: Defines data structures used throughout the application
- `utils`: Provides utility functions
## Prerequisites
- Go 1.16 or later
- PostgreSQL
- [golang-migrate](https://github.com/golang-migrate/migrate)
## Setup
1. Clone the repository
2. Install dependencies:
```
go mod tidy
```
3. Set up the database:
- Create a PostgreSQL database
- Update the `DatabaseURL` in your configuration file
4. Run database migrations:
```
migrate -source file://database/migrations -database "postgresql://agpt_user:pass123@localhost:5432/apgt_marketplace?sslmode=disable" up
```
## Running the Application
To run the application in development mode with hot reloading:
```
air
```
For production, build and run the binary:
```
go build -o market-api
./market-api
```
## Testing
Run tests with coverage:
```
go test -cover ./...
```
## Code Formatting
Format the code using:
```
gofmt -w .
```
## Database Migrations
Create a new migration:
```
migrate create -ext sql -dir database/migrations -seq <migration_name>
```
Apply migrations:
```
migrate -source file://database/migrations -database "postgresql://user:password@localhost:5432/dbname?sslmode=disable" up
```
Revert the last migration:
```
migrate -source file://database/migrations -database "postgresql://user:password@localhost:5432/dbname?sslmode=disable" down 1
```
## API Endpoints
The API provides various endpoints for agent management, user authentication, and administrative tasks. Some key endpoints include:
- `/api/agents`: Get list of agents
- `/api/agents/:agent_id`: Get agent details
- `/api/agents/submit`: Submit a new agent
- `/api/admin/*`: Various administrative endpoints (requires admin authentication)
Refer to the `main.go` file for a complete list of endpoints and their corresponding handlers.
# Swagger Documentation
This project uses `gin-swagger` and `Swaggo` tools for automatic generation of API documentation in OpenAPI (Swagger) format. The documentation is based on comments added to the code using Swagger annotations.
To view and interact with the generated Swagger documentation, follow these steps:
1. Run your Gin server.
2. Access the Swagger UI by navigating to `http://localhost:8015/docs/index.html` in your web browser.
Alternatively, you can view the raw OpenAPI specification at `http://localhost:8015/docs/doc.json`.
## Regenerating Swagger Documentation
If you make changes to your codebase and want to regenerate the Swagger documentation, follow these steps:
1. Run the `swag init` command in your project directory to create a new `docs.go` file (or update an existing one) with Swagger documentation comments based on your code:
```bash
swag init -g main.go
```
Replace `main.go` with the name of your main Go source file.
3. Run your Gin server, and access the updated Swagger UI at `http://localhost:8015/docs/index.html`. You should see your documentation reflecting the latest changes in your codebase.

View File

@@ -0,0 +1,7 @@
ServerAddress: ":8015"
DatabaseURL: "postgresql://agpt_user:pass123@localhost:5433/agpt_marketplace?connect_timeout=60"
JWTSecret: "Z86RsQ+nhSk+A8ODJX1kQA11JCk9nlw8n+MRdSgmR+P1sMPTTDG1rjBTwj7Ucjb3TRHSVxkCNPgXISmzU/vMkA=="
JWTAlgorithm: "HS256"
CORSAllowOrigins:
- "http://localhost:3000"
- "http://127.0.0.1:3000"

View File

@@ -0,0 +1,63 @@
package config
import (
"fmt"
"github.com/spf13/viper"
"go.uber.org/zap"
)
type Config struct {
ServerAddress string `mapstructure:"serveraddress"`
DatabaseURL string `mapstructure:"databaseurl"`
AuthEnabled bool `mapstructure:"authenabled"`
JWTSecret string `mapstructure:"jwtsecret"`
JWTAlgorithm string `mapstructure:"jwtalgorithm"`
CORSAllowOrigins []string `mapstructure:"corsalloworigins"`
}
func Load(configFile ...string) (*Config, error) {
logger := zap.L().With(zap.String("function", "Load"))
if len(configFile) > 0 {
viper.SetConfigFile(configFile[0])
} else {
viper.SetConfigName("config")
viper.SetConfigType("yaml")
viper.AddConfigPath(".")
}
viper.SetEnvPrefix("AGPT")
viper.AutomaticEnv()
if err := viper.ReadInConfig(); err != nil {
logger.Error("Failed to read config file", zap.Error(err))
return nil, err
}
var config Config
if err := viper.Unmarshal(&config); err != nil {
logger.Error("Failed to unmarshal config", zap.Error(err))
return nil, err
}
// Validate required fields
if config.ServerAddress == "" {
logger.Error("ServerAddress is required")
return nil, fmt.Errorf("ServerAddress is required")
}
if config.DatabaseURL == "" {
logger.Error("DatabaseURL is required")
return nil, fmt.Errorf("DatabaseURL is required")
}
if config.JWTSecret == "" {
logger.Error("JWTSecret is required")
return nil, fmt.Errorf("JWTSecret is required")
}
if config.JWTAlgorithm == "" {
logger.Error("JWTAlgorithm is required")
return nil, fmt.Errorf("JWTAlgorithm is required")
}
return &config, nil
}

View File

@@ -0,0 +1,78 @@
package config
import (
"os"
"testing"
"github.com/stretchr/testify/assert"
)
func TestLoadValidConfig(t *testing.T) {
// Create a temporary config file for testing
tempFile, err := os.CreateTemp("", "test-config*.yaml")
if err != nil {
t.Fatalf("Failed to create temp file: %v", err)
}
defer os.Remove(tempFile.Name())
// Write test configuration to the temp file
testConfig := []byte(`
serveraddress: ":8080"
databaseurl: "postgres://user:pass@localhost:5432/testdb"
authenabled: true
jwtsecret: "test-secret"
jwtalgorithm: "HS256"
`)
if _, err := tempFile.Write(testConfig); err != nil {
t.Fatalf("Failed to write to temp file: %v", err)
}
tempFile.Close()
// Test the Load function with a specific config file
config, err := Load(tempFile.Name())
assert.NoError(t, err)
assert.NotNil(t, config)
// Verify the loaded configuration
assert.Equal(t, ":8080", config.ServerAddress)
assert.Equal(t, "postgres://user:pass@localhost:5432/testdb", config.DatabaseURL)
assert.True(t, config.AuthEnabled)
assert.Equal(t, "test-secret", config.JWTSecret)
assert.Equal(t, "HS256", config.JWTAlgorithm)
}
func TestLoadDefaultConfigFile(t *testing.T) {
// Test with default config file (should fail in test environment)
config, err := Load()
assert.Error(t, err)
assert.Nil(t, config)
}
func TestLoadMissingConfigFile(t *testing.T) {
// Test with missing config file
config, err := Load("non_existent_config.yaml")
assert.Error(t, err)
assert.Nil(t, config)
}
func TestLoadInvalidConfigFormat(t *testing.T) {
// Create a temporary config file for testing
tempFile, err := os.CreateTemp("", "test-config*.yaml")
if err != nil {
t.Fatalf("Failed to create temp file: %v", err)
}
defer os.Remove(tempFile.Name())
// Test with invalid config format
invalidConfig := []byte(`
serveraddress: ":8080"
databaseurl: 123 # Invalid type, should be string
`)
if err := os.WriteFile(tempFile.Name(), invalidConfig, 0644); err != nil {
t.Fatalf("Failed to write invalid config: %v", err)
}
config, err := Load(tempFile.Name())
assert.Error(t, err)
assert.Nil(t, config)
}

View File

@@ -0,0 +1,805 @@
package database
import (
"context"
"fmt"
"strings"
"time"
"github.com/google/uuid"
"github.com/jackc/pgx/v5"
"github.com/jackc/pgx/v5/pgxpool"
"github.com/swiftyos/market/config"
"github.com/swiftyos/market/models"
"github.com/swiftyos/market/utils"
"go.uber.org/zap"
)
func NewDB(cfg *config.Config) (*pgxpool.Pool, error) {
return pgxpool.New(context.Background(), cfg.DatabaseURL)
}
func GetAgents(ctx context.Context, db *pgxpool.Pool, logger *zap.Logger, page int, pageSize int, name *string, keywords *string, categories *string) ([]models.Agent, error) {
logger = logger.With(zap.String("function", "GetAgents")).With(zap.String("file", "db.go"))
logger.Debug("Query parameters",
zap.Int("page", page),
zap.Int("pageSize", pageSize),
zap.String("name", utils.StringOrNil(name)),
zap.String("keywords", utils.StringOrNil(keywords)),
zap.String("categories", utils.StringOrNil(categories)))
query := `
SELECT "id", "name", "description", "author", "keywords", "categories", "graph" FROM "Agents"
WHERE "submissionStatus" = 'APPROVED'
AND ($3::text IS NULL OR name ILIKE '%' || $3 || '%')
AND ($4::text IS NULL OR $4 = ANY(keywords))
AND ($5::text IS NULL OR $5 = ANY(categories))
ORDER BY "createdAt" DESC
LIMIT $1 OFFSET $2
`
rows, err := db.Query(ctx, query, pageSize, (page-1)*pageSize, name, keywords, categories)
if err != nil {
logger.Error("Error querying agents", zap.Error(err))
return nil, err
}
defer rows.Close()
var agents []models.Agent
for rows.Next() {
var agent models.Agent
err := rows.Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
)
if err != nil {
logger.Error("Error scanning agent", zap.Error(err))
return nil, err
}
agents = append(agents, agent)
}
if err := rows.Err(); err != nil {
logger.Error("Error iterating over agents", zap.Error(err))
return nil, err
}
logger.Info("Found agents", zap.Int("count", len(agents)))
if agents == nil {
agents = []models.Agent{}
}
return agents, err
}
func SubmitAgent(ctx context.Context, db *pgxpool.Pool, request models.AddAgentRequest, user interface{}) (*models.AgentWithMetadata, error) {
logger := zap.L().With(zap.String("function", "SubmitAgent"))
logger.Info("Submitting new agent")
// Generate a new UUID for the agent
agentID := uuid.New().String()
// Create the Agent struct
agent := models.Agent{
ID: agentID,
Name: request.Graph.Name,
Description: request.Graph.Description,
Author: request.Author,
Keywords: request.Keywords,
Categories: request.Categories,
Graph: request.Graph,
}
// Create the AgentWithMetadata struct
agentWithMetadata := models.AgentWithMetadata{
Agent: agent,
Version: 1,
CreatedAt: time.Now(),
UpdatedAt: time.Now(),
SubmissionDate: time.Now(),
SubmissionStatus: models.SubmissionStatusPending,
}
// Start a transaction
tx, err := db.Begin(ctx)
if err != nil {
logger.Error("Failed to begin transaction", zap.Error(err))
return nil, err
}
defer tx.Rollback(ctx)
// Insert the agent into the database
_, err = tx.Exec(ctx, `
INSERT INTO "Agents" (id, name, description, author, keywords, categories, graph, version, created_at, updated_at, submission_date, submission_status)
VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11, $12)
`, agentWithMetadata.ID, agentWithMetadata.Name, agentWithMetadata.Description, agentWithMetadata.Author,
agentWithMetadata.Keywords, agentWithMetadata.Categories, agentWithMetadata.Graph,
agentWithMetadata.Version, agentWithMetadata.CreatedAt, agentWithMetadata.UpdatedAt,
agentWithMetadata.SubmissionDate, agentWithMetadata.SubmissionStatus)
if err != nil {
logger.Error("Failed to insert agent", zap.Error(err))
return nil, err
}
// Commit the transaction
err = tx.Commit(ctx)
if err != nil {
logger.Error("Failed to commit transaction", zap.Error(err))
return nil, err
}
logger.Info("Successfully submitted new agent", zap.String("agentID", agentID))
return &agentWithMetadata, nil
}
func GetAgentDetails(ctx context.Context, db *pgxpool.Pool, agentID string) (*models.AgentWithMetadata, error) {
logger := zap.L().With(zap.String("function", "GetAgentDetails"))
query := `
SELECT id, name, description, author, keywords, categories, graph, version, created_at, updated_at, submission_date, submission_status
FROM "Agents"
WHERE id = $1
`
var agent models.AgentWithMetadata
err := db.QueryRow(ctx, query, agentID).Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
&agent.Version,
&agent.CreatedAt,
&agent.UpdatedAt,
&agent.SubmissionDate,
&agent.SubmissionStatus,
)
if err != nil {
if err == pgx.ErrNoRows {
logger.Error("Agent not found", zap.String("agentID", agentID))
return nil, fmt.Errorf("agent not found")
}
logger.Error("Error querying agent details", zap.Error(err))
return nil, err
}
logger.Info("Agent details retrieved", zap.String("agentID", agentID))
return &agent, nil
}
func IncrementDownloadCount(ctx context.Context, db *pgxpool.Pool, agentID string) error {
logger := zap.L().With(zap.String("function", "IncrementDownloadCount"))
query := `
UPDATE "Agents"
SET download_count = download_count + 1
WHERE id = $1
`
_, err := db.Exec(ctx, query, agentID)
if err != nil {
logger.Error("Failed to increment download count", zap.Error(err), zap.String("agentID", agentID))
return err
}
logger.Info("Download count incremented", zap.String("agentID", agentID))
return nil
}
func GetAgentFile(ctx context.Context, db *pgxpool.Pool, agentID string) (*models.AgentFile, error) {
logger := zap.L().With(zap.String("function", "GetAgentFile"))
query := `
SELECT id, name, graph
FROM "Agents"
WHERE id = $1
`
var agentFile models.AgentFile
err := db.QueryRow(ctx, query, agentID).Scan(
&agentFile.ID,
&agentFile.Name,
&agentFile.Graph,
)
if err != nil {
if err == pgx.ErrNoRows {
logger.Error("Agent not found", zap.String("agentID", agentID))
return nil, fmt.Errorf("agent not found")
}
logger.Error("Error querying agent file", zap.Error(err))
return nil, err
}
logger.Info("Agent file retrieved", zap.String("agentID", agentID))
return &agentFile, nil
}
func GetTopAgentsByDownloads(ctx context.Context, db *pgxpool.Pool, page, pageSize int) ([]models.AgentWithDownloads, int, error) {
logger := zap.L().With(zap.String("function", "GetTopAgentsByDownloads"))
offset := (page - 1) * pageSize
query := `
SELECT a.id, a.name, a.description, a.author, a.keywords, a.categories, a.graph, at.downloads
FROM "Agents" a
JOIN "AnalyticsTracker" at ON a.id = at.agent_id
WHERE a.submission_status = 'APPROVED'
ORDER BY at.downloads DESC
LIMIT $1 OFFSET $2
`
rows, err := db.Query(ctx, query, pageSize, offset)
if err != nil {
logger.Error("Failed to query top agents", zap.Error(err))
return nil, 0, err
}
defer rows.Close()
var agents []models.AgentWithDownloads
for rows.Next() {
var agent models.AgentWithDownloads
err := rows.Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
&agent.Downloads,
)
if err != nil {
logger.Error("Failed to scan agent row", zap.Error(err))
return nil, 0, err
}
agents = append(agents, agent)
}
var totalCount int
err = db.QueryRow(ctx, `SELECT COUNT(*) FROM "Agents" WHERE submission_status = 'APPROVED'`).Scan(&totalCount)
if err != nil {
logger.Error("Failed to get total count", zap.Error(err))
return nil, 0, err
}
logger.Info("Top agents retrieved", zap.Int("count", len(agents)))
return agents, totalCount, nil
}
func GetFeaturedAgents(ctx context.Context, db *pgxpool.Pool, category string, page, pageSize int) ([]models.Agent, int, error) {
logger := zap.L().With(zap.String("function", "GetFeaturedAgents"))
offset := (page - 1) * pageSize
query := `
SELECT a.id, a.name, a.description, a.author, a.keywords, a.categories, a.graph
FROM "Agents" a
JOIN "FeaturedAgent" fa ON a.id = fa.agent_id
WHERE $1 = ANY(fa.featured_categories) AND fa.is_active = true AND a.submission_status = 'APPROVED'
ORDER BY a.created_at DESC
LIMIT $2 OFFSET $3
`
rows, err := db.Query(ctx, query, category, pageSize, offset)
if err != nil {
logger.Error("Failed to query featured agents", zap.Error(err))
return nil, 0, err
}
defer rows.Close()
var agents []models.Agent
for rows.Next() {
var agent models.Agent
err := rows.Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
)
if err != nil {
logger.Error("Failed to scan featured agent row", zap.Error(err))
return nil, 0, err
}
agents = append(agents, agent)
}
var totalCount int
err = db.QueryRow(ctx, `SELECT COUNT(*) FROM "FeaturedAgent" fa JOIN "Agents" a ON fa.agent_id = a.id WHERE $1 = ANY(fa.featured_categories) AND fa.is_active = true AND a.submission_status = 'APPROVED'`, category).Scan(&totalCount)
if err != nil {
logger.Error("Failed to get total count of featured agents", zap.Error(err))
return nil, 0, err
}
logger.Info("Featured agents retrieved", zap.Int("count", len(agents)))
return agents, totalCount, nil
}
func Search(ctx context.Context, db *pgxpool.Pool, query string, categories []string, page, pageSize int, sortBy, sortOrder string) ([]models.AgentWithRank, error) {
logger := zap.L().With(zap.String("function", "Search"))
offset := (page - 1) * pageSize
categoryFilter := ""
if len(categories) > 0 {
categoryConditions := make([]string, len(categories))
for i, cat := range categories {
categoryConditions[i] = fmt.Sprintf("'%s' = ANY(a.categories)", cat)
}
categoryFilter = "AND (" + strings.Join(categoryConditions, " OR ") + ")"
}
orderByClause := ""
switch sortBy {
case "createdAt", "updatedAt":
orderByClause = fmt.Sprintf(`a."%s" %s, rank DESC`, sortBy, sortOrder)
case "name":
orderByClause = fmt.Sprintf(`a.name %s, rank DESC`, sortOrder)
default:
orderByClause = `rank DESC, a."createdAt" DESC`
}
sqlQuery := fmt.Sprintf(`
WITH query AS (
SELECT to_tsquery(string_agg(lexeme || ':*', ' & ' ORDER BY positions)) AS q
FROM unnest(to_tsvector($1))
)
SELECT
a.id,
a.created_at,
a.updated_at,
a.version,
a.name,
LEFT(a.description, 500) AS description,
a.author,
a.keywords,
a.categories,
a.graph,
a.submission_status,
a.submission_date,
ts_rank(CAST(a.search AS tsvector), query.q) AS rank
FROM "Agents" a, query
WHERE a.submission_status = 'APPROVED' %s
ORDER BY %s
LIMIT $2
OFFSET $3
`, categoryFilter, orderByClause)
rows, err := db.Query(ctx, sqlQuery, query, pageSize, offset)
if err != nil {
logger.Error("Failed to execute search query", zap.Error(err))
return nil, err
}
defer rows.Close()
var agents []models.AgentWithRank
for rows.Next() {
var agent models.AgentWithRank
err := rows.Scan(
&agent.ID,
&agent.CreatedAt,
&agent.UpdatedAt,
&agent.Version,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
&agent.SubmissionStatus,
&agent.SubmissionDate,
&agent.Rank,
)
if err != nil {
logger.Error("Failed to scan search result row", zap.Error(err))
return nil, err
}
agents = append(agents, agent)
}
logger.Info("Search completed", zap.Int("results", len(agents)))
return agents, nil
}
func CreateAgentInstalledEvent(ctx context.Context, db *pgxpool.Pool, eventData models.InstallTracker) error {
logger := zap.L().With(zap.String("function", "CreateAgentInstalledEvent"))
logger.Info("Creating agent installed event")
query := `
INSERT INTO install_tracker (marketplace_agent_id, installed_agent_id, installation_location)
VALUES ($1, $2, $3)
`
_, err := db.Exec(ctx, query,
eventData.MarketplaceAgentID,
eventData.InstalledAgentID,
eventData.InstallationLocation,
)
if err != nil {
logger.Error("Failed to create agent installed event", zap.Error(err))
return fmt.Errorf("failed to create agent installed event: %w", err)
}
logger.Info("Agent installed event created successfully")
return nil
}
// Admin Queries
func CreateAgentEntry(ctx context.Context, db *pgxpool.Pool, agent models.Agent) (models.Agent, error) {
logger := zap.L().With(zap.String("function", "CreateAgentEntry"))
logger.Info("Creating agent entry")
query := `
INSERT INTO agents (id, name, description, author, keywords, categories, graph)
VALUES ($1, $2, $3, $4, $5, $6, $7)
RETURNING id, name, description, author, keywords, categories, graph
`
var createdAgent models.Agent
err := db.QueryRow(ctx, query,
agent.ID,
agent.Name,
agent.Description,
agent.Author,
agent.Keywords,
agent.Categories,
agent.Graph,
).Scan(
&createdAgent.ID,
&createdAgent.Name,
&createdAgent.Description,
&createdAgent.Author,
&createdAgent.Keywords,
&createdAgent.Categories,
&createdAgent.Graph,
)
if err != nil {
logger.Error("Failed to create agent entry", zap.Error(err))
return models.Agent{}, err
}
logger.Info("Agent entry created successfully", zap.String("agentID", agent.ID))
return createdAgent, nil
}
func SetAgentFeatured(ctx context.Context, db *pgxpool.Pool, agentID string, isActive bool, featuredCategories []string) (*models.FeaturedAgent, error) {
logger := zap.L().With(zap.String("function", "SetAgentFeatured"))
logger.Info("Setting agent featured status", zap.String("agentID", agentID), zap.Bool("isActive", isActive))
// Check if the agent exists
var exists bool
err := db.QueryRow(ctx, `SELECT EXISTS(SELECT 1 FROM "Agents" WHERE id = $1)`, agentID).Scan(&exists)
if err != nil {
logger.Error("Failed to check if agent exists", zap.Error(err))
return nil, fmt.Errorf("failed to check if agent exists: %w", err)
}
if !exists {
return nil, fmt.Errorf("agent with ID %s not found", agentID)
}
var query string
var args []interface{}
if isActive {
// Set the agent as featured
query = `
INSERT INTO "FeaturedAgent" (agent_id, featured_categories, is_active)
VALUES ($1, $2, $3)
ON CONFLICT (agent_id) DO UPDATE
SET featured_categories = $2, is_active = $3
RETURNING agent_id, featured_categories, is_active
`
args = []interface{}{agentID, featuredCategories, isActive}
} else {
// Unset the agent as featured
query = `
DELETE FROM "FeaturedAgent"
WHERE agent_id = $1
RETURNING agent_id, featured_categories, is_active
`
args = []interface{}{agentID}
}
var featuredAgent models.FeaturedAgent
err = db.QueryRow(ctx, query, args...).Scan(
&featuredAgent.AgentID,
&featuredAgent.FeaturedCategories,
&featuredAgent.IsActive,
)
if err != nil {
if err == pgx.ErrNoRows && !isActive {
logger.Info("Agent was not featured, no action needed", zap.String("agentID", agentID))
return nil, nil
}
logger.Error("Failed to set agent featured status", zap.Error(err))
return nil, fmt.Errorf("failed to set agent featured status: %w", err)
}
if isActive {
logger.Info("Agent set as featured successfully", zap.String("agentID", agentID))
} else {
logger.Info("Agent unset as featured successfully", zap.String("agentID", agentID))
}
return &featuredAgent, nil
}
func GetAgentFeatured(ctx context.Context, db *pgxpool.Pool, agentID string) (*models.FeaturedAgent, error) {
logger := zap.L().With(zap.String("function", "GetAgentFeatured"))
logger.Info("Getting featured agent", zap.String("agentID", agentID))
query := `
SELECT agent_id, featured_categories, is_active
FROM "FeaturedAgent"
WHERE agent_id = $1
`
var featuredAgent models.FeaturedAgent
err := db.QueryRow(ctx, query, agentID).Scan(
&featuredAgent.AgentID,
&featuredAgent.FeaturedCategories,
&featuredAgent.IsActive,
)
if err == pgx.ErrNoRows {
return nil, nil
}
if err != nil {
logger.Error("Failed to get featured agent", zap.Error(err))
return nil, fmt.Errorf("failed to get featured agent: %w", err)
}
logger.Info("Featured agent retrieved successfully", zap.String("agentID", agentID))
return &featuredAgent, nil
}
func RemoveFeaturedCategory(ctx context.Context, db *pgxpool.Pool, agentID string, category string) (*models.FeaturedAgent, error) {
logger := zap.L().With(zap.String("function", "RemoveFeaturedCategory"))
logger.Info("Removing featured category", zap.String("agentID", agentID), zap.String("category", category))
query := `
UPDATE "FeaturedAgent"
SET featured_categories = array_remove(featured_categories, $1)
WHERE agent_id = $2
RETURNING agent_id, featured_categories, is_active
`
var featuredAgent models.FeaturedAgent
err := db.QueryRow(ctx, query, category, agentID).Scan(
&featuredAgent.AgentID,
&featuredAgent.FeaturedCategories,
&featuredAgent.IsActive,
)
if err == pgx.ErrNoRows {
return nil, nil
}
if err != nil {
logger.Error("Failed to remove featured category", zap.Error(err))
return nil, fmt.Errorf("failed to remove featured category: %w", err)
}
logger.Info("Featured category removed successfully", zap.String("agentID", agentID), zap.String("category", category))
return &featuredAgent, nil
}
func GetNotFeaturedAgents(ctx context.Context, db *pgxpool.Pool, page, pageSize int) ([]models.Agent, error) {
logger := zap.L().With(zap.String("function", "GetNotFeaturedAgents"))
logger.Info("Getting not featured agents", zap.Int("page", page), zap.Int("pageSize", pageSize))
offset := (page - 1) * pageSize
query := `
SELECT a.id, a.name, a.description, a.author, a.keywords, a.categories, a.graph
FROM "Agents" a
LEFT JOIN "FeaturedAgent" fa ON a.id = fa.agent_id
WHERE (fa.agent_id IS NULL OR fa.featured_categories = '{}')
AND a.submission_status = 'APPROVED'
ORDER BY a.created_at DESC
LIMIT $1 OFFSET $2
`
rows, err := db.Query(ctx, query, pageSize, offset)
if err != nil {
logger.Error("Failed to query not featured agents", zap.Error(err))
return nil, err
}
defer rows.Close()
var agents []models.Agent
for rows.Next() {
var agent models.Agent
err := rows.Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
)
if err != nil {
logger.Error("Failed to scan not featured agent row", zap.Error(err))
return nil, err
}
agents = append(agents, agent)
}
logger.Info("Not featured agents retrieved", zap.Int("count", len(agents)))
return agents, nil
}
func GetAgentSubmissions(ctx context.Context, db *pgxpool.Pool, page, pageSize int, name, keyword, category *string, sortBy, sortOrder string) ([]models.AgentWithMetadata, int, error) {
logger := zap.L().With(zap.String("function", "GetAgentSubmissions"))
logger.Info("Getting agent submissions", zap.Int("page", page), zap.Int("pageSize", pageSize))
offset := (page - 1) * pageSize
query := `
SELECT a.id, a.name, a.description, a.author, a.keywords, a.categories, a.graph, a.created_at, a.updated_at, a.version, a.submission_status, a.submission_review_date, a.submission_review_comments
FROM "Agents" a
WHERE a.submission_status = 'PENDING'
`
args := []interface{}{}
argCount := 1
if name != nil {
query += fmt.Sprintf(" AND a.name ILIKE $%d", argCount)
args = append(args, "%"+*name+"%")
argCount++
}
if keyword != nil {
query += fmt.Sprintf(" AND $%d = ANY(a.keywords)", argCount)
args = append(args, *keyword)
argCount++
}
if category != nil {
query += fmt.Sprintf(" AND $%d = ANY(a.categories)", argCount)
args = append(args, *category)
argCount++
}
// Add sorting
query += fmt.Sprintf(" ORDER BY a.%s %s", sortBy, sortOrder)
// Add pagination
query += fmt.Sprintf(" LIMIT $%d OFFSET $%d", argCount, argCount+1)
args = append(args, pageSize, offset)
rows, err := db.Query(ctx, query, args...)
if err != nil {
logger.Error("Failed to query agent submissions", zap.Error(err))
return nil, 0, err
}
defer rows.Close()
var agents []models.AgentWithMetadata
for rows.Next() {
var agent models.AgentWithMetadata
err := rows.Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
&agent.CreatedAt,
&agent.UpdatedAt,
&agent.Version,
&agent.SubmissionStatus,
&agent.SubmissionReviewDate,
&agent.SubmissionReviewComments,
)
if err != nil {
logger.Error("Failed to scan agent submission row", zap.Error(err))
return nil, 0, err
}
agents = append(agents, agent)
}
// Get total count
countQuery := `SELECT COUNT(*) FROM "Agents" WHERE submission_status = 'PENDING'`
var totalCount int
err = db.QueryRow(ctx, countQuery).Scan(&totalCount)
if err != nil {
logger.Error("Failed to get total count of agent submissions", zap.Error(err))
return nil, 0, err
}
logger.Info("Agent submissions retrieved", zap.Int("count", len(agents)))
return agents, totalCount, nil
}
func ReviewSubmission(ctx context.Context, db *pgxpool.Pool, agentID string, version int, status models.SubmissionStatus, comments *string) (*models.AgentWithMetadata, error) {
logger := zap.L().With(zap.String("function", "ReviewSubmission"))
logger.Info("Reviewing agent submission", zap.String("agentID", agentID), zap.Int("version", version))
query := `
UPDATE "Agents"
SET submission_status = $1,
submission_review_date = NOW(),
submission_review_comments = $2
WHERE id = $3 AND version = $4
RETURNING id, name, description, author, keywords, categories, graph, created_at, updated_at, version, submission_status, submission_review_date, submission_review_comments
`
var agent models.AgentWithMetadata
err := db.QueryRow(ctx, query, status, comments, agentID, version).Scan(
&agent.ID,
&agent.Name,
&agent.Description,
&agent.Author,
&agent.Keywords,
&agent.Categories,
&agent.Graph,
&agent.CreatedAt,
&agent.UpdatedAt,
&agent.Version,
&agent.SubmissionStatus,
&agent.SubmissionReviewDate,
&agent.SubmissionReviewComments,
)
if err != nil {
if err == pgx.ErrNoRows {
logger.Error("Agent submission not found", zap.String("agentID", agentID), zap.Int("version", version))
return nil, fmt.Errorf("agent submission not found")
}
logger.Error("Failed to review agent submission", zap.Error(err))
return nil, err
}
logger.Info("Agent submission reviewed successfully", zap.String("agentID", agentID), zap.Int("version", version))
return &agent, nil
}
func GetAllCategories(ctx context.Context, db *pgxpool.Pool) ([]string, error) {
logger := zap.L().With(zap.String("function", "GetAllCategories"))
logger.Info("Getting all categories")
query := `
SELECT DISTINCT unnest(categories) AS category
FROM "Agents"
ORDER BY category
`
rows, err := db.Query(ctx, query)
if err != nil {
logger.Error("Failed to query categories", zap.Error(err))
return nil, err
}
defer rows.Close()
var categories []string
for rows.Next() {
var category string
err := rows.Scan(&category)
if err != nil {
logger.Error("Failed to scan category row", zap.Error(err))
return nil, err
}
categories = append(categories, category)
}
logger.Info("Categories retrieved", zap.Int("count", len(categories)))
return categories, nil
}

View File

@@ -0,0 +1,22 @@
-- Drop foreign key constraints
ALTER TABLE "AnalyticsTracker" DROP CONSTRAINT IF EXISTS "AnalyticsTracker_agentId_fkey";
ALTER TABLE "InstallTracker" DROP CONSTRAINT IF EXISTS "InstallTracker_marketplaceAgentId_fkey";
ALTER TABLE "FeaturedAgent" DROP CONSTRAINT IF EXISTS "FeaturedAgent_agentId_fkey";
-- Drop indexes
DROP INDEX IF EXISTS "FeaturedAgent_agentId_key";
DROP INDEX IF EXISTS "FeaturedAgent_id_key";
DROP INDEX IF EXISTS "InstallTracker_marketplaceAgentId_installedAgentId_key";
DROP INDEX IF EXISTS "AnalyticsTracker_agentId_key";
DROP INDEX IF EXISTS "AnalyticsTracker_id_key";
DROP INDEX IF EXISTS "Agents_id_key";
-- Drop tables
DROP TABLE IF EXISTS "FeaturedAgent";
DROP TABLE IF EXISTS "InstallTracker";
DROP TABLE IF EXISTS "AnalyticsTracker";
DROP TABLE IF EXISTS "Agents";
-- Drop enums
DROP TYPE IF EXISTS "InstallationLocation";
DROP TYPE IF EXISTS "SubmissionStatus";

View File

@@ -0,0 +1,86 @@
-- CreateEnum
CREATE TYPE "SubmissionStatus" AS ENUM ('PENDING', 'APPROVED', 'REJECTED');
-- CreateEnum
CREATE TYPE "InstallationLocation" AS ENUM ('LOCAL', 'CLOUD');
-- CreateTable
CREATE TABLE "Agents" (
"id" UUID NOT NULL DEFAULT gen_random_uuid(),
"version" INTEGER NOT NULL DEFAULT 1,
"createdAt" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
"updatedAt" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
"submissionDate" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
"submissionReviewDate" TIMESTAMP(3),
"submissionStatus" "SubmissionStatus" NOT NULL DEFAULT 'PENDING',
"submissionReviewComments" TEXT,
"name" TEXT,
"description" TEXT,
"author" TEXT,
"keywords" TEXT[],
"categories" TEXT[],
"search" tsvector DEFAULT ''::tsvector,
"graph" JSONB NOT NULL,
CONSTRAINT "Agents_pkey" PRIMARY KEY ("id","version")
);
-- CreateTable
CREATE TABLE "AnalyticsTracker" (
"id" UUID NOT NULL DEFAULT gen_random_uuid(),
"agentId" UUID NOT NULL,
"views" INTEGER NOT NULL,
"downloads" INTEGER NOT NULL,
CONSTRAINT "AnalyticsTracker_pkey" PRIMARY KEY ("id")
);
-- CreateTable
CREATE TABLE "InstallTracker" (
"id" UUID NOT NULL DEFAULT gen_random_uuid(),
"marketplaceAgentId" UUID NOT NULL,
"installedAgentId" UUID NOT NULL,
"installationLocation" "InstallationLocation" NOT NULL,
"createdAt" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
CONSTRAINT "InstallTracker_pkey" PRIMARY KEY ("id")
);
-- CreateTable
CREATE TABLE "FeaturedAgent" (
"id" UUID NOT NULL DEFAULT gen_random_uuid(),
"agentId" UUID NOT NULL,
"isActive" BOOLEAN NOT NULL DEFAULT false,
"featuredCategories" TEXT[],
"createdAt" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
"updatedAt" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
CONSTRAINT "FeaturedAgent_pkey" PRIMARY KEY ("id")
);
-- CreateIndex
CREATE UNIQUE INDEX "Agents_id_key" ON "Agents"("id");
-- CreateIndex
CREATE UNIQUE INDEX "AnalyticsTracker_id_key" ON "AnalyticsTracker"("id");
-- CreateIndex
CREATE UNIQUE INDEX "AnalyticsTracker_agentId_key" ON "AnalyticsTracker"("agentId");
-- CreateIndex
CREATE UNIQUE INDEX "InstallTracker_marketplaceAgentId_installedAgentId_key" ON "InstallTracker"("marketplaceAgentId", "installedAgentId");
-- CreateIndex
CREATE UNIQUE INDEX "FeaturedAgent_id_key" ON "FeaturedAgent"("id");
-- CreateIndex
CREATE UNIQUE INDEX "FeaturedAgent_agentId_key" ON "FeaturedAgent"("agentId");
-- AddForeignKey
ALTER TABLE "AnalyticsTracker" ADD CONSTRAINT "AnalyticsTracker_agentId_fkey" FOREIGN KEY ("agentId") REFERENCES "Agents"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
-- AddForeignKey
ALTER TABLE "InstallTracker" ADD CONSTRAINT "InstallTracker_marketplaceAgentId_fkey" FOREIGN KEY ("marketplaceAgentId") REFERENCES "Agents"("id") ON DELETE RESTRICT ON UPDATE CASCADE;
-- AddForeignKey
ALTER TABLE "FeaturedAgent" ADD CONSTRAINT "FeaturedAgent_agentId_fkey" FOREIGN KEY ("agentId") REFERENCES "Agents"("id") ON DELETE RESTRICT ON UPDATE CASCADE;

View File

@@ -0,0 +1,11 @@
-- Remove sample data from FeaturedAgent table
DELETE FROM "FeaturedAgent" WHERE "agentId" IN ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', 'eaa773b1-5efa-485f-b2f0-2e05bae6d297');
-- Remove sample data from InstallTracker table
DELETE FROM "InstallTracker" WHERE "marketplaceAgentId" IN ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', 'eaa773b1-5efa-485f-b2f0-2e05bae6d297', 'b47e40a7-ad5f-4b29-9eac-abd5b728f19a', 'a4d3598f-6180-4e6d-96bf-6e15c3de05a9', '9f332ff3-4c74-4f5b-9838-65938a06711f');
-- Remove sample data from AnalyticsTracker table
DELETE FROM "AnalyticsTracker" WHERE "agentId" IN ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', 'eaa773b1-5efa-485f-b2f0-2e05bae6d297', 'b47e40a7-ad5f-4b29-9eac-abd5b728f19a', 'a4d3598f-6180-4e6d-96bf-6e15c3de05a9', '9f332ff3-4c74-4f5b-9838-65938a06711f');
-- Remove sample data from Agents table
DELETE FROM "Agents" WHERE "id" IN ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', 'eaa773b1-5efa-485f-b2f0-2e05bae6d297', 'b47e40a7-ad5f-4b29-9eac-abd5b728f19a', 'a4d3598f-6180-4e6d-96bf-6e15c3de05a9', '9f332ff3-4c74-4f5b-9838-65938a06711f');

View File

@@ -0,0 +1,86 @@
-- Sample data for Agents table (10 agents)
INSERT INTO "Agents" ("id", "name", "description", "author", "keywords", "categories", "graph", "submissionStatus")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', 'AI Recruiter', 'An AI-powered tool that assists HR teams with talent acquisition, screening, and shortlisting.', 'Author1', ARRAY['recruitment', 'HR'], ARRAY['human resources', 'talent management'], '{"key": "value"}', 'APPROVED');
INSERT INTO "Agents" ("id", "name", "description", "author", "keywords", "categories", "graph", "submissionStatus")
VALUES ('3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', 'Customer Service Bot', 'A chatbot that provides 24/7 support and assistance to customers, handling common inquiries and issues.', 'Author2', ARRAY['customer service', 'chatbot'], ARRAY['customer experience', 'artificial intelligence'], '{"key": "value"}', 'APPROVED');
INSERT INTO "Agents" ("id", "name", "description", "author", "keywords", "categories", "graph", "submissionStatus")
VALUES ('eaa773b1-5efa-485f-b2f0-2e05bae6d297', 'Financial Advisor', 'An AI-powered financial advisor that offers personalized investment recommendations and portfolio management.', 'Author3', ARRAY['finance', 'investment'], ARRAY['wealth management', 'artificial intelligence'], '{"key": "value"}', 'APPROVED');
INSERT INTO "Agents" ("id", "name", "description", "author", "keywords", "categories", "graph", "submissionStatus")
VALUES ('b47e40a7-ad5f-4b29-9eac-abd5b728f19a', 'AI Content Writer', 'An AI-powered tool that generates high-quality content for websites, blogs, and marketing materials.', 'Author4', ARRAY['content writing', 'AI'], ARRAY['marketing', 'artificial intelligence'], '{"key": "value"}', 'APPROVED');
INSERT INTO "Agents" ("id", "name", "description", "author", "keywords", "categories", "graph", "submissionStatus")
VALUES ('a4d3598f-6180-4e6d-96bf-6e15c3de05a9', 'AI Image Generator', 'An AI-powered tool that creates realistic images based on text prompts.', 'Author5', ARRAY['image generation', 'AI'], ARRAY['marketing', 'artificial intelligence'], '{"key": "value"}', 'APPROVED');
INSERT INTO "Agents" ("id", "name", "description", "author", "keywords", "categories", "graph")
VALUES ('9f332ff3-4c74-4f5b-9838-65938a06711f', 'AI Video Editor', 'An AI-powered tool that edits and enhances videos with advanced AI algorithms.', 'Author6', ARRAY['video editing', 'AI'], ARRAY['marketing', 'artificial intelligence'], '{"key": "value"}');
-- Sample data for AnalyticsTracker table (10 agents)
INSERT INTO "AnalyticsTracker" ("agentId", "views", "downloads")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', 200, 80);
INSERT INTO "AnalyticsTracker" ("agentId", "views", "downloads")
VALUES ('3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', 150, 60);
INSERT INTO "AnalyticsTracker" ("agentId", "views", "downloads")
VALUES ('eaa773b1-5efa-485f-b2f0-2e05bae6d297', 100, 40);
INSERT INTO "AnalyticsTracker" ("agentId", "views", "downloads")
VALUES ('b47e40a7-ad5f-4b29-9eac-abd5b728f19a', 120, 50);
INSERT INTO "AnalyticsTracker" ("agentId", "views", "downloads")
VALUES ('a4d3598f-6180-4e6d-96bf-6e15c3de05a9', 130, 55);
INSERT INTO "AnalyticsTracker" ("agentId", "views", "downloads")
VALUES ('9f332ff3-4c74-4f5b-9838-65938a06711f', 140, 60);
-- Sample data for InstallTracker table (10 agents)
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac11', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac12', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac13', 'LOCAL');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac14', 'LOCAL');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac15', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac16', 'LOCAL');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', '244f809e-1eee-4a36-a49b-ac2db008ac17', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', '244f809e-1eee-4a36-a49b-ac2db008ac18', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('eaa773b1-5efa-485f-b2f0-2e05bae6d297', '244f809e-1eee-4a36-a49b-ac2db008ac19', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('b47e40a7-ad5f-4b29-9eac-abd5b728f19a', '244f809e-1eee-4a36-a49b-ac2db008ac20', 'LOCAL');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('a4d3598f-6180-4e6d-96bf-6e15c3de05a9', '244f809e-1eee-4a36-a49b-ac2db008ac22', 'CLOUD');
INSERT INTO "InstallTracker" ("marketplaceAgentId", "installedAgentId", "installationLocation")
VALUES ('9f332ff3-4c74-4f5b-9838-65938a06711f', '244f809e-1eee-4a36-a49b-ac2db008ac21', 'CLOUD');
-- Sample data for FeaturedAgent table (3 featured agents)
INSERT INTO "FeaturedAgent" ("agentId", "isActive", "featuredCategories")
VALUES ('b609e5fd-c992-4be9-b68f-afc1980f93c0', true, ARRAY['human resources', 'talent management']);
INSERT INTO "FeaturedAgent" ("agentId", "isActive", "featuredCategories")
VALUES ('3b6d8f75-99d3-41e3-b484-4b2c5f835f5b', true, ARRAY['customer experience', 'artificial intelligence']);
INSERT INTO "FeaturedAgent" ("agentId", "isActive", "featuredCategories")
VALUES ('eaa773b1-5efa-485f-b2f0-2e05bae6d297', true, ARRAY['wealth management', 'artificial intelligence']);

View File

@@ -0,0 +1,47 @@
services:
postgres:
image: ankane/pgvector:latest
environment:
- POSTGRES_USER=agpt_user
- POSTGRES_PASSWORD=pass123
- POSTGRES_DB=agpt_marketplace
healthcheck:
test: pg_isready -U $$POSTGRES_USER -d $$POSTGRES_DB
interval: 10s
timeout: 5s
retries: 5
ports:
- "5433:5432"
market:
build:
context: .
dockerfile: Dockerfile
ports:
- "8015:8015"
depends_on:
postgres:
condition: service_healthy
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:8015/metrics"]
interval: 30s
timeout: 10s
retries: 3
prometheus:
image: prom/prometheus:latest
volumes:
- ./prometheus.yml:/etc/prometheus/prometheus.yml
command:
- '--config.file=/etc/prometheus/prometheus.yml'
ports:
- "9090:9090"
depends_on:
- market
grafana:
image: grafana/grafana:latest
ports:
- "9091:3000"
depends_on:
- prometheus

View File

@@ -0,0 +1,631 @@
// Package docs Code generated by swaggo/swag. DO NOT EDIT
package docs
import "github.com/swaggo/swag"
const docTemplate = `{
"schemes": {{ marshal .Schemes }},
"swagger": "2.0",
"info": {
"description": "{{escape .Description}}",
"title": "{{.Title}}",
"contact": {},
"version": "{{.Version}}"
},
"host": "{{.Host}}",
"basePath": "{{.BasePath}}",
"paths": {
"/agent/featured/{agent_id}": {
"get": {
"description": "Get the featured agent for a specific category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Agent Featured",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
},
"post": {
"description": "Set an agent as featured in a specific category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Set Agent Featured",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
},
"delete": {
"description": "Unset an agent as featured in a specific category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Unset Agent Featured",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agent/not-featured": {
"get": {
"description": "Get a list of agents that are not featured",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Not Featured Agents",
"parameters": [
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "page_size",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agent/submissions": {
"get": {
"description": "Get a list of agent submissions",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Agent Submissions",
"parameters": [
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "page_size",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agent/submissions/{agent_id}": {
"post": {
"description": "Review an agent submission",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Review Submission",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Status",
"name": "status",
"in": "query",
"required": true
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agents": {
"get": {
"description": "Get Agents",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Get Agents",
"parameters": [
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "pageSize",
"in": "query"
},
{
"type": "string",
"description": "Agent Name",
"name": "name",
"in": "query"
},
{
"type": "array",
"items": {
"type": "string"
},
"collectionFormat": "csv",
"description": "Keywords",
"name": "keywords",
"in": "query"
},
{
"type": "array",
"items": {
"type": "string"
},
"collectionFormat": "csv",
"description": "Categories",
"name": "categories",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"post": {
"description": "Submit an agent for review",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Submit Agent",
"parameters": [
{
"description": "Agent details",
"name": "agent",
"in": "body",
"required": true,
"schema": {
"$ref": "#/definitions/models.AddAgentRequest"
}
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agents/featured": {
"get": {
"description": "Get featured agents based on category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Get Featured Agents",
"parameters": [
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
},
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "pageSize",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
}
},
"/agents/search": {
"get": {
"description": "Search for agents based on query and categories",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Search Agents",
"parameters": [
{
"type": "string",
"description": "Search query",
"name": "q",
"in": "query",
"required": true
},
{
"type": "array",
"items": {
"type": "string"
},
"collectionFormat": "csv",
"description": "Categories",
"name": "categories",
"in": "query"
},
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "pageSize",
"in": "query"
},
{
"type": "string",
"description": "Sort by",
"name": "sortBy",
"in": "query"
},
{
"type": "string",
"description": "Sort order",
"name": "sortOrder",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
}
},
"/agents/{id}": {
"get": {
"description": "Get details of a specific agent by ID",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Get Agent Details",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "id",
"in": "path",
"required": true
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agents/{id}/download": {
"get": {
"description": "Download an agent file by ID",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Download Agent File",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "id",
"in": "path",
"required": true
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/categories": {
"get": {
"description": "Get a list of categories",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Categories",
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"type": "string"
}
}
}
}
}
}
},
"definitions": {
"models.AddAgentRequest": {
"type": "object",
"properties": {
"author": {
"type": "string"
},
"categories": {
"type": "array",
"items": {
"type": "string"
}
},
"graph": {
"$ref": "#/definitions/models.Graph"
},
"keywords": {
"type": "array",
"items": {
"type": "string"
}
}
}
},
"models.Agent": {
"type": "object",
"properties": {
"author": {
"type": "string"
},
"categories": {
"type": "array",
"items": {
"type": "string"
}
},
"description": {
"type": "string"
},
"graph": {
"$ref": "#/definitions/models.Graph"
},
"id": {
"type": "string"
},
"keywords": {
"type": "array",
"items": {
"type": "string"
}
},
"name": {
"type": "string"
}
}
},
"models.Graph": {
"type": "object",
"properties": {
"description": {
"type": "string"
},
"name": {
"type": "string"
}
}
}
}
}`
// SwaggerInfo holds exported Swagger Info so clients can modify it
var SwaggerInfo = &swag.Spec{
Version: "",
Host: "",
BasePath: "",
Schemes: []string{},
Title: "",
Description: "",
InfoInstanceName: "swagger",
SwaggerTemplate: docTemplate,
LeftDelim: "{{",
RightDelim: "}}",
}
func init() {
swag.Register(SwaggerInfo.InstanceName(), SwaggerInfo)
}

View File

@@ -0,0 +1,602 @@
{
"swagger": "2.0",
"info": {
"contact": {}
},
"paths": {
"/agent/featured/{agent_id}": {
"get": {
"description": "Get the featured agent for a specific category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Agent Featured",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
},
"post": {
"description": "Set an agent as featured in a specific category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Set Agent Featured",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
},
"delete": {
"description": "Unset an agent as featured in a specific category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Unset Agent Featured",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agent/not-featured": {
"get": {
"description": "Get a list of agents that are not featured",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Not Featured Agents",
"parameters": [
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "page_size",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agent/submissions": {
"get": {
"description": "Get a list of agent submissions",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Agent Submissions",
"parameters": [
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "page_size",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agent/submissions/{agent_id}": {
"post": {
"description": "Review an agent submission",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Review Submission",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "agent_id",
"in": "path",
"required": true
},
{
"type": "string",
"description": "Status",
"name": "status",
"in": "query",
"required": true
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agents": {
"get": {
"description": "Get Agents",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Get Agents",
"parameters": [
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "pageSize",
"in": "query"
},
{
"type": "string",
"description": "Agent Name",
"name": "name",
"in": "query"
},
{
"type": "array",
"items": {
"type": "string"
},
"collectionFormat": "csv",
"description": "Keywords",
"name": "keywords",
"in": "query"
},
{
"type": "array",
"items": {
"type": "string"
},
"collectionFormat": "csv",
"description": "Categories",
"name": "categories",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"post": {
"description": "Submit an agent for review",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Submit Agent",
"parameters": [
{
"description": "Agent details",
"name": "agent",
"in": "body",
"required": true,
"schema": {
"$ref": "#/definitions/models.AddAgentRequest"
}
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agents/featured": {
"get": {
"description": "Get featured agents based on category",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Get Featured Agents",
"parameters": [
{
"type": "string",
"description": "Category",
"name": "category",
"in": "query"
},
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "pageSize",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
}
},
"/agents/search": {
"get": {
"description": "Search for agents based on query and categories",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Search Agents",
"parameters": [
{
"type": "string",
"description": "Search query",
"name": "q",
"in": "query",
"required": true
},
{
"type": "array",
"items": {
"type": "string"
},
"collectionFormat": "csv",
"description": "Categories",
"name": "categories",
"in": "query"
},
{
"type": "integer",
"description": "Page number",
"name": "page",
"in": "query"
},
{
"type": "integer",
"description": "Page size",
"name": "pageSize",
"in": "query"
},
{
"type": "string",
"description": "Sort by",
"name": "sortBy",
"in": "query"
},
{
"type": "string",
"description": "Sort order",
"name": "sortOrder",
"in": "query"
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
}
},
"/agents/{id}": {
"get": {
"description": "Get details of a specific agent by ID",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Get Agent Details",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "id",
"in": "path",
"required": true
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/agents/{id}/download": {
"get": {
"description": "Download an agent file by ID",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Agents"
],
"summary": "Download Agent File",
"parameters": [
{
"type": "string",
"description": "Agent ID",
"name": "id",
"in": "path",
"required": true
}
],
"responses": {
"200": {
"description": "OK",
"schema": {
"$ref": "#/definitions/models.Agent"
}
}
}
}
},
"/categories": {
"get": {
"description": "Get a list of categories",
"consumes": [
"application/json"
],
"produces": [
"application/json"
],
"tags": [
"Admin"
],
"summary": "Get Categories",
"responses": {
"200": {
"description": "OK",
"schema": {
"type": "array",
"items": {
"type": "string"
}
}
}
}
}
}
},
"definitions": {
"models.AddAgentRequest": {
"type": "object",
"properties": {
"author": {
"type": "string"
},
"categories": {
"type": "array",
"items": {
"type": "string"
}
},
"graph": {
"$ref": "#/definitions/models.Graph"
},
"keywords": {
"type": "array",
"items": {
"type": "string"
}
}
}
},
"models.Agent": {
"type": "object",
"properties": {
"author": {
"type": "string"
},
"categories": {
"type": "array",
"items": {
"type": "string"
}
},
"description": {
"type": "string"
},
"graph": {
"$ref": "#/definitions/models.Graph"
},
"id": {
"type": "string"
},
"keywords": {
"type": "array",
"items": {
"type": "string"
}
},
"name": {
"type": "string"
}
}
},
"models.Graph": {
"type": "object",
"properties": {
"description": {
"type": "string"
},
"name": {
"type": "string"
}
}
}
}
}

View File

@@ -0,0 +1,395 @@
definitions:
models.AddAgentRequest:
properties:
author:
type: string
categories:
items:
type: string
type: array
graph:
$ref: '#/definitions/models.Graph'
keywords:
items:
type: string
type: array
type: object
models.Agent:
properties:
author:
type: string
categories:
items:
type: string
type: array
description:
type: string
graph:
$ref: '#/definitions/models.Graph'
id:
type: string
keywords:
items:
type: string
type: array
name:
type: string
type: object
models.Graph:
properties:
description:
type: string
name:
type: string
type: object
info:
contact: {}
paths:
/agent/featured/{agent_id}:
delete:
consumes:
- application/json
description: Unset an agent as featured in a specific category
parameters:
- description: Agent ID
in: path
name: agent_id
required: true
type: string
- description: Category
in: query
name: category
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Unset Agent Featured
tags:
- Admin
get:
consumes:
- application/json
description: Get the featured agent for a specific category
parameters:
- description: Agent ID
in: path
name: agent_id
required: true
type: string
- description: Category
in: query
name: category
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Get Agent Featured
tags:
- Admin
post:
consumes:
- application/json
description: Set an agent as featured in a specific category
parameters:
- description: Agent ID
in: path
name: agent_id
required: true
type: string
- description: Category
in: query
name: category
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Set Agent Featured
tags:
- Admin
/agent/not-featured:
get:
consumes:
- application/json
description: Get a list of agents that are not featured
parameters:
- description: Page number
in: query
name: page
type: integer
- description: Page size
in: query
name: page_size
type: integer
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Get Not Featured Agents
tags:
- Admin
/agent/submissions:
get:
consumes:
- application/json
description: Get a list of agent submissions
parameters:
- description: Page number
in: query
name: page
type: integer
- description: Page size
in: query
name: page_size
type: integer
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Get Agent Submissions
tags:
- Admin
/agent/submissions/{agent_id}:
post:
consumes:
- application/json
description: Review an agent submission
parameters:
- description: Agent ID
in: path
name: agent_id
required: true
type: string
- description: Status
in: query
name: status
required: true
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Review Submission
tags:
- Admin
/agents:
get:
consumes:
- application/json
description: Get Agents
parameters:
- description: Page number
in: query
name: page
type: integer
- description: Page size
in: query
name: pageSize
type: integer
- description: Agent Name
in: query
name: name
type: string
- collectionFormat: csv
description: Keywords
in: query
items:
type: string
name: keywords
type: array
- collectionFormat: csv
description: Categories
in: query
items:
type: string
name: categories
type: array
produces:
- application/json
responses:
"200":
description: OK
schema:
items:
$ref: '#/definitions/models.Agent'
type: array
summary: Get Agents
tags:
- Agents
post:
consumes:
- application/json
description: Submit an agent for review
parameters:
- description: Agent details
in: body
name: agent
required: true
schema:
$ref: '#/definitions/models.AddAgentRequest'
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Submit Agent
tags:
- Agents
/agents/{id}:
get:
consumes:
- application/json
description: Get details of a specific agent by ID
parameters:
- description: Agent ID
in: path
name: id
required: true
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Get Agent Details
tags:
- Agents
/agents/{id}/download:
get:
consumes:
- application/json
description: Download an agent file by ID
parameters:
- description: Agent ID
in: path
name: id
required: true
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
$ref: '#/definitions/models.Agent'
summary: Download Agent File
tags:
- Agents
/agents/featured:
get:
consumes:
- application/json
description: Get featured agents based on category
parameters:
- description: Category
in: query
name: category
type: string
- description: Page number
in: query
name: page
type: integer
- description: Page size
in: query
name: pageSize
type: integer
produces:
- application/json
responses:
"200":
description: OK
schema:
items:
$ref: '#/definitions/models.Agent'
type: array
summary: Get Featured Agents
tags:
- Agents
/agents/search:
get:
consumes:
- application/json
description: Search for agents based on query and categories
parameters:
- description: Search query
in: query
name: q
required: true
type: string
- collectionFormat: csv
description: Categories
in: query
items:
type: string
name: categories
type: array
- description: Page number
in: query
name: page
type: integer
- description: Page size
in: query
name: pageSize
type: integer
- description: Sort by
in: query
name: sortBy
type: string
- description: Sort order
in: query
name: sortOrder
type: string
produces:
- application/json
responses:
"200":
description: OK
schema:
items:
$ref: '#/definitions/models.Agent'
type: array
summary: Search Agents
tags:
- Agents
/categories:
get:
consumes:
- application/json
description: Get a list of categories
produces:
- application/json
responses:
"200":
description: OK
schema:
items:
type: string
type: array
summary: Get Categories
tags:
- Admin
swagger: "2.0"

84
rnd/rest-api-go/go.mod Normal file
View File

@@ -0,0 +1,84 @@
module github.com/swiftyos/market
go 1.23.1
require (
github.com/Depado/ginprom v1.8.1
github.com/gin-contrib/cors v1.7.2
github.com/gin-contrib/gzip v1.0.1
github.com/gin-contrib/zap v1.1.4
github.com/gin-gonic/gin v1.10.0
github.com/golang-jwt/jwt/v4 v4.5.0
github.com/google/uuid v1.4.0
github.com/jackc/pgx/v5 v5.7.1
github.com/spf13/viper v1.19.0
github.com/stretchr/testify v1.9.0
github.com/swaggo/files v1.0.1
github.com/swaggo/gin-swagger v1.6.0
github.com/swaggo/swag v1.16.3
go.uber.org/zap v1.27.0
)
require (
github.com/KyleBanks/depth v1.2.1 // indirect
github.com/beorn7/perks v1.0.1 // indirect
github.com/bytedance/sonic v1.12.2 // indirect
github.com/bytedance/sonic/loader v0.2.0 // indirect
github.com/cespare/xxhash/v2 v2.2.0 // indirect
github.com/cloudwego/base64x v0.1.4 // indirect
github.com/cloudwego/iasm v0.2.0 // indirect
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc // indirect
github.com/fsnotify/fsnotify v1.7.0 // indirect
github.com/gabriel-vasile/mimetype v1.4.5 // indirect
github.com/gin-contrib/sse v0.1.0 // indirect
github.com/go-openapi/jsonpointer v0.21.0 // indirect
github.com/go-openapi/jsonreference v0.21.0 // indirect
github.com/go-openapi/spec v0.21.0 // indirect
github.com/go-openapi/swag v0.23.0 // indirect
github.com/go-playground/locales v0.14.1 // indirect
github.com/go-playground/universal-translator v0.18.1 // indirect
github.com/go-playground/validator/v10 v10.22.1 // indirect
github.com/goccy/go-json v0.10.3 // indirect
github.com/hashicorp/hcl v1.0.0 // indirect
github.com/jackc/pgpassfile v1.0.0 // indirect
github.com/jackc/pgservicefile v0.0.0-20240606120523-5a60cdf6a761 // indirect
github.com/jackc/puddle/v2 v2.2.2 // indirect
github.com/josharian/intern v1.0.0 // indirect
github.com/json-iterator/go v1.1.12 // indirect
github.com/klauspost/cpuid/v2 v2.2.8 // indirect
github.com/leodido/go-urn v1.4.0 // indirect
github.com/magiconair/properties v1.8.7 // indirect
github.com/mailru/easyjson v0.7.7 // indirect
github.com/mattn/go-isatty v0.0.20 // indirect
github.com/matttproud/golang_protobuf_extensions/v2 v2.0.0 // indirect
github.com/mitchellh/mapstructure v1.5.0 // indirect
github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd // indirect
github.com/modern-go/reflect2 v1.0.2 // indirect
github.com/pelletier/go-toml/v2 v2.2.3 // indirect
github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect
github.com/prometheus/client_golang v1.18.0 // indirect
github.com/prometheus/client_model v0.5.0 // indirect
github.com/prometheus/common v0.45.0 // indirect
github.com/prometheus/procfs v0.12.0 // indirect
github.com/sagikazarmark/locafero v0.6.0 // indirect
github.com/sagikazarmark/slog-shim v0.1.0 // indirect
github.com/sourcegraph/conc v0.3.0 // indirect
github.com/spf13/afero v1.11.0 // indirect
github.com/spf13/cast v1.7.0 // indirect
github.com/spf13/pflag v1.0.5 // indirect
github.com/subosito/gotenv v1.6.0 // indirect
github.com/twitchyliquid64/golang-asm v0.15.1 // indirect
github.com/ugorji/go/codec v1.2.12 // indirect
go.uber.org/multierr v1.11.0 // indirect
golang.org/x/arch v0.10.0 // indirect
golang.org/x/crypto v0.27.0 // indirect
golang.org/x/exp v0.0.0-20240909161429-701f63a606c0 // indirect
golang.org/x/net v0.29.0 // indirect
golang.org/x/sync v0.8.0 // indirect
golang.org/x/sys v0.25.0 // indirect
golang.org/x/text v0.18.0 // indirect
golang.org/x/tools v0.25.0 // indirect
google.golang.org/protobuf v1.34.2 // indirect
gopkg.in/ini.v1 v1.67.0 // indirect
gopkg.in/yaml.v3 v3.0.1 // indirect
)

217
rnd/rest-api-go/go.sum Normal file
View File

@@ -0,0 +1,217 @@
github.com/Depado/ginprom v1.8.1 h1:lrQTddbRqlHq1j6SpJDySDumJlR7FEybzdX0PS3HXPc=
github.com/Depado/ginprom v1.8.1/go.mod h1:9Z+ahPJLSeMndDfnDTfiuBn2SKVAuL2yvihApWzof9A=
github.com/KyleBanks/depth v1.2.1 h1:5h8fQADFrWtarTdtDudMmGsC7GPbOAu6RVB3ffsVFHc=
github.com/KyleBanks/depth v1.2.1/go.mod h1:jzSb9d0L43HxTQfT+oSA1EEp2q+ne2uh6XgeJcm8brE=
github.com/appleboy/gofight/v2 v2.1.2 h1:VOy3jow4vIK8BRQJoC/I9muxyYlJ2yb9ht2hZoS3rf4=
github.com/appleboy/gofight/v2 v2.1.2/go.mod h1:frW+U1QZEdDgixycTj4CygQ48yLTUhplt43+Wczp3rw=
github.com/beorn7/perks v1.0.1 h1:VlbKKnNfV8bJzeqoa4cOKqO6bYr3WgKZxO8Z16+hsOM=
github.com/beorn7/perks v1.0.1/go.mod h1:G2ZrVWU2WbWT9wwq4/hrbKbnv/1ERSJQ0ibhJ6rlkpw=
github.com/bytedance/sonic v1.12.2 h1:oaMFuRTpMHYLpCntGca65YWt5ny+wAceDERTkT2L9lg=
github.com/bytedance/sonic v1.12.2/go.mod h1:B8Gt/XvtZ3Fqj+iSKMypzymZxw/FVwgIGKzMzT9r/rk=
github.com/bytedance/sonic/loader v0.1.1/go.mod h1:ncP89zfokxS5LZrJxl5z0UJcsk4M4yY2JpfqGeCtNLU=
github.com/bytedance/sonic/loader v0.2.0 h1:zNprn+lsIP06C/IqCHs3gPQIvnvpKbbxyXQP1iU4kWM=
github.com/bytedance/sonic/loader v0.2.0/go.mod h1:ncP89zfokxS5LZrJxl5z0UJcsk4M4yY2JpfqGeCtNLU=
github.com/cespare/xxhash/v2 v2.2.0 h1:DC2CZ1Ep5Y4k3ZQ899DldepgrayRUGE6BBZ/cd9Cj44=
github.com/cespare/xxhash/v2 v2.2.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs=
github.com/cloudwego/base64x v0.1.4 h1:jwCgWpFanWmN8xoIUHa2rtzmkd5J2plF/dnLS6Xd/0Y=
github.com/cloudwego/base64x v0.1.4/go.mod h1:0zlkT4Wn5C6NdauXdJRhSKRlJvmclQ1hhJgA0rcu/8w=
github.com/cloudwego/iasm v0.2.0 h1:1KNIy1I1H9hNNFEEH3DVnI4UujN+1zjpuk6gwHLTssg=
github.com/cloudwego/iasm v0.2.0/go.mod h1:8rXZaNYT2n95jn+zTI1sDr+IgcD2GVs0nlbbQPiEFhY=
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc h1:U9qPSI2PIWSS1VwoXQT9A3Wy9MM3WgvqSxFWenqJduM=
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
github.com/frankban/quicktest v1.14.6 h1:7Xjx+VpznH+oBnejlPUj8oUpdxnVs4f8XU8WnHkI4W8=
github.com/frankban/quicktest v1.14.6/go.mod h1:4ptaffx2x8+WTWXmUCuVU6aPUX1/Mz7zb5vbUoiM6w0=
github.com/fsnotify/fsnotify v1.7.0 h1:8JEhPFa5W2WU7YfeZzPNqzMP6Lwt7L2715Ggo0nosvA=
github.com/fsnotify/fsnotify v1.7.0/go.mod h1:40Bi/Hjc2AVfZrqy+aj+yEI+/bRxZnMJyTJwOpGvigM=
github.com/gabriel-vasile/mimetype v1.4.5 h1:J7wGKdGu33ocBOhGy0z653k/lFKLFDPJMG8Gql0kxn4=
github.com/gabriel-vasile/mimetype v1.4.5/go.mod h1:ibHel+/kbxn9x2407k1izTA1S81ku1z/DlgOW2QE0M4=
github.com/gin-contrib/cors v1.7.2 h1:oLDHxdg8W/XDoN/8zamqk/Drgt4oVZDvaV0YmvVICQw=
github.com/gin-contrib/cors v1.7.2/go.mod h1:SUJVARKgQ40dmrzgXEVxj2m7Ig1v1qIboQkPDTQ9t2E=
github.com/gin-contrib/gzip v1.0.1 h1:HQ8ENHODeLY7a4g1Au/46Z92bdGFl74OhxcZble9WJE=
github.com/gin-contrib/gzip v1.0.1/go.mod h1:njt428fdUNRvjuJf16tZMYZ2Yl+WQB53X5wmhDwXvC4=
github.com/gin-contrib/sse v0.1.0 h1:Y/yl/+YNO8GZSjAhjMsSuLt29uWRFHdHYUb5lYOV9qE=
github.com/gin-contrib/sse v0.1.0/go.mod h1:RHrZQHXnP2xjPF+u1gW/2HnVO7nvIa9PG3Gm+fLHvGI=
github.com/gin-contrib/zap v1.1.4 h1:xvxTybg6XBdNtcQLH3Tf0lFr4vhDkwzgLLrIGlNTqIo=
github.com/gin-contrib/zap v1.1.4/go.mod h1:7lgEpe91kLbeJkwBTPgtVBy4zMa6oSBEcvj662diqKQ=
github.com/gin-gonic/gin v1.10.0 h1:nTuyha1TYqgedzytsKYqna+DfLos46nTv2ygFy86HFU=
github.com/gin-gonic/gin v1.10.0/go.mod h1:4PMNQiOhvDRa013RKVbsiNwoyezlm2rm0uX/T7kzp5Y=
github.com/go-openapi/jsonpointer v0.21.0 h1:YgdVicSA9vH5RiHs9TZW5oyafXZFc6+2Vc1rr/O9oNQ=
github.com/go-openapi/jsonpointer v0.21.0/go.mod h1:IUyH9l/+uyhIYQ/PXVA41Rexl+kOkAPDdXEYns6fzUY=
github.com/go-openapi/jsonreference v0.21.0 h1:Rs+Y7hSXT83Jacb7kFyjn4ijOuVGSvOdF2+tg1TRrwQ=
github.com/go-openapi/jsonreference v0.21.0/go.mod h1:LmZmgsrTkVg9LG4EaHeY8cBDslNPMo06cago5JNLkm4=
github.com/go-openapi/spec v0.21.0 h1:LTVzPc3p/RzRnkQqLRndbAzjY0d0BCL72A6j3CdL9ZY=
github.com/go-openapi/spec v0.21.0/go.mod h1:78u6VdPw81XU44qEWGhtr982gJ5BWg2c0I5XwVMotYk=
github.com/go-openapi/swag v0.23.0 h1:vsEVJDUo2hPJ2tu0/Xc+4noaxyEffXNIs3cOULZ+GrE=
github.com/go-openapi/swag v0.23.0/go.mod h1:esZ8ITTYEsH1V2trKHjAN8Ai7xHb8RV+YSZ577vPjgQ=
github.com/go-playground/assert/v2 v2.2.0 h1:JvknZsQTYeFEAhQwI4qEt9cyV5ONwRHC+lYKSsYSR8s=
github.com/go-playground/assert/v2 v2.2.0/go.mod h1:VDjEfimB/XKnb+ZQfWdccd7VUvScMdVu0Titje2rxJ4=
github.com/go-playground/locales v0.14.1 h1:EWaQ/wswjilfKLTECiXz7Rh+3BjFhfDFKv/oXslEjJA=
github.com/go-playground/locales v0.14.1/go.mod h1:hxrqLVvrK65+Rwrd5Fc6F2O76J/NuW9t0sjnWqG1slY=
github.com/go-playground/universal-translator v0.18.1 h1:Bcnm0ZwsGyWbCzImXv+pAJnYK9S473LQFuzCbDbfSFY=
github.com/go-playground/universal-translator v0.18.1/go.mod h1:xekY+UJKNuX9WP91TpwSH2VMlDf28Uj24BCp08ZFTUY=
github.com/go-playground/validator/v10 v10.22.1 h1:40JcKH+bBNGFczGuoBYgX4I6m/i27HYW8P9FDk5PbgA=
github.com/go-playground/validator/v10 v10.22.1/go.mod h1:dbuPbCMFw/DrkbEynArYaCwl3amGuJotoKCe95atGMM=
github.com/goccy/go-json v0.10.3 h1:KZ5WoDbxAIgm2HNbYckL0se1fHD6rz5j4ywS6ebzDqA=
github.com/goccy/go-json v0.10.3/go.mod h1:oq7eo15ShAhp70Anwd5lgX2pLfOS3QCiwU/PULtXL6M=
github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg=
github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0=
github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI=
github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY=
github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg=
github.com/google/uuid v1.4.0 h1:MtMxsa51/r9yyhkyLsVeVt0B+BGQZzpQiTQ4eHZ8bc4=
github.com/google/uuid v1.4.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
github.com/hashicorp/hcl v1.0.0 h1:0Anlzjpi4vEasTeNFn2mLJgTSwt0+6sfsiTG8qcWGx4=
github.com/hashicorp/hcl v1.0.0/go.mod h1:E5yfLk+7swimpb2L/Alb/PJmXilQ/rhwaUYs4T20WEQ=
github.com/jackc/pgpassfile v1.0.0 h1:/6Hmqy13Ss2zCq62VdNG8tM1wchn8zjSGOBJ6icpsIM=
github.com/jackc/pgpassfile v1.0.0/go.mod h1:CEx0iS5ambNFdcRtxPj5JhEz+xB6uRky5eyVu/W2HEg=
github.com/jackc/pgservicefile v0.0.0-20240606120523-5a60cdf6a761 h1:iCEnooe7UlwOQYpKFhBabPMi4aNAfoODPEFNiAnClxo=
github.com/jackc/pgservicefile v0.0.0-20240606120523-5a60cdf6a761/go.mod h1:5TJZWKEWniPve33vlWYSoGYefn3gLQRzjfDlhSJ9ZKM=
github.com/jackc/pgx/v5 v5.7.1 h1:x7SYsPBYDkHDksogeSmZZ5xzThcTgRz++I5E+ePFUcs=
github.com/jackc/pgx/v5 v5.7.1/go.mod h1:e7O26IywZZ+naJtWWos6i6fvWK+29etgITqrqHLfoZA=
github.com/jackc/puddle/v2 v2.2.2 h1:PR8nw+E/1w0GLuRFSmiioY6UooMp6KJv0/61nB7icHo=
github.com/jackc/puddle/v2 v2.2.2/go.mod h1:vriiEXHvEE654aYKXXjOvZM39qJ0q+azkZFrfEOc3H4=
github.com/josharian/intern v1.0.0 h1:vlS4z54oSdjm0bgjRigI+G1HpF+tI+9rE5LLzOg8HmY=
github.com/josharian/intern v1.0.0/go.mod h1:5DoeVV0s6jJacbCEi61lwdGj/aVlrQvzHFFd8Hwg//Y=
github.com/json-iterator/go v1.1.12 h1:PV8peI4a0ysnczrg+LtxykD8LfKY9ML6u2jnxaEnrnM=
github.com/json-iterator/go v1.1.12/go.mod h1:e30LSqwooZae/UwlEbR2852Gd8hjQvJoHmT4TnhNGBo=
github.com/klauspost/cpuid/v2 v2.0.9/go.mod h1:FInQzS24/EEf25PyTYn52gqo7WaD8xa0213Md/qVLRg=
github.com/klauspost/cpuid/v2 v2.2.8 h1:+StwCXwm9PdpiEkPyzBXIy+M9KUb4ODm0Zarf1kS5BM=
github.com/klauspost/cpuid/v2 v2.2.8/go.mod h1:Lcz8mBdAVJIBVzewtcLocK12l3Y+JytZYpaMropDUws=
github.com/knz/go-libedit v1.10.1/go.mod h1:MZTVkCWyz0oBc7JOWP3wNAzd002ZbM/5hgShxwh4x8M=
github.com/kr/pretty v0.3.1 h1:flRD4NNwYAUpkphVc1HcthR4KEIFJ65n8Mw5qdRn3LE=
github.com/kr/pretty v0.3.1/go.mod h1:hoEshYVHaxMs3cyo3Yncou5ZscifuDolrwPKZanG3xk=
github.com/kr/text v0.2.0 h1:5Nx0Ya0ZqY2ygV366QzturHI13Jq95ApcVaJBhpS+AY=
github.com/kr/text v0.2.0/go.mod h1:eLer722TekiGuMkidMxC/pM04lWEeraHUUmBw8l2grE=
github.com/leodido/go-urn v1.4.0 h1:WT9HwE9SGECu3lg4d/dIA+jxlljEa1/ffXKmRjqdmIQ=
github.com/leodido/go-urn v1.4.0/go.mod h1:bvxc+MVxLKB4z00jd1z+Dvzr47oO32F/QSNjSBOlFxI=
github.com/magiconair/properties v1.8.7 h1:IeQXZAiQcpL9mgcAe1Nu6cX9LLw6ExEHKjN0VQdvPDY=
github.com/magiconair/properties v1.8.7/go.mod h1:Dhd985XPs7jluiymwWYZ0G4Z61jb3vdS329zhj2hYo0=
github.com/mailru/easyjson v0.7.7 h1:UGYAvKxe3sBsEDzO8ZeWOSlIQfWFlxbzLZe7hwFURr0=
github.com/mailru/easyjson v0.7.7/go.mod h1:xzfreul335JAWq5oZzymOObrkdz5UnU4kGfJJLY9Nlc=
github.com/mattn/go-isatty v0.0.20 h1:xfD0iDuEKnDkl03q4limB+vH+GxLEtL/jb4xVJSWWEY=
github.com/mattn/go-isatty v0.0.20/go.mod h1:W+V8PltTTMOvKvAeJH7IuucS94S2C6jfK/D7dTCTo3Y=
github.com/matttproud/golang_protobuf_extensions/v2 v2.0.0 h1:jWpvCLoY8Z/e3VKvlsiIGKtc+UG6U5vzxaoagmhXfyg=
github.com/matttproud/golang_protobuf_extensions/v2 v2.0.0/go.mod h1:QUyp042oQthUoa9bqDv0ER0wrtXnBruoNd7aNjkbP+k=
github.com/mitchellh/mapstructure v1.5.0 h1:jeMsZIYE/09sWLaz43PL7Gy6RuMjD2eJVyuac5Z2hdY=
github.com/mitchellh/mapstructure v1.5.0/go.mod h1:bFUtVrKA4DC2yAKiSyO/QUcy7e+RRV2QTWOzhPopBRo=
github.com/modern-go/concurrent v0.0.0-20180228061459-e0a39a4cb421/go.mod h1:6dJC0mAP4ikYIbvyc7fijjWJddQyLn8Ig3JB5CqoB9Q=
github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd h1:TRLaZ9cD/w8PVh93nsPXa1VrQ6jlwL5oN8l14QlcNfg=
github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd/go.mod h1:6dJC0mAP4ikYIbvyc7fijjWJddQyLn8Ig3JB5CqoB9Q=
github.com/modern-go/reflect2 v1.0.2 h1:xBagoLtFs94CBntxluKeaWgTMpvLxC4ur3nMaC9Gz0M=
github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjYzDa0/r8luk=
github.com/pelletier/go-toml/v2 v2.2.3 h1:YmeHyLY8mFWbdkNWwpr+qIL2bEqT0o95WSdkNHvL12M=
github.com/pelletier/go-toml/v2 v2.2.3/go.mod h1:MfCQTFTvCcUyyvvwm1+G6H/jORL20Xlb6rzQu9GuUkc=
github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U=
github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
github.com/prometheus/client_golang v1.18.0 h1:HzFfmkOzH5Q8L8G+kSJKUx5dtG87sewO+FoDDqP5Tbk=
github.com/prometheus/client_golang v1.18.0/go.mod h1:T+GXkCk5wSJyOqMIzVgvvjFDlkOQntgjkJWKrN5txjA=
github.com/prometheus/client_model v0.5.0 h1:VQw1hfvPvk3Uv6Qf29VrPF32JB6rtbgI6cYPYQjL0Qw=
github.com/prometheus/client_model v0.5.0/go.mod h1:dTiFglRmd66nLR9Pv9f0mZi7B7fk5Pm3gvsjB5tr+kI=
github.com/prometheus/common v0.45.0 h1:2BGz0eBc2hdMDLnO/8n0jeB3oPrt2D08CekT0lneoxM=
github.com/prometheus/common v0.45.0/go.mod h1:YJmSTw9BoKxJplESWWxlbyttQR4uaEcGyv9MZjVOJsY=
github.com/prometheus/procfs v0.12.0 h1:jluTpSng7V9hY0O2R9DzzJHYb2xULk9VTR1V1R/k6Bo=
github.com/prometheus/procfs v0.12.0/go.mod h1:pcuDEFsWDnvcgNzo4EEweacyhjeA9Zk3cnaOZAZEfOo=
github.com/rogpeppe/go-internal v1.11.0 h1:cWPaGQEPrBb5/AsnsZesgZZ9yb1OQ+GOISoDNXVBh4M=
github.com/rogpeppe/go-internal v1.11.0/go.mod h1:ddIwULY96R17DhadqLgMfk9H9tvdUzkipdSkR5nkCZA=
github.com/sagikazarmark/locafero v0.6.0 h1:ON7AQg37yzcRPU69mt7gwhFEBwxI6P9T4Qu3N51bwOk=
github.com/sagikazarmark/locafero v0.6.0/go.mod h1:77OmuIc6VTraTXKXIs/uvUxKGUXjE1GbemJYHqdNjX0=
github.com/sagikazarmark/slog-shim v0.1.0 h1:diDBnUNK9N/354PgrxMywXnAwEr1QZcOr6gto+ugjYE=
github.com/sagikazarmark/slog-shim v0.1.0/go.mod h1:SrcSrq8aKtyuqEI1uvTDTK1arOWRIczQRv+GVI1AkeQ=
github.com/sourcegraph/conc v0.3.0 h1:OQTbbt6P72L20UqAkXXuLOj79LfEanQ+YQFNpLA9ySo=
github.com/sourcegraph/conc v0.3.0/go.mod h1:Sdozi7LEKbFPqYX2/J+iBAM6HpqSLTASQIKqDmF7Mt0=
github.com/spf13/afero v1.11.0 h1:WJQKhtpdm3v2IzqG8VMqrr6Rf3UYpEF239Jy9wNepM8=
github.com/spf13/afero v1.11.0/go.mod h1:GH9Y3pIexgf1MTIWtNGyogA5MwRIDXGUr+hbWNoBjkY=
github.com/spf13/cast v1.7.0 h1:ntdiHjuueXFgm5nzDRdOS4yfT43P5Fnud6DH50rz/7w=
github.com/spf13/cast v1.7.0/go.mod h1:ancEpBxwJDODSW/UG4rDrAqiKolqNNh2DX3mk86cAdo=
github.com/spf13/pflag v1.0.5 h1:iy+VFUOCP1a+8yFto/drg2CJ5u0yRoB7fZw3DKv/JXA=
github.com/spf13/pflag v1.0.5/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg=
github.com/spf13/viper v1.19.0 h1:RWq5SEjt8o25SROyN3z2OrDB9l7RPd3lwTWU8EcEdcI=
github.com/spf13/viper v1.19.0/go.mod h1:GQUN9bilAbhU/jgc1bKs99f/suXKeUMct8Adx5+Ntkg=
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
github.com/stretchr/objx v0.4.0/go.mod h1:YvHI0jy2hoMjB+UWwv71VJQ9isScKT/TqJzVSSt89Yw=
github.com/stretchr/objx v0.5.0/go.mod h1:Yh+to48EsGEfYuaHDzXPcE3xhTkx73EhmCGUpEOglKo=
github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI=
github.com/stretchr/testify v1.7.0/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
github.com/stretchr/testify v1.7.1/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
github.com/stretchr/testify v1.8.0/go.mod h1:yNjHg4UonilssWZ8iaSj1OCr/vHnekPRkoO+kdMU+MU=
github.com/stretchr/testify v1.8.1/go.mod h1:w2LPCIKwWwSfY2zedu0+kehJoqGctiVI29o6fzry7u4=
github.com/stretchr/testify v1.9.0 h1:HtqpIVDClZ4nwg75+f6Lvsy/wHu+3BoSGCbBAcpTsTg=
github.com/stretchr/testify v1.9.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8C91i36aY=
github.com/subosito/gotenv v1.6.0 h1:9NlTDc1FTs4qu0DDq7AEtTPNw6SVm7uBMsUCUjABIf8=
github.com/subosito/gotenv v1.6.0/go.mod h1:Dk4QP5c2W3ibzajGcXpNraDfq2IrhjMIvMSWPKKo0FU=
github.com/swaggo/files v1.0.1 h1:J1bVJ4XHZNq0I46UU90611i9/YzdrF7x92oX1ig5IdE=
github.com/swaggo/files v1.0.1/go.mod h1:0qXmMNH6sXNf+73t65aKeB+ApmgxdnkQzVTAj2uaMUg=
github.com/swaggo/gin-swagger v1.6.0 h1:y8sxvQ3E20/RCyrXeFfg60r6H0Z+SwpTjMYsMm+zy8M=
github.com/swaggo/gin-swagger v1.6.0/go.mod h1:BG00cCEy294xtVpyIAHG6+e2Qzj/xKlRdOqDkvq0uzo=
github.com/swaggo/swag v1.16.3 h1:PnCYjPCah8FK4I26l2F/KQ4yz3sILcVUN3cTlBFA9Pg=
github.com/swaggo/swag v1.16.3/go.mod h1:DImHIuOFXKpMFAQjcC7FG4m3Dg4+QuUgUzJmKjI/gRk=
github.com/twitchyliquid64/golang-asm v0.15.1 h1:SU5vSMR7hnwNxj24w34ZyCi/FmDZTkS4MhqMhdFk5YI=
github.com/twitchyliquid64/golang-asm v0.15.1/go.mod h1:a1lVb/DtPvCB8fslRZhAngC2+aY1QWCk3Cedj/Gdt08=
github.com/ugorji/go/codec v1.2.12 h1:9LC83zGrHhuUA9l16C9AHXAqEV/2wBQ4nkvumAE65EE=
github.com/ugorji/go/codec v1.2.12/go.mod h1:UNopzCgEMSXjBc6AOMqYvWC1ktqTAfzJZUZgYf6w6lg=
github.com/yuin/goldmark v1.4.13/go.mod h1:6yULJ656Px+3vBD8DxQVa3kxgyrAnzto9xy5taEt/CY=
go.uber.org/goleak v1.3.0 h1:2K3zAYmnTNqV73imy9J1T3WC+gmCePx2hEGkimedGto=
go.uber.org/goleak v1.3.0/go.mod h1:CoHD4mav9JJNrW/WLlf7HGZPjdw8EucARQHekz1X6bE=
go.uber.org/multierr v1.11.0 h1:blXXJkSxSSfBVBlC76pxqeO+LN3aDfLQo+309xJstO0=
go.uber.org/multierr v1.11.0/go.mod h1:20+QtiLqy0Nd6FdQB9TLXag12DsQkrbs3htMFfDN80Y=
go.uber.org/zap v1.27.0 h1:aJMhYGrd5QSmlpLMr2MftRKl7t8J8PTZPA732ud/XR8=
go.uber.org/zap v1.27.0/go.mod h1:GB2qFLM7cTU87MWRP2mPIjqfIDnGu+VIO4V/SdhGo2E=
golang.org/x/arch v0.10.0 h1:S3huipmSclq3PJMNe76NGwkBR504WFkQ5dhzWzP8ZW8=
golang.org/x/arch v0.10.0/go.mod h1:FEVrYAQjsQXMVJ1nsMoVVXPZg6p2JE2mx8psSWTDQys=
golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w=
golang.org/x/crypto v0.0.0-20210921155107-089bfa567519/go.mod h1:GvvjBRRGRdwPK5ydBHafDWAxML/pGHZbMvKqRZ5+Abc=
golang.org/x/crypto v0.27.0 h1:GXm2NjJrPaiv/h1tb2UH8QfgC/hOf/+z0p6PT8o1w7A=
golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt70=
golang.org/x/exp v0.0.0-20240909161429-701f63a606c0 h1:e66Fs6Z+fZTbFBAxKfP3PALWBtpfqks2bwGcexMxgtk=
golang.org/x/exp v0.0.0-20240909161429-701f63a606c0/go.mod h1:2TbTHSBQa924w8M6Xs1QcRcFwyucIwBGpK1p2f1YFFY=
golang.org/x/mod v0.6.0-dev.0.20220419223038-86c51ed26bb4/go.mod h1:jJ57K6gSWd91VN4djpZkiMVwK6gcyfeH4XE8wZrZaV4=
golang.org/x/mod v0.21.0 h1:vvrHzRwRfVKSiLrG+d4FMl/Qi4ukBCE6kZlTUkDYRT0=
golang.org/x/mod v0.21.0/go.mod h1:6SkKJ3Xj0I0BrPOZoBy3bdMptDDU9oJrpohJ3eWZ1fY=
golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20220722155237-a158d28d115b/go.mod h1:XRhObCWvk6IyKnWLug+ECip1KBveYUHfp+8e9klMJ9c=
golang.org/x/net v0.7.0/go.mod h1:2Tu9+aMcznHK/AK1HMvgo6xiTLG5rD5rZLDS+rp2Bjs=
golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo=
golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0=
golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
golang.org/x/sync v0.0.0-20220722155255-886fb9371eb4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
golang.org/x/sync v0.8.0 h1:3NFvSEYkUoMifnESzZl15y791HH1qU2xm6eCJU5ZPXQ=
golang.org/x/sync v0.8.0/go.mod h1:Czt+wKu1gCyEFDUtn0jG5QVvpJ6rzVqr5aXyt9drQfk=
golang.org/x/sys v0.0.0-20190215142949-d0b11bdaac8a/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
golang.org/x/sys v0.0.0-20201119102817-f84b799fce68/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
golang.org/x/sys v0.0.0-20210615035016-665e8c7367d1/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
golang.org/x/sys v0.0.0-20220520151302-bc2c85ada10a/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
golang.org/x/sys v0.0.0-20220722155257-8c9f86f7a55f/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
golang.org/x/sys v0.5.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
golang.org/x/sys v0.6.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34=
golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA=
golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo=
golang.org/x/term v0.0.0-20210927222741-03fcf44c2211/go.mod h1:jbD1KX2456YbFQfuXm/mYQcufACuNUgVhRMnK/tPxf8=
golang.org/x/term v0.5.0/go.mod h1:jMB1sMXY+tzblOD4FWmEbocvup2/aLOaQEp7JmGp78k=
golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
golang.org/x/text v0.3.7/go.mod h1:u+2+/6zg+i71rQMx5EYifcz6MCKuco9NR6JIITiCfzQ=
golang.org/x/text v0.7.0/go.mod h1:mrYo+phRRbMaCq/xk9113O4dZlRixOauAjOtrjsXDZ8=
golang.org/x/text v0.18.0 h1:XvMDiNzPAl0jr17s6W9lcaIhGUfUORdGCNsuLmPG224=
golang.org/x/text v0.18.0/go.mod h1:BuEKDfySbSR4drPmRPG/7iBdf8hvFMuRexcpahXilzY=
golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ=
golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo=
golang.org/x/tools v0.1.12/go.mod h1:hNGJHUnrk76NpqgfD5Aqm5Crs+Hm0VOH/i9J2+nxYbc=
golang.org/x/tools v0.25.0 h1:oFU9pkj/iJgs+0DT+VMHrx+oBKs/LJMV+Uvg78sl+fE=
golang.org/x/tools v0.25.0/go.mod h1:/vtpO8WL1N9cQC3FN5zPqb//fRXskFHbLKk4OW1Q7rg=
golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg=
google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw=
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk=
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c/go.mod h1:JHkPIbrfpd72SG/EVd6muEfDQjcINNoR0C8j2r3qZ4Q=
gopkg.in/ini.v1 v1.67.0 h1:Dgnx+6+nfE+IfzjUEISNeydPJh9AXNNsWbGP9KzCsOA=
gopkg.in/ini.v1 v1.67.0/go.mod h1:pNLf8WUiyNEtQjuu5G5vTm06TEv9tsIgeAvK8hOrP4k=
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
gopkg.in/yaml.v3 v3.0.1/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
nullprogram.com/x/optparse v1.0.0/go.mod h1:KdyPE+Igbe0jQUrVfMqDMeJQIJZEuyV7pjYmp6pbG50=

View File

@@ -0,0 +1,282 @@
package handlers
import (
"net/http"
"strconv"
"github.com/gin-gonic/gin"
"github.com/jackc/pgx/v5/pgxpool"
"github.com/swiftyos/market/database"
"github.com/swiftyos/market/models"
"go.uber.org/zap"
)
func requireAdminUser() gin.HandlerFunc {
return func(c *gin.Context) {
user, exists := c.Get("user")
if !exists {
c.AbortWithStatusJSON(http.StatusUnauthorized, gin.H{"error": "User not found in context"})
return
}
userModel, ok := user.(models.User)
if !ok {
c.AbortWithStatusJSON(http.StatusInternalServerError, gin.H{"error": "Invalid user model"})
return
}
if userModel.Role != "admin" {
c.AbortWithStatusJSON(http.StatusForbidden, gin.H{"error": "Admin access required"})
return
}
c.Next()
}
}
// @BasePath /api/v1/marketplace/admin
// CreateAgentEntry godoc
// @Summary Create Agent Entry
// @Description Create a new agent entry
// @Tags Admin
// @Accept json
// @Produce json
// @Param request body models.AddAgentRequest true "Agent details"
// @Success 200 {object} models.Agent
// @Router /agents [post]
func CreateAgentEntry(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
var request models.AddAgentRequest
if err := c.ShouldBindJSON(&request); err != nil {
c.JSON(http.StatusBadRequest, gin.H{"error": err.Error()})
return
}
agent, err := database.CreateAgentEntry(c.Request.Context(), db, models.Agent{
Name: request.Graph.Name,
Description: request.Graph.Description,
Author: request.Author,
Keywords: request.Keywords,
Categories: request.Categories,
Graph: request.Graph,
})
if err != nil {
c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()})
return
}
c.JSON(http.StatusOK, agent)
}
}
// SetAgentFeatured godoc
// @Summary Set Agent Featured
// @Description Set an agent as featured in a specific category
// @Tags Admin
// @Accept json
// @Produce json
// @Param agent_id path string true "Agent ID"
// @Param category query string false "Category"
// @Success 200 {object} models.Agent
// @Router /agent/featured/{agent_id} [post]
func SetAgentFeatured(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
agentID := c.Param("agent_id")
categories := c.QueryArray("categories")
if len(categories) == 0 {
categories = []string{"featured"}
}
featuredAgent, err := database.SetAgentFeatured(c.Request.Context(), db, agentID, true, categories)
if err != nil {
c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()})
return
}
c.JSON(http.StatusOK, featuredAgent)
}
}
// GetAgentFeatured godoc
// @Summary Get Agent Featured
// @Description Get the featured agent for a specific category
// @Tags Admin
// @Accept json
// @Produce json
// @Param agent_id path string true "Agent ID"
// @Param category query string false "Category"
// @Success 200 {object} models.Agent
// @Router /agent/featured/{agent_id} [get]
func GetAgentFeatured(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
agentID := c.Param("agent_id")
featuredAgent, err := database.GetAgentFeatured(c.Request.Context(), db, agentID)
if err != nil {
c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()})
return
}
if featuredAgent == nil {
c.JSON(http.StatusNotFound, gin.H{"message": "Featured agent not found"})
return
}
c.JSON(http.StatusOK, featuredAgent)
}
}
// UnsetAgentFeatured godoc
// @Summary Unset Agent Featured
// @Description Unset an agent as featured in a specific category
// @Tags Admin
// @Accept json
// @Produce json
// @Param agent_id path string true "Agent ID"
// @Param category query string false "Category"
// @Success 200 {object} models.Agent
// @Router /agent/featured/{agent_id} [delete]
func UnsetAgentFeatured(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
agentID := c.Param("agent_id")
category := c.DefaultQuery("category", "featured")
featuredAgent, err := database.RemoveFeaturedCategory(c.Request.Context(), db, agentID, category)
if err != nil {
c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()})
return
}
if featuredAgent == nil {
c.JSON(http.StatusNotFound, gin.H{"message": "Featured agent not found"})
return
}
c.JSON(http.StatusOK, featuredAgent)
}
}
// GetNotFeaturedAgents godoc
// @Summary Get Not Featured Agents
// @Description Get a list of agents that are not featured
// @Tags Admin
// @Accept json
// @Produce json
// @Param page query int false "Page number"
// @Param page_size query int false "Page size"
// @Success 200 {object} models.Agent
// @Router /agent/not-featured [get]
func GetNotFeaturedAgents(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
page, _ := strconv.Atoi(c.DefaultQuery("page", "1"))
pageSize, _ := strconv.Atoi(c.DefaultQuery("page_size", "10"))
agents, err := database.GetNotFeaturedAgents(c.Request.Context(), db, page, pageSize)
if err != nil {
c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()})
return
}
c.JSON(http.StatusOK, gin.H{
"agents": agents,
"total_count": len(agents),
"page": page,
"page_size": pageSize,
})
}
}
// GetAgentSubmissions godoc
// @Summary Get Agent Submissions
// @Description Get a list of agent submissions
// @Tags Admin
// @Accept json
// @Produce json
// @Param page query int false "Page number"
// @Param page_size query int false "Page size"
// @Success 200 {object} models.Agent
// @Router /agent/submissions [get]
func GetAgentSubmissions(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
// TODO: Implement GetAgentSubmissions
c.JSON(http.StatusNotImplemented, gin.H{"message": "Not Implemented: GetAgentSubmissions"})
}
}
// ReviewSubmission godoc
// @Summary Review Submission
// @Description Review an agent submission
// @Tags Admin
// @Accept json
// @Produce json
// @Param agent_id path string true "Agent ID"
// @Param status query string true "Status"
// @Success 200 {object} models.Agent
// @Router /agent/submissions/{agent_id} [post]
func ReviewSubmission(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
requireAdminUser()(c)
if c.IsAborted() {
return
}
// TODO: Implement ReviewSubmission
c.JSON(http.StatusNotImplemented, gin.H{"message": "Not Implemented: ReviewSubmission"})
}
}
// GetCategories godoc
// @Summary Get Categories
// @Description Get a list of categories
// @Tags Admin
// @Accept json
// @Produce json
// @Success 200 {array} string
// @Router /categories [get]
func GetCategories(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
if c.IsAborted() {
return
}
categories, err := database.GetAllCategories(c.Request.Context(), db)
if err != nil {
c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()})
return
}
c.JSON(http.StatusOK, categories)
}
}

View File

@@ -0,0 +1,369 @@
package handlers
import (
"fmt"
"net/http"
"strconv"
"github.com/gin-gonic/gin"
"github.com/jackc/pgx/v5/pgxpool"
"go.uber.org/zap"
"github.com/swiftyos/market/database"
"github.com/swiftyos/market/models"
"github.com/swiftyos/market/utils"
)
// @BasePath /api/v1/marketplace
// GetAgents godoc
// @Summary Get Agents
// @Schemes
// @Description Get Agents
// @Tags Agents
// @Accept json
// @Produce json
// @Param page query int false "Page number"
// @Param pageSize query int false "Page size"
// @Param name query string false "Agent Name"
// @Param keywords query []string false "Keywords"
// @Param categories query []string false "Categories"
// @Success 200 {array} models.Agent
// @Router /agents [get]
func GetAgents(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "GetAgents")).With(zap.String("file", "handlers/agents.go"))
logger.Info("Get Agents Request Started")
// Get pagination parameters from context
page := getPageFromContext(c.Request.Context())
pageSize := getPageSizeFromContext(c.Request.Context())
// Get filter parameters from context
name := getNameFromContext(c.Request.Context())
keywords := getKeywordsFromContext(c.Request.Context())
categories := getCategoriesFromContext(c.Request.Context())
logger.Debug("Request parameters",
zap.Int("page", page),
zap.Int("pageSize", pageSize),
zap.String("name", utils.StringOrNil(name)),
zap.String("keywords", utils.StringOrNil(keywords)),
zap.String("categories", utils.StringOrNil(categories)))
agents, err := database.GetAgents(c.Request.Context(), db, log_ctx, page, pageSize, name, keywords, categories)
if err != nil {
logger.Error("Database requested returned error!", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to fetch agents"})
return
}
c.JSON(http.StatusOK, agents)
}
}
// @Summary Submit Agent
// @Description Submit an agent for review
// @Tags Agents
// @Accept json
// @Produce json
// @Param agent body models.AddAgentRequest true "Agent details"
// @Success 200 {object} models.Agent
// @Router /agents [post]
func SubmitAgent(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "SubmitAgent"))
var request models.AddAgentRequest
logger.Debug("Add Agent Request body", zap.Any("request", request))
if err := c.ShouldBindJSON(&request); err != nil {
c.JSON(http.StatusBadRequest, gin.H{"error": err.Error()})
return
}
user, exists := c.Get("user")
if !exists {
c.JSON(http.StatusUnauthorized, gin.H{"error": "User not authenticated"})
return
}
agent, err := database.SubmitAgent(c.Request.Context(), db, request, user)
if err != nil {
logger.Error("Failed to submit agent", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to submit agent"})
return
}
c.JSON(http.StatusOK, agent)
}
}
// @Summary Get Agent Details
// @Description Get details of a specific agent by ID
// @Tags Agents
// @Accept json
// @Produce json
// @Param id path string true "Agent ID"
// @Success 200 {object} models.Agent
// @Router /agents/{id} [get]
func GetAgentDetails(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "GetAgentDetails"))
agentID := c.Param("id")
logger.Debug("Agent ID", zap.String("agentID", agentID))
if agentID == "" {
logger.Error("Agent ID is required")
c.JSON(http.StatusBadRequest, gin.H{"error": "Agent ID is required"})
return
}
agent, err := database.GetAgentDetails(c.Request.Context(), db, agentID)
if err != nil {
if err.Error() == "agent not found" {
logger.Error("Agent not found", zap.String("agentID", agentID))
c.JSON(http.StatusNotFound, gin.H{"error": "Agent not found"})
return
}
logger.Error("Failed to fetch agent details", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to fetch agent details"})
return
}
c.JSON(http.StatusOK, agent)
}
}
// @Summary Download Agent
// @Description Download an agent by ID
// @Tags Agents
// @Accept json
// @Produce json
// @Param id path string true "Agent ID"
// @Success 200 {object} models.Agent
// @Router /agents/{id}/download [get]
func DownloadAgent(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "DownloadAgent"))
agentID := c.Param("id")
if agentID == "" {
logger.Error("Agent ID is required")
c.JSON(http.StatusBadRequest, gin.H{"error": "Agent ID is required"})
return
}
agent, err := database.GetAgentDetails(c.Request.Context(), db, agentID)
if err != nil {
if err.Error() == "agent not found" {
logger.Error("Agent not found", zap.String("agentID", agentID))
c.JSON(http.StatusNotFound, gin.H{"error": "Agent not found"})
return
}
logger.Error("Failed to fetch agent details", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to fetch agent details"})
return
}
err = database.IncrementDownloadCount(c.Request.Context(), db, agentID)
if err != nil {
logger.Error("Failed to increment download count", zap.Error(err))
// Continue with the download even if the count update fails
}
c.JSON(http.StatusOK, agent)
}
}
// @Summary Download Agent File
// @Description Download an agent file by ID
// @Tags Agents
// @Accept json
// @Produce json
// @Param id path string true "Agent ID"
// @Success 200 {object} models.Agent
// @Router /agents/{id}/download [get]
func DownloadAgentFile(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "DownloadAgentFile"))
agentID := c.Param("id")
if agentID == "" {
logger.Error("Agent ID is required")
c.JSON(http.StatusBadRequest, gin.H{"error": "Agent ID is required"})
return
}
agentFile, err := database.GetAgentFile(c.Request.Context(), db, agentID)
if err != nil {
if err.Error() == "agent not found" {
logger.Error("Agent not found", zap.String("agentID", agentID))
c.JSON(http.StatusNotFound, gin.H{"error": "Agent not found"})
return
}
logger.Error("Failed to fetch agent file", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to fetch agent file"})
return
}
err = database.IncrementDownloadCount(c.Request.Context(), db, agentID)
if err != nil {
logger.Error("Failed to increment download count", zap.Error(err))
// Continue with the download even if the count update fails
}
fileName := fmt.Sprintf("agent_%s.json", agentID)
c.Header("Content-Disposition", fmt.Sprintf("attachment; filename=%s", fileName))
c.JSON(http.StatusOK, agentFile.Graph)
}
}
func TopAgentsByDownloads(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "TopAgentsByDownloads"))
logger.Info("Handling request for top agents by downloads")
page, err := strconv.Atoi(c.DefaultQuery("page", "1"))
if err != nil || page < 1 {
logger.Error("Invalid page number", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid page number"})
return
}
pageSize, err := strconv.Atoi(c.DefaultQuery("pageSize", "10"))
if err != nil || pageSize < 1 {
logger.Error("Invalid page size", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid page size"})
return
}
agents, totalCount, err := database.GetTopAgentsByDownloads(c.Request.Context(), db, page, pageSize)
if err != nil {
logger.Error("Failed to fetch top agents", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to fetch top agents"})
return
}
logger.Info("Successfully fetched top agents", zap.Int("count", len(agents)), zap.Int("totalCount", totalCount))
c.JSON(http.StatusOK, gin.H{
"agents": agents,
"totalCount": totalCount,
"page": page,
"pageSize": pageSize,
})
}
}
// @Summary Get Featured Agents
// @Description Get featured agents based on category
// @Tags Agents
// @Accept json
// @Produce json
// @Param category query string false "Category"
// @Param page query int false "Page number"
// @Param pageSize query int false "Page size"
// @Success 200 {array} models.Agent
// @Router /agents/featured [get]
func GetFeaturedAgents(db *pgxpool.Pool, logger *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := logger.With(zap.String("function", "GetFeaturedAgents"))
logger.Info("Handling request for featured agents")
category := c.Query("category")
if category == "" {
logger.Debug("No category specified, fetching all featured agents")
} else {
logger.Debug("Fetching featured agents for category", zap.String("category", category))
}
page, err := strconv.Atoi(c.DefaultQuery("page", "1"))
if err != nil || page < 1 {
logger.Error("Invalid page number", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid page number"})
return
}
pageSize, err := strconv.Atoi(c.DefaultQuery("pageSize", "10"))
if err != nil || pageSize < 1 {
logger.Error("Invalid page size", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid page size"})
return
}
agents, totalCount, err := database.GetFeaturedAgents(c.Request.Context(), db, category, page, pageSize)
if err != nil {
logger.Error("Failed to fetch featured agents", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to fetch featured agents"})
return
}
logger.Info("Successfully fetched featured agents", zap.Int("count", len(agents)), zap.Int("totalCount", totalCount))
c.JSON(http.StatusOK, gin.H{
"agents": agents,
"totalCount": totalCount,
"page": page,
"pageSize": pageSize,
})
}
}
// @Summary Search Agents
// @Description Search for agents based on query and categories
// @Tags Agents
// @Accept json
// @Produce json
// @Param q query string true "Search query"
// @Param categories query []string false "Categories"
// @Param page query int false "Page number"
// @Param pageSize query int false "Page size"
// @Param sortBy query string false "Sort by"
// @Param sortOrder query string false "Sort order"
// @Success 200 {array} models.Agent
// @Router /agents/search [get]
func SearchAgents(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := log_ctx.With(zap.String("function", "Search"))
logger.Info("Handling search request")
query := c.Query("q")
if query == "" {
logger.Error("Search query is required")
c.JSON(http.StatusBadRequest, gin.H{"error": "Search query is required"})
return
}
categories := c.QueryArray("categories")
page, err := strconv.Atoi(c.DefaultQuery("page", "1"))
if err != nil || page < 1 {
logger.Error("Invalid page number", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid page number"})
return
}
pageSize, err := strconv.Atoi(c.DefaultQuery("pageSize", "10"))
if err != nil || pageSize < 1 {
logger.Error("Invalid page size", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid page size"})
return
}
sortBy := c.DefaultQuery("sortBy", "rank")
sortOrder := c.DefaultQuery("sortOrder", "DESC")
agents, err := database.Search(c.Request.Context(), db, query, categories, page, pageSize, sortBy, sortOrder)
if err != nil {
logger.Error("Failed to perform search", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to perform search"})
return
}
logger.Info("Successfully performed search", zap.Int("resultCount", len(agents)))
c.JSON(http.StatusOK, gin.H{
"agents": agents,
"page": page,
"pageSize": pageSize,
})
}
}

View File

@@ -0,0 +1,41 @@
package handlers
import (
"net/http"
"github.com/gin-gonic/gin"
"github.com/jackc/pgx/v5/pgxpool"
"github.com/swiftyos/market/database"
"go.uber.org/zap"
"github.com/swiftyos/market/models"
)
func AgentInstalled(db *pgxpool.Pool, log_ctx *zap.Logger) gin.HandlerFunc {
return func(c *gin.Context) {
logger := zap.L().With(zap.String("function", "AgentInstalled"))
var eventData models.InstallTracker
if err := c.ShouldBindJSON(&eventData); err != nil {
logger.Error("Failed to bind JSON", zap.Error(err))
c.JSON(http.StatusBadRequest, gin.H{"error": "Invalid request body"})
return
}
err := database.CreateAgentInstalledEvent(c.Request.Context(), db, models.InstallTracker{
MarketplaceAgentID: eventData.MarketplaceAgentID,
InstalledAgentID: eventData.InstalledAgentID,
InstallationLocation: eventData.InstallationLocation,
})
if err != nil {
logger.Error("Failed to process agent installed event", zap.Error(err))
c.JSON(http.StatusInternalServerError, gin.H{"error": "Failed to process agent installed event"})
return
}
logger.Info("Agent installed event processed successfully",
zap.String("marketplaceAgentID", eventData.MarketplaceAgentID),
zap.String("installedAgentID", eventData.InstalledAgentID),
zap.String("installationLocation", string(eventData.InstallationLocation)))
c.JSON(http.StatusOK, gin.H{"message": "Agent installed event processed successfully"})
}
}

View File

@@ -0,0 +1,95 @@
package handlers
import (
"context"
"strconv"
)
const pageKey string = "page"
func getPageFromContext(ctx context.Context) int {
defaultPage := 1
if ctx == nil {
return defaultPage
}
pageValue := ctx.Value(pageKey)
if pageValue == nil {
return defaultPage
}
// Type assertion to check if the value is an int
if page, ok := pageValue.(int); ok {
if page < 1 {
return defaultPage
}
return page
}
// If it's not an int, try to convert from string
if pageStr, ok := pageValue.(string); ok {
page, err := strconv.Atoi(pageStr)
if err != nil || page < 1 {
return defaultPage
}
return page
}
return defaultPage
}
const pageSizeKey string = "page_size"
func getPageSizeFromContext(ctx context.Context) int {
pageSizeValue := ctx.Value(pageSizeKey)
if pageSizeValue == nil {
return 10
}
if pageSizeValue, ok := pageSizeValue.(int); ok {
if pageSizeValue < 1 {
return 10
}
return pageSizeValue
}
return 10
}
const nameKey string = "name"
func getNameFromContext(ctx context.Context) *string {
nameValue := ctx.Value(nameKey)
if nameValue == nil {
return nil
}
if nameValue, ok := nameValue.(string); ok {
return &nameValue
}
return nil
}
const keywordsKey string = "keywords"
func getKeywordsFromContext(ctx context.Context) *string {
keywordsValue := ctx.Value(keywordsKey)
if keywordsValue == nil {
return nil
}
if keywordsValue, ok := keywordsValue.(string); ok {
return &keywordsValue
}
return nil
}
const categoriesKey string = "categories"
func getCategoriesFromContext(ctx context.Context) *string {
categoriesValue := ctx.Value(categoriesKey)
if categoriesValue == nil {
return nil
}
if categoriesValue, ok := categoriesValue.(string); ok {
return &categoriesValue
}
return nil
}

View File

@@ -0,0 +1,151 @@
package handlers
import (
"context"
"testing"
"github.com/stretchr/testify/assert"
"go.uber.org/zap"
"go.uber.org/zap/zaptest"
)
func TestGetPageFromContext_ValidPage(t *testing.T) {
ctx := context.WithValue(context.Background(), pageKey, 5)
result := getPageFromContext(ctx)
assert.Equal(t, 5, result)
}
func TestGetPageFromContext_InvalidPageZero(t *testing.T) {
ctx := context.WithValue(context.Background(), pageKey, 0)
result := getPageFromContext(ctx)
assert.Equal(t, 1, result)
}
func TestGetPageFromContext_NoPageValue(t *testing.T) {
ctx := context.Background()
result := getPageFromContext(ctx)
assert.Equal(t, 1, result)
}
func TestGetPageFromContext_InvalidPageNegative(t *testing.T) {
ctx := context.WithValue(context.Background(), pageKey, -1)
result := getPageFromContext(ctx)
assert.Equal(t, 1, result)
}
func TestGetPageFromContext_InvalidType(t *testing.T) {
ctx := context.WithValue(context.Background(), pageKey, "not an int")
result := getPageFromContext(ctx)
assert.Equal(t, 1, result)
}
func TestGetPageSizeFromContext_ValidPageSize(t *testing.T) {
ctx := context.WithValue(context.Background(), pageSizeKey, 20)
result := getPageSizeFromContext(ctx)
assert.Equal(t, 20, result)
}
func TestGetPageSizeFromContext_InvalidPageSizeNegative(t *testing.T) {
ctx := context.WithValue(context.Background(), pageSizeKey, -1)
result := getPageSizeFromContext(ctx)
assert.Equal(t, 10, result)
}
func TestGetPageSizeFromContext_InvalidPageSizeZero(t *testing.T) {
ctx := context.WithValue(context.Background(), pageSizeKey, 0)
result := getPageSizeFromContext(ctx)
assert.Equal(t, 10, result)
}
func TestGetPageSizeFromContext_NoPageSizeValue(t *testing.T) {
ctx := context.Background()
result := getPageSizeFromContext(ctx)
assert.Equal(t, 10, result)
}
func TestGetPageSizeFromContext_InvalidType(t *testing.T) {
ctx := context.WithValue(context.Background(), pageSizeKey, "not an int")
result := getPageSizeFromContext(ctx)
assert.Equal(t, 10, result)
}
func TestGetNameFromContext_ValidName(t *testing.T) {
ctx := context.WithValue(context.Background(), nameKey, "Test Name")
result := getNameFromContext(ctx)
assert.Equal(t, strPtr("Test Name"), result)
}
func TestGetNameFromContext_EmptyString(t *testing.T) {
ctx := context.WithValue(context.Background(), nameKey, "")
result := getNameFromContext(ctx)
assert.Equal(t, strPtr(""), result)
}
func TestGetNameFromContext_NoNameValue(t *testing.T) {
ctx := context.Background()
result := getNameFromContext(ctx)
assert.Nil(t, result)
}
func TestGetNameFromContext_InvalidType(t *testing.T) {
ctx := context.WithValue(context.Background(), nameKey, 123)
result := getNameFromContext(ctx)
assert.Nil(t, result)
}
func TestGetKeywordsFromContext_ValidKeywords(t *testing.T) {
ctx := context.WithValue(context.Background(), keywordsKey, "keyword1,keyword2")
result := getKeywordsFromContext(ctx)
assert.Equal(t, strPtr("keyword1,keyword2"), result)
}
func TestGetKeywordsFromContext_EmptyString(t *testing.T) {
ctx := context.WithValue(context.Background(), keywordsKey, "")
result := getKeywordsFromContext(ctx)
assert.Equal(t, strPtr(""), result)
}
func TestGetKeywordsFromContext_NoKeywordsValue(t *testing.T) {
ctx := context.Background()
result := getKeywordsFromContext(ctx)
assert.Nil(t, result)
}
func TestGetKeywordsFromContext_InvalidType(t *testing.T) {
ctx := context.WithValue(context.Background(), keywordsKey, 123)
result := getKeywordsFromContext(ctx)
assert.Nil(t, result)
}
func TestGetCategoriesFromContext_ValidCategories(t *testing.T) {
ctx := context.WithValue(context.Background(), categoriesKey, "category1,category2")
result := getCategoriesFromContext(ctx)
assert.Equal(t, strPtr("category1,category2"), result)
}
func TestGetCategoriesFromContext_EmptyString(t *testing.T) {
ctx := context.WithValue(context.Background(), categoriesKey, "")
result := getCategoriesFromContext(ctx)
assert.Equal(t, strPtr(""), result)
}
func TestGetCategoriesFromContext_NoCategoriesValue(t *testing.T) {
ctx := context.Background()
result := getCategoriesFromContext(ctx)
assert.Nil(t, result)
}
func TestGetCategoriesFromContext_InvalidType(t *testing.T) {
ctx := context.WithValue(context.Background(), categoriesKey, 123)
result := getCategoriesFromContext(ctx)
assert.Nil(t, result)
}
func strPtr(s string) *string {
return &s
}
func init() {
logger := zaptest.NewLogger(nil)
zap.ReplaceGlobals(logger)
}

View File

@@ -0,0 +1,14 @@
package handlers
import (
"github.com/gin-gonic/gin"
"github.com/swiftyos/market/models"
)
func GetUserFromContext(c *gin.Context) (models.User, bool) {
user, exists := c.Get("user")
if !exists {
return models.User{}, false
}
return user.(models.User), true
}

View File

@@ -0,0 +1,45 @@
package handlers
import (
"testing"
"github.com/gin-gonic/gin"
"github.com/stretchr/testify/assert"
"github.com/swiftyos/market/models"
)
func TestGetUserFromContext(t *testing.T) {
t.Run("User exists in context", func(t *testing.T) {
// Create a new gin context
c, _ := gin.CreateTestContext(nil)
// Create a test user
testUser := models.User{
UserID: "123",
Role: "admin",
Email: "test@example.com",
}
// Set the user in the context
c.Set("user", testUser)
// Call the function
user, exists := GetUserFromContext(c)
// Assert the results
assert.True(t, exists)
assert.Equal(t, testUser, user)
})
t.Run("User does not exist in context", func(t *testing.T) {
// Create a new gin context
c, _ := gin.CreateTestContext(nil)
// Call the function
user, exists := GetUserFromContext(c)
// Assert the results
assert.False(t, exists)
assert.Equal(t, models.User{}, user)
})
}

116
rnd/rest-api-go/main.go Normal file
View File

@@ -0,0 +1,116 @@
package main
import (
"log"
"net/http"
"time"
"github.com/Depado/ginprom"
"github.com/gin-contrib/cors"
"github.com/gin-contrib/zap"
"github.com/gin-gonic/gin"
swaggerfiles "github.com/swaggo/files"
ginSwagger "github.com/swaggo/gin-swagger"
"github.com/swiftyos/market/config"
"github.com/swiftyos/market/database"
docs "github.com/swiftyos/market/docs"
"github.com/swiftyos/market/handlers"
"github.com/swiftyos/market/middleware"
"github.com/swiftyos/market/utils"
"go.uber.org/zap"
)
func main() {
// Initialize configuration
cfg, err := config.Load()
if err != nil {
log.Fatalf("Failed to load config: %v", err)
}
// Initialize logger
logger := utils.NewLogger(cfg)
// Initialize database connection
db, err := database.NewDB(cfg)
if err != nil {
logger.Fatal("Failed to connect to database", zap.Error(err))
}
// Initialize Gin router
r := gin.New()
// Set the port
port := cfg.ServerAddress
if port == "" {
port = "8080" // Default port if not specified in config
}
r.Run(":" + port)
p := ginprom.New(
ginprom.Engine(r),
ginprom.Subsystem("gin"),
ginprom.Path("/metrics"),
)
r.Use(p.Instrument())
// Use middleware
r.Use(ginzap.Ginzap(logger, time.RFC1123, true))
r.Use(ginzap.RecoveryWithZap(logger, true))
r.Use(middleware.Gzip())
// Update CORS configuration
corsConfig := cors.DefaultConfig()
if len(cfg.CORSAllowOrigins) > 0 {
corsConfig.AllowOrigins = cfg.CORSAllowOrigins
} else {
corsConfig.AllowOrigins = []string{"*"} // Fallback to allow all origins if not specified
}
corsConfig.AllowHeaders = append(corsConfig.AllowHeaders, "Authorization")
corsConfig.AllowMethods = []string{"GET", "POST", "PUT", "PATCH", "DELETE", "OPTIONS"}
corsConfig.AllowCredentials = true
r.Use(cors.New(corsConfig))
// Route welcome
r.GET("/", func(c *gin.Context) {
c.String(http.StatusOK, "Welcome to the Marketplace API")
})
docs.SwaggerInfo.BasePath = "/api/v1/market/"
// Setup routes
// [Error] Request header field Authorization is not allowed by Access-Control-Allow-Headers.
// [Error] Fetch API cannot load http://localhost:8015/api/v1/market/featured/agents?page=1&page_size=10 due to access control checks.
// [Error] Failed to load resource: Request header field Authorization is not allowed by Access-Control-Allow-Headers. (agents, line 0)
api := r.Group("/api/v1/market/")
{
agents := api.Group("/agents")
{
agents.GET("", handlers.GetAgents(db, logger))
agents.GET("/:agent_id", handlers.GetAgentDetails(db, logger))
agents.GET("/:agent_id/download", handlers.DownloadAgent(db, logger))
agents.GET("/:agent_id/download-file", handlers.DownloadAgentFile(db, logger))
agents.GET("/top-downloads", handlers.TopAgentsByDownloads(db, logger))
agents.GET("/featured", handlers.GetFeaturedAgents(db, logger))
agents.GET("/search", handlers.SearchAgents(db, logger))
agents.POST("/submit", middleware.Auth(cfg), handlers.SubmitAgent(db, logger))
}
// Admin routes
admin := api.Group("/admin")
{
admin.POST("/agent", middleware.Auth(cfg), handlers.CreateAgentEntry(db, logger))
admin.POST("/agent/featured/:agent_id", middleware.Auth(cfg), handlers.SetAgentFeatured(db, logger))
admin.GET("/agent/featured/:agent_id", middleware.Auth(cfg), handlers.GetAgentFeatured(db, logger))
admin.DELETE("/agent/featured/:agent_id", middleware.Auth(cfg), handlers.UnsetAgentFeatured(db, logger))
admin.GET("/agent/not-featured", middleware.Auth(cfg), handlers.GetNotFeaturedAgents(db, logger))
admin.GET("/agent/submissions", middleware.Auth(cfg), handlers.GetAgentSubmissions(db, logger))
admin.POST("/agent/submissions", middleware.Auth(cfg), handlers.ReviewSubmission(db, logger))
}
api.GET("/categories", handlers.GetCategories(db, logger))
// Analytics routes
api.POST("/agent-installed", handlers.AgentInstalled(db, logger))
}
r.GET("/docs/*any", ginSwagger.WrapHandler(swaggerfiles.Handler))
// Start server
if err := r.Run(cfg.ServerAddress); err != nil {
logger.Fatal("Failed to start server", zap.Error(err))
}
}

View File

@@ -0,0 +1,83 @@
package middleware
import (
"errors"
"net/http"
"strings"
"github.com/gin-gonic/gin"
"github.com/golang-jwt/jwt/v4"
"github.com/swiftyos/market/config"
"github.com/swiftyos/market/models"
)
func Auth(cfg *config.Config) gin.HandlerFunc {
return func(c *gin.Context) {
if !cfg.AuthEnabled {
// This handles the case when authentication is disabled
defaultUser := models.User{
UserID: "3e53486c-cf57-477e-ba2a-cb02dc828e1a",
Role: "admin",
}
c.Set("user", defaultUser)
c.Next()
return
}
authHeader := c.GetHeader("Authorization")
if authHeader == "" {
c.AbortWithStatusJSON(http.StatusUnauthorized, gin.H{"error": "Authorization header is missing"})
return
}
tokenString := strings.TrimPrefix(authHeader, "Bearer ")
token, err := parseJWTToken(tokenString, cfg.JWTSecret)
if err != nil {
c.AbortWithStatusJSON(http.StatusUnauthorized, gin.H{"error": err.Error()})
return
}
claims, ok := token.Claims.(jwt.MapClaims)
if !ok || !token.Valid {
c.AbortWithStatusJSON(http.StatusUnauthorized, gin.H{"error": "Invalid token"})
return
}
user, err := verifyUser(claims, false) // Pass 'true' for admin-only routes
if err != nil {
c.AbortWithStatusJSON(http.StatusUnauthorized, gin.H{"error": err.Error()})
return
}
c.Set("user", user)
c.Next()
}
}
func verifyUser(payload jwt.MapClaims, adminOnly bool) (models.User, error) {
user, err := models.NewUserFromPayload(payload)
if err != nil {
return models.User{}, err
}
if adminOnly && user.Role != "admin" {
return models.User{}, errors.New("Admin access required")
}
return user, nil
}
func parseJWTToken(tokenString string, secret string) (*jwt.Token, error) {
token, err := jwt.Parse(tokenString, func(token *jwt.Token) (interface{}, error) {
if _, ok := token.Method.(*jwt.SigningMethodHMAC); !ok {
return nil, errors.New("unexpected signing method")
}
return []byte(secret), nil
})
if err != nil {
return nil, err
}
return token, nil
}

View File

@@ -0,0 +1,213 @@
package middleware
import (
"net/http/httptest"
"testing"
"github.com/gin-gonic/gin"
"github.com/golang-jwt/jwt/v4"
"github.com/stretchr/testify/assert"
"github.com/swiftyos/market/config"
"github.com/swiftyos/market/models"
)
func TestVerifyUser(t *testing.T) {
tests := []struct {
name string
payload jwt.MapClaims
adminOnly bool
wantUser models.User
wantErr bool
}{
{
name: "Valid user",
payload: jwt.MapClaims{
"sub": "test-user",
"email": "test@example.com",
"role": "user",
},
adminOnly: false,
wantUser: models.User{
UserID: "test-user",
Email: "test@example.com",
Role: "user",
},
wantErr: false,
},
{
name: "Valid admin",
payload: jwt.MapClaims{
"sub": "admin-user",
"email": "admin@example.com",
"role": "admin",
},
adminOnly: true,
wantUser: models.User{
UserID: "admin-user",
Email: "admin@example.com",
Role: "admin",
},
wantErr: false,
},
{
name: "Non-admin accessing admin-only route",
payload: jwt.MapClaims{
"sub": "test-user",
"email": "test@example.com",
"role": "user",
},
adminOnly: true,
wantUser: models.User{},
wantErr: true,
},
{
name: "Missing sub claim",
payload: jwt.MapClaims{},
adminOnly: false,
wantUser: models.User{},
wantErr: true,
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
gotUser, err := verifyUser(tt.payload, tt.adminOnly)
if tt.wantErr {
assert.Error(t, err)
} else {
assert.NoError(t, err)
assert.Equal(t, tt.wantUser, gotUser)
}
})
}
}
func TestParseJWTToken(t *testing.T) {
secret := "test-secret"
tests := []struct {
name string
tokenString string
wantErr bool
}{
{
name: "Valid token",
tokenString: createValidToken(secret),
wantErr: false,
},
{
name: "Invalid token",
tokenString: "invalid.token.string",
wantErr: true,
},
{
name: "Empty token",
tokenString: "",
wantErr: true,
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
token, err := parseJWTToken(tt.tokenString, secret)
if tt.wantErr {
assert.Error(t, err)
assert.Nil(t, token)
} else {
assert.NoError(t, err)
assert.NotNil(t, token)
assert.True(t, token.Valid)
}
})
}
}
func createValidToken(secret string) string {
token := jwt.NewWithClaims(jwt.SigningMethodHS256, jwt.MapClaims{
"sub": "test-user",
"email": "test@example.com",
"role": "user",
})
tokenString, _ := token.SignedString([]byte(secret))
return tokenString
}
func TestAuth(t *testing.T) {
cfg := &config.Config{
JWTSecret: "test-secret",
AuthEnabled: true,
}
tests := []struct {
name string
authHeader string
expectedUser models.User
expectedError bool
}{
{
name: "Valid token",
authHeader: "Bearer " + createValidToken(cfg.JWTSecret),
expectedUser: models.User{
UserID: "test-user",
Email: "test@example.com",
Role: "user",
},
expectedError: false,
},
{
name: "Invalid token",
authHeader: "Bearer invalid.token.string",
expectedUser: models.User{},
expectedError: true,
},
{
name: "Missing auth header",
authHeader: "",
expectedUser: models.User{},
expectedError: true,
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
// Create a mock gin.Context
c, _ := gin.CreateTestContext(httptest.NewRecorder())
c.Request = httptest.NewRequest("GET", "/", nil)
c.Request.Header.Set("Authorization", tt.authHeader)
// Call the Auth middleware
Auth(cfg)(c)
// Check the results
if tt.expectedError {
assert.True(t, c.IsAborted())
} else {
assert.False(t, c.IsAborted())
user, exists := c.Get("user")
assert.True(t, exists)
assert.Equal(t, tt.expectedUser, user.(models.User))
}
})
}
}
func TestAuthDisabled(t *testing.T) {
cfg := &config.Config{
JWTSecret: "test-secret",
AuthEnabled: false,
}
// Create a mock gin.Context
c, _ := gin.CreateTestContext(httptest.NewRecorder())
c.Request = httptest.NewRequest("GET", "/", nil)
Auth(cfg)(c)
assert.False(t, c.IsAborted())
user, exists := c.Get("user")
assert.True(t, exists)
assert.Equal(t, models.User{
UserID: "3e53486c-cf57-477e-ba2a-cb02dc828e1a",
Role: "admin",
}, user.(models.User))
}

View File

@@ -0,0 +1,10 @@
package middleware
import (
"github.com/gin-contrib/gzip"
"github.com/gin-gonic/gin"
)
func Gzip() gin.HandlerFunc {
return gzip.Gzip(gzip.DefaultCompression)
}

View File

@@ -0,0 +1,104 @@
package models
import (
"time"
)
// Agent represents the basic agent information
type Agent struct {
ID string `json:"id"`
Name string `json:"name"`
Description string `json:"description"`
Author string `json:"author"`
Keywords []string `json:"keywords"`
Categories []string `json:"categories"`
Graph Graph `json:"graph"`
}
// Graph represents the graph structure of an agent
type Graph struct {
Name string `json:"name"`
Description string `json:"description"`
// Add other fields as needed
}
// AddAgentRequest represents the request structure for adding a new agent
type AddAgentRequest struct {
Graph Graph `json:"graph"`
Author string `json:"author"`
Keywords []string `json:"keywords"`
Categories []string `json:"categories"`
}
// SubmissionStatus represents the status of an agent submission
type SubmissionStatus string
const (
SubmissionStatusPending SubmissionStatus = "PENDING"
SubmissionStatusApproved SubmissionStatus = "APPROVED"
SubmissionStatusRejected SubmissionStatus = "REJECTED"
)
// AgentWithMetadata extends Agent with additional metadata
type AgentWithMetadata struct {
Agent
Version int `json:"version"`
CreatedAt time.Time `json:"createdAt"`
UpdatedAt time.Time `json:"updatedAt"`
SubmissionDate time.Time `json:"submissionDate"`
SubmissionReviewDate *time.Time `json:"submissionReviewDate,omitempty"`
SubmissionStatus SubmissionStatus `json:"submissionStatus"`
SubmissionReviewComments *string `json:"submissionReviewComments,omitempty"`
}
// AgentWithRank extends AgentWithMetadata with a rank field for search results
type AgentWithRank struct {
AgentWithMetadata
Rank float64 `json:"rank"`
}
type AgentWithDownloads struct {
AgentWithMetadata
Downloads int `json:"downloads"`
}
// AnalyticsTracker represents analytics data for an agent
type AnalyticsTracker struct {
ID string `json:"id"`
AgentID string `json:"agentId"`
Views int `json:"views"`
Downloads int `json:"downloads"`
}
// InstallationLocation represents the location where an agent is installed
type InstallationLocation string
const (
InstallationLocationLocal InstallationLocation = "LOCAL"
InstallationLocationCloud InstallationLocation = "CLOUD"
)
// InstallTracker represents installation data for an agent
type InstallTracker struct {
ID string `json:"id"`
MarketplaceAgentID string `json:"marketplaceAgentId"`
InstalledAgentID string `json:"installedAgentId"`
InstallationLocation InstallationLocation `json:"installationLocation"`
CreatedAt time.Time `json:"createdAt"`
}
// FeaturedAgent represents a featured agent in the marketplace
type FeaturedAgent struct {
ID string `json:"id"`
AgentID string `json:"agentId"`
IsActive bool `json:"isActive"`
FeaturedCategories []string `json:"featuredCategories"`
CreatedAt time.Time `json:"createdAt"`
UpdatedAt time.Time `json:"updatedAt"`
}
type AgentFile struct {
ID string `json:"id"`
Name string `json:"name"`
Graph interface{} `json:"graph"`
}

View File

@@ -0,0 +1,176 @@
package models
import (
"encoding/json"
"testing"
"time"
"github.com/stretchr/testify/assert"
)
func TestAgentJSON(t *testing.T) {
jsonStr := `{
"id": "test-id",
"name": "Test Agent",
"description": "A test agent",
"author": "Test Author",
"keywords": ["test", "agent"],
"categories": ["testing"],
"graph": {
"name": "Test Graph",
"description": "A test graph"
}
}`
var agent Agent
err := json.Unmarshal([]byte(jsonStr), &agent)
assert.NoError(t, err)
assert.Equal(t, "test-id", agent.ID)
assert.Equal(t, "Test Agent", agent.Name)
assert.Equal(t, "A test agent", agent.Description)
assert.Equal(t, "Test Author", agent.Author)
assert.Equal(t, []string{"test", "agent"}, agent.Keywords)
assert.Equal(t, []string{"testing"}, agent.Categories)
assert.Equal(t, "Test Graph", agent.Graph.Name)
assert.Equal(t, "A test graph", agent.Graph.Description)
}
func TestGraphJSON(t *testing.T) {
jsonStr := `{
"name": "Test Graph",
"description": "A test graph"
}`
var graph Graph
err := json.Unmarshal([]byte(jsonStr), &graph)
assert.NoError(t, err)
assert.Equal(t, "Test Graph", graph.Name)
assert.Equal(t, "A test graph", graph.Description)
}
func TestAddAgentRequestJSON(t *testing.T) {
jsonStr := `{
"graph": {
"name": "Test Graph",
"description": "A test graph"
},
"author": "Test Author",
"keywords": ["test", "agent"],
"categories": ["testing"]
}`
var request AddAgentRequest
err := json.Unmarshal([]byte(jsonStr), &request)
assert.NoError(t, err)
assert.Equal(t, "Test Graph", request.Graph.Name)
assert.Equal(t, "A test graph", request.Graph.Description)
assert.Equal(t, "Test Author", request.Author)
assert.Equal(t, []string{"test", "agent"}, request.Keywords)
assert.Equal(t, []string{"testing"}, request.Categories)
}
func TestAgentWithMetadataJSON(t *testing.T) {
now := time.Now().UTC().Round(time.Second)
jsonStr := `{
"id": "test-id",
"name": "Test Agent",
"description": "A test agent",
"author": "Test Author",
"keywords": ["test", "agent"],
"categories": ["testing"],
"graph": {
"name": "Test Graph",
"description": "A test graph"
},
"version": 1,
"createdAt": "` + now.Format(time.RFC3339) + `",
"updatedAt": "` + now.Format(time.RFC3339) + `",
"submissionDate": "` + now.Format(time.RFC3339) + `",
"submissionStatus": "PENDING"
}`
var agent AgentWithMetadata
err := json.Unmarshal([]byte(jsonStr), &agent)
assert.NoError(t, err)
assert.Equal(t, "test-id", agent.ID)
assert.Equal(t, "Test Agent", agent.Name)
assert.Equal(t, "A test agent", agent.Description)
assert.Equal(t, "Test Author", agent.Author)
assert.Equal(t, []string{"test", "agent"}, agent.Keywords)
assert.Equal(t, []string{"testing"}, agent.Categories)
assert.Equal(t, "Test Graph", agent.Graph.Name)
assert.Equal(t, "A test graph", agent.Graph.Description)
assert.Equal(t, 1, agent.Version)
assert.Equal(t, now, agent.CreatedAt)
assert.Equal(t, now, agent.UpdatedAt)
assert.Equal(t, now, agent.SubmissionDate)
assert.Equal(t, SubmissionStatusPending, agent.SubmissionStatus)
assert.Nil(t, agent.SubmissionReviewDate)
assert.Nil(t, agent.SubmissionReviewComments)
}
func TestAnalyticsTrackerJSON(t *testing.T) {
jsonStr := `{
"id": "tracker-id",
"agentId": "agent-id",
"views": 100,
"downloads": 50
}`
var tracker AnalyticsTracker
err := json.Unmarshal([]byte(jsonStr), &tracker)
assert.NoError(t, err)
assert.Equal(t, "tracker-id", tracker.ID)
assert.Equal(t, "agent-id", tracker.AgentID)
assert.Equal(t, 100, tracker.Views)
assert.Equal(t, 50, tracker.Downloads)
}
func TestInstallTrackerJSON(t *testing.T) {
now := time.Now().UTC().Round(time.Second)
jsonStr := `{
"id": "install-id",
"marketplaceAgentId": "marketplace-agent-id",
"installedAgentId": "installed-agent-id",
"installationLocation": "LOCAL",
"createdAt": "` + now.Format(time.RFC3339) + `"
}`
var tracker InstallTracker
err := json.Unmarshal([]byte(jsonStr), &tracker)
assert.NoError(t, err)
assert.Equal(t, "install-id", tracker.ID)
assert.Equal(t, "marketplace-agent-id", tracker.MarketplaceAgentID)
assert.Equal(t, "installed-agent-id", tracker.InstalledAgentID)
assert.Equal(t, InstallationLocationLocal, tracker.InstallationLocation)
assert.Equal(t, now, tracker.CreatedAt)
}
func TestFeaturedAgentJSON(t *testing.T) {
now := time.Now().UTC().Round(time.Second)
jsonStr := `{
"id": "featured-id",
"agentId": "agent-id",
"isActive": true,
"featuredCategories": ["category1", "category2"],
"createdAt": "` + now.Format(time.RFC3339) + `",
"updatedAt": "` + now.Format(time.RFC3339) + `"
}`
var featured FeaturedAgent
err := json.Unmarshal([]byte(jsonStr), &featured)
assert.NoError(t, err)
assert.Equal(t, "featured-id", featured.ID)
assert.Equal(t, "agent-id", featured.AgentID)
assert.True(t, featured.IsActive)
assert.Equal(t, []string{"category1", "category2"}, featured.FeaturedCategories)
assert.Equal(t, now, featured.CreatedAt)
assert.Equal(t, now, featured.UpdatedAt)
}

View File

@@ -0,0 +1,28 @@
package models
import (
"fmt"
"github.com/golang-jwt/jwt/v4"
)
type User struct {
UserID string `json:"user_id"`
Email string `json:"email"`
Role string `json:"role"`
}
func NewUserFromPayload(claims jwt.MapClaims) (User, error) {
userID, ok := claims["sub"].(string)
if !ok {
return User{}, fmt.Errorf("invalid or missing 'sub' claim")
}
email, _ := claims["email"].(string)
role, _ := claims["role"].(string)
return User{
UserID: userID,
Email: email,
Role: role,
}, nil
}

View File

@@ -0,0 +1,62 @@
package models
import (
"encoding/json"
"testing"
"github.com/golang-jwt/jwt/v4"
"github.com/stretchr/testify/assert"
)
func TestNewUserFromPayload(t *testing.T) {
testCases := []struct {
name string
payload string
expectedUser User
expectedError bool
}{
{
name: "Valid payload",
payload: `{"sub": "123", "email": "test@example.com", "role": "user"}`,
expectedUser: User{
UserID: "123",
Email: "test@example.com",
Role: "user",
},
expectedError: false,
},
{
name: "Missing sub claim",
payload: `{"email": "test@example.com", "role": "user"}`,
expectedUser: User{},
expectedError: true,
},
{
name: "Missing optional claims",
payload: `{"sub": "456"}`,
expectedUser: User{
UserID: "456",
Email: "",
Role: "",
},
expectedError: false,
},
}
for _, tc := range testCases {
t.Run(tc.name, func(t *testing.T) {
var claims jwt.MapClaims
err := json.Unmarshal([]byte(tc.payload), &claims)
assert.NoError(t, err)
user, err := NewUserFromPayload(claims)
if tc.expectedError {
assert.Error(t, err)
} else {
assert.NoError(t, err)
assert.Equal(t, tc.expectedUser, user)
}
})
}
}

View File

@@ -0,0 +1,11 @@
global:
scrape_interval: 15s
evaluation_interval: 15s
scrape_configs:
- job_name: 'market'
static_configs:
- targets: ['market:8015']
metrics_path: '/metrics'
scheme: 'http'

View File

@@ -0,0 +1,27 @@
package utils
import (
"github.com/swiftyos/market/config"
"go.uber.org/zap"
"go.uber.org/zap/zapcore"
"os"
)
func NewLogger(cfg *config.Config) *zap.Logger {
encoderConfig := zap.NewProductionEncoderConfig()
encoderConfig.EncodeTime = zapcore.ISO8601TimeEncoder
consoleEncoder := zapcore.NewConsoleEncoder(encoderConfig)
consoleWriter := zapcore.AddSync(os.Stdout)
core := zapcore.NewCore(consoleEncoder, consoleWriter, zap.InfoLevel)
logger := zap.New(core)
return logger
}
func StringOrNil(s *string) string {
if s == nil {
return "nil"
}
return *s
}