mirror of
https://github.com/googleapis/genai-toolbox.git
synced 2026-01-08 15:14:00 -05:00
Add parameter `embeddedBy` field to support vector embedding & semantic search. Major change in `internal/util/parameters/parameters.go` This PR only adds vector formatter for the postgressql tool. Other tools requiring vector formatting may not work with embeddedBy. Second part of the Semantic Search support. First part: https://github.com/googleapis/genai-toolbox/pull/2121
315 lines
12 KiB
Go
315 lines
12 KiB
Go
// Copyright 2025 Google LLC
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package bigqueryexecutesql
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"strings"
|
|
|
|
bigqueryapi "cloud.google.com/go/bigquery"
|
|
yaml "github.com/goccy/go-yaml"
|
|
"github.com/googleapis/genai-toolbox/internal/embeddingmodels"
|
|
"github.com/googleapis/genai-toolbox/internal/sources"
|
|
bigqueryds "github.com/googleapis/genai-toolbox/internal/sources/bigquery"
|
|
"github.com/googleapis/genai-toolbox/internal/tools"
|
|
bqutil "github.com/googleapis/genai-toolbox/internal/tools/bigquery/bigquerycommon"
|
|
"github.com/googleapis/genai-toolbox/internal/util"
|
|
"github.com/googleapis/genai-toolbox/internal/util/parameters"
|
|
bigqueryrestapi "google.golang.org/api/bigquery/v2"
|
|
)
|
|
|
|
const kind string = "bigquery-execute-sql"
|
|
|
|
func init() {
|
|
if !tools.Register(kind, newConfig) {
|
|
panic(fmt.Sprintf("tool kind %q already registered", kind))
|
|
}
|
|
}
|
|
|
|
func newConfig(ctx context.Context, name string, decoder *yaml.Decoder) (tools.ToolConfig, error) {
|
|
actual := Config{Name: name}
|
|
if err := decoder.DecodeContext(ctx, &actual); err != nil {
|
|
return nil, err
|
|
}
|
|
return actual, nil
|
|
}
|
|
|
|
type compatibleSource interface {
|
|
BigQueryClient() *bigqueryapi.Client
|
|
BigQuerySession() bigqueryds.BigQuerySessionProvider
|
|
BigQueryWriteMode() string
|
|
UseClientAuthorization() bool
|
|
IsDatasetAllowed(projectID, datasetID string) bool
|
|
BigQueryAllowedDatasets() []string
|
|
RetrieveClientAndService(tools.AccessToken) (*bigqueryapi.Client, *bigqueryrestapi.Service, error)
|
|
RunSQL(context.Context, *bigqueryapi.Client, string, string, []bigqueryapi.QueryParameter, []*bigqueryapi.ConnectionProperty) (any, error)
|
|
}
|
|
|
|
type Config struct {
|
|
Name string `yaml:"name" validate:"required"`
|
|
Kind string `yaml:"kind" validate:"required"`
|
|
Source string `yaml:"source" validate:"required"`
|
|
Description string `yaml:"description" validate:"required"`
|
|
AuthRequired []string `yaml:"authRequired"`
|
|
}
|
|
|
|
// validate interface
|
|
var _ tools.ToolConfig = Config{}
|
|
|
|
func (cfg Config) ToolConfigKind() string {
|
|
return kind
|
|
}
|
|
|
|
func (cfg Config) Initialize(srcs map[string]sources.Source) (tools.Tool, error) {
|
|
// verify source exists
|
|
rawS, ok := srcs[cfg.Source]
|
|
if !ok {
|
|
return nil, fmt.Errorf("no source named %q configured", cfg.Source)
|
|
}
|
|
|
|
// verify the source is compatible
|
|
s, ok := rawS.(compatibleSource)
|
|
if !ok {
|
|
return nil, fmt.Errorf("invalid source for %q tool: source %q not compatible", kind, cfg.Source)
|
|
}
|
|
|
|
var sqlDescriptionBuilder strings.Builder
|
|
switch s.BigQueryWriteMode() {
|
|
case bigqueryds.WriteModeBlocked:
|
|
sqlDescriptionBuilder.WriteString("The SQL to execute. In 'blocked' mode, only SELECT statements are allowed; other statement types will fail.")
|
|
case bigqueryds.WriteModeProtected:
|
|
sqlDescriptionBuilder.WriteString("The SQL to execute. Only SELECT statements and writes to the session's temporary dataset are allowed (e.g., `CREATE TEMP TABLE ...`).")
|
|
default: // WriteModeAllowed
|
|
sqlDescriptionBuilder.WriteString("The SQL to execute.")
|
|
}
|
|
|
|
allowedDatasets := s.BigQueryAllowedDatasets()
|
|
if len(allowedDatasets) > 0 {
|
|
if len(allowedDatasets) == 1 {
|
|
datasetFQN := allowedDatasets[0]
|
|
parts := strings.Split(datasetFQN, ".")
|
|
if len(parts) < 2 {
|
|
return nil, fmt.Errorf("expected allowedDataset to have at least 2 parts (project.dataset): %s", datasetFQN)
|
|
}
|
|
datasetID := parts[1]
|
|
sqlDescriptionBuilder.WriteString(fmt.Sprintf(" The query must only access the `%s` dataset. "+
|
|
"To query a table within this dataset (e.g., `my_table`), "+
|
|
"qualify it with the dataset id (e.g., `%s.my_table`).", datasetFQN, datasetID))
|
|
} else {
|
|
datasetIDs := []string{}
|
|
for _, ds := range allowedDatasets {
|
|
datasetIDs = append(datasetIDs, fmt.Sprintf("`%s`", ds))
|
|
}
|
|
sqlDescriptionBuilder.WriteString(fmt.Sprintf(" The query must only access datasets from the following list: %s.", strings.Join(datasetIDs, ", ")))
|
|
}
|
|
}
|
|
|
|
sqlParameter := parameters.NewStringParameter("sql", sqlDescriptionBuilder.String())
|
|
dryRunParameter := parameters.NewBooleanParameterWithDefault(
|
|
"dry_run",
|
|
false,
|
|
"If set to true, the query will be validated and information about the execution will be returned "+
|
|
"without running the query. Defaults to false.",
|
|
)
|
|
params := parameters.Parameters{sqlParameter, dryRunParameter}
|
|
mcpManifest := tools.GetMcpManifest(cfg.Name, cfg.Description, cfg.AuthRequired, params, nil)
|
|
|
|
// finish tool setup
|
|
t := Tool{
|
|
Config: cfg,
|
|
Parameters: params,
|
|
manifest: tools.Manifest{Description: cfg.Description, Parameters: params.Manifest(), AuthRequired: cfg.AuthRequired},
|
|
mcpManifest: mcpManifest,
|
|
}
|
|
return t, nil
|
|
}
|
|
|
|
// validate interface
|
|
var _ tools.Tool = Tool{}
|
|
|
|
type Tool struct {
|
|
Config
|
|
Parameters parameters.Parameters `yaml:"parameters"`
|
|
manifest tools.Manifest
|
|
mcpManifest tools.McpManifest
|
|
}
|
|
|
|
func (t Tool) ToConfig() tools.ToolConfig {
|
|
return t.Config
|
|
}
|
|
|
|
func (t Tool) Invoke(ctx context.Context, resourceMgr tools.SourceProvider, params parameters.ParamValues, accessToken tools.AccessToken) (any, error) {
|
|
source, err := tools.GetCompatibleSource[compatibleSource](resourceMgr, t.Source, t.Name, t.Kind)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
paramsMap := params.AsMap()
|
|
sql, ok := paramsMap["sql"].(string)
|
|
if !ok {
|
|
return nil, fmt.Errorf("unable to cast sql parameter %s", paramsMap["sql"])
|
|
}
|
|
dryRun, ok := paramsMap["dry_run"].(bool)
|
|
if !ok {
|
|
return nil, fmt.Errorf("unable to cast dry_run parameter %s", paramsMap["dry_run"])
|
|
}
|
|
|
|
bqClient, restService, err := source.RetrieveClientAndService(accessToken)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var connProps []*bigqueryapi.ConnectionProperty
|
|
var session *bigqueryds.Session
|
|
if source.BigQueryWriteMode() == bigqueryds.WriteModeProtected {
|
|
session, err = source.BigQuerySession()(ctx)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to get BigQuery session for protected mode: %w", err)
|
|
}
|
|
connProps = []*bigqueryapi.ConnectionProperty{
|
|
{Key: "session_id", Value: session.ID},
|
|
}
|
|
}
|
|
|
|
dryRunJob, err := bqutil.DryRunQuery(ctx, restService, bqClient.Project(), bqClient.Location, sql, nil, connProps)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("query validation failed: %w", err)
|
|
}
|
|
|
|
statementType := dryRunJob.Statistics.Query.StatementType
|
|
|
|
switch source.BigQueryWriteMode() {
|
|
case bigqueryds.WriteModeBlocked:
|
|
if statementType != "SELECT" {
|
|
return nil, fmt.Errorf("write mode is 'blocked', only SELECT statements are allowed")
|
|
}
|
|
case bigqueryds.WriteModeProtected:
|
|
if dryRunJob.Configuration != nil && dryRunJob.Configuration.Query != nil {
|
|
if dest := dryRunJob.Configuration.Query.DestinationTable; dest != nil && dest.DatasetId != session.DatasetID {
|
|
return nil, fmt.Errorf("protected write mode only supports SELECT statements, or write operations in the anonymous "+
|
|
"dataset of a BigQuery session, but destination was %q", dest.DatasetId)
|
|
}
|
|
}
|
|
}
|
|
|
|
if len(source.BigQueryAllowedDatasets()) > 0 {
|
|
switch statementType {
|
|
case "CREATE_SCHEMA", "DROP_SCHEMA", "ALTER_SCHEMA":
|
|
return nil, fmt.Errorf("dataset-level operations like '%s' are not allowed when dataset restrictions are in place", statementType)
|
|
case "CREATE_FUNCTION", "CREATE_TABLE_FUNCTION", "CREATE_PROCEDURE":
|
|
return nil, fmt.Errorf("creating stored routines ('%s') is not allowed when dataset restrictions are in place, as their contents cannot be safely analyzed", statementType)
|
|
case "CALL":
|
|
return nil, fmt.Errorf("calling stored procedures ('%s') is not allowed when dataset restrictions are in place, as their contents cannot be safely analyzed", statementType)
|
|
}
|
|
|
|
// Use a map to avoid duplicate table names.
|
|
tableIDSet := make(map[string]struct{})
|
|
|
|
// Get all tables from the dry run result. This is the most reliable method.
|
|
queryStats := dryRunJob.Statistics.Query
|
|
if queryStats != nil {
|
|
for _, tableRef := range queryStats.ReferencedTables {
|
|
tableIDSet[fmt.Sprintf("%s.%s.%s", tableRef.ProjectId, tableRef.DatasetId, tableRef.TableId)] = struct{}{}
|
|
}
|
|
if tableRef := queryStats.DdlTargetTable; tableRef != nil {
|
|
tableIDSet[fmt.Sprintf("%s.%s.%s", tableRef.ProjectId, tableRef.DatasetId, tableRef.TableId)] = struct{}{}
|
|
}
|
|
if tableRef := queryStats.DdlDestinationTable; tableRef != nil {
|
|
tableIDSet[fmt.Sprintf("%s.%s.%s", tableRef.ProjectId, tableRef.DatasetId, tableRef.TableId)] = struct{}{}
|
|
}
|
|
}
|
|
|
|
var tableNames []string
|
|
if len(tableIDSet) > 0 {
|
|
for tableID := range tableIDSet {
|
|
tableNames = append(tableNames, tableID)
|
|
}
|
|
} else if statementType != "SELECT" {
|
|
// If dry run yields no tables, fall back to the parser for non-SELECT statements
|
|
// to catch unsafe operations like EXECUTE IMMEDIATE.
|
|
parsedTables, parseErr := bqutil.TableParser(sql, source.BigQueryClient().Project())
|
|
if parseErr != nil {
|
|
// If parsing fails (e.g., EXECUTE IMMEDIATE), we cannot guarantee safety, so we must fail.
|
|
return nil, fmt.Errorf("could not parse tables from query to validate against allowed datasets: %w", parseErr)
|
|
}
|
|
tableNames = parsedTables
|
|
}
|
|
|
|
for _, tableID := range tableNames {
|
|
parts := strings.Split(tableID, ".")
|
|
if len(parts) == 3 {
|
|
projectID, datasetID := parts[0], parts[1]
|
|
if !source.IsDatasetAllowed(projectID, datasetID) {
|
|
return nil, fmt.Errorf("query accesses dataset '%s.%s', which is not in the allowed list", projectID, datasetID)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if dryRun {
|
|
if dryRunJob != nil {
|
|
jobJSON, err := json.MarshalIndent(dryRunJob, "", " ")
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to marshal dry run job to JSON: %w", err)
|
|
}
|
|
return string(jobJSON), nil
|
|
}
|
|
// This case should not be reached, but as a fallback, we return a message.
|
|
return "Dry run was requested, but no job information was returned.", nil
|
|
}
|
|
|
|
// Log the query executed for debugging.
|
|
logger, err := util.LoggerFromContext(ctx)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("error getting logger: %s", err)
|
|
}
|
|
logger.DebugContext(ctx, fmt.Sprintf("executing `%s` tool query: %s", kind, sql))
|
|
return source.RunSQL(ctx, bqClient, sql, statementType, nil, connProps)
|
|
}
|
|
|
|
func (t Tool) ParseParams(data map[string]any, claims map[string]map[string]any) (parameters.ParamValues, error) {
|
|
return parameters.ParseParams(t.Parameters, data, claims)
|
|
}
|
|
|
|
func (t Tool) EmbedParams(ctx context.Context, paramValues parameters.ParamValues, embeddingModelsMap map[string]embeddingmodels.EmbeddingModel) (parameters.ParamValues, error) {
|
|
return parameters.EmbedParams(ctx, t.Parameters, paramValues, embeddingModelsMap, nil)
|
|
}
|
|
|
|
func (t Tool) Manifest() tools.Manifest {
|
|
return t.manifest
|
|
}
|
|
|
|
func (t Tool) McpManifest() tools.McpManifest {
|
|
return t.mcpManifest
|
|
}
|
|
|
|
func (t Tool) Authorized(verifiedAuthServices []string) bool {
|
|
return tools.IsAuthorized(t.AuthRequired, verifiedAuthServices)
|
|
}
|
|
|
|
func (t Tool) RequiresClientAuthorization(resourceMgr tools.SourceProvider) (bool, error) {
|
|
source, err := tools.GetCompatibleSource[compatibleSource](resourceMgr, t.Source, t.Name, t.Kind)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return source.UseClientAuthorization(), nil
|
|
}
|
|
|
|
func (t Tool) GetAuthTokenHeaderName(resourceMgr tools.SourceProvider) (string, error) {
|
|
return "Authorization", nil
|
|
}
|