mirror of
https://github.com/googleapis/genai-toolbox.git
synced 2026-02-04 12:15:09 -05:00
This PR introduces a significant update to the Toolbox configuration file format, which is one of the primary **breaking changes** required for the implementation of the Advanced Control Plane. # Summary of Changes The configuration schema has been updated to enforce resource isolation and facilitate atomic, incremental updates. * Resource Isolation: Resource definitions are now separated into individual blocks, using a distinct structure for each resource type (Source, Tool, Toolset, etc.). This improves readability, management, and auditing of configuration files. * Field Name Modification: Internal field names have been modified to align with declarative methodologies. Specifically, the configuration now separates kind (general resource type, e.g., Source) from type (specific implementation, e.g., Postgres). # User Impact Existing tools.yaml configuration files are now in an outdated format. Users must eventually update their files to the new YAML format. # Mitigation & Compatibility Backward compatibility is maintained during this transition to ensure no immediate user action is required for existing files. * Immediate Backward Compatibility: The source code includes a pre-processing layer that automatically detects outdated configuration files (v1 format) and converts them to the new v2 format under the hood. * [COMING SOON] Migration Support: The new toolbox migrate subcommand will be introduced to allow users to automatically convert their old configuration files to the latest format. # Example Example for config file v2: ``` kind: sources name: my-pg-instance type: cloud-sql-postgres project: my-project region: my-region instance: my-instance database: my_db user: my_user password: my_pass --- kind: authServices name: my-google-auth type: google clientId: testing-id --- kind: tools name: example_tool type: postgres-sql source: my-pg-instance description: some description statement: SELECT * FROM SQL_STATEMENT; parameters: - name: country type: string description: some description --- kind: tools name: example_tool_2 type: postgres-sql source: my-pg-instance description: returning the number one statement: SELECT 1; --- kind: toolsets name: example_toolset tools: - example_tool ``` --------- Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Co-authored-by: Averi Kitsch <akitsch@google.com>
173 lines
4.3 KiB
Go
173 lines
4.3 KiB
Go
// Copyright 2025 Google LLC
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package sqlite
|
|
|
|
import (
|
|
"context"
|
|
"database/sql"
|
|
"encoding/json"
|
|
"fmt"
|
|
|
|
"github.com/goccy/go-yaml"
|
|
"github.com/googleapis/genai-toolbox/internal/sources"
|
|
"github.com/googleapis/genai-toolbox/internal/util/orderedmap"
|
|
"go.opentelemetry.io/otel/trace"
|
|
_ "modernc.org/sqlite" // Pure Go SQLite driver
|
|
)
|
|
|
|
const SourceType string = "sqlite"
|
|
|
|
// validate interface
|
|
var _ sources.SourceConfig = Config{}
|
|
|
|
func init() {
|
|
if !sources.Register(SourceType, newConfig) {
|
|
panic(fmt.Sprintf("source type %q already registered", SourceType))
|
|
}
|
|
}
|
|
|
|
func newConfig(ctx context.Context, name string, decoder *yaml.Decoder) (sources.SourceConfig, error) {
|
|
actual := Config{Name: name}
|
|
if err := decoder.DecodeContext(ctx, &actual); err != nil {
|
|
return nil, err
|
|
}
|
|
return actual, nil
|
|
}
|
|
|
|
type Config struct {
|
|
Name string `yaml:"name" validate:"required"`
|
|
Type string `yaml:"type" validate:"required"`
|
|
Database string `yaml:"database" validate:"required"` // Path to SQLite database file
|
|
}
|
|
|
|
func (r Config) SourceConfigType() string {
|
|
return SourceType
|
|
}
|
|
|
|
func (r Config) Initialize(ctx context.Context, tracer trace.Tracer) (sources.Source, error) {
|
|
db, err := initSQLiteConnection(ctx, tracer, r.Name, r.Database)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("unable to create db connection: %w", err)
|
|
}
|
|
|
|
err = db.PingContext(context.Background())
|
|
if err != nil {
|
|
return nil, fmt.Errorf("unable to connect successfully: %w", err)
|
|
}
|
|
|
|
s := &Source{
|
|
Config: r,
|
|
Db: db,
|
|
}
|
|
return s, nil
|
|
}
|
|
|
|
var _ sources.Source = &Source{}
|
|
|
|
type Source struct {
|
|
Config
|
|
Db *sql.DB
|
|
}
|
|
|
|
func (s *Source) SourceType() string {
|
|
return SourceType
|
|
}
|
|
|
|
func (s *Source) ToConfig() sources.SourceConfig {
|
|
return s.Config
|
|
}
|
|
|
|
func (s *Source) SQLiteDB() *sql.DB {
|
|
return s.Db
|
|
}
|
|
|
|
func (s *Source) RunSQL(ctx context.Context, statement string, params []any) (any, error) {
|
|
// Execute the SQL query with parameters
|
|
rows, err := s.SQLiteDB().QueryContext(ctx, statement, params...)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("unable to execute query: %w", err)
|
|
}
|
|
defer rows.Close()
|
|
|
|
// Get column names
|
|
cols, err := rows.Columns()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("unable to get column names: %w", err)
|
|
}
|
|
|
|
// The sqlite driver does not support ColumnTypes, so we can't get the
|
|
// underlying database type of the columns. We'll have to rely on the
|
|
// generic `any` type and then handle the JSON data separately.
|
|
rawValues := make([]any, len(cols))
|
|
values := make([]any, len(cols))
|
|
for i := range rawValues {
|
|
values[i] = &rawValues[i]
|
|
}
|
|
|
|
// Prepare the result slice
|
|
var out []any
|
|
for rows.Next() {
|
|
if err := rows.Scan(values...); err != nil {
|
|
return nil, fmt.Errorf("unable to scan row: %w", err)
|
|
}
|
|
|
|
// Create a map for this row
|
|
row := orderedmap.Row{}
|
|
for i, name := range cols {
|
|
val := rawValues[i]
|
|
// Handle nil values
|
|
if val == nil {
|
|
row.Add(name, nil)
|
|
continue
|
|
}
|
|
// Handle JSON data
|
|
if jsonString, ok := val.(string); ok {
|
|
var unmarshaledData any
|
|
if json.Unmarshal([]byte(jsonString), &unmarshaledData) == nil {
|
|
row.Add(name, unmarshaledData)
|
|
continue
|
|
}
|
|
}
|
|
// Store the value in the map
|
|
row.Add(name, val)
|
|
}
|
|
out = append(out, row)
|
|
}
|
|
|
|
if err = rows.Err(); err != nil {
|
|
return nil, fmt.Errorf("error iterating rows: %w", err)
|
|
}
|
|
|
|
return out, nil
|
|
}
|
|
|
|
func initSQLiteConnection(ctx context.Context, tracer trace.Tracer, name, dbPath string) (*sql.DB, error) {
|
|
//nolint:all // Reassigned ctx
|
|
ctx, span := sources.InitConnectionSpan(ctx, tracer, SourceType, name)
|
|
defer span.End()
|
|
|
|
// Open database connection
|
|
db, err := sql.Open("sqlite", dbPath)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("sql.Open: %w", err)
|
|
}
|
|
|
|
// Set some reasonable defaults for SQLite
|
|
db.SetMaxOpenConns(1) // SQLite only supports one writer at a time
|
|
db.SetMaxIdleConns(1)
|
|
|
|
return db, nil
|
|
}
|