Files
genai-toolbox/docs/en/resources/tools/serverless-spark/serverless-spark-get-batch.md
Yuan Teoh 293c1d6889 feat!: update configuration file v2 (#2369)
This PR introduces a significant update to the Toolbox configuration
file format, which is one of the primary **breaking changes** required
for the implementation of the Advanced Control Plane.

# Summary of Changes
The configuration schema has been updated to enforce resource isolation
and facilitate atomic, incremental updates.
* Resource Isolation: Resource definitions are now separated into
individual blocks, using a distinct structure for each resource type
(Source, Tool, Toolset, etc.). This improves readability, management,
and auditing of configuration files.
* Field Name Modification: Internal field names have been modified to
align with declarative methodologies. Specifically, the configuration
now separates kind (general resource type, e.g., Source) from type
(specific implementation, e.g., Postgres).

# User Impact
Existing tools.yaml configuration files are now in an outdated format.
Users must eventually update their files to the new YAML format.

# Mitigation & Compatibility
Backward compatibility is maintained during this transition to ensure no
immediate user action is required for existing files.
* Immediate Backward Compatibility: The source code includes a
pre-processing layer that automatically detects outdated configuration
files (v1 format) and converts them to the new v2 format under the hood.
* [COMING SOON] Migration Support: The new toolbox migrate subcommand
will be introduced to allow users to automatically convert their old
configuration files to the latest format.

# Example
Example for config file v2:
```
kind: sources
name: my-pg-instance
type: cloud-sql-postgres
project: my-project
region: my-region
instance: my-instance
database: my_db
user: my_user
password: my_pass
---
kind: authServices
name: my-google-auth
type: google
clientId: testing-id
---
kind: tools
name: example_tool
type: postgres-sql
source: my-pg-instance
description: some description
statement: SELECT * FROM SQL_STATEMENT;
parameters:
- name: country
  type: string
  description: some description
---
kind: tools
name: example_tool_2
type: postgres-sql
source: my-pg-instance
description: returning the number one
statement: SELECT 1;
---
kind: toolsets
name: example_toolset
tools:
- example_tool
```

---------

Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com>
Co-authored-by: Averi Kitsch <akitsch@google.com>
2026-01-27 16:58:43 -08:00

3.0 KiB

title, type, weight, description, aliases
title type weight description aliases
serverless-spark-get-batch docs 1 A "serverless-spark-get-batch" tool gets a single Spark batch from the source.
/resources/tools/serverless-spark-get-batch

serverless-spark-get-batch

The serverless-spark-get-batch tool allows you to retrieve a specific Serverless Spark batch job. It's compatible with the following sources:

serverless-spark-list-batches accepts the following parameters:

  • name: The short name of the batch, e.g. for projects/my-project/locations/us-central1/my-batch, pass my-batch.

The tool gets the project and location from the source configuration.

Example

kind: tools
name: get_my_batch
type: serverless-spark-get-batch
source: my-serverless-spark-source
description: Use this tool to get a serverless spark batch.

Response Format

The response contains the full Batch object as defined in the API spec, plus additional fields consoleUrl and logsUrl where a human can go for more detailed information.

{
  "batch": {
    "createTime": "2025-10-10T15:15:21.303146Z",
    "creator": "alice@example.com",
    "labels": {
      "goog-dataproc-batch-uuid": "aaaaaaaa-bbbb-cccc-dddd-eeeeeeeeeeee",
      "goog-dataproc-location": "us-central1"
    },
    "name": "projects/google.com:hadoop-cloud-dev/locations/us-central1/batches/alice-20251010-abcd",
    "operation": "projects/google.com:hadoop-cloud-dev/regions/us-central1/operations/11111111-2222-3333-4444-555555555555",
    "runtimeConfig": {
      "properties": {
        "spark:spark.driver.cores": "4",
        "spark:spark.driver.memory": "12200m"
      }
    },
    "sparkBatch": {
      "jarFileUris": [
        "file:///usr/lib/spark/examples/jars/spark-examples.jar"
      ],
      "mainClass": "org.apache.spark.examples.SparkPi"
    },
    "state": "SUCCEEDED",
    "stateHistory": [
      {
        "state": "PENDING",
        "stateStartTime": "2025-10-10T15:15:21.303146Z"
      },
      {
        "state": "RUNNING",
        "stateStartTime": "2025-10-10T15:16:41.291747Z"
      }
    ],
    "stateTime": "2025-10-10T15:17:21.265493Z",
    "uuid": "aaaaaaaa-bbbb-cccc-dddd-eeeeeeeeeeee"
  },
  "consoleUrl": "https://console.cloud.google.com/dataproc/batches/...",
  "logsUrl": "https://console.cloud.google.com/logs/viewer?..."
}

Reference

field type required description
type string true Must be "serverless-spark-get-batch".
source string true Name of the source the tool should use.
description string true Description of the tool that is passed to the LLM.
authRequired string[] false List of auth services required to invoke this tool