Getting Started

Understand the core concepts of MCP Toolbox, explore integration strategies, and learn how to architect your AI agent connections.

Before you spin up your server and start writing code, it is helpful to understand the different ways you can utilize the Toolbox within your architecture.

This guide breaks down the core methodologies for using MCP Toolbox, how to think about your tool configurations, and the different ways your applications can connect to it.

Prebuilt vs. Custom Configs

MCP Toolbox provides two main approaches for tools: prebuilt and custom.

Prebuilt tools are ready to use out of the box. For example, a tool like postgres-execute-sql has fixed parameters and always works the same way, allowing the agent to execute arbitrary SQL. While these are convenient, they are typically only safe when a developer is in the loop (e.g., during prototyping, developing, or debugging).

For application use cases, you need to be wary of security risks such as prompt injection or data poisoning. Allowing an LLM to execute arbitrary queries in production is highly dangerous.

To secure your application, you should use custom tools to suit your specific schema and application needs. Creating a custom tool restricts the agent’s capabilities to only what is necessary. For example, you can use the postgres-sql tool to define a specific action. This typically involves:

  • Prepared Statements: Writing a SQL query ahead of time and letting the agent only fill in specific basic parameters.

Build-Time vs. Runtime Implementation

A key architectural benefit of the MCP Toolbox is flexibility in how and when your AI clients learn about their available tools. Understanding this distinction helps you choose the right integration path.

Build-Time

In this model, the available tools and their schemas are established when the client initializes.

  • How it works: The client launches or connects to the MCP Toolbox server, reads the available tools once, and keeps them static for the session.
  • Best for: IDEs and CLI tools

Runtime

In this model, your application dynamically requests the latest tools from the Toolbox server on the fly.

  • How it works: Your application code actively calls the server at runtime to fetch the latest toolsets and their schemas.
  • Best for: AI Agents and Custom Applications.

Usage Methodologies: How to Connect

Being built on the Model Context Protocol (MCP), MCP Toolbox is framework-agnostic. You can connect to it in three main ways:

  • IDE Integrations: Connect your local Toolbox server directly to MCP-compatible development environments.
  • CLI Tools: Use command-line interfaces like the Gemini CLI to interact with your databases using natural language directly from your terminal.
  • Application Integration (Client SDKs): If you are building custom AI agents, you can use our Client SDKs to pull tools directly into your application code. We provide native support for major orchestration frameworks including LangChain, LlamaIndex, Genkit, and more across Python, JavaScript/TypeScript, and Go.

Ready to dive in? Here are some of the most popular paths to getting your first agent up and running:

  • Python SDK Quickstart: Build a custom agent from scratch using our native Python client. This is the go-to choice for developers wanting full control over their application logic and orchestration.

  • MCP Client Quickstart: Plug your databases directly into the MCP ecosystem. Perfect for a setup that works instantly with existing MCP-compatible clients and various IDEs.

Tip

These are just a few starting points. For a complete list of tutorials, language-specific samples (Go, JS/TS, etc.), and advanced usage, explore the full Build with MCP Toolbox section.

Next Steps

Now that you understand the high-level concepts, it’s time to build!

Learn how to configure your custom MCP Toolbox Server.




Last modified March 12, 2026: gemini review (c686f472844)