
MCP with PydanticAI
Last Updated on April 23, 2025 by Editorial Team
Author(s): Barrett Studdard
Originally published on Towards AI.
Building a basic MCP server and interacting with PydanticAI
In my prior article on building a streaming approach with Pydantic AI, I built a pattern around streaming with API calls to Anthropic. In this article, weβll look to expand to use Pydantic AI MCP Clients.
Before implementing a connection to a MCP server via Pydantic AI, letβs review what MCP is at a very high level as well as implement a simple server with Fast API.
At a high level, an MCP server allows for a standardized way to define how an LLM interacts with tools. Instead of defining tools on a one-off basis in our LLM application, we can utilize prebuilt or custom servers that expose tools. This allows for both reusability for servers we may build ourselves or plugging into various vendor or open source MCP servers β preventing us from reinventing the wheel when we want to use a new tool.
For more information, Iβd recommend reading through Anthropicβs release post, the model context protocol site, and browsing through the python sdk github repo.
For our MCP server, weβll define one very basic tool β getting a userβs name. This allows us to hardcode a name and verify the LLM is picking up the information… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI