As AI workflows become more modular and service-driven, integrating local LLMs into existing infrastructure becomes essential. One way to achieve this is through the Model Context Protocol (MCP), which allows services and models to expose structured, discoverable metadata. In this post, we demonstrate how to connect an MCP server with