Close Menu
    Facebook X (Twitter) Instagram
    • Privacy Policy
    • Terms Of Service
    • Social Media Disclaimer
    • DMCA Compliance
    • Anti-Spam Policy
    Facebook X (Twitter) Instagram
    Fintech Fetch
    • Home
    • Crypto News
      • Bitcoin
      • Ethereum
      • Altcoins
      • Blockchain
      • DeFi
    • AI News
    • Stock News
    • Learn
      • AI for Beginners
      • AI Tips
      • Make Money with AI
    • Reviews
    • Tools
      • Best AI Tools
      • Crypto Market Cap List
      • Stock Market Overview
      • Market Heatmap
    • Contact
    Fintech Fetch
    Home»AI News»How to Build a Stateless, Secure, and Asynchronous MCP-Style Protocol for Scalable Agent Workflows
    How to Build a Stateless, Secure, and Asynchronous MCP-Style Protocol for Scalable Agent Workflows
    AI News

    How to Build a Stateless, Secure, and Asynchronous MCP-Style Protocol for Scalable Agent Workflows

    January 14, 20264 Mins Read
    Share
    Facebook Twitter LinkedIn Pinterest Email
    ledger

    In this tutorial, we build a clean, advanced demonstration of modern MCP design by focusing on three core ideas: stateless communication, strict SDK-level validation, and asynchronous, long-running operations. We implement a minimal MCP-like protocol using structured envelopes, signed requests, and Pydantic-validated tools to show how agents and services can interact safely without relying on persistent sessions.

    import asyncio, time, json, uuid, hmac, hashlib
    from dataclasses import dataclass
    from typing import Any, Dict, Optional, Literal, List
    from pydantic import BaseModel, Field, ValidationError, ConfigDict

    def _now_ms():
    return int(time.time() * 1000)

    def _uuid():
    return str(uuid.uuid4())

    def _canonical_json(obj):
    return json.dumps(obj, separators=(“,”, “:”), sort_keys=True).encode()

    aistudios

    def _hmac_hex(secret, payload):
    return hmac.new(secret, _canonical_json(payload), hashlib.sha256).hexdigest()

    We set up the core utilities required across the entire system, including time helpers, UUID generation, canonical JSON serialization, and cryptographic signing. We ensure that all requests and responses can be deterministically signed and verified using HMAC.

    class MCPEnvelope(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    v: Literal[“mcp/0.1”] = “mcp/0.1″
    request_id: str = Field(default_factory=_uuid)
    ts_ms: int = Field(default_factory=_now_ms)
    client_id: str
    server_id: str
    tool: str
    args: Dict[str, Any] = Field(default_factory=dict)
    nonce: str = Field(default_factory=_uuid)
    signature: str

    class MCPResponse(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    v: Literal[“mcp/0.1”] = “mcp/0.1”
    request_id: str
    ts_ms: int = Field(default_factory=_now_ms)
    ok: bool
    server_id: str
    status: Literal[“ok”, “accepted”, “running”, “done”, “error”]
    result: Optional[Dict[str, Any]] = None
    error: Optional[str] = None
    signature: str

    We define the structured MCP envelope and response formats that every interaction follows. We enforce strict schemas using Pydantic to guarantee that malformed or unexpected fields are rejected early. It ensures consistent contracts between clients and servers, which is critical for SDK standardization.

    class ServerIdentityOut(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    server_id: str
    fingerprint: str
    capabilities: Dict[str, Any]

    class BatchSumIn(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    numbers: List[float] = Field(min_length=1)

    class BatchSumOut(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    count: int
    total: float

    class StartLongTaskIn(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    seconds: int = Field(ge=1, le=20)
    payload: Dict[str, Any] = Field(default_factory=dict)

    class PollJobIn(BaseModel):
    model_config = ConfigDict(extra=”forbid”)
    job_id: str

    We declare the validated input and output models for each tool exposed by the server. We use Pydantic constraints to clearly express what each tool accepts and returns. It makes tool behavior predictable and safe, even when invoked by LLM-driven agents.

    @dataclass
    class JobState:
    job_id: str
    status: str
    result: Optional[Dict[str, Any]] = None
    error: Optional[str] = None

    class MCPServer:
    def __init__(self, server_id, secret):
    self.server_id = server_id
    self.secret = secret
    self.jobs = {}
    self.tasks = {}

    def _fingerprint(self):
    return hashlib.sha256(self.secret).hexdigest()[:16]

    async def handle(self, env_dict, client_secret):
    env = MCPEnvelope(**env_dict)
    payload = env.model_dump()
    sig = payload.pop(“signature”)
    if _hmac_hex(client_secret, payload) != sig:
    return {“error”: “bad signature”}

    if env.tool == “server_identity”:
    out = ServerIdentityOut(
    server_id=self.server_id,
    fingerprint=self._fingerprint(),
    capabilities={“async”: True, “stateless”: True},
    )
    resp = MCPResponse(
    request_id=env.request_id,
    ok=True,
    server_id=self.server_id,
    status=”ok”,
    result=out.model_dump(),
    signature=””,
    )

    elif env.tool == “batch_sum”:
    args = BatchSumIn(**env.args)
    out = BatchSumOut(count=len(args.numbers), total=sum(args.numbers))
    resp = MCPResponse(
    request_id=env.request_id,
    ok=True,
    server_id=self.server_id,
    status=”ok”,
    result=out.model_dump(),
    signature=””,
    )

    elif env.tool == “start_long_task”:
    args = StartLongTaskIn(**env.args)
    jid = _uuid()
    self.jobs[jid] = JobState(jid, “running”)

    async def run():
    await asyncio.sleep(args.seconds)
    self.jobs[jid].status = “done”
    self.jobs[jid].result = args.payload

    self.tasks[jid] = asyncio.create_task(run())
    resp = MCPResponse(
    request_id=env.request_id,
    ok=True,
    server_id=self.server_id,
    status=”accepted”,
    result={“job_id”: jid},
    signature=””,
    )

    elif env.tool == “poll_job”:
    args = PollJobIn(**env.args)
    job = self.jobs[args.job_id]
    resp = MCPResponse(
    request_id=env.request_id,
    ok=True,
    server_id=self.server_id,
    status=job.status,
    result=job.result,
    signature=””,
    )

    payload = resp.model_dump()
    resp.signature = _hmac_hex(self.secret, payload)
    return resp.model_dump()

    We implement the stateless MCP server along with its async task management logic. We handle request verification, tool dispatch, and long-running job execution without relying on session state. By returning job identifiers and allowing polling, we demonstrate non-blocking, scalable task execution.

    class MCPClient:
    def __init__(self, client_id, secret, server):
    self.client_id = client_id
    self.secret = secret
    self.server = server

    async def call(self, tool, args=None):
    env = MCPEnvelope(
    client_id=self.client_id,
    server_id=self.server.server_id,
    tool=tool,
    args=args or {},
    signature=””,
    ).model_dump()
    env[“signature”] = _hmac_hex(self.secret, {k: v for k, v in env.items() if k != “signature”})
    return await self.server.handle(env, self.secret)

    async def demo():
    server_secret = b”server_secret”
    client_secret = b”client_secret”
    server = MCPServer(“mcp-server-001”, server_secret)
    client = MCPClient(“client-001”, client_secret, server)

    print(await client.call(“server_identity”))
    print(await client.call(“batch_sum”, {“numbers”: [1, 2, 3]}))

    start = await client.call(“start_long_task”, {“seconds”: 2, “payload”: {“task”: “demo”}})
    jid = start[“result”][“job_id”]

    while True:
    poll = await client.call(“poll_job”, {“job_id”: jid})
    if poll[“status”] == “done”:
    print(poll)
    break
    await asyncio.sleep(0.5)

    await demo()

    We build a lightweight stateless client that signs each request and interacts with the server through structured envelopes. We demonstrate synchronous calls, input validation failures, and asynchronous task polling in a single flow. It shows how clients can reliably consume MCP-style services in real agent pipelines.

    In conclusion, we showed how MCP evolves from a simple tool-calling interface into a robust protocol suitable for real-world systems. We started tasks asynchronously and poll for results without blocking execution, enforce clear contracts through schema validation, and rely on stateless, signed messages to preserve security and flexibility. Together, these patterns demonstrate how modern MCP-style systems support reliable, enterprise-ready agent workflows while remaining simple, transparent, and easy to extend.

    frase
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Fintech Fetch Editorial Team
    • Website

    Related Posts

    Nvidia, Groq and the limestone race to real-time AI: Why enterprises win or lose here

    Nvidia, Groq and the limestone race to real-time AI: Why enterprises win or lose here

    February 15, 2026
    AI forecasting model targets healthcare resource efficiency

    AI forecasting model targets healthcare resource efficiency

    February 14, 2026
    Exa AI Introduces Exa Instant: A Sub-200ms Neural Search Engine Designed to Eliminate Bottlenecks for Real-Time Agentic Workflows

    Exa AI Introduces Exa Instant: A Sub-200ms Neural Search Engine Designed to Eliminate Bottlenecks for Real-Time Agentic Workflows

    February 13, 2026
    Using synthetic biology and AI to address global antimicrobial resistance threat | MIT News

    Using synthetic biology and AI to address global antimicrobial resistance threat | MIT News

    February 11, 2026
    Add A Comment

    Comments are closed.

    Join our email newsletter and get news & updates into your inbox for free.


    Privacy Policy

    Thanks! We sent confirmation message to your inbox.

    livechat
    Latest Posts
    Tomasz Stańczak Steps Down as Ethereum Foundation Co-Executive Director

    Tomasz Stańczak Resigns from His Position as Co-Executive Director of the Ethereum Foundation

    February 16, 2026
    Upcoming Dividend Run For ALL?

    Anticipated Dividend Surge for Everyone?

    February 16, 2026
    Nvidia, Groq and the limestone race to real-time AI: Why enterprises win or lose here

    Nvidia, Groq and the limestone race to real-time AI: Why enterprises win or lose here

    February 15, 2026
    Bitcoin

    Bitcoin Whales Are Leaving the Profit Zone — and the Situation May Deteriorate Further

    February 15, 2026
    Trump Media Files Bitcoin, Ether and Cronos Crypto ETFs with SEC

    Trump Media Submits Bitcoin, Ether, and Cronos Crypto ETFs to SEC

    February 15, 2026
    synthesia
    LEGAL INFORMATION
    • Privacy Policy
    • Terms Of Service
    • Social Media Disclaimer
    • DMCA Compliance
    • Anti-Spam Policy
    Top Insights
    Wall Street Opens 2026 with Green Candles, Will Crypto Follow?

    The Influence of Quantum Computing on Bitcoin’s Value Assessment

    February 16, 2026
    Apollo Partners With Morpho To Support Lending Infrastructure

    Apollo Partners With Morpho To Support Lending Infrastructure

    February 16, 2026
    coinbase
    Facebook X (Twitter) Instagram Pinterest
    © 2026 FintechFetch.com - All rights reserved.

    Type above and press Enter to search. Press Esc to cancel.