<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>AI with VD — Notes from Vishvdeep Dasadiya</title>
    <link>https://aiwithvd.com/#/blog</link>
    <atom:link href="https://aiwithvd.com/rss.xml" rel="self" type="application/rss+xml" />
    <description>Field notes on machine learning, generative AI, agentic systems, and production MLOps from Vishvdeep Dasadiya.</description>
    <language>en-us</language>
    <lastBuildDate>Mon, 13 Apr 2026 00:00:00 GMT</lastBuildDate>
    <item>
      <title>Building a Production-Ready LangChain DeepAgent with SSE Streaming</title>
      <link>https://aiwithvd.com/#/blog/building-langchain-deepagent-sse-streaming</link>
      <guid isPermaLink="false">https://aiwithvd.com/#/blog/building-langchain-deepagent-sse-streaming</guid>
      <pubDate>Mon, 13 Apr 2026 00:00:00 GMT</pubDate>
      <description>How I built a production AI agent API using LangGraph, Ollama, and FastAPI with real-time SSE streaming, Redis rate limiting, and multi-turn session memory.</description>
      <category>LangChain</category>
      <category>LangGraph</category>
      <category>FastAPI</category>
      <category>Ollama</category>
      <category>SSE</category>
      <category>AI Agents</category>
      <category>Production</category>
    </item>
    <item>
      <title>OpenClaw: A Self-Hosted AI Assistant with Ollama, Telegram &amp; Discord</title>
      <link>https://aiwithvd.com/#/blog/openclaw-self-hosted-ai-assistant-guide</link>
      <guid isPermaLink="false">https://aiwithvd.com/#/blog/openclaw-self-hosted-ai-assistant-guide</guid>
      <pubDate>Mon, 13 Apr 2026 00:00:00 GMT</pubDate>
      <description>Install OpenClaw, wire Anthropic/OpenAI/Google or a local Ollama model, control it from Telegram and Discord, extend it with skills.sh, and turn it into a business gateway for lead qualification, customer support, and agent-ecosystem management.</description>
      <category>OpenClaw</category>
      <category>Ollama</category>
      <category>AI Agents</category>
      <category>Self-Hosted</category>
      <category>Telegram</category>
      <category>Discord</category>
      <category>Skills</category>
      <category>CRM</category>
      <category>Leads</category>
      <category>Customer Support</category>
      <category>MCP</category>
    </item>
    <item>
      <title>Building a Production-Ready Document Extraction Pipeline with FastAPI and MinerU</title>
      <link>https://aiwithvd.com/#/blog/building-document-extraction-pipeline-fastapi-ollama</link>
      <guid isPermaLink="false">https://aiwithvd.com/#/blog/building-document-extraction-pipeline-fastapi-ollama</guid>
      <pubDate>Thu, 09 Apr 2026 00:00:00 GMT</pubDate>
      <description>How I built a high-performance document extraction system using FastAPI, Celery, MinerU, and LLMs. Learn the architecture, challenges, and lessons from deploying AI-powered document processing at scale.</description>
      <category>FastAPI</category>
      <category>MinerU</category>
      <category>Ollama</category>
      <category>OCR</category>
      <category>Document AI</category>
      <category>LLM</category>
      <category>Celery</category>
      <category>Production</category>
    </item>
    <item>
      <title>Getting Started with LLM Agents in Production</title>
      <link>https://aiwithvd.com/#/blog/getting-started-with-llm-agents</link>
      <guid isPermaLink="false">https://aiwithvd.com/#/blog/getting-started-with-llm-agents</guid>
      <pubDate>Thu, 09 Apr 2026 00:00:00 GMT</pubDate>
      <description>A practical guide to building and deploying LLM-powered agents using LangChain and FastAPI, with lessons from real-world implementations.</description>
      <category>LLM</category>
      <category>LangChain</category>
      <category>FastAPI</category>
      <category>AI Agents</category>
      <category>Production</category>
    </item>
  </channel>
</rss>
