<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>MCP on onseok</title><link>https://onseok.github.io/tags/mcp/</link><description>Recent content in MCP on onseok</description><generator>Hugo</generator><language>en-us</language><copyright>© {year} onseok</copyright><lastBuildDate>Tue, 31 Mar 2026 00:00:00 +0900</lastBuildDate><atom:link href="https://onseok.github.io/tags/mcp/index.xml" rel="self" type="application/rss+xml"/><item><title>Building a Production RAG System: From Hybrid Search to Agentic Retrieval</title><link>https://onseok.github.io/posts/building-production-rag-system/</link><pubDate>Tue, 31 Mar 2026 00:00:00 +0900</pubDate><guid>https://onseok.github.io/posts/building-production-rag-system/</guid><description>&lt;p&gt;There&amp;rsquo;s no shortage of RAG tutorials online, but most stop at the &amp;ldquo;hello world&amp;rdquo; stage: embed some documents, throw them into a vector database, retrieve top-k, and feed them to an LLM. That works for demos. It falls apart in production.&lt;/p&gt;
&lt;p&gt;I&amp;rsquo;ve been building a RAG system at work — one that searches across thousands of internal documents (engineering issues, SDK source code, design specs, technical docs) and serves results to LLM agents via &lt;a href="https://modelcontextprotocol.io/"&gt;MCP&lt;/a&gt;. I can&amp;rsquo;t go into the specifics of the product or infrastructure, but the technical challenges and lessons are universal enough to share.&lt;/p&gt;</description></item></channel></rss>