<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>RAG on Rachid Youven Zeghlache</title><link>https://youvenz.github.io/tags/rag/</link><description>Recent content in RAG on Rachid Youven Zeghlache</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Wed, 04 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://youvenz.github.io/tags/rag/index.xml" rel="self" type="application/rss+xml"/><item><title>What is RAG? Retrieval Augmented Generation Explained</title><link>https://youvenz.github.io/blog/2026-03-04-what-is-rag-retrieval-augmented-generation-explained/</link><pubDate>Wed, 04 Mar 2026 00:00:00 +0000</pubDate><guid>https://youvenz.github.io/blog/2026-03-04-what-is-rag-retrieval-augmented-generation-explained/</guid><description>&lt;h2 id="rag-explained--how-to-give-your-llm-a-memory-without-retraining"&gt;RAG Explained — How to Give Your LLM a Memory Without Retraining&lt;/h2&gt;
&lt;p&gt;You&amp;rsquo;ve probably noticed that ChatGPT doesn&amp;rsquo;t know about events from last week, or that your company&amp;rsquo;s fine-tuned model can&amp;rsquo;t answer questions about your internal documentation. Most people assume the solution is retraining the model with new data—an expensive, time-consuming process requiring GPU clusters and ML expertise.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;There&amp;rsquo;s a better way.&lt;/strong&gt; LLMs don&amp;rsquo;t actually need to &amp;ldquo;learn&amp;rdquo; new information to use it effectively. They just need access to it at the right moment. That&amp;rsquo;s the insight behind &lt;strong&gt;RAG (Retrieval Augmented Generation)&lt;/strong&gt;, and it&amp;rsquo;s why you&amp;rsquo;re seeing it everywhere from customer support bots to research assistants.&lt;/p&gt;</description></item><item><title>Agentic AI for Health</title><link>https://youvenz.github.io/research/agentic-ai-for-health/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://youvenz.github.io/research/agentic-ai-for-health/</guid><description>&lt;h2 id="overview"&gt;Overview&lt;/h2&gt;
&lt;p&gt;Agentic AI refers to systems where a large language model (LLM) acts not just as a question-answering endpoint, but as an autonomous agent — planning, calling tools, retrieving information, and completing multi-step tasks without constant human intervention. In healthcare, this paradigm opens the door to AI systems that can assist clinical workflows, automate research tasks, and synthesise evidence at scale.&lt;/p&gt;
&lt;h2 id="frameworks-i-work-with"&gt;Frameworks I work with&lt;/h2&gt;
&lt;h3 id="langgraph"&gt;LangGraph&lt;/h3&gt;
&lt;p&gt;LangGraph models agent workflows as stateful directed graphs, enabling complex multi-step reasoning chains with loops, conditional branches, and human-in-the-loop checkpoints. This is particularly useful for clinical decision support pipelines that require evidence retrieval → reasoning → recommendation → validation cycles.&lt;/p&gt;</description></item></channel></rss>