<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - Retrieval-Augmented Generation - Presentations</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ Retrieval-Augmented Generation Presentations feed</description>
    <item>
      <title>Presentation: Deploy MultiModal RAG Systems with vLLM</title>
      <link>https://www.infoq.com/presentations/rag-vllm/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Retrieval-Augmented+Generation-presentations</link>
      <description>&lt;img src="https://res.infoq.com/presentations/rag-vllm/en/mediumimage/stephen-batifol-medium-1758016495689.jpg"/&gt;&lt;p&gt;Stephen Batifol discusses building and optimizing self-hosted, multimodal RAG systems. He breaks down vector search, nearest neighbor indexes (FLAT, IVF, HNSW), and the critical role of choosing the right embedding model. He then explains vLLM inference optimization (paged attention, quantization) and uses Mistral's Pixtral to detail multimodal large language model architecture.&lt;/p&gt; &lt;i&gt;By Stephen Batifol&lt;/i&gt;</description>
      <category>QCon London 2025</category>
      <category>Transcripts</category>
      <category>Large language models</category>
      <category>Retrieval-Augmented Generation</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>presentation</category>
      <pubDate>Fri, 10 Oct 2025 14:12:00 GMT</pubDate>
      <guid>https://www.infoq.com/presentations/rag-vllm/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Retrieval-Augmented+Generation-presentations</guid>
      <dc:creator>Stephen Batifol</dc:creator>
      <dc:date>2025-10-10T14:12:00Z</dc:date>
      <dc:identifier>/presentations/rag-vllm/en</dc:identifier>
    </item>
  </channel>
</rss>
