<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - Model - Presentations</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ Model Presentations feed</description>
    <item>
      <title>Presentation: The Data Backbone of LLM Systems</title>
      <link>https://www.infoq.com/presentations/llm-data-code-model-prompt/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Model-presentations</link>
      <description>&lt;img src="https://res.infoq.com/presentations/llm-data-code-model-prompt/en/mediumimage/paul-iusztin-medium-1756298947850.jpeg"/&gt;&lt;p&gt;Drawing from his 8 years of experience in AI, Paul Iusztin breaks down the core components of a scalable architecture, emphasizing the importance of RAG. He shares practical patterns, including the Feature Training Inference architecture, and provides a detailed use case for creating a "Second Brain" AI assistant, covering everything from data pipelines to observability and agentic layers.&lt;/p&gt; &lt;i&gt;By Paul Iusztin&lt;/i&gt;</description>
      <category>Model</category>
      <category>Data</category>
      <category>Large language models</category>
      <category>Transcripts</category>
      <category>QCon London 2025</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>presentation</category>
      <pubDate>Wed, 10 Sep 2025 13:40:00 GMT</pubDate>
      <guid>https://www.infoq.com/presentations/llm-data-code-model-prompt/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Model-presentations</guid>
      <dc:creator>Paul Iusztin</dc:creator>
      <dc:date>2025-09-10T13:40:00Z</dc:date>
      <dc:identifier>/presentations/llm-data-code-model-prompt/en</dc:identifier>
    </item>
  </channel>
</rss>
