<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - Frameworks - Articles</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ Frameworks Articles feed</description>
    <item>
      <title>Article: Disaggregation in Large Language Models: the Next Evolution in AI Infrastructure</title>
      <link>https://www.infoq.com/articles/llms-evolution-ai-infrastructure/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Frameworks-articles</link>
      <description>&lt;img src="https://res.infoq.com/articles/llms-evolution-ai-infrastructure/en/headerimage/llms-evolution-ai-infrastructure-header-1758096401785.jpg"/&gt;&lt;p&gt;Large Language Model (LLM) inference faces a fundamental challenge: the same hardware that excels at processing input prompts struggles with generating responses, and vice versa. Disaggregated serving architectures solve this by separating these distinct computational phases, delivering throughput improvements and better resource utilization while reducing costs.&lt;/p&gt; &lt;i&gt;By Anat Heilper&lt;/i&gt;</description>
      <category>Artificial Intelligence</category>
      <category>Frameworks</category>
      <category>Infrastructure</category>
      <category>Large language models</category>
      <category>Model</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>article</category>
      <pubDate>Mon, 29 Sep 2025 11:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/articles/llms-evolution-ai-infrastructure/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Frameworks-articles</guid>
      <dc:creator>Anat Heilper</dc:creator>
      <dc:date>2025-09-29T11:00:00Z</dc:date>
      <dc:identifier>/articles/llms-evolution-ai-infrastructure/en</dc:identifier>
    </item>
  </channel>
</rss>
