<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - GPU - News</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ GPU News feed</description>
    <item>
      <title>Google New TPU Generation is Specifically Designed for Agents and SOTA Model Training</title>
      <link>https://www.infoq.com/news/2026/05/google-8th-tpu-generation/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=GPU-news</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/05/google-8th-tpu-generation/en/headerimage/google-8th-gen-tpus-1778060595193.jpeg"/&gt;&lt;p&gt;Google has unvelied a new generation of Tensor Processing Units (TPUs), featuring two specialized chips designed to accelerate model training and agent workflows, which require continuous, multi-step reasoning, and action loops distributed across multiple models. The new TPUs deliver better performance, memory, and energy efficiency, the company says.&lt;/p&gt; &lt;i&gt;By Sergio De Simone&lt;/i&gt;</description>
      <category>Agents</category>
      <category>Google</category>
      <category>Large language models</category>
      <category>GPU</category>
      <category>Development</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Wed, 06 May 2026 10:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/05/google-8th-tpu-generation/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=GPU-news</guid>
      <dc:creator>Sergio De Simone</dc:creator>
      <dc:date>2026-05-06T10:00:00Z</dc:date>
      <dc:identifier>/news/2026/05/google-8th-tpu-generation/en</dc:identifier>
    </item>
    <item>
      <title>Cloudflare Builds High-Performance Infrastructure for Running LLMs</title>
      <link>https://www.infoq.com/news/2026/05/cloudflare-llm-infrastructure/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=GPU-news</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/05/cloudflare-llm-infrastructure/en/headerimage/generatedHeaderImage-1776661318905.jpg"/&gt;&lt;p&gt;Cloudflare has recently announced new infrastructure designed to run large AI language models across its global network. As these models rely on costly hardware and must handle large volumes of incoming and outgoing text, Cloudflare separates the model's input processing and output generation onto different optimized systems.&lt;/p&gt; &lt;i&gt;By Renato Losio&lt;/i&gt;</description>
      <category>Optimization</category>
      <category>Big Data Infrastructure</category>
      <category>AI Architecture</category>
      <category>Cloudflare</category>
      <category>Large language models</category>
      <category>GPU</category>
      <category>Development</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Sun, 03 May 2026 10:58:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/05/cloudflare-llm-infrastructure/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=GPU-news</guid>
      <dc:creator>Renato Losio</dc:creator>
      <dc:date>2026-05-03T10:58:00Z</dc:date>
      <dc:identifier>/news/2026/05/cloudflare-llm-infrastructure/en</dc:identifier>
    </item>
  </channel>
</rss>
