<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - Optimization</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ Optimization feed</description>
    <item>
      <title>Netflix Serves 84% of Query Results from Cache with Interval-Aware Caching in Apache Druid</title>
      <link>https://www.infoq.com/news/2026/05/netflix-druid-interval-cache/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Optimization</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/05/netflix-druid-interval-cache/en/headerimage/generatedHeaderImage-1777092326529.jpg"/&gt;&lt;p&gt;Netflix improves Apache Druid performance with interval aware caching, serving 84% of analytics results from cache and reducing query load by 33%. The system decomposes rolling window queries into reusable time segments, enabling partial cache reuse and recomputation only for recent data. At scale, it reduces scan volume, improves P90 latency, and optimizes real time analytics workloads.&lt;/p&gt; &lt;i&gt;By Leela Kumili&lt;/i&gt;</description>
      <category>Data Analytics</category>
      <category>Observability</category>
      <category>Caching</category>
      <category>Distributed Systems</category>
      <category>Optimization</category>
      <category>Apache</category>
      <category>Time Series Data</category>
      <category>Development</category>
      <category>Architecture &amp; Design</category>
      <category>news</category>
      <pubDate>Mon, 11 May 2026 14:36:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/05/netflix-druid-interval-cache/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Optimization</guid>
      <dc:creator>Leela Kumili</dc:creator>
      <dc:date>2026-05-11T14:36:00Z</dc:date>
      <dc:identifier>/news/2026/05/netflix-druid-interval-cache/en</dc:identifier>
    </item>
    <item>
      <title>OpenAI Introduces Websocket-Based Execution Mode to Reduce Latency in Agentic Workflows</title>
      <link>https://www.infoq.com/news/2026/05/openai-websocket-responses-api/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Optimization</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/05/openai-websocket-responses-api/en/headerimage/generatedHeaderImage-1777845282531.jpg"/&gt;&lt;p&gt;OpenAI introduces a WebSocket-based execution mode for its Responses API to improve agentic workflow performance in coding agents and real-time AI systems. The update reduces latency by up to 40 percent by replacing HTTP request-response cycles with persistent connections, improving streaming, tool execution, and multi-step orchestration in production-scale AI systems.&lt;/p&gt; &lt;i&gt;By Leela Kumili&lt;/i&gt;</description>
      <category>OpenAI</category>
      <category>Realtime API</category>
      <category>API</category>
      <category>Workflow Foundation</category>
      <category>AI Architecture</category>
      <category>Optimization</category>
      <category>Orchestration</category>
      <category>WebSocket</category>
      <category>Large language models</category>
      <category>Low Latency</category>
      <category>Distributed Systems</category>
      <category>Agents</category>
      <category>Artificial Intelligence</category>
      <category>AI Assisted Coding</category>
      <category>SDK</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>Development</category>
      <category>Architecture &amp; Design</category>
      <category>news</category>
      <pubDate>Thu, 07 May 2026 14:48:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/05/openai-websocket-responses-api/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Optimization</guid>
      <dc:creator>Leela Kumili</dc:creator>
      <dc:date>2026-05-07T14:48:00Z</dc:date>
      <dc:identifier>/news/2026/05/openai-websocket-responses-api/en</dc:identifier>
    </item>
    <item>
      <title>Cloudflare Builds High-Performance Infrastructure for Running LLMs</title>
      <link>https://www.infoq.com/news/2026/05/cloudflare-llm-infrastructure/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Optimization</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/05/cloudflare-llm-infrastructure/en/headerimage/generatedHeaderImage-1776661318905.jpg"/&gt;&lt;p&gt;Cloudflare has recently announced new infrastructure designed to run large AI language models across its global network. As these models rely on costly hardware and must handle large volumes of incoming and outgoing text, Cloudflare separates the model's input processing and output generation onto different optimized systems.&lt;/p&gt; &lt;i&gt;By Renato Losio&lt;/i&gt;</description>
      <category>Large language models</category>
      <category>GPU</category>
      <category>Big Data Infrastructure</category>
      <category>AI Architecture</category>
      <category>Optimization</category>
      <category>Cloudflare</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>Development</category>
      <category>news</category>
      <pubDate>Sun, 03 May 2026 10:58:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/05/cloudflare-llm-infrastructure/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Optimization</guid>
      <dc:creator>Renato Losio</dc:creator>
      <dc:date>2026-05-03T10:58:00Z</dc:date>
      <dc:identifier>/news/2026/05/cloudflare-llm-infrastructure/en</dc:identifier>
    </item>
  </channel>
</rss>
