<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ feed</description>
    <item>
      <title>Google’s TurboQuant Compression May Support Faster Inference, Same Accuracy on Less Capable Hardware</title>
      <link>https://www.infoq.com/news/2026/04/turboquant-compression-kv-cache/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/turboquant-compression-kv-cache/en/headerimage/generatedHeaderImage-1776265077411.jpg"/&gt;&lt;p&gt;Google Research unveiled TurboQuant, a novel quantization algorithm that compresses large language models’ Key-Value caches by up to 6x. With 3.5-bit compression, near-zero accuracy loss, and no retraining needed, it allows developers to run massive context windows on significantly more modest hardware than previously required. Early community benchmarks confirm significant efficiency gains.&lt;/p&gt; &lt;i&gt;By Bruno Couriol&lt;/i&gt;</description>
      <category>Optimization</category>
      <category>Compression</category>
      <category>Performance</category>
      <category>Large language models</category>
      <category>Development</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Wed, 15 Apr 2026 16:53:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/turboquant-compression-kv-cache/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Bruno Couriol</dc:creator>
      <dc:date>2026-04-15T16:53:00Z</dc:date>
      <dc:identifier>/news/2026/04/turboquant-compression-kv-cache/en</dc:identifier>
    </item>
    <item>
      <title>Presentation: Empower Your Developers: How Open Source Dependencies Risk Management Can Unlock Innovation</title>
      <link>https://www.infoq.com/presentations/open-source-dependencies/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/presentations/open-source-dependencies/en/mediumimage/celine-pypaert-medium-1775047335370.jpeg"/&gt;&lt;p&gt;Celine Pypaert discusses the ubiquitous nature of open-source software and shares a blueprint for securing modern applications. She explains how to prioritize high-risk vulnerabilities using exploitability data, the role of Software Bill of Materials (SBOM), and the importance of bridging the gap between DevOps and Security through clear accountability and automated governance.&lt;/p&gt; &lt;i&gt;By Celine Pypaert&lt;/i&gt;</description>
      <category>QCon London 2025</category>
      <category>Risk Management</category>
      <category>Transcripts</category>
      <category>Open Source</category>
      <category>Culture &amp; Methods</category>
      <category>presentation</category>
      <pubDate>Wed, 15 Apr 2026 12:50:00 GMT</pubDate>
      <guid>https://www.infoq.com/presentations/open-source-dependencies/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Celine Pypaert</dc:creator>
      <dc:date>2026-04-15T12:50:00Z</dc:date>
      <dc:identifier>/presentations/open-source-dependencies/en</dc:identifier>
    </item>
    <item>
      <title>Zendesk Says AI Makes Code Abundant, Shifting the Bottleneck to “Absorption Capacity”</title>
      <link>https://www.infoq.com/news/2026/04/zendesk-absorption-capacity/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/zendesk-absorption-capacity/en/headerimage/Zendesk-Absorption-Capacity-Header-1776167359787.jpeg"/&gt;&lt;p&gt;Zendesk argues that GenAI shifts the bottleneck in software delivery from writing code to “absorption capacity”, which is the organisation’s ability to define problems clearly, integrate changes into the wider system, and turn implementation into reliable value. As code becomes abundant, architectural coherence, review capacity, and delivery flow become the main constraints.&lt;/p&gt; &lt;i&gt;By Eran Stiller&lt;/i&gt;</description>
      <category>Netflix</category>
      <category>Software Development Lifecycle</category>
      <category>Architecture Decision Records</category>
      <category>AI Assisted Coding</category>
      <category>Generative AI</category>
      <category>Architecture &amp; Design</category>
      <category>Culture &amp; Methods</category>
      <category>news</category>
      <pubDate>Wed, 15 Apr 2026 12:30:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/zendesk-absorption-capacity/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Eran Stiller</dc:creator>
      <dc:date>2026-04-15T12:30:00Z</dc:date>
      <dc:identifier>/news/2026/04/zendesk-absorption-capacity/en</dc:identifier>
    </item>
    <item>
      <title>Claude Code Used to Find Remotely Exploitable Linux Kernel Vulnerability Hidden for 23 Years</title>
      <link>https://www.infoq.com/news/2026/04/claude-code-linux-vulnerability/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/claude-code-linux-vulnerability/en/headerimage/generatedHeaderImage-1775817662497.jpg"/&gt;&lt;p&gt;Anthropic researcher Nicholas Carlini used Claude Code to find a remotely exploitable heap buffer overflow in the Linux kernel's NFS driver, undiscovered for 23 years. Five kernel vulnerabilities have been confirmed so far. Linux kernel maintainers report that AI bug reports have recently shifted from slop to legitimate findings, with security lists now receiving 5-10 valid reports daily.&lt;/p&gt; &lt;i&gt;By Steef-Jan Wiggers&lt;/i&gt;</description>
      <category>Linux</category>
      <category>Artificial Intelligence</category>
      <category>Claude</category>
      <category>Anthropic</category>
      <category>Development</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Wed, 15 Apr 2026 09:36:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/claude-code-linux-vulnerability/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Steef-Jan Wiggers</dc:creator>
      <dc:date>2026-04-15T09:36:00Z</dc:date>
      <dc:identifier>/news/2026/04/claude-code-linux-vulnerability/en</dc:identifier>
    </item>
    <item>
      <title>Article: Using AWS Lambda Extensions to Run Post-Response Telemetry Flush</title>
      <link>https://www.infoq.com/articles/lambda-extension-deferred-flush/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/articles/lambda-extension-deferred-flush/en/headerimage/lambda-extension-deferred-flush-header-1775648097720.jpg"/&gt;&lt;p&gt;At Lead Bank, synchronous telemetry flushing caused intermittent exporter stalls to become user-facing 504 gateway timeouts. By leveraging AWS Lambda's Extensions API and goroutine chaining in Go, flush work is moved off the response path, returning responses immediately while preserving full observability without telemetry loss.&lt;/p&gt; &lt;i&gt;By Melvin Philips&lt;/i&gt;</description>
      <category>Serverless</category>
      <category>AWS</category>
      <category>API Gateway</category>
      <category>API</category>
      <category>HTTP</category>
      <category>Cloud</category>
      <category>AWS Lambda</category>
      <category>Development</category>
      <category>DevOps</category>
      <category>article</category>
      <pubDate>Wed, 15 Apr 2026 09:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/articles/lambda-extension-deferred-flush/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Melvin Philips</dc:creator>
      <dc:date>2026-04-15T09:00:00Z</dc:date>
      <dc:identifier>/articles/lambda-extension-deferred-flush/en</dc:identifier>
    </item>
    <item>
      <title>New Rowhammer Attacks on NVIDIA GPUs Enable Full System Takeover</title>
      <link>https://www.infoq.com/news/2026/04/rowhammer-attacks-nvidia/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/rowhammer-attacks-nvidia/en/headerimage/generatedHeaderImage-1775830147456.jpg"/&gt;&lt;p&gt;Security researchers have demonstrated a new class of Rowhammer attacks targeting NVIDIA GPUs that can escalate from memory corruption to full system compromise, marking a significant shift in hardware-level security risks.&lt;/p&gt; &lt;i&gt;By Craig Risi&lt;/i&gt;</description>
      <category>Application Security</category>
      <category>Security Vulnerabilities</category>
      <category>Cloud Security</category>
      <category>DevOps</category>
      <category>news</category>
      <pubDate>Tue, 14 Apr 2026 12:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/rowhammer-attacks-nvidia/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Craig Risi</dc:creator>
      <dc:date>2026-04-14T12:00:00Z</dc:date>
      <dc:identifier>/news/2026/04/rowhammer-attacks-nvidia/en</dc:identifier>
    </item>
    <item>
      <title>Anthropic Paper Examines Behavioral Impact of Emotion-Like Mechanisms in LLMs</title>
      <link>https://www.infoq.com/news/2026/04/anthropic-paper-llms/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/anthropic-paper-llms/en/headerimage/generatedHeaderImage-1776164869421.jpg"/&gt;&lt;p&gt;A recent paper from Anthropic examines how large language models internally represent concepts related to emotions and how these representations influence behavior. The work is part of the company’s interpretability research and focuses on analyzing internal activations in Claude Sonnet 4.5 to understand the mechanisms behind model responses better.&lt;/p&gt; &lt;i&gt;By Robert Krzaczyński&lt;/i&gt;</description>
      <category>Claude</category>
      <category>Anthropic</category>
      <category>Large language models</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Tue, 14 Apr 2026 11:35:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/anthropic-paper-llms/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Robert Krzaczyński</dc:creator>
      <dc:date>2026-04-14T11:35:00Z</dc:date>
      <dc:identifier>/news/2026/04/anthropic-paper-llms/en</dc:identifier>
    </item>
    <item>
      <title>Presentation: Platform Engineering: Lessons from the Rise and Fall of eBay Velocity</title>
      <link>https://www.infoq.com/presentations/platform-engineering-lessons/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/presentations/platform-engineering-lessons/en/mediumimage/randy-shoup-medium-1775637120944.jpg"/&gt;&lt;p&gt;Randy Shoup discusses the "Velocity Initiative," a transformation that doubled engineering productivity and modernized eBay’s DORA metrics. He shares the technical playbook used to scale 4,500 services while explaining why even elite engineering execution can’t save a company hampered by waterfall planning, risk aversion, and a "pathological" culture of fear.&lt;/p&gt; &lt;i&gt;By Randy Shoup&lt;/i&gt;</description>
      <category>Best Practices</category>
      <category>Platform Engineering</category>
      <category>Transcripts</category>
      <category>Case Study</category>
      <category>QCon San Francisco 2025</category>
      <category>Architecture &amp; Design</category>
      <category>presentation</category>
      <pubDate>Tue, 14 Apr 2026 11:17:00 GMT</pubDate>
      <guid>https://www.infoq.com/presentations/platform-engineering-lessons/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Randy Shoup</dc:creator>
      <dc:date>2026-04-14T11:17:00Z</dc:date>
      <dc:identifier>/presentations/platform-engineering-lessons/en</dc:identifier>
    </item>
    <item>
      <title>Article: Beyond One-Click: Designing an Enterprise-Grade Observability Extension for Docker</title>
      <link>https://www.infoq.com/articles/enterprise-grade-observability-extension-docker/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/articles/enterprise-grade-observability-extension-docker/en/headerimage/enterprise-grade-observability-extension-docker-header-1775560652994.jpg"/&gt;&lt;p&gt;Docker Extensions boost developer speed but create a "visibility gap" by isolating telemetry. To meet enterprise needs, extensions must act as bridges to centralized platforms. This article details how to use OpenTelemetry, policy-as-code, and encryption to build secure pipelines. Learn to balance developer productivity with the governance required for scalable, compliant observability.&lt;/p&gt; &lt;i&gt;By Pragya Keshap&lt;/i&gt;</description>
      <category>Observability</category>
      <category>Docker</category>
      <category>DevOps</category>
      <category>article</category>
      <pubDate>Tue, 14 Apr 2026 09:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/articles/enterprise-grade-observability-extension-docker/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Pragya Keshap</dc:creator>
      <dc:date>2026-04-14T09:00:00Z</dc:date>
      <dc:identifier>/articles/enterprise-grade-observability-extension-docker/en</dc:identifier>
    </item>
    <item>
      <title>Airbnb Migrates High-Volume Metrics Pipeline to OpenTelemetry</title>
      <link>https://www.infoq.com/news/2026/04/airbnd-opentelemetry-vmagent/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/airbnd-opentelemetry-vmagent/en/headerimage/generatedHeaderImage-1776092205754.jpg"/&gt;&lt;p&gt;Airbnb's observability engineering team has published details of a large-scale migration away from StatsD and a proprietary Veneur-based aggregation pipeline toward a modern, open-source metrics stack built on OpenTelemetry Protocol (OTLP), the OpenTelemetry Collector, and VictoriaMetrics' vmagent. The resulting system now ingests over 100 million samples per second in production.&lt;/p&gt; &lt;i&gt;By Claudio Masolo&lt;/i&gt;</description>
      <category>Observability</category>
      <category>OpenTelemetry</category>
      <category>DevOps</category>
      <category>news</category>
      <pubDate>Tue, 14 Apr 2026 08:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/airbnd-opentelemetry-vmagent/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Claudio Masolo</dc:creator>
      <dc:date>2026-04-14T08:00:00Z</dc:date>
      <dc:identifier>/news/2026/04/airbnd-opentelemetry-vmagent/en</dc:identifier>
    </item>
    <item>
      <title>Google Released Gemma 4 with a Focus On Local-First, On-Device AI Inference</title>
      <link>https://www.infoq.com/news/2026/04/gemma-4-android-ai-inference/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/gemma-4-android-ai-inference/en/headerimage/gemma-4-android-inference-1776112478269.jpeg"/&gt;&lt;p&gt;With the release of Gemma 4, Google aims to enable local, agentic AI for Android development through a family of models designed to support the entire software lifecycle, from coding to production.&lt;/p&gt; &lt;i&gt;By Sergio De Simone&lt;/i&gt;</description>
      <category>Mobile</category>
      <category>Agents</category>
      <category>Android Studio</category>
      <category>Google+</category>
      <category>Large language models</category>
      <category>Android</category>
      <category>Development</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Mon, 13 Apr 2026 21:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/gemma-4-android-ai-inference/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Sergio De Simone</dc:creator>
      <dc:date>2026-04-13T21:00:00Z</dc:date>
      <dc:identifier>/news/2026/04/gemma-4-android-ai-inference/en</dc:identifier>
    </item>
    <item>
      <title>Lyft Scales Global Localization Using AI and Human-in-the-Loop Review</title>
      <link>https://www.infoq.com/news/2026/04/lyft-ai-localization-pipeline/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/04/lyft-ai-localization-pipeline/en/headerimage/generatedHeaderImage-1775411926263.jpg"/&gt;&lt;p&gt;Lyft has implemented an AI-driven localization system to accelerate translations of its app and web content. Using a dual-path pipeline with large language models and human review, the system processes most content in minutes, improves international release speed, ensures brand consistency, and handles complex cases like regional idioms and legal messaging efficiently.&lt;/p&gt; &lt;i&gt;By Leela Kumili&lt;/i&gt;</description>
      <category>i18n</category>
      <category>localization</category>
      <category>Data Pipelines</category>
      <category>Batch Processing</category>
      <category>Web</category>
      <category>Internationalization</category>
      <category>Automation</category>
      <category>Translation</category>
      <category>Large language models</category>
      <category>Real Time</category>
      <category>App Engine</category>
      <category>Development</category>
      <category>Architecture &amp; Design</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Mon, 13 Apr 2026 13:45:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/04/lyft-ai-localization-pipeline/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Leela Kumili</dc:creator>
      <dc:date>2026-04-13T13:45:00Z</dc:date>
      <dc:identifier>/news/2026/04/lyft-ai-localization-pipeline/en</dc:identifier>
    </item>
    <item>
      <title>Presentation: Reimagining Platform Engagement with Graph Neural Networks</title>
      <link>https://www.infoq.com/presentations/graph-neural-networks/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/presentations/graph-neural-networks/en/mediumimage/Mariia-Bulycheva-medium-1775048997053.jpeg"/&gt;&lt;p&gt;Mariia Bulycheva discusses the transition from classic deep learning to GNNs for Zalando's landing page. She explains the complexities of converting user logs into heterogeneous graphs, the "message passing" training process, and the technical pitfalls of graph data leakage. She shares how a hybrid architecture solved inference latency, delivering contextual embeddings to a downstream model.&lt;/p&gt; &lt;i&gt;By Mariia Bulycheva&lt;/i&gt;</description>
      <category>Machine Learning</category>
      <category>Transcripts</category>
      <category>Neural Networks</category>
      <category>Graph Database</category>
      <category>InfoQ Dev Summit Munich 2025</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>presentation</category>
      <pubDate>Mon, 13 Apr 2026 13:23:00 GMT</pubDate>
      <guid>https://www.infoq.com/presentations/graph-neural-networks/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Mariia Bulycheva</dc:creator>
      <dc:date>2026-04-13T13:23:00Z</dc:date>
      <dc:identifier>/presentations/graph-neural-networks/en</dc:identifier>
    </item>
    <item>
      <title>Article: The Spring Team on Spring Framework 7 and Spring Boot 4</title>
      <link>https://www.infoq.com/articles/spring-team-spring-7-boot-4/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/articles/spring-team-spring-7-boot-4/en/headerimage/spring-team-spring-7-boot-4-header-1775634533622.jpg"/&gt;&lt;p&gt;InfoQ recently spoke with key members of the Spring team about the significant architectural and functional advancements in Spring Framework 7 and Spring Boot 4. This conversation explores the strategic shift toward core resilience by integrating features such as retry and concurrency throttling directly into the framework, alongside the performance benefits of modularizing auto-configurations.&lt;/p&gt; &lt;i&gt;By Karsten Silz, Phil Webb, Sam Brannen, Rossen Stoyanchev, Mark Pollack, Martin Lippert, Michael Minella&lt;/i&gt;</description>
      <category>Spring Framework</category>
      <category>AI Coding</category>
      <category>Java</category>
      <category>Spring Boot</category>
      <category>Virtual Panel</category>
      <category>Development</category>
      <category>article</category>
      <pubDate>Mon, 13 Apr 2026 11:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/articles/spring-team-spring-7-boot-4/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Karsten Silz, Phil Webb, Sam Brannen, Rossen Stoyanchev, Mark Pollack, Martin Lippert, Michael Minella</dc:creator>
      <dc:date>2026-04-13T11:00:00Z</dc:date>
      <dc:identifier>/articles/spring-team-spring-7-boot-4/en</dc:identifier>
    </item>
    <item>
      <title>Podcast: How SBOMs and Engineering Discipline Can Help You Avoid Trivy’s Compromise</title>
      <link>https://www.infoq.com/podcasts/help-avoid-trivy-compromise/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</link>
      <description>&lt;img src="https://res.infoq.com/podcasts/help-avoid-trivy-compromise/en/smallimage/the-infoq-podcast-logo-thumbnail-1775549272964.jpg"/&gt;&lt;p&gt;Viktor Peterson, part of the CISA task force working on SBOM blueprints and co-founder of sbomify, explores the shifting landscape of software supply chain security as the EU's Cyber Resilience Act (CRA) comes into force, a "GDPR moment" for the industry.&lt;/p&gt; &lt;i&gt;By Viktor Peterson&lt;/i&gt;</description>
      <category>The InfoQ Podcast</category>
      <category>Software Development</category>
      <category>Security</category>
      <category>Software Supply Chain</category>
      <category>Compliance</category>
      <category>Architecture &amp; Design</category>
      <category>DevOps</category>
      <category>podcast</category>
      <pubDate>Mon, 13 Apr 2026 11:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/podcasts/help-avoid-trivy-compromise/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=global</guid>
      <dc:creator>Viktor Peterson</dc:creator>
      <dc:date>2026-04-13T11:00:00Z</dc:date>
      <dc:identifier>/podcasts/help-avoid-trivy-compromise/en</dc:identifier>
    </item>
  </channel>
</rss>
