<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - Transformer models</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ Transformer models feed</description>
    <item>
      <title>Transformers v5 Introduces a More Modular and Interoperable Core</title>
      <link>https://www.infoq.com/news/2025/12/transformers-hugging-face/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Transformer+models</link>
      <description>&lt;img src="https://res.infoq.com/news/2025/12/transformers-hugging-face/en/headerimage/generatedHeaderImage-1765901416110.jpg"/&gt;&lt;p&gt;Hugging Face has released the first candidate for Transformers v5, marking a significant evolution from v4 five years ago. The library has grown from a specialized model toolkit to a critical resource in AI development, achieving over three million installations daily and more than 1.2 billion total installs.&lt;/p&gt; &lt;i&gt;By Robert Krzaczyński&lt;/i&gt;</description>
      <category>Transformer models</category>
      <category>Python</category>
      <category>Hugging Face</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Tue, 16 Dec 2025 16:45:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2025/12/transformers-hugging-face/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Transformer+models</guid>
      <dc:creator>Robert Krzaczyński</dc:creator>
      <dc:date>2025-12-16T16:45:00Z</dc:date>
      <dc:identifier>/news/2025/12/transformers-hugging-face/en</dc:identifier>
    </item>
  </channel>
</rss>
