<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - Model Fine Tuning</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ Model Fine Tuning feed</description>
    <item>
      <title>Google Introduces TranslateGemma Open Models for Multilingual Translation</title>
      <link>https://www.infoq.com/news/2026/01/google-translategemma-models/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Model+Fine+Tuning</link>
      <description>&lt;img src="https://res.infoq.com/news/2026/01/google-translategemma-models/en/headerimage/generatedHeaderImage-1769531019828.jpg"/&gt;&lt;p&gt;Google has released TranslateGemma, a set of open translation models based on the Gemma 3 architecture, offering 4B, 12B, and 27B parameter variants designed to support machine translation across 55 languages and to run on platforms ranging from mobile and edge devices to consumer hardware and cloud accelerators.&lt;/p&gt; &lt;i&gt;By Daniel Dominguez&lt;/i&gt;</description>
      <category>Gemma</category>
      <category>Natural Language Processing</category>
      <category>Google</category>
      <category>Model Fine Tuning</category>
      <category>Translation</category>
      <category>Large language models</category>
      <category>Reinforcement Learning</category>
      <category>Gemini</category>
      <category>Artificial Intelligence</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>news</category>
      <pubDate>Wed, 28 Jan 2026 10:16:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2026/01/google-translategemma-models/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=Model+Fine+Tuning</guid>
      <dc:creator>Daniel Dominguez</dc:creator>
      <dc:date>2026-01-28T10:16:00Z</dc:date>
      <dc:identifier>/news/2026/01/google-translategemma-models/en</dc:identifier>
    </item>
  </channel>
</rss>
