<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:dc="http://purl.org/dc/elements/1.1/" version="2.0">
  <channel>
    <title>InfoQ - scikit-learn</title>
    <link>https://www.infoq.com</link>
    <description>InfoQ scikit-learn feed</description>
    <item>
      <title>Machine Learning Systems Vulnerable to Specific Attacks</title>
      <link>https://www.infoq.com/news/2022/08/machine-learning-vulnerabilities/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=scikit-learn</link>
      <description>&lt;img src="https://res.infoq.com/news/2022/08/machine-learning-vulnerabilities/en/headerimage/machine-learning-vulnerabilities-1660594757326.jpeg"/&gt;&lt;p&gt;The growing number of organizations creating and deploying machine learning solutions raises concerns as to their intrinsic security, argues the NCC Group in a recent whitepaper (Practical Attacks on Machine Learning Systems).&lt;/p&gt; &lt;i&gt;By Sergio De Simone&lt;/i&gt;</description>
      <category>TensorFlow</category>
      <category>Machine Learning</category>
      <category>Security Vulnerabilities</category>
      <category>scikit-learn</category>
      <category>Python</category>
      <category>AI, ML &amp; Data Engineering</category>
      <category>Development</category>
      <category>news</category>
      <pubDate>Mon, 15 Aug 2022 22:00:00 GMT</pubDate>
      <guid>https://www.infoq.com/news/2022/08/machine-learning-vulnerabilities/?utm_campaign=infoq_content&amp;utm_source=infoq&amp;utm_medium=feed&amp;utm_term=scikit-learn</guid>
      <dc:creator>Sergio De Simone</dc:creator>
      <dc:date>2022-08-15T22:00:00Z</dc:date>
      <dc:identifier>/news/2022/08/machine-learning-vulnerabilities/en</dc:identifier>
    </item>
  </channel>
</rss>
