<?xml version='1.0'?><rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:georss="http://www.georss.org/georss" xmlns:atom="http://www.w3.org/2005/Atom" >
<channel>
	<title><![CDATA[PublMe - Space: Posted Reaction by PublMe bot in PublMe]]></title>
	<link>https://publme.space/reactions/v/46237</link>
	<atom:link href="https://publme.space/reactions/v/46237" rel="self" type="application/rss+xml" />
	<description><![CDATA[]]></description>
	
	<item>
	<guid isPermaLink="true">https://publme.space/reactions/v/46237</guid>
	<pubDate>Sat, 26 Oct 2024 22:49:16 +0200</pubDate>
	<link>https://publme.space/reactions/v/46237</link>
	<title><![CDATA[Posted Reaction by PublMe bot in PublMe]]></title>
	<description><![CDATA[
<p>OpenAI’s Whisper transcription tool has hallucination issues, researchers say</p>
<p>Software engineers, developers, and academic researchers have serious concerns about transcriptions from OpenAI’s Whisper, according to a report in the Associated Press. While there’s been no shortage of discussion around generative AI’s tendency to hallucinate — basically, to make stuff up — it’s a bit surprising that this is an issue in transcription, where you’d […]</p><p>© 2024 TechCrunch. All rights reserved. For personal use only.</p>]]></description>
	<dc:creator>PublMe bot</dc:creator>
</item>

</channel>
</rss>