<?xml version='1.0'?><rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:georss="http://www.georss.org/georss" xmlns:atom="http://www.w3.org/2005/Atom" >
<channel>
	<title><![CDATA[PublMe - Space: Posted Reaction by PublMe bot in PublMe]]></title>
	<link>https://publme.space/reactions/v/39424</link>
	<atom:link href="https://publme.space/reactions/v/39424" rel="self" type="application/rss+xml" />
	<description><![CDATA[]]></description>
	
	<item>
	<guid isPermaLink="true">https://publme.space/reactions/v/39424</guid>
	<pubDate>Wed, 15 May 2024 20:30:21 +0200</pubDate>
	<link>https://publme.space/reactions/v/39424</link>
	<title><![CDATA[Posted Reaction by PublMe bot in PublMe]]></title>
	<description><![CDATA[
<p>How AI Large Language Models Work, Explained Without Math</p>
<div><img width="800" height="484" src="https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?w=800" alt="" srcset="https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg 3000w, https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?resize=250, 151 250w, https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?resize=400, 242 400w, https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?resize=800, 484 800w, https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?resize=1536, 929 1536w, https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?resize=2048, 1239 2048w" data-attachment-id="310494" data-permalink="https://hackaday.com/ros-3/" data-orig-file="https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg" data-orig-size="3000,1815" data-comments-opened="1" data-image-meta="{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;1&quot;}" data-image-title="ROS" data-image-description="" data-image-caption="" data-medium-file="https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?w=400" data-large-file="https://hackaday.com/wp-content/uploads/2018/05/ros1.jpg?w=800"></div><p>Large Language Models (LLMs ) are everywhere, but how exactly do they work under the hood? [Miguel Grinberg] provides <a rel="nofollow" href="https://blog.miguelgrinberg.com/post/how-llms-work-explained-without-math" target="_blank">a great explanation of the inner workings of LLMs</a> in simple (but not simplistic) terms that eschews the low-level mathematics of <em>how</em> they work in favor of laying bare <em>what</em> it is they do.</p><p>At their heart, LLMs are prediction machines that work on tokens (small groups of letters and punctuation) and are as a result capable of great feats of human-seeming communication. Most technical-minded people understand that LLMs have no idea what they are saying, and this peek at their inner workings will make that abundantly clear.</p><p>Be sure to also review an illustrated guide to <a rel="nofollow" href="https://hackaday.com/2022/10/24/how-the-art-generating-ai-of-stable-diffusion-works/">how image-generating AIs work</a>. And if a peek under the hood of LLMs left you hungry for more low-level details, check out our coverage of <a rel="nofollow" href="https://hackaday.com/2024/04/28/train-a-gpt-2-llm-using-only-pure-c-code/">training a GPT-2 LLM using pure C code</a>.</p>]]></description>
	<dc:creator>PublMe bot</dc:creator>
</item>

</channel>
</rss>