<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Standout Systems by Teodora: AI in 60 Seconds]]></title><description><![CDATA[Complex AI concepts explained clearly - by someone who builds these systems.]]></description><link>https://teodoracoach.substack.com/s/ai-in-60-seconds</link><image><url>https://substackcdn.com/image/fetch/$s_!SPkR!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea332445-8e4c-465e-af9c-2a97c29d2b6e_796x796.png</url><title>Standout Systems by Teodora: AI in 60 Seconds</title><link>https://teodoracoach.substack.com/s/ai-in-60-seconds</link></image><generator>Substack</generator><lastBuildDate>Sun, 05 Apr 2026 08:27:10 GMT</lastBuildDate><atom:link href="https://teodoracoach.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Teodora Szasz]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[dora@teodora.coach]]></webMaster><itunes:owner><itunes:email><![CDATA[dora@teodora.coach]]></itunes:email><itunes:name><![CDATA[Dr Teodora Szasz]]></itunes:name></itunes:owner><itunes:author><![CDATA[Dr Teodora Szasz]]></itunes:author><googleplay:owner><![CDATA[dora@teodora.coach]]></googleplay:owner><googleplay:email><![CDATA[dora@teodora.coach]]></googleplay:email><googleplay:author><![CDATA[Dr Teodora Szasz]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[King - Man + Woman = Queen. How AI Learned the Geometry of Meaning.]]></title><description><![CDATA[Watch now | AI in 60 Seconds - Episode 9]]></description><link>https://teodoracoach.substack.com/p/king-man-woman-queen-how-ai-learned</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/king-man-woman-queen-how-ai-learned</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Thu, 26 Mar 2026 17:04:05 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/192152206/2781d69591a638b745ff99cc55fd445e.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Watch the video above. Below: the part that still blows my mind after years of working in AI.</p><p>Quick recap. Episode 8: everything gets chopped into tokens. But a TOKEN is just a symbol - &#8220;cat&#8221; or &#8220;un&#8221; or a 16&#215;16 image patch. Symbols mean nothing to a computer.</p><p>So how does the model go from symbols to understanding?</p><p><strong>Embeddings</strong>. And they are wild.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Symbols in, meaning out</h2><p>An embedding converts a token into a long list of numbers. For GPT-5, each token becomes a vector of roughly 12,000 numbers. Think of those numbers as coordinates - a position and direction in a very high-dimensional space.</p><p>The critical property: <strong>similar meaning = pointing in the same direction.</strong></p><p>Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p><p>&#8220;Happy&#8221; and &#8220;joyful&#8221; end up pointing nearly the same way. <br>&#8220;Happy&#8221; and &#8220;carburetor&#8221; point in completely different directions. <br>&#8220;Espresso&#8221; and &#8220;cappuccino&#8221; - almost parallel. <br>&#8220;Espresso&#8221; and &#8220;democracy&#8221; - nothing in common.</p><p>The model learns these positions during training. Nobody hand-places them. They emerge from billions of next-word predictions. Words that appear in similar contexts develop similar embeddings. Because if &#8220;happy&#8221; and &#8220;joyful&#8221; can replace each other in most sentences, the model learns to give them similar directions in embedding space.</p><p><strong>Meaning becomes geography.</strong></p><h2>How the model measures &#8220;similar&#8221;</h2><p>So if similar things point the same way - how does the model actually measure that?</p><p><strong>Cosine similarity</strong>. It&#8217;s simpler than it sounds. You can also find a deep dive on cosine similarity in the RAG for Healthcare series: </p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;68bdb8cd-fe04-4533-80a0-93a75b210fa8&quot;,&quot;caption&quot;:&quot;On Day 6, you learned that cosine similarity measures how &#8220;close&#8221; two pieces of text are in meaning. You used it to retrieve clinical guideline chunks. You saw scores like 0.92 (highly relevant) and 0.71 (borderline).&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Day 15: Cosine Similarity Deep Dive - The Math Made Simple With a Patient Analogy&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:422922028,&quot;name&quot;:&quot;Dr Teodora Szasz&quot;,&quot;bio&quot;:&quot;PhD &#183; Staff ML Scientist &#183; 10+ patents &#183; FDA-cleared algorithms &#183; Women in AI Award winner. I coach AI/ML professionals at teodora.coach because I know what it takes to stand out&#8212;I'm still doing it. Actionable career systems, daily.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/423bf25c-ef09-42cf-9c8b-5c5238d0e359_2736x2736.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2026-03-21T14:11:40.415Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89c53364-df25-480b-9744-61af2ace7ff9_2713x1344.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://teodoracoach.substack.com/p/day-15-cosine-similarity-deep-dive&quot;,&quot;section_name&quot;:&quot;The Responsible AI Clinic&quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:191418584,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:1,&quot;comment_count&quot;:0,&quot;publication_id&quot;:7189588,&quot;publication_name&quot;:&quot;Standout Systems by Teodora&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!SPkR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea332445-8e4c-465e-af9c-2a97c29d2b6e_796x796.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>Imagine two arrows starting from the same point. Cosine similarity measures the angle between them:</p><ul><li><p>If they point in the same direction, angle close to zero, the similarity score is close to 1. </p></li><li><p>If they point in perpendicular directions, the score is close to 0. </p></li><li><p>If they point in opposite directions, the score is close to -1.</p></li></ul><p>That&#8217;s it. No complicated formula needed to understand it. </p><p>Just: <strong>same direction = similar meaning.</strong></p><p>Think of a compass:</p><ul><li><p>&#8220;Happy&#8221; and &#8220;joyful&#8221; are both pointing roughly northeast. </p></li><li><p>&#8220;Happy&#8221; and &#8220;sad&#8221; point in opposite directions. </p></li><li><p>&#8220;Happy&#8221; and &#8220;refrigerator&#8221; point in directions that have nothing to do with each other.</p></li></ul><p>The elegant part: cosine similarity doesn&#8217;t care about magnitude (how &#8220;far&#8221; a point is from the origin). Only the direction matters. <br><strong>A whispered &#8220;happy&#8221; and a shouted &#8220;HAPPY&#8221; might have different magnitudes, but they point the same way. The meaning is the same.</strong></p><p>This single measurement (the angle between two vectors) is the foundation of how AI decides whether two things are related. Every similarity comparison, every search-by-meaning, every &#8220;find me something like this&#8221; starts with cosine similarity on embeddings.</p><p>Remember this. It comes back in a big way when we talk about semantic search and RAG.</p><h2>Now here&#8217;s where it gets wild</h2><p>In 2013, researchers at Google trained a simple word embedding model called <a href="https://en.wikipedia.org/wiki/Word2vec">Word2Vec</a> and decided to visualize the results. What they found stunned the field.</p><p><strong>The vector from &#8220;king&#8221; to &#8220;queen&#8221; was almost identical to the vector from &#8220;man&#8221; to &#8220;woman.&#8221;</strong></p><p>King - Man + Woman = Queen. Actual vector arithmetic. Actually works.</p><p>But it wasn&#8217;t just gender. They found parallel structures everywhere:</p><p>&#8220;France&#8221; &#8594; &#8220;Paris&#8221; has the same direction as &#8220;Japan&#8221; &#8594; &#8220;Tokyo.&#8221; Country-to-capital is a consistent direction in embedding space.</p><p>&#8220;Walk&#8221; &#8594; &#8220;walked&#8221; has the same direction as &#8220;swim&#8221; &#8594; &#8220;swam.&#8221; Verb tense is a direction.</p><p>&#8220;Big&#8221; &#8594; &#8220;bigger&#8221; has the same direction as &#8220;small&#8221; &#8594; &#8220;smaller.&#8221; Comparative degree is a direction.</p><p><strong>Nobody programmed any of this.</strong> The model was just trained to predict nearby words. And from that simple task, it discovered that meaning has geometry. That abstract relationships - gender, tense, geography, degree - are directions you can measure, compare, and compute with.</p><p>This was the moment the field realized: <strong>something deep is happening in these representations. Not just pattern matching. Something that looks like structured understanding of the world.</strong></p><h2>This connects everything</h2><p>If you have been following the series, embeddings are the thread that ties it all together.</p><p><strong>Self-supervised learning</strong> (<a href="https://teodoracoach.substack.com/p/the-most-important-idea-in-ai-youve">Episode 6</a>): the model predicts the next word and learns meaning. Embeddings are WHERE that meaning gets stored - as directions in high-dimensional space.</p><p><strong>Latent prediction</strong> (<a href="https://teodoracoach.substack.com/p/predict-pixels-or-predict-meaning">Episode 7</a>): when EchoJEPA predicts in latent space instead of pixel space, it&#8217;s predicting embeddings &#8212;-points in a space where similar cardiac states point in the same direction, and where noise has been smoothed away by the EMA target encoder. <br><br><strong>The &#8220;latent space&#8221; IS an embedding space</strong>. And when we measure whether the model&#8217;s prediction is good, we&#8217;re essentially asking: does this prediction point in the same direction as the target? That&#8217;s cosine similarity at work.</p><p><strong>Attention</strong> (<a href="https://teodoracoach.substack.com/p/attention-the-engine-inside-every">Episode 2</a>): when a word &#8220;attends to&#8221; another word, it&#8217;s comparing their embeddings - asking &#8220;how relevant is your meaning to mine?&#8221; <strong>The Query and Key vectors in the attention mechanism are embeddings, and the attention score between them is related to their similarity in that space.</strong></p><p><strong>Embeddings aren&#8217;t just a step in the pipeline. They are the medium in which the entire model thinks.</strong></p><h2>The practical superpower: search by meaning</h2><p>Here&#8217;s the part that turns all of this from fascinating theory into something you can use.</p><p>Once meaning is geometry - once every piece of text, every image, every question is a direction in space - <strong>you can SEARCH by meaning. Not by keyword matching. By actual semantic similarity.</strong></p><p><strong>Embed a question. Embed a million documents. Compute the cosine similarity between your question and every document</strong>. The highest scores - the documents pointing most closely in the same direction as your question - are your most relevant results.</p><p>Search &#8220;how to fix a broken heart&#8221; and the model knows whether you mean cardiology or a breakup - because the surrounding context shifts the embedding to a completely different region of space. A keyword search returns both. A semantic search understands which one you need.</p><p>This is the foundation of how AI retrieves knowledge from databases, how smart search works, how AI assistants find the right information to answer your question. We&#8217;ll go deep on this in a future episode - but it all starts here, with embeddings and cosine similarity.</p><h2>The one thing to remember</h2><p>Embeddings turn symbols into directions in space where meaning has geometry.</p><p>Similar things point the same way. Relationships are directions. And the angle between two embeddings (<strong>cosine similarity</strong>) tells you how related they are.</p><p>Once meaning is math, you can compute with it, search it, and build everything modern AI does on top of it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jfAO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jfAO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!jfAO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!jfAO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!jfAO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jfAO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png" width="510" height="278.408203125" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1024,&quot;resizeWidth&quot;:510,&quot;bytes&quot;:1144510,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/192152206?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jfAO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 424w, https://substackcdn.com/image/fetch/$s_!jfAO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 848w, https://substackcdn.com/image/fetch/$s_!jfAO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 1272w, https://substackcdn.com/image/fetch/$s_!jfAO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13d33a53-bf0a-4f16-a343-1b252c011eed_1024x559.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>King &#8722; Man + Woman = Queen.</strong></p><p>Nobody taught it that. The model discovered the shape of meaning.</p><div><hr></div><p><em>Next up: We have talked about tokens going IN to the Transformer. But there are actually two fundamentally different machines inside: one that reads and one that writes. <br><strong>What&#8217;s an encoder? What&#8217;s a decoder? And why does it matter?</strong></em></p><p><em>I&#8217;m Teodora - AI/ML scientist. The embedding space of heart ultrasounds is where I spend most of my research time. Subscribe to Standout Systems for more.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Tokens: AI's Real Alphabet (And Why It Explains So Many AI Quirks)]]></title><description><![CDATA[AI in 60 Seconds - Episode 8 Watch the video above. Below: the small thing that explains some of AI's biggest weirdnesses.]]></description><link>https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Thu, 19 Mar 2026 17:03:42 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/191387239/1de870db3eab0641d03d2b2cdfa3b243.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>I have a confession. For seven episodes, I have been casually saying things like &#8220;words go into the Transformer&#8221; and &#8220;image patches enter the model.&#8221; It was a useful simplification. But today it&#8217;s time for the truth.</p><p>Transformers don&#8217;t see words. They see tokens. And this one detail - how data gets chopped up before the model ever touches it - explains an absurd number of things about AI that people find confusing, frustrating, or outright weird.</p><h2>What is a token?</h2><p>A token is the smallest unit a Transformer processes. Think of it as AI&#8217;s alphabet - except it&#8217;s not letters, and it&#8217;s not words. It&#8217;s something in between.</p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading Standout Systems by Teodora! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p>For text, the intuitive assumption is: one word = one token. And for common, short words, that&#8217;s true. &#8220;The&#8221; is one token. &#8220;Cat&#8221; is one token. &#8220;Run&#8221; is one token. &#8220;Paris&#8221; is one token.</p><p>But less common words get broken apart. &#8220;Unbelievable&#8221; becomes three tokens: &#8220;un&#8221; + &#8220;believ&#8221; + &#8220;able.&#8221; &#8220;Echocardiography&#8221; - a word I type daily - probably becomes four or five tokens. &#8220;Spaghettification&#8221; - the physics term for what a black hole does to you - might be six or seven.</p><p>The model splits text using an algorithm called Byte Pair Encoding (BPE). You start with individual characters and iteratively merge the most common pairs. After enough merging, you end up with a vocabulary - typically 30,000 to 100,000 tokens - that represents the most efficient way to encode the text the model was trained on.</p><p>Common patterns survive as single tokens. Rare ones get decomposed into familiar subparts. It&#8217;s actually similar to how your brain reads - you recognize frequent words instantly as whole units, but slow down and break unfamiliar words into syllables.</p><h2>The quirks this explains</h2><p>This sounds like boring preprocessing. It&#8217;s not. It explains things you&#8217;ve probably noticed and wondered about.</p><p><strong>&#8220;How many r&#8217;s in strawberry?&#8221;</strong></p><p>This is maybe the most famous AI failure. Ask almost any language model, and it&#8217;ll often get it wrong - saying 2 instead of 3. The internet loves mocking this. But the reason is simple: &#8220;strawberry&#8221; is tokenized as one or two tokens. The model never sees individual letters. It literally cannot count what it cannot see.</p><p>It&#8217;s like asking you how many times the letter &#8216;e&#8217; appears in &#8220;nevertheless&#8221; - but you can only see the word as a single shape, not as individual characters. You&#8217;d probably have to think about it too.</p><p><strong>AI struggles with arithmetic</strong></p><p>Ask an AI to add 1,847 + 3,926 and it might stumble. Not because it&#8217;s bad at math, but because those numbers don&#8217;t enter the model as clean digits. &#8220;1847&#8221; might be tokenized as &#8220;18&#8221; + &#8220;47&#8221; or &#8220;1&#8221; + &#8220;847&#8221; depending on the tokenizer. The model is trying to do math on broken-up number chunks. It&#8217;s like someone handed you an addition problem but scrambled which digits go together.</p><p>This is why newer models use special techniques for math - some process numbers digit by digit, effectively re-tokenizing for arithmetic tasks.</p><p><strong>Why AI sounds smarter in English</strong></p><p>Here&#8217;s a fairness issue hiding in tokenization that most people never think about.</p><p>Tokenizers are trained on data - and most tokenizer training data is heavily English. This means English text is tokenized efficiently: about 1.3 tokens per word on average. Common English words and patterns map to single tokens.</p><p>But other languages - Japanese, Chinese, Korean, Hindi, Arabic, and many more - are tokenized much less efficiently. The same meaning might require two to three times as many tokens.</p><p>This has cascading consequences. A 128,000-token context window holds roughly 100,000 words of English. But it might hold only 40,000-50,000 words worth of Japanese content. Same model, same price, dramatically less capacity.</p><p>It also means the model has &#8220;seen&#8221; each Japanese token far fewer times during training than each English token, so its understanding is thinner. Response quality is lower. Translation is worse. Reasoning is weaker.</p><p>The tokenizer - this seemingly boring text preprocessing step - creates a structural disadvantage for billions of non-English speakers. It&#8217;s one of those responsible AI issues that hides in the infrastructure, invisible unless you know where to look.</p><h2>Tokens beyond text</h2><p>If you&#8217;ve been following the series, you already understand visual tokenization - you just didn&#8217;t know that&#8217;s what it was called.</p><p>In Episode 3, I explained that Vision Transformers cut an image into a grid of patches - say 16&#215;16 pixels each. Each patch IS a token. It gets projected into a numerical representation and enters the Transformer alongside positional information so the model knows where the patch came from in the original image.</p><p>For video - our domain - each space-time patch is a token. A small cube spanning a few pixels across a few consecutive frames. In EchoJEPA, every heart ultrasound video becomes a sequence of hundreds of visual tokens. Attention runs across all of them - spatial and temporal - letting the model track how cardiac structures move over time.</p><p>For audio, the equivalent is spectral patches - small windows of the audio frequency spectrum become tokens.</p><p>The unifying principle: regardless of modality, step one is always the same. Chop the raw input into tokens. Give each token a numerical representation. Feed the sequence into the Transformer. This is the universal interface. Text tokens, image tokens, video tokens, audio tokens - the Transformer doesn&#8217;t care. It just processes sequences.</p><h2>The tokenization design choice</h2><p>Here&#8217;s something that connects back to Episode 7. When we talked about pixel prediction vs. latent prediction in EchoJEPA, we were really talking about two different levels of tokenization.</p><p>The MAE approach tokenizes at the pixel level - each patch token represents raw pixel values. The JEPA approach tokenizes at the latent level - each token represents a meaningful abstraction where noise has been smoothed away.</p><p>Same video. Same patches. But what those tokens CONTAIN - raw pixels or cleaned-up representations - changes what the model learns. The token is just the container. What you put inside determines everything.</p><h2>The one thing to remember</h2><p>A token is AI&#8217;s fundamental unit of perception. Everything - words, numbers, image patches, video frames, audio segments - gets converted into tokens before a Transformer can process it.</p><p>And the way you tokenize shapes everything downstream: what the model can see (individual letters? no), how much context it can hold (depends on your language), who it works well for (English speakers, mostly), and what it struggles with (counting, arithmetic, anything that requires seeing what the tokenizer hid).</p><p>Next time an AI does something weird, before blaming the model - ask yourself: what did the tokens look like?</p><div><hr></div><p><em>Next up: OK so we have tokens. But a token is still just a symbol - &#8220;un&#8221; or &#8220;believ&#8221; or a 16&#215;16 image patch. The Transformer speaks math. How does a word become a number? How does that number carry MEANING? That&#8217;s EMBEDDINGS - and it&#8217;s where the magic truly lives. Wednesday.</em></p><p><em>I&#8217;m Teodora - AI/ML scientist. Follow Standout Systems for more AI in about 60 seconds.</em></p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading Standout Systems by Teodora! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://teodoracoach.substack.com/p/tokens-ais-real-alphabet-and-why?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p></p>]]></content:encoded></item><item><title><![CDATA[Predict Pixels or Predict Meaning? The Decision That Defined Our Heart AI]]></title><description><![CDATA[AI in 60 Seconds - Episode 7]]></description><link>https://teodoracoach.substack.com/p/predict-pixels-or-predict-meaning</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/predict-pixels-or-predict-meaning</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Sat, 14 Mar 2026 17:12:39 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/190538839/f74517fe39fa5b0f4b9de546fe8a6675.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p><em>Watch the video above. <br><br>Below: the full story of a design choice that changed our AI model&#8217;s performance by 26.7% - and what it teaches about building AI that actually works in the real world.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://teodoracoach.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Last episode I explained <a href="https://teodoracoach.substack.com/p/the-most-important-idea-in-ai-youve">self-supervised learning</a> - hide part of the data, predict what&#8217;s missing, learn from the process. Simple, elegant, powerful.</p><p>But I left you with a teaser: WHAT you predict matters as much as the prediction itself.</p><p>This episode is the payoff. And it&#8217;s personal - this is the single design decision we are most proud of in building <a href="https://echojepa.com/">EchoJEPA</a>.</p><h2>Two art students and a photograph</h2><p>Let me start with an analogy.</p><p>Imagine two art students. You show both of them a photograph with a section cut out, and you ask them to fill in the missing area.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H8XV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H8XV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 424w, https://substackcdn.com/image/fetch/$s_!H8XV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 848w, https://substackcdn.com/image/fetch/$s_!H8XV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 1272w, https://substackcdn.com/image/fetch/$s_!H8XV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H8XV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png" width="1011" height="385" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:385,&quot;width&quot;:1011,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:902067,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/190538839?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb3f942e1-b465-42dc-a3dd-4e9ad7cd814e_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H8XV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 424w, https://substackcdn.com/image/fetch/$s_!H8XV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 848w, https://substackcdn.com/image/fetch/$s_!H8XV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 1272w, https://substackcdn.com/image/fetch/$s_!H8XV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4756b430-72bc-4e1b-93f0-84c430c463c0_1011x385.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Student A</strong> is meticulous. They study the surrounding pixels, the exact color values, the grain of the paper, every tiny imperfection - and try to reproduce the missing region as a pixel-perfect copy. If there&#8217;s a smudge on the lens, they reproduce the smudge. If there&#8217;s film grain, they reproduce the grain.</p><p><strong>Student B</strong> takes a different approach. They look at the scene and think: &#8220;OK, there&#8217;s a tree here, the light source is to the left, the leaves are this shade of green, there&#8217;s a shadow at this angle.&#8221; Then they fill in the missing region based on their understanding of what&#8217;s IN the image - the objects, the structure, the relationships.</p><p><strong>Student A works in pixel space</strong> - predicting the literal surface values. Student B works in latent space - predicting the meaning behind the surface.</p><p>For a clean, high-quality photograph, both students do well. The difference is barely noticeable.</p><p>But now give them a noisy image. And watch what happens.</p><h2>The noise problem in ultrasound</h2><p><strong>Echocardiography</strong> (heart ultrasound) is not like a photograph. Every ultrasound image contains speckle noise: a shimmering, grainy texture that&#8217;s fundamental to how ultrasound physics works. Sound waves bounce off tissue, interfere with each other, and create patterns that look random.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!N50T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!N50T!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 424w, https://substackcdn.com/image/fetch/$s_!N50T!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 848w, https://substackcdn.com/image/fetch/$s_!N50T!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!N50T!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!N50T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg" width="520" height="351.6033755274262" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:641,&quot;width&quot;:948,&quot;resizeWidth&quot;:520,&quot;bytes&quot;:80475,&quot;alt&quot;:&quot;Echocardiogram (echo) - Dysautonomia-MVP Center - Paula ...&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Echocardiogram (echo) - Dysautonomia-MVP Center - Paula ..." title="Echocardiogram (echo) - Dysautonomia-MVP Center - Paula ..." srcset="https://substackcdn.com/image/fetch/$s_!N50T!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 424w, https://substackcdn.com/image/fetch/$s_!N50T!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 848w, https://substackcdn.com/image/fetch/$s_!N50T!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!N50T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1a9dbeb-1b77-4465-b602-398e3a49b06b_948x641.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And crucially: <strong>speckle IS random</strong>. It&#8217;s different in every single frame, even when the heart underneath is in the exact same position doing the exact same thing. Frame 1 and frame 2 might show the same ventricle, but the speckle pattern is completely different.</p><p>This creates a problem for <strong>Student A - the pixel predictor</strong>.</p><p>When you ask a model to predict the exact pixel values of a masked region in an ultrasound image, the model MUST reproduce the speckle noise. It&#8217;s part of the pixels. The model can&#8217;t distinguish &#8220;these pixels represent the heart wall&#8221; from &#8220;these pixels represent random speckle.&#8221; It&#8217;s all just numbers.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>So the model dedicates significant capacity - significant learning - to modeling the patterns in randomness. It tries to learn the structure of noise.</p><p>It&#8217;s like asking someone to memorize TV static. There&#8217;s nothing to memorize. But the model will try.</p><p>The result: a model that&#8217;s mediocre at understanding hearts because it spent too much of its learning budget on understanding noise.</p><h2>Student B: predict in latent space</h2><p>Our solution was to become Student B. Don&#8217;t predict pixels. Predict in latent space: compressed, meaningful representations that capture what&#8217;s in the image rather than what the image literally looks like.</p><p>This is the JEPA approach: Joint-Embedding Predictive Architecture.</p><p>Here&#8217;s how it works mechanically. We have two encoders:</p><p><strong>The main encoder</strong>: this is the model we&#8217;re training. It processes the visible (unmasked) parts of the video and tries to predict representations of the masked parts.</p><p><strong>The target encoder</strong>: this processes the full video (including the masked parts) and produces the &#8220;correct answer&#8221; the main encoder is trying to match.</p><p>The target encoder is an EMA (Exponential Moving Average) of the main encoder. Instead of being trained directly, it&#8217;s updated as a running average of the main encoder&#8217;s weights over thousands of training steps.</p><p>This detail sounds minor. It&#8217;s the whole game.</p><h2>Why averaging kills noise and preserves signal</h2><p>Here&#8217;s the key insight: and once you see it, you can&#8217;t unsee it.</p><p>Speckle noise is random. In frame 1, a particular pixel might be bright due to noise. In frame 2, that same pixel might be dark. In frame 3, medium. Over thousands of frames, the noise at any given location varies randomly around some average value.</p><p>When the EMA target encoder averages over thousands of training steps, each step processing different frames with different noise patterns, the random variations cancel out. Noise that&#8217;s different every time averages toward zero.</p><p>But the heart is consistent. The left ventricle wall is in the same place, frame after frame. It contracts and relaxes in the same pattern. The valve opens and closes at the same position. Consistent structure, repeated across every frame, accumulates through averaging. It gets reinforced.</p><p>The EMA target encoder produces representations where the noise has been naturally smoothed away and the anatomy has been naturally amplified. Clean, stable descriptions of cardiac structure - for free, through the mathematics of averaging.</p><p>When we train the main encoder to predict these clean targets from noisy inputs, it learns: &#8220;ignore the noise, find the structure.&#8221; It learns to extract meaning from messy data. It learns hearts.</p><h2>The evidence</h2><p>Theory is nice. Numbers are better.</p><p><strong>Accuracy:</strong> On estimating left ventricular ejection fraction (the single most important measurement in clinical cardiology) our <a href="https://echojepa.com/">JEPA approach</a> outperformed the Masked Autoencoder (pixel prediction) baseline by 26.7%. Not a marginal improvement. A generational leap.</p><p><strong>Robustness:</strong> We deliberately degraded image quality to simulate real-world conditions - poor acoustic windows, probe movement, patient habitus issues. Our model showed 2.3% performance degradation. The best pixel-level baseline showed 16.8%. Our approach was 86% less sensitive to image quality.</p><p><strong>Transfer:</strong> Trained entirely on adult echocardiography. Tested on pediatric hearts (different size, different proportions, faster heart rates). Zero additional training. Mean absolute error: 4.32 (ours) versus 5.10 (baseline). The foundation transferred to a population it had never seen.</p><p><strong>Attention visualization:</strong> When we mapped where the model focuses its attention, it consistently localized on cardiac structures - chambers, valves, walls - not image artifacts or noise. The model learned anatomy. Evidence that it learned the right thing.<br></p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;cf48a32a-c24f-4274-96cf-ac07377bd266&quot;,&quot;duration&quot;:null}"></div><h2>The broader principle</h2><p>This extends far beyond heart ultrasound. Any time your data has noise, irrelevant variation, or messy real-world conditions (which is almost all real-world data) the choice of what you predict matters enormously.</p><p>Manufacturing images have lighting variation. Satellite images have atmospheric distortion. Audio recordings have background noise. Medical images have scanner-specific artifacts.</p><p>In every case, predicting raw pixel values forces the model to model the mess. Predicting in a cleaned-up latent space focuses the model on what matters.</p><p>It&#8217;s a design philosophy: don&#8217;t ask the model to learn everything. Ask it to learn what&#8217;s important.</p><h2>The one thing to remember</h2><p>The space you predict in determines what the model learns.</p><p>Predict pixels and you learn the surface - including all the noise, artifacts, and irrelevant variation. Predict meaning and you learn the structure - the anatomy, the patterns, the relationships that actually matter.</p><p>Sometimes the most important decision in building an AI system isn&#8217;t the architecture, the dataset, or the compute budget. It&#8217;s what you ask the model to do.</p><div><hr></div><p><em>Next up: We keep saying &#8220;words go in&#8221; and &#8220;patches go in.&#8221; But HOW does data actually enter a Transformer? What IS a token? The answer is simpler than you think - and it explains some weird things about AI you have probably noticed.</em></p><p><em>I&#8217;m Teodora - AI/ML scientist and co-author on EchoJEPA. This design decision - latent prediction over pixel prediction - is the one I&#8217;m most proud of. Subscribe to Standout Systems for more.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[The Most Important Idea in AI You've Probably Never Heard Of]]></title><description><![CDATA[AI in 60 Seconds - Episode 6]]></description><link>https://teodoracoach.substack.com/p/the-most-important-idea-in-ai-youve</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/the-most-important-idea-in-ai-youve</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Sat, 28 Feb 2026 18:03:13 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/189328075/6fdaac03f5f4cb95a5aeae1f370ebc07.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Watch the video above. <br><br>Below: Why self-supervised learning changed everything. And why your brain has been doing it your whole life.<br><br><a href="https://teodoracoach.substack.com/p/when-ai-surprises-its-own-creators">Last episode</a> I told you about emergence: <strong>how foundation models develop abilities nobody programmed</strong>. Arithmetic appearing from nowhere. Cardiac anatomy discovered without labels.</p><p>But I skipped the question that makes all of that possible: <strong>If nobody labeled the data, how did the model learn?</strong></p><p>The answer is called <strong>self-supervised learning.</strong> It&#8217;s arguably the single most important idea in modern AI. And it starts with something your brain has been doing since the day you were born.</p><h2>You already do this</h2><p>Right now, as you read this sentence, your brain is predicting. Before your eyes reach each word, your brain has already generated a probabilistic guess about what&#8217;s coming next. When the word matches your prediction - you read faster, barely registering it. When it doesn&#8217;t match, you slow down. You notice. You update your understanding.</p><p>That mismatch - prediction versus reality - is the <strong>learning signal</strong>.</p><p>You do this constantly. Walking through a doorway, your brain predicts the room on the other side. Listening to music, it predicts the next note. Having a conversation, it predicts how the sentence ends. You are a prediction machine, and the errors are how you learn about the world.</p><p>Self-supervised learning takes this principle and formalizes it for AI. Take your data. Hide part of it. Ask the model to predict the hidden part. <strong>If it gets good at predicting, it must have learned something real.</strong></p><p>The data provides its own supervision. No external labels needed. No humans in the loop. Just: hide, predict, learn.</p><h2>Three flavors, one principle</h2><p>The beauty is that this works for ANY type of data. The format changes. The principle never does.</p><p><strong>Text: predict the next word.</strong></p><p>&#8220;The capital of France is ___.&#8221; Paris. &#8220;After the rain, the streets were ___.&#8221; Wet. &#8220;E equals MC ___.&#8221; Squared.</p><p>GPT is trained this way. Every large language model is. <strong>Hide the next word, predict it, learn from billions of sentences</strong>. By getting good at this prediction task, the model learns grammar, facts, reasoning, common sense, humor, cultural references. None of which were in the training objective. The only objective was: <strong>predict the next word</strong>.</p><p>I wrote about this in an earlier article: <a href="https://teodoracoach.substack.com/p/why-your-phones-autocomplete-is-terrible?utm_source=publication-search">the &#8220;sophisticated autocomplete&#8221; framing</a>. And it&#8217;s still the most accurate description. But autocomplete at this scale, with this much data, produces something that looks astonishingly like understanding.</p><p><strong>Images: predict the missing patch.</strong></p><p>Take a photograph. Mask out 75% of it: random patches removed, like a jigsaw puzzle with most pieces missing. Ask the model to predict what was there. If it can fill in the missing eye in a face, the missing wheel on a car, the missing branch on a tree: it must have learned what those objects look like, how they&#8217;re structured, how they relate to each other.</p><p>This is how <a href="https://ai.meta.com/research/publications/masked-autoencoders-as-spatiotemporal-learners/">Meta&#8217;s MAE (Masked Autoencoder)</a> works. Mask aggressively. Predict. Learn visual representations from millions of unlabeled images.</p><p><strong>Audio: predict the missing chunk.</strong></p><p>Remove a segment from a speech recording or a song. Ask the model to predict what goes in the gap. If it can reconstruct the missing melody, the missing phonemes, the missing rhythm, it learned the structure of sound.</p><p>This is how speech models like <a href="https://ai.meta.com/research/impact/wav2vec/">wav2vec</a> learn. Same principle, different modality.</p><p><strong>Video: predict across space AND time.</strong></p><p>This is our domain. Take a heart ultrasound video. Mask out small space-time regions - cubes spanning a few pixels across a few frames. Ask: what&#8217;s missing?</p><p>If the model can predict what a masked region of a beating heart looks like in frame 15 based on the surrounding spatial context and the temporal progression from frames 1 through 14 - it must have learned how hearts look, how they move, what&#8217;s normal, and what&#8217;s not.</p><p>One principle. Every modality. <strong>HIDE, PREDICT, LEARN</strong>.</p><h2>Why this single idea changed everything</h2><p>Before self-supervised learning, AI hit a wall. The wall wasn&#8217;t data: data was everywhere. The wall was labels.</p><p>Training a supervised model to detect heart disease requires a cardiologist to annotate each image. At scale, 18 million videos, that means thousands of doctors working for years. Millions of dollars. The bottleneck was never the data itself. It was the human effort required to make the data usable.</p><p>Self-supervised learning removed the bottleneck entirely.</p><p>Instead of requiring humans to label every example, the model creates its own training signal from the raw data. <strong>And suddenly, ALL data becomes training data</strong>. Every book ever written. Every medical image in every hospital archive. Every video ever recorded. Every audio file, every satellite image, every code repository.</p><p>This is why foundation models exploded. Not because of a new architecture - <a href="https://teodoracoach.substack.com/p/what-is-a-transformer-and-why-you">Transformers</a> had existed since 2017. Not because of new hardware - GPUs had been improving steadily. But because <strong>self-supervised learning gave us a way to USE all the data</strong>. The data was always there. The method to learn from it changed.</p><p>The internet wasn&#8217;t a dataset until self-supervised learning made it one. Hospital archives weren&#8217;t training data until we learned to train without labels.</p><h2>The version we built: what happens in latent space</h2><p>Now, there&#8217;s a nuance here that sets up next episode&#8217;s topic.</p><p>When you do self-supervised learning on images, you have a choice. You can predict the exact pixel values of the masked region (<a href="https://arxiv.org/abs/2111.06377">the MAE approach</a>). Or you can predict something more abstract: a compressed representation in latent space (<a href="https://ai.meta.com/blog/yann-lecun-ai-model-i-jepa/">the JEPA approach</a>).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Wiil!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Wiil!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!Wiil!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!Wiil!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!Wiil!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Wiil!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png" width="506" height="284.625" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:506,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Wiil!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!Wiil!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!Wiil!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!Wiil!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6106224a-b296-48d0-915c-043502e9e3c3_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">JEPA&#8217;s architect, Meta&#8217;s Chief AI Scientist, Yann LeCun.</figcaption></figure></div><p>For our echocardiography model, this choice mattered enormously. Ultrasound has speckle noise - random grainy texture that&#8217;s different every frame. Predicting exact pixels means predicting noise. Predicting in latent space means the noise gets averaged away and the model focuses on consistent structure - the actual anatomy.</p><p>We will go deep on this next episode. For now, just know: <strong>WHAT you predict in self-supervised learning matters as much as the prediction task itself.</strong></p><h2>What the model discovered</h2><p>Here&#8217;s the part that still gives me chills.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;95cea153-e707-4f4a-8c8b-b8c8729a1fef&quot;,&quot;duration&quot;:null}"></div><p>We trained <a href="https://echojepa.com/">EchoJEPA</a> on 18 million echocardiography videos. The only task: predict masked space-time regions in latent space. We never told it what a ventricle is. Never showed it an anatomy textbook. Never labeled a single frame.</p><p>After training, we visualized where the model&#8217;s attention focuses. It had organized itself around the left ventricle, the right ventricle, the valve positions, the wall boundaries. It discovered the functional anatomy of the heart - from raw video alone.</p><p>Then we tested it on pediatric hearts - completely different from the adult hearts it trained on. Different size, different proportions, faster rates. Zero additional training. And it worked. The foundation transferred.</p><p>The model taught itself hearts. By predicting missing pieces. From 18 million unlabeled videos.</p><p><strong>Hide. Predict.</strong> <strong>Learn</strong>. That&#8217;s it. And from that simple idea - everything in modern AI follows.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ndYl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ndYl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 424w, https://substackcdn.com/image/fetch/$s_!ndYl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 848w, https://substackcdn.com/image/fetch/$s_!ndYl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 1272w, https://substackcdn.com/image/fetch/$s_!ndYl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ndYl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png" width="945" height="517" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:517,&quot;width&quot;:945,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:977530,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/189328075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88a6c217-f0c6-47a1-b7e8-17dee3e3cd39_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ndYl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 424w, https://substackcdn.com/image/fetch/$s_!ndYl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 848w, https://substackcdn.com/image/fetch/$s_!ndYl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 1272w, https://substackcdn.com/image/fetch/$s_!ndYl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F429e3f6e-f125-4022-88e1-ca5e1a6366a4_945x517.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The one thing to remember</h2><p><strong>Self-supervised learning: your data is the teacher. </strong>Hide part of it, predict it back, and understanding emerges.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YDLL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YDLL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 424w, https://substackcdn.com/image/fetch/$s_!YDLL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 848w, https://substackcdn.com/image/fetch/$s_!YDLL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 1272w, https://substackcdn.com/image/fetch/$s_!YDLL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YDLL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png" width="945" height="516" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:516,&quot;width&quot;:945,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1009112,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/189328075?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F76a8d2c4-6720-4520-8555-d0787193053e_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YDLL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 424w, https://substackcdn.com/image/fetch/$s_!YDLL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 848w, https://substackcdn.com/image/fetch/$s_!YDLL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 1272w, https://substackcdn.com/image/fetch/$s_!YDLL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed4c8e45-3f3e-4948-a341-f6bf8d662a96_945x516.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It&#8217;s how GPT learned language. It&#8217;s how Vision Transformers learned to see. It&#8217;s how our model learned cardiac anatomy. And it&#8217;s the reason foundation models exist - because self-supervised learning unlocked all the data that was sitting there, waiting to become knowledge.</p><div><hr></div><p><em>Next up: when the model predicts the missing piece, <strong>WHAT should it predict? Exact pixels - or something deeper?</strong> That choice is the most important design decision we made in <a href="https://echojepa.com/">EchoJEPA</a>. And it changed everything.</em></p><p><em>I&#8217;m Teodora: AI/ML scientist. I build foundation models that teach themselves from unlabeled clinical data. Follow Standout Systems for more AI in about 60 seconds.</em></p>]]></content:encoded></item><item><title><![CDATA[When AI Surprises Its Own Creators]]></title><description><![CDATA[AI in 60 Seconds - Episode 5]]></description><link>https://teodoracoach.substack.com/p/when-ai-surprises-its-own-creators</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/when-ai-surprises-its-own-creators</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Thu, 26 Feb 2026 18:05:19 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/188945885/0af3c8602e9c940db5fca1fcafbccd21.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Watch the video above. <br>Below: the part of AI that genuinely keeps researchers up at night - in a good way and a bad way.<br><br>Last episode I explained the three steps of LLM training: pretraining on massive data, fine-tuning to be helpful, and RLHF to learn values. Straightforward enough.</p><p>But here&#8217;s what I left out: something happens at scale that nobody fully understands. And it&#8217;s the reason the term &#8220;foundation model&#8221; exists - and the reason some very smart people are both thrilled and worried.</p><h2>The arithmetic mystery</h2><p>In 2022, researchers at Google published a paper that documented something strange.</p><p>They took language models of increasing size - same architecture, same training data, just more parameters - and tested them on a bunch of tasks. For most tasks, performance improved gradually. A little bigger, a little better. Expected.</p><p>But for some tasks, performance did something else entirely.</p><p>At one billion parameters: near-zero accuracy on three-digit arithmetic. The model can&#8217;t add 472 + 385. At ten billion: still near-zero. Still can&#8217;t do it. At sixty billion parameters: suddenly 80%+ accuracy. The model can now do arithmetic. Not a little. A lot.</p><p>The jump wasn&#8217;t gradual. It was a phase transition - like water going from liquid to gas. One moment the ability isn&#8217;t there. The next moment it is.</p><p>This happened with task after task. Multi-step logical reasoning: nothing, nothing, nothing, suddenly yes. Translating between language pairs that weren&#8217;t explicitly in the training data: nothing, nothing, suddenly yes. Understanding sarcasm and irony: same pattern.</p><p>Nobody programmed these abilities. The training objective was always the same: predict the next word. But at sufficient scale, capabilities EMERGED that weren&#8217;t in the training signal.</p><p>Researchers call these emergent capabilities. And they are the defining feature of foundation models.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9yS8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9yS8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 424w, https://substackcdn.com/image/fetch/$s_!9yS8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 848w, https://substackcdn.com/image/fetch/$s_!9yS8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 1272w, https://substackcdn.com/image/fetch/$s_!9yS8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9yS8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png" width="464" height="242.6422018348624" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:513,&quot;width&quot;:981,&quot;resizeWidth&quot;:464,&quot;bytes&quot;:1160051,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/188945885?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc204d53e-66b9-4d67-8870-e3d46aaf4887_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9yS8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 424w, https://substackcdn.com/image/fetch/$s_!9yS8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 848w, https://substackcdn.com/image/fetch/$s_!9yS8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 1272w, https://substackcdn.com/image/fetch/$s_!9yS8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee45a5fe-17a7-4eed-a903-5afca7e9c3c5_981x513.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>What actually IS a foundation model?</h2><p>The term was coined in a 2021 Stanford paper, and the definition is specific. A foundation model has two properties:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://teodoracoach.substack.com/subscribe?"><span>Subscribe now</span></a></p><p><strong>Scale.</strong> Not trained on a curated dataset for a single task. Trained on an enormous, broad dataset - the internet, millions of images, millions of videos. The scale is what enables emergence. You can&#8217;t get emergent capabilities from a small model; there aren&#8217;t enough parameters for the complex internal representations to form.</p><p><strong>Generality.</strong> Because it&#8217;s trained on broad data with a general objective (predict the next word, predict the missing patch), the model develops general-purpose understanding that can be adapted to virtually any downstream task. One model, many applications. That&#8217;s the &#8220;foundation&#8221; - the base everything else is built on.</p><p>GPT-4 is a foundation model. So is Gemini. Apple&#8217;s AFM. Claude. Llama. And - in a different domain - EchoJEPA.</p><h2>&#8220;But wait - emergence might be a mirage&#8221;</h2><p>I want to be honest here because this is an active debate in the field, and I think you deserve the real picture.</p><p>In 2023, a group of Stanford researchers published a paper arguing that emergence might be partially an illusion of measurement. Their argument: if you measure model performance using metrics that have sharp thresholds (like exact-match accuracy - either the answer is exactly right or it&#8217;s wrong), you&#8217;ll see sudden jumps. But if you use smoother metrics (like how close the model gets to the right answer), the improvement looks more gradual.</p><p>In other words: maybe the model IS getting better at arithmetic at 10 billion parameters - just not well enough to get the exact right answer. And at 60 billion, it crosses the threshold of &#8220;close enough to be exactly right.&#8221; The phase transition might be in the metric, not the model.</p><p>This is a legitimate critique. And it&#8217;s probably partially true for some capabilities.</p><p>But not all. Some capabilities, like the model spontaneously developing the ability to perform tasks in languages it saw very little of during training, or learning to use tools it was never explicitly trained to use  are harder to explain as measurement artifacts. Something IS happening at scale that goes beyond gradual improvement. We just don&#8217;t fully understand what.</p><p>And that honesty - &#8220;we don&#8217;t fully understand this&#8221; - is important. Because it&#8217;s the reason responsible AI work matters so much. If we can&#8217;t predict what capabilities will emerge at the next scale, we also can&#8217;t predict what FAILURES will emerge.</p><h2>I saw emergence in hearts</h2><p>This isn&#8217;t just a language model phenomenon. I saw it firsthand with EchoJEPA.</p><p>We trained a Vision Transformer on 18 million echocardiography videos. The training objective: predict masked regions in latent space. We never taught it anatomy. Never labeled a heart structure. Never told it what a ventricle is, where valves are, what normal motion looks like.</p><p>But when we visualized the model&#8217;s attention patterns - where it focuses when processing a video - it had learned to localize on cardiac chambers. On valve positions. On wall motion. The internal representations organized themselves around anatomically meaningful structures.</p><p>That&#8217;s emergence. The model discovered cardiac anatomy because understanding anatomy is useful for predicting what a missing region of heart video looks like. Nobody programmed that understanding. It emerged from the task and the scale.</p><p>And then something even more unexpected. We trained exclusively on adult echocardiography. Adults. Full-grown hearts with adult proportions. Then we tested on pediatric hearts - smaller, different chamber proportions, faster heart rates. Zero additional training. Zero pediatric data.</p><p>It worked. Mean absolute error of 4.32 compared to 5.10 for the best baseline. The foundation model had learned something general enough about cardiac anatomy and motion that it transferred across patient populations it had never seen.</p><p>That&#8217;s the power of a foundation. At sufficient scale, it doesn&#8217;t just memorize - it understands. And understanding transfers.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6CGZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6CGZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 424w, https://substackcdn.com/image/fetch/$s_!6CGZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 848w, https://substackcdn.com/image/fetch/$s_!6CGZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 1272w, https://substackcdn.com/image/fetch/$s_!6CGZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6CGZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png" width="528" height="254.74639175257732" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:468,&quot;width&quot;:970,&quot;resizeWidth&quot;:528,&quot;bytes&quot;:1043020,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/188945885?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b643664-9eb9-4bf6-b63b-0b648e2ad320_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6CGZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 424w, https://substackcdn.com/image/fetch/$s_!6CGZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 848w, https://substackcdn.com/image/fetch/$s_!6CGZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 1272w, https://substackcdn.com/image/fetch/$s_!6CGZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c8db26-68f7-4f77-9c91-d6983fe6e4dd_970x468.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The honest picture: exciting and concerning</h2><p>I want to be transparent about both sides because this is where a lot of AI commentary goes wrong - people either say &#8220;emergence is magical and AI is about to solve everything&#8221; or &#8220;emergence is terrifying and we should stop.&#8221;</p><p><strong>What&#8217;s exciting:</strong> Foundation models develop capabilities nobody explicitly designed, which means they can solve problems we haven&#8217;t thought to train them for. EchoJEPA&#8217;s zero-shot transfer to pediatric hearts wasn&#8217;t planned - it was discovered. This kind of unexpected generalization is how foundation models could accelerate medical research, scientific discovery, and creative work in ways we can&#8217;t fully anticipate.</p><p><strong>What&#8217;s concerning:</strong> If we can&#8217;t predict what capabilities emerge at the next scale, we also can&#8217;t predict what failure modes emerge. A model might develop the ability to produce convincing misinformation as an emergent capability of getting really good at language. It might develop biases that only appear at scale because they require complex reasoning to manifest. This is why responsible AI evaluation can&#8217;t be a one-time checklist - it needs to be continuous and proactive, testing for capabilities and failures we haven&#8217;t seen yet.</p><p>In my world - healthcare AI - this tension is very real. The same emergence that lets EchoJEPA discover cardiac anatomy without labels could, in theory, also learn subtle biases in the data - like patterns correlated with which patients get referred for imaging in the first place. That&#8217;s why our evaluation includes subgroup analysis across demographics, multi-site validation, and deliberate stress testing. You can&#8217;t trust emergence blindly. You have to verify what it learned.</p><h2>The one thing to remember</h2><p>A foundation model isn&#8217;t just a big model. It&#8217;s a model where scale produces something qualitatively new - capabilities that weren&#8217;t programmed, weren&#8217;t predicted, and aren&#8217;t fully understood.</p><p>That&#8217;s what makes them powerful. And that&#8217;s what makes Responsible AI work essential.</p><div><hr></div><p><em>Next up: Foundation models learn from massive data. But nobody labeled that data. How does a model teach itself? Self-supervised learning - the trick that makes everything we discussed today possible. Stay tuned. </em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>I&#8217;m Teodora - AI/ML scientist. I build foundation models for healthcare where emergence is both the breakthrough and the thing that keeps me carefully checking everything twice. Subscribe to Standout Systems for more.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://teodoracoach.substack.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[How Is an LLM Actually Trained? (It's Not Just "Read the Internet")]]></title><description><![CDATA[AI in 60 Seconds - Episode 4]]></description><link>https://teodoracoach.substack.com/p/how-is-an-llm-actually-trained-its</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/how-is-an-llm-actually-trained-its</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Tue, 24 Feb 2026 18:12:25 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/188865308/94339210daf328dee3e22dbc971b73a5.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Watch the video above. Below: the full story - and why Step 3 is the one that keeps me up at night.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><br>I saw a post this week that got a lot of engagement. <br>The gist: AI was trained on the internet, the same internet full of misinformation and conspiracy theories, and now we&#8217;re asking it to help us make decisions.</p><p>It&#8217;s a fair point. But it&#8217;s incomplete. And the incomplete version is actually more dangerous than the full picture - because it makes people either dismiss AI entirely or trust it blindly, when neither is appropriate.</p><p>So here&#8217;s the full picture. Three steps. Each one fundamentally changes what the model is.</p><h2>Step 1: Pretraining - &#8220;Read everything, predict the next word&#8221;</h2><p>This is the step that post was describing. And yes - it&#8217;s exactly as wild as it sounds.</p><p>You take an enormous amount of text. Books. Wikipedia. News articles. Academic papers. Reddit threads. Code repositories. Forum posts. Some filtered, some not. Trillions of tokens.</p><p>Then you give the model a single task: given the beginning of a sentence, predict the next word.</p><p>&#8220;The capital of France is ___.&#8221; &#8594; Paris. &#8220;Photosynthesis converts sunlight into ___.&#8221; &#8594; energy. &#8220;The earth is ___.&#8221; &#8594; ... well, it depends on context.</p><p>If the model gets good at prediction across trillions of examples, something remarkable happens. It doesn&#8217;t just memorize sequences - it learns grammar, facts, reasoning patterns, even rudimentary common sense. All from the pressure of predicting what comes next.</p><p>I wrote about this in a <a href="https://teodoracoach.substack.com/p/why-your-phones-autocomplete-is-terrible?utm_source=publication-search">previous article</a> &#8212; the &#8220;sophisticated autocomplete&#8221; framing. And I still think it&#8217;s the most honest description of what&#8217;s happening. The model is a prediction engine. But prediction at this scale produces something that looks a lot like understanding.</p><p>Here&#8217;s the critical thing about this step though: <strong>after pretraining, you don&#8217;t have ChatGPT.</strong> You have a base model. And a base model will happily write you a conspiracy theory with the same fluency as a physics textbook. It learned the patterns of ALL the text - the knowledge and the nonsense. It doesn&#8217;t know which is which.</p><p>The base model is not the model you talk to.</p><h2>Step 2: Fine-tuning - &#8220;Learn to be helpful&#8221;</h2><p>The base model is brilliant but useless as an assistant. Ask it &#8220;What&#8217;s the capital of France?&#8221; and it might continue writing a Wikipedia paragraph instead of answering your question. It doesn&#8217;t know it&#8217;s supposed to be <em>helping</em> you. It just predicts text.</p><p>Fine-tuning fixes this. Human trainers write thousands of example conversations:</p><p><em>User: What&#8217;s the capital of France?</em> <em>Good response: The capital of France is Paris.</em></p><p><em>User: Can you help me write an email?</em> <em>Good response: Of course! What&#8217;s the context and tone you&#8217;re going for?</em></p><p>The model learns from these examples. It learns the FORMAT of being an assistant - answer questions directly, be concise, ask for clarification when needed, stay helpful.</p><p>This step transforms the model from a general text predictor into something that actually tries to respond usefully. But it still has a problem: it doesn&#8217;t have strong judgment about what it SHOULD and SHOULDN&#8217;T say.</p><h2>Step 3: RLHF - &#8220;Learn right from wrong&#8221;</h2><p>This is the step most people don&#8217;t know about. And it&#8217;s the most important one for responsible AI.</p><p>RLHF stands for Reinforcement Learning from Human Feedback. Here&#8217;s how it works:</p><p>The model generates two different responses to the same prompt. Human evaluators look at both and choose: which one is better? More helpful? More accurate? Safer? Thousands of these comparisons.</p><p>From all those human judgments, you train a reward model - a separate AI that learns to predict which responses humans would prefer. Then you use that reward model as a coach: the main model generates a response, the reward model says &#8220;good&#8221; or &#8220;not good,&#8221; and over millions of iterations, the model learns to consistently produce preferred outputs.</p><p>This is where the model learns: don&#8217;t repeat conspiracy theories as fact. Don&#8217;t generate instructions for harmful activities. Acknowledge when you&#8217;re uncertain. Refuse requests that could cause harm.</p><p>Is it perfect? No. The quality of RLHF depends entirely on the quality of the human evaluation - who the evaluators are, what guidelines they follow, what biases they bring. If the evaluation only covers English, the model won&#8217;t be equally safe in Japanese. If the evaluators have blind spots, the model inherits them.</p><p>This is exactly the kind of problem I work on. Building robust evaluation pipelines, ensuring diverse representation in the evaluation process, measuring fairness across demographics and languages. The alignment step is where responsible AI engineering matters most.</p><h2>Why the three-step picture matters</h2><p><strong>That viral post implied: AI learned from garbage, therefore AI outputs garbage.</strong></p><p><strong>The reality is more nuanced. The base model learned from everything - knowledge and nonsense alike. Fine-tuning taught it to be an assistant. RLHF taught it values and judgment.</strong></p><p><strong>The model that read the internet is not the model you talk to.</strong> Three transformations happened in between.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jaEd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jaEd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 424w, https://substackcdn.com/image/fetch/$s_!jaEd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 848w, https://substackcdn.com/image/fetch/$s_!jaEd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 1272w, https://substackcdn.com/image/fetch/$s_!jaEd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jaEd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png" width="586" height="215.32974137931035" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:341,&quot;width&quot;:928,&quot;resizeWidth&quot;:586,&quot;bytes&quot;:508504,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/188865308?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc51f5090-531b-47d4-8ea6-bd4e4af9f01e_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jaEd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 424w, https://substackcdn.com/image/fetch/$s_!jaEd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 848w, https://substackcdn.com/image/fetch/$s_!jaEd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 1272w, https://substackcdn.com/image/fetch/$s_!jaEd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1076710b-7b2c-445e-8854-8cf1cae06032_928x341.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Does this mean current LLMs are perfectly safe? Absolutely not. RLHF can be gamed. Safety training can be bypassed. The models still hallucinate - confidently stating things that aren&#8217;t true. The evaluation pipelines have gaps.</p><p>But understanding the full training process matters because it tells you WHERE the problems live:</p><p><strong>Hallucination?</strong> Mostly a pretraining problem - the model learned to always sound confident because confident text was in the training data.</p><p><strong>Bias?</strong> Could be in any of the three steps - biased training data, biased fine-tuning examples, or biased human evaluators in RLHF.</p><p><strong>Safety failures?</strong> Usually a gap in Step 3 - an adversarial prompt the safety training didn&#8217;t anticipate.</p><p>When you understand where problems come from, you stop saying &#8220;AI is garbage&#8221; or &#8220;AI is magic&#8221; and start asking the right question: <strong>where specifically is this model likely to fail, and how do I account for that?</strong></p><p>That&#8217;s AI literacy. And that&#8217;s a career superpower.</p><h2>The one thing to remember</h2><p>Three steps. Pretraining gives the model knowledge (and nonsense). Fine-tuning gives it manners. RLHF gives it judgment.</p><p>The internet trained the base model. Humans trained the assistant.</p><div><hr></div><p><em>I&#8217;m Teodora - AI/ML scientist. I build AI systems where the gap between Steps 1 and 3 can mean the difference between a correct diagnosis and a missed one. Subscribe to Standout Systems for more.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Same Trick, But for Images: Vision Transformers Explained]]></title><description><![CDATA[Watch now | AI in 60 Seconds - Episode 3]]></description><link>https://teodoracoach.substack.com/p/same-trick-but-for-images-vision</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/same-trick-but-for-images-vision</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Sun, 22 Feb 2026 18:03:15 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/188689871/cdbf8ba9352badc06056f2b95d78af92.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Watch the video above for the 120+ second version. Below: the part that didn't fit.<br><br>Here&#8217;s the thing that makes the Transformer architecture genuinely beautiful: it doesn&#8217;t care what you feed it.</p><p>Words? Sure. Image patches? Also works. Audio chunks? Yep. Video frames? That too.</p><p>The last two episodes covered how <a href="https://teodoracoach.substack.com/p/what-is-a-transformer-and-why-you">Transformers</a> work on text - <a href="https://teodoracoach.substack.com/p/attention-the-engine-inside-every">self-attention</a> letting every word figure out which other words matter. Today: how the exact same idea works on images.</p><h2>The patch trick</h2><p>A Transformer expects a sequence. Text is already a sequence - words in a row. Images aren&#8217;t. They&#8217;re a 2D grid of pixels.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://teodoracoach.substack.com/subscribe?"><span>Subscribe now</span></a></p><p>So you make them a sequence. Take an image, divide it into a grid of small patches - say 16&#215;16 pixels each. Flatten the grid into a row. Now you have a sequence of patches, just like a sequence of words.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Sfg_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Sfg_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 424w, https://substackcdn.com/image/fetch/$s_!Sfg_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 848w, https://substackcdn.com/image/fetch/$s_!Sfg_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 1272w, https://substackcdn.com/image/fetch/$s_!Sfg_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Sfg_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png" width="997" height="310" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:310,&quot;width&quot;:997,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:482608,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/188689871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F326fcce2-8a2d-4243-9e2a-7fb42f09e0e4_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Sfg_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 424w, https://substackcdn.com/image/fetch/$s_!Sfg_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 848w, https://substackcdn.com/image/fetch/$s_!Sfg_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 1272w, https://substackcdn.com/image/fetch/$s_!Sfg_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F529015ae-801d-421b-a3ec-753b2cb6e24d_997x310.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Each patch gets embedded - converted into a numerical representation the model can work with. Then you feed the whole sequence into a standard Transformer and let self-attention run.</p><p>That&#8217;s it. That&#8217;s a Vision Transformer - ViT. Published by Google in 2020. Same attention mechanism from the 2017 &#8220;Attention Is All You Need&#8221; paper, applied to image patches instead of words.</p><h2>What attention sees in an image</h2><p>This is where it gets interesting. When attention runs across patches of an image, each patch learns to look at the patches that are most relevant to it.</p><p>A patch containing a dog&#8217;s ear attends to the patch containing the other ear and the face. A patch showing sky doesn&#8217;t attend much to a patch showing a shoe. The model builds an understanding of the image by figuring out which parts relate to which other parts - no one programs these relationships. They emerge from the data.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9tZb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9tZb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 424w, https://substackcdn.com/image/fetch/$s_!9tZb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 848w, https://substackcdn.com/image/fetch/$s_!9tZb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 1272w, https://substackcdn.com/image/fetch/$s_!9tZb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9tZb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png" width="1015" height="444" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:444,&quot;width&quot;:1015,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:919173,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/188689871?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd678a7b5-d47c-4ad7-b62e-921954421d17_1024x559.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9tZb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 424w, https://substackcdn.com/image/fetch/$s_!9tZb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 848w, https://substackcdn.com/image/fetch/$s_!9tZb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 1272w, https://substackcdn.com/image/fetch/$s_!9tZb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01f788ab-788f-4fa5-8bfe-125d2ba43c36_1015x444.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And because attention is global - every patch can attend to every other patch regardless of distance - the model handles long-range relationships naturally. The top-left corner of an image can directly attend to the bottom-right. In a traditional convolutional network, that kind of long-range connection requires many stacked layers. In a Vision Transformer, it&#8217;s one attention computation.</p><h2>From images to video: space-time patches</h2><p>Here&#8217;s where it clicked for me personally.</p><p>If you can cut a 2D image into patches, you can cut a 3D video into small cubes - patches that span both space and time. These are sometimes called tubelets. Each tubelet captures a small region of a few consecutive frames.</p><p>Feed those into a Transformer, run attention, and the model can track how things move and change across time. A patch showing the left ventricle wall in frame 1 attends to the same region in frame 10 - learning motion patterns.</p><p>This is exactly what we built in <a href="https://substack.com/@teodora485587/note/p-187210316">EchoJEPA</a>. A Vision Transformer with 1.1 Billion parameters (ViT-G), processing heart ultrasound videos as space-time patches. Self-attention across all patches, across all frames. The model learned cardiac anatomy and motion from 18 million videos without any human labels telling it what a heart looks like.</p><p>And here&#8217;s the part that still amazes me: when we visualize where the attention focuses, it localizes on actual cardiac structures. Not image noise, not artifacts - the heart. The model discovered anatomy through attention alone.</p><h2>Why one architecture for everything matters</h2><p>This is the deeper point. The fact that the same Transformer architecture &#8212; same attention, same scaling - works for text AND images AND video means you can build models that understand multiple modalities at once.</p><p>Feed an image and a question into the same model. The text tokens attend to the image patches. The image patches attend to the text. That&#8217;s how Vision-Language Models work - the kind Apple uses for Apple Intelligence features that understand both what you see and what you type.</p><p>One architecture. Any input. That&#8217;s not just convenient engineering. It&#8217;s a conceptual breakthrough.</p><h2>The one thing to remember</h2><p>Cut it into patches. Run attention. Whether it&#8217;s a sentence, a photograph, or a heart ultrasound video - the Transformer doesn&#8217;t care. It just learns what to pay attention to.</p><div><hr></div><p><em>I&#8217;m Teodora - AI/ML scientist. I co-built a foundation model for cardiac imaging using the exact architecture I just described. Subscribe to Standout Systems for more.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Attention - The Engine Inside Every AI You Use]]></title><description><![CDATA[AI in 60 Seconds - Episode 2. Your Transformer has one trick. It's called attention. Here's how it works.]]></description><link>https://teodoracoach.substack.com/p/attention-the-engine-inside-every</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/attention-the-engine-inside-every</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Fri, 20 Feb 2026 18:03:19 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/188436525/eef1ec8c556d9ec28311abb67ad2428b.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Watch the video above for the 150-second version. Below: the stuff I couldn't fit.<br><br>Last episode I explained what a Transformer is - an architecture that sees the entire input at once instead of reading one word at a time.</p><p>But I glossed over the most important part. <em>How</em> does it know what to pay attention to?</p><p>The answer is a single mechanism. It&#8217;s called self-attention. And once you understand it, you understand the engine inside every major AI system today.</p><h2>The crowded room</h2><p>Here&#8217;s how I think about it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mR7Z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mR7Z!,w_424,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 424w, https://substackcdn.com/image/fetch/$s_!mR7Z!,w_848,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 848w, https://substackcdn.com/image/fetch/$s_!mR7Z!,w_1272,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 1272w, https://substackcdn.com/image/fetch/$s_!mR7Z!,w_1456,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mR7Z!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif" width="480" height="270" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:270,&quot;width&quot;:480,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:9147326,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/gif&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://teodoracoach.substack.com/i/188436525?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mR7Z!,w_424,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 424w, https://substackcdn.com/image/fetch/$s_!mR7Z!,w_848,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 848w, https://substackcdn.com/image/fetch/$s_!mR7Z!,w_1272,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 1272w, https://substackcdn.com/image/fetch/$s_!mR7Z!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38ebbb25-608f-43b8-8f10-03e9ecfc435f_480x270.gif 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>You walk into a crowded room with a specific question in mind. Everyone&#8217;s wearing a name tag describing what they know. You scan the tags, find the people who match your question, and go listen to what they have to say.</p><p>That&#8217;s self-attention in three steps:</p><p><strong>Query</strong> - your question. Each word generates one. The word &#8220;bank&#8221; might ask: &#8220;am I about money or about rivers?&#8221;</p><p><strong>Key</strong> - the name tag. Every other word advertises what it contains. The word &#8220;deposit&#8221; holds up a key that says &#8220;finance.&#8221; The word &#8220;river&#8221; holds up a key that says &#8220;nature.&#8221;</p><p><strong>Value</strong> - the actual information. Once &#8220;bank&#8221; finds that &#8220;deposit&#8221; is highly relevant (high attention score), it collects the information &#8220;deposit&#8221; carries. Now &#8220;bank&#8221; has been enriched with financial context.</p><p>Every word does this with every other word, simultaneously. That&#8217;s the &#8220;self&#8221; in self-attention - the sentence attending to itself.</p><h2>Multiple spotlights, not just one</h2><p>If this happened once, it would be useful but limited. One question per word, one type of relationship.</p><p>So Transformers run multiple attention computations in parallel. Each one is called an <strong>attention head</strong>, and each head learns to look for a different kind of relationship:</p><p>One head might learn grammar - which verb belongs to which subject, even when they&#8217;re far apart.</p><p>Another might learn meaning - grouping words by topic, connecting &#8220;climate&#8221; in paragraph one to &#8220;emissions&#8221; in paragraph three.</p><p>Another might track references - figuring out that &#8220;she&#8221; in sentence five refers to &#8220;Dr. Chen&#8221; in sentence two.</p><p>Together, it&#8217;s like having multiple spotlights in that crowded room, each one illuminating a different dimension of relationships. That&#8217;s <strong>multi-head attention</strong> - and it&#8217;s why Transformers understand language so much better than anything that came before.</p><h2>Why this matters outside of text</h2><p>Here&#8217;s what excites me as someone who builds AI for healthcare. The attention mechanism doesn&#8217;t care whether the input is words, image patches, or frames of a video.</p><p>When we built <a href="http://echojepa.com/">EchoJEPA </a>- our foundation model for reading heart ultrasounds - the model uses attention to figure out which regions of the heart are relevant to each other across frames. The left ventricle wall in frame 10 &#8220;attends to&#8221; its position in frame 1 to track motion. Same mechanism, completely different domain.</p><p>And here&#8217;s a bonus: those attention patterns are interpretable. When we visualize where the model focuses, we can see it localizing on actual cardiac anatomy - not image artifacts. That&#8217;s not just cool engineering. For clinical AI, it&#8217;s evidence that the model learned the right thing.</p><h2>The one-liner</h2><p>Every word asks &#8220;who matters to me?&#8221;, finds the answer, and collects the information. Multiply that by several parallel spotlights. That&#8217;s attention.</p><p>Or as the paper that started it all said: Attention Is All You Need.</p><div><hr></div><p><em>I&#8217;m Teodora - AI/ML scientist, building AI systems in healthcare. Subscribe to Standout Systems for more.<br></em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Read more about why I created this series and my background as an AI scientist here: </p><div class="comment" data-attrs="{&quot;url&quot;:&quot;https://open.substack.com/home&quot;,&quot;commentId&quot;:215967569,&quot;comment&quot;:{&quot;id&quot;:215967569,&quot;date&quot;:&quot;2026-02-17T22:27:37.034Z&quot;,&quot;edited_at&quot;:null,&quot;body&quot;:&quot;Everyone talks about AI. Transformers, foundation models, RLHF, diffusion models - these terms are everywhere.\n\nBut clear explanations? Rare.\n\nI am starting something: AI in 60 Seconds. One concept per video. No jargon for the sake of jargon - just real understanding from someone who actually builds these systems.\n\nI'm an AI/ML scientist. I co-built a foundation model trained on 18 Million heart videos. I built an AI that measures bias in children's books at scale. I want to make these concepts accessible - because they're not as complicated as people make them sound.\n\nFirst real episode drops this week. Follow along.&quot;,&quot;body_json&quot;:{&quot;type&quot;:&quot;doc&quot;,&quot;attrs&quot;:{&quot;schemaVersion&quot;:&quot;v1&quot;},&quot;content&quot;:[{&quot;type&quot;:&quot;paragraph&quot;,&quot;content&quot;:[{&quot;type&quot;:&quot;text&quot;,&quot;text&quot;:&quot;Everyone talks about AI. Transformers, foundation models, RLHF, diffusion models - these terms are everywhere.&quot;}]},{&quot;type&quot;:&quot;paragraph&quot;,&quot;content&quot;:[{&quot;type&quot;:&quot;text&quot;,&quot;text&quot;:&quot;But clear explanations? Rare.&quot;}]},{&quot;type&quot;:&quot;paragraph&quot;,&quot;content&quot;:[{&quot;type&quot;:&quot;text&quot;,&quot;text&quot;:&quot;I am starting something: AI in 60 Seconds. One concept per video. No jargon for the sake of jargon - just real understanding from someone who actually builds these systems.&quot;}]},{&quot;type&quot;:&quot;paragraph&quot;,&quot;content&quot;:[{&quot;type&quot;:&quot;text&quot;,&quot;text&quot;:&quot;I'm an AI/ML scientist. I co-built a foundation model trained on 18 Million heart videos. I built an AI that measures bias in children's books at scale. I want to make these concepts accessible - because they're not as complicated as people make them sound.&quot;}]},{&quot;type&quot;:&quot;paragraph&quot;,&quot;content&quot;:[{&quot;type&quot;:&quot;text&quot;,&quot;text&quot;:&quot;First real episode drops this week. Follow along.&quot;}]}]},&quot;restacks&quot;:0,&quot;reaction_count&quot;:0,&quot;attachments&quot;:[{&quot;id&quot;:&quot;3e9d3d6a-33e6-4efe-82ee-ee560089a3a3&quot;,&quot;user_id&quot;:422922028,&quot;comment_id&quot;:215967569,&quot;type&quot;:&quot;video&quot;,&quot;media_upload_id&quot;:&quot;4586ab40-1e31-46b5-8292-7a92ba4dce3b&quot;,&quot;mediaUpload&quot;:{&quot;id&quot;:&quot;4586ab40-1e31-46b5-8292-7a92ba4dce3b&quot;,&quot;name&quot;:&quot;&quot;,&quot;created_at&quot;:&quot;2026-02-17T22:25:44.437Z&quot;,&quot;uploaded_at&quot;:&quot;2026-02-17T22:26:16.961Z&quot;,&quot;publication_id&quot;:null,&quot;state&quot;:&quot;transcoded&quot;,&quot;post_id&quot;:null,&quot;user_id&quot;:422922028,&quot;duration&quot;:96.7,&quot;height&quot;:2048,&quot;width&quot;:1152,&quot;thumbnail_id&quot;:1,&quot;preview_start&quot;:null,&quot;preview_duration&quot;:null,&quot;media_type&quot;:&quot;video&quot;,&quot;primary_file_size&quot;:302726308,&quot;is_mux&quot;:true,&quot;mux_asset_id&quot;:&quot;Yh8ExJTFkcxXtWTG7O0002GLp43tqLfVTc66Xe01UvN9II&quot;,&quot;mux_playback_id&quot;:&quot;9P701WRgTt1IUVCZilm7VKmWxf8LJn1Ilj4cCPE6HX4c&quot;,&quot;mux_preview_asset_id&quot;:null,&quot;mux_preview_playback_id&quot;:null,&quot;mux_rendition_quality&quot;:&quot;high&quot;,&quot;mux_preview_rendition_quality&quot;:null,&quot;explicit&quot;:false,&quot;copyright_infringement&quot;:null,&quot;src_media_upload_id&quot;:null,&quot;live_stream_id&quot;:null}}],&quot;name&quot;:&quot;Dr Teodora Szasz&quot;,&quot;user_id&quot;:422922028,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/423bf25c-ef09-42cf-9c8b-5c5238d0e359_2736x2736.jpeg&quot;,&quot;user_bestseller_tier&quot;:null,&quot;userStatus&quot;:{&quot;bestsellerTier&quot;:null,&quot;subscriberTier&quot;:null,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:null,&quot;paidPublicationIds&quot;:[],&quot;subscriber&quot;:null}}}" data-component-name="CommentPlaceholder"></div><p></p>]]></content:encoded></item><item><title><![CDATA[What IS a Transformer? (And Why You Should Care)]]></title><description><![CDATA[Watch the video above for the 60+ second version. Below: the stuff I couldn't fit in 60 seconds.]]></description><link>https://teodoracoach.substack.com/p/what-is-a-transformer-and-why-you</link><guid isPermaLink="false">https://teodoracoach.substack.com/p/what-is-a-transformer-and-why-you</guid><dc:creator><![CDATA[Dr Teodora Szasz]]></dc:creator><pubDate>Wed, 18 Feb 2026 18:05:46 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/188343249/fb67446f00da45bd21e85aac1d953bd7.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Welcome to the first episode of <strong>AI in 60 Seconds</strong>. Today, we are tackling the backbone of modern AI: <strong>Transformers</strong>.</p><p>I started this series because while AI terms are everywhere, clear explanations are rare. I want to cut through the jargon and offer real understanding from the perspective of someone who actually builds these systems.</p><p>Read more about why I created this series and my background as an AI scientist here: <a href="https://substack.com/@teodora485587/note/c-215967569">https://substack.com/@teodora485587/note/c-215967569</a><br><br>OK so here&#8217;s a fun fact. The architecture behind ChatGPT, Gemini, Claude, Apple Intelligence, and basically every AI you&#8217;ve touched this week - it all comes from one 2017 paper called &#8220;<a href="https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf">Attention Is All You Need</a>.&#8221;</p><p>Eight authors. Eight pages. Changed everything.</p><p>So what did they actually figure out?</p><h2>The old way was painfully slow</h2><p>Before Transformers, AI read text the way you&#8217;d read a book if you could only see one word at a time and your memory faded with every step.</p><p>That&#8217;s what Recurrent Neural Networks (RNNs) did. Word by word. Sequentially. By the time the model reached the end of a paragraph, it had mostly forgotten the beginning.</p><p>People tried to fix this with Long Short-Term Memory (LSTM)s - basically giving the model a notebook to jot down important things. Better memory, same problem: still sequential. Still slow. Still couldn&#8217;t look at two distant words at the same time.</p><h2>The breakthrough: just look at everything at once</h2><p>The Transformer said: what if we stop reading one word at a time entirely?</p><p>Instead, it sees the entire input simultaneously and uses something called <strong>self-attention</strong> to figure out which words are relevant to which other words. Every word can attend to every other word, in parallel.</p><p>The word &#8220;bank&#8221; doesn&#8217;t stay generic. By the time it passes through the Transformer, it <em>knows</em> whether it means a river bank or a financial bank - because it looked at every other word in the sentence to figure that out.</p><p><strong>The analogy I keep coming back to:</strong> going from reading one word at a time with a fading memory... to seeing the entire page at once.</p><h2>Why did this win?</h2><p>Three reasons:</p><p><strong>It&#8217;s parallel.</strong> Every word processed at the same time = you can throw massive GPU clusters at it. RNNs couldn&#8217;t do this. That sequential bottleneck was a dealbreaker at scale.</p><p><strong>It scales.</strong> More data + more parameters = better results, in a way that keeps working. That&#8217;s why we went from millions to billions to trillions of parameters. The architecture doesn&#8217;t break.</p><p><strong>It works for everything.</strong> Text, images, audio, video. The same attention mechanism that helps a language model understand sentences helps a Vision Transformer understand images. I know this firsthand - when we built EchoJEPA, a foundation model for reading heart ultrasounds trained on 18 million videos, we used a Vision Transformer as the backbone. Same core idea, completely different domain.</p><h2>The one thing to remember</h2><p>A Transformer sees everything at once and learns what to pay attention to.</p><p>That&#8217;s the whole insight. And it&#8217;s why every major AI system in 2025 is built on it.</p><div><hr></div><p><em>I&#8217;m Teodora - AI/ML scientist, building AI systems in healthcare. Follow Standout Systems for more.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://teodoracoach.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Standout Systems by Teodora is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>