<!DOCTYPE html><html lang="en" xmlns="http://www.w3.org/1999/xhtml" xmlns:v="urn:schemas-microsoft-com:vml" xmlns:o="urn:schemas-microsoft-com:office:office" style="font-size:16px;"><head></head><head><meta charset="utf-8"/><!--[if !mso]><!--><meta http-equiv="X-UA-Compatible" content="IE=edge"/><!--<![endif]--><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="x-apple-disable-message-reformatting"/><meta name="format-detection" content="telephone=no,address=no,email=no,date=no,url=no"/><meta name="color-scheme" content="light"/><meta name="supported-color-schemes" content="light"/><title>Anthropic's Research On The Biology of a LLM</title><!--[if mso]><xml><o:OfficeDocumentSettings><o:AllowPNG/><o:PixelsPerInch>96</o:PixelsPerInch></o:OfficeDocumentSettings></xml><![endif]--><style> :root { color-scheme: light; supported-color-schemes: light; } body { margin: 0; padding: 0; min-width: 100%!important; -ms-text-size-adjust: 100% !important; -webkit-transform: scale(1) !important; -webkit-text-size-adjust: 100% !important; -webkit-font-smoothing: antialiased !important; } .body { word-wrap: normal; word-spacing:normal; } table.mso { width: 100%; border-collapse: collapse; padding: 0; table-layout: fixed; } img { border: 0; outline: none; } table { mso-table-lspace: 0px; mso-table-rspace: 0px; } td, a, span { mso-line-height-rule: exactly; } #root [x-apple-data-detectors=true], a[x-apple-data-detectors=true], #MessageViewBody a { color: inherit !important; text-decoration: inherit !important; font-size: inherit !important; font-family: inherit !important; font-weight: inherit !important; line-height: inherit !important; } span.MsoHyperlink { color: inherit !important; mso-style-priority: 99 !important; } span.MsoHyperlinkFollowed { color: inherit !important; mso-style-priority: 99 !important; } .a { background-color:#dedede; } .b { background-color:#2a2a2a; } .c { background-color:#ffffff; } .d { background-color:#fff0c8; } .d2 { background-color:#FFFFFF; } .d3 { background-color:#FFFFFF; } h1 a { text-decoration:none;color:#2A2A2A !important; } h2 a { text-decoration:none;color:#2A2A2A !important; } h3 a { text-decoration:none;color:#2A2A2A !important; } h4 a { text-decoration:none;color:#2A2A2A !important; } h5 a { text-decoration:none;color:#2A2A2A !important; } h6 a { text-decoration:none;color:#2A2A2A !important; } h1, h1 a, h2, h2 a, h3, h3 a, h4, h4 a, h5, h5 a, h6, h6 a, ul, li, ol, p, p a { margin: 0;padding: 0; } h1 { font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif;font-weight:700;font-size:28px;color:#2A2A2A;line-height:42px;padding-bottom:4px;padding-top:16px;mso-margin-top-alt:16px;mso-margin-bottom-alt:4px } h2 { font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif;font-weight:700;font-size:24px;color:#2A2A2A;line-height:36px;padding-bottom:4px;padding-top:16px;mso-margin-top-alt:16px;mso-margin-bottom-alt:4px } h3 { font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif;font-weight:400;font-size:20px;color:#2A2A2A;line-height:30px;padding-bottom:4px;padding-top:16px;mso-margin-top-alt:16px;mso-margin-bottom-alt:4px } h4 { font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif;font-weight:400;font-size:18px;color:#2A2A2A;line-height:27px;padding-bottom:4px;padding-top:16px;mso-margin-top-alt:16px;mso-margin-bottom-alt:4px } h5 { font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif;font-weight:400;font-size:16px;color:#2A2A2A;line-height:24px;padding-bottom:4px;padding-top:16px;mso-margin-top-alt:16px;mso-margin-bottom-alt:4px } h6 { font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif;font-weight:400;font-size:14px;color:#2A2A2A;line-height:21px;padding-bottom:4px;padding-top:16px;mso-margin-top-alt:16px;mso-margin-bottom-alt:4px } p { font-family:'Georgia','Times New Roman',serif;font-weight:400;color:#2D2D2D;font-size:16px;line-height:24px;padding-bottom:8px;padding-top:8px;mso-margin-top-alt:8px;mso-margin-bottom-alt:8px; } p a, .e a, ul a, li a, .h a, .h2 a, .h3 a { word-break:break-word;color:#2C81E5 !important;text-decoration:none;font-style:italic; } p a span, .e a span, ul a span, li a span { color: inherit } p .bold { font-weight:bold;color:#2D2D2D; } p span[style*="font-size"] { line-height: 1.6; } .f p { font-size:12px;line-height:15px;color:#2D2D2D;padding:0; } .f p a { color:#2D2D2D !important; } .g p { font-family:'Helvetica',Arial,sans-serif;font-size:14px;line-height:20px;font-weight:normal;margin:0; } .g p a { text-decoration: underline; } .i p { font-family:'Helvetica',Arial,sans-serif;line-height:23px;font-size:15px;color:#2D2D2D; } .i p a { color:#2D2D2D !important; } .i2 p { font-family:'Helvetica',Arial,sans-serif;line-height:23px;font-size:15px;color:#2D2D2D; } .i2 p a { color:#2D2D2D !important; } .i3 p { font-family:'Helvetica',Arial,sans-serif;line-height:43px;font-size:24px;color:#2D2D2D; } .i3 p a { color:#2D2D2D !important; } .h p a { color:#595959 !important; } .h2 p a { color:#595959 !important; } .h3 p a { color:#595959 !important; } .f p a, .i p a, .i2 p a, .i3 p a, .h p a, .h2 p a, .h3 p a { text-decoration:underline; } .j { border-top:3px solid #ffeb2d; } .k p { padding-left:15px;padding-bottom:0px;padding-top:6px;mso-margin-top-alt:6px;mso-margin-bottom-alt:0px;mso-margin-left-alt:15px; } .o { background-color:#FFFFFF;border:1px solid #F1F1F1;border-radius:5px; } .o p { font-family:'Helvetica',Arial,sans-serif;padding:0px;margin:0px; } .l p, .l p a { font-size:14px;line-height:20px;font-weight: bold;color:#2D2D2D;padding-bottom:6px;mso-margin-bottom-alt:6px;text-decoration:none; } .m p, .m p a { font-size:13px;line-height:18px;font-weight:400;color:#2D2D2D;padding-bottom:6px;mso-margin-bottom-alt:6px;text-decoration:none; } .n p, .n p a { font-size:12px;line-height:17px;font-weight:400;color:#2D2D2D;padding-bottom:6px;mso-margin-bottom-alt:6px;text-decoration:none; } .p { background-color:#FFFFFF;max-width:520px;border:1px solid #E1E8ED;border:1px solid rgba(80, 80, 80, 0.3);border-radius:5px; } .q { font-size:16px;font-family:Helvetica,Roboto,Calibri,sans-serif !important;border:1px solid #e1e8ed;border:1px solid rgba(80, 80, 80, 0.3);border-radius:10px;background-color:#FFFFFF; } .q p { font-size:16px;font-family:system-ui,Helvetica,Roboto,Calibri,sans-serif !important;color:#222222;padding:4px 0; } .r { border:1px solid #E1E8ED !important;border-radius:5px; } .s p { font-size: 14px; line-height: 17px; font-weight: 400; color: #697882; text-decoration: none; } .t p { font-family:'Helvetica',Arial,sans-serif;font-size:12px;line-height:18px;font-weight:400;color:#000000;font-style:italic;padding:4px 0px 0px;} .v { border-radius:10px;border:solid 0px #DFD150;background-color:#2C81E5;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;color:#FFFFFF; } .v a { text-decoration:none;display:block;color:#FFFFFF; } .w p { font-size:12px;line-height:15px;font-weight:400;color:#FFFFFF; } .w p a { text-decoration: underline !important;color:#FFFFFF !important; } ul { font-family:'Helvetica',Arial,sans-serif;margin:0px 0px 0px 25px !important;padding:0px !important;color:#2D2D2D;line-height:24px;list-style:disc;font-size:16px; } ul > li { font-family:'Helvetica',Arial,sans-serif;margin:10px 0px 0px 0px !important;padding: 0px 0px 0px 0px !important; color: #2D2D2D; list-style:disc; } ol { font-family:'Helvetica',Arial,sans-serif;margin: 0px 0px 0px 25px !important;padding:0px !important;color:#2D2D2D;line-height:24px;list-style:decimal;font-size:16px; } ol > li { font-family:'Helvetica',Arial,sans-serif;margin:10px 0px 0px 0px !important;padding: 0px 0px 0px 0px !important; color: #2D2D2D; list-style:decimal; } .e h3, .e p, .e span { padding-bottom:0px;padding-top:0px;mso-margin-top-alt:0px;mso-margin-bottom-alt:0px; } .e span, .e li { font-family:'Helvetica',Arial,sans-serif;font-size:16px;color:#2D2D2D;line-height:24px; } .rec { font-family: ui-sans-serif, system-ui, -apple-system, BlinkMacSystemFont, "Segoe UI", Roboto, "Helvetica Neue", Arial, "Noto Sans", sans-serif, "Apple Color Emoji", "Segoe UI Emoji", "Segoe UI Symbol", "Noto Color Emoji" !important; } .rec__button:hover { background-color: #f9fafb !important; } .copyright a {color: inherit !important; text-decoration: none !important; font-size: inherit !important; font-family: inherit !important; font-weight: inherit !important; line-height: inherit !important;} .txt_social p { padding: 0; word-break: break-all; } .table, .table-c, .table-h { border: 1px solid #C0C0C0; } .table-c { padding:5px; background-color:#FFFFFF; } .table-c p { color: #2D2D2D; font-family:'Helvetica',Arial,sans-serif !important;overflow-wrap: break-word; } .table-h { padding:5px; background-color:#F1F1F1; } .table-h p { color: #2A2A2A; font-family:'Trebuchet MS','Lucida Grande',Tahoma,sans-serif !important;overflow-wrap: break-word; } @media only screen and (max-width:667px) { .aa { width: 100% !important; } .bb img { width: 100% !important; height: auto !important; max-width: none !important; } .cc { padding: 0px 8px !important; } .ee { padding-top:10px !important;padding-bottom:10px !important; } .ff ul, .ff ol { margin: 0px 0px 0px 10px !important;padding: 0px !important; } .ff li { margin:10px 0px 0px 10px !important; } .r {height:140px !important;} .s p { font-size:13px !important;line-height:15px !important; } .mob-hide {display:none !important;} .mob-stack {display:block !important;width:100% !important;} .mob-w-full {width:100% !important;} .mob-block {display:block !important;} .embed-img {padding:0px 0px 12px 0px !important;} .socialShare {padding-top:15px !important;} .rec { padding-left:15px!important;padding-right:15px!important; } .bodyWrapper { padding:7px 4px 7px 4px !important; } .social-mobile {float:left !important;margin-top:10px !important;} } @media screen and (max-width: 480px) { u + .a .gg { width: 100% !important; width: 100vw !important; } .tok-heart { padding-top:75% !important; } .tok-play { padding-top: 250px !important; } } @media screen and (max-width: 320px) { .tok-heart { padding-top:65% !important; } } .u { border: 1px solid #CACACA !important; border-radius: 2px !important; background-color: #ffffff !important; padding: 0px 13px 0px 13px !important; font-family:ui-sans-serif,system-ui,-apple-system,BlinkMacSystemFont,"Segoe UI",Roboto,"Helvetica Neue",Arial,"Noto Sans",sans-serif !important;font-size: 12px !important; color: #767676 !important; } .u a { text-decoration: none; display: block !important; color: #767676 !important; margin: 0px !important; } .u span, .u img { color: #767676 !important;margin:0px !important; max-height:32px !important;background-color:#ffffff !important; } </style><!--[if mso]><style type="text/css"> sup { font-size: 100% !important;vertical-align: .5em !important;mso-text-raise: -1.5% !important;line-height: 0 !important; } ul { margin-left:0px !important; margin-right:10px !important; margin-top:20px !important; margin-bottom:20px !important; } ul li { margin-left: 0px !important; mso-special-format: decimal; } ol { margin-left:0px !important; margin-right:10px !important; margin-top:20px !important; margin-bottom:20px !important; } ol li { margin-left: 0px !important; mso-special-format: decimal; } li.listItem { margin-left:15px !important; margin-top:0px !important; } .paddingDesktop { padding: 10px 0 !important; } .edm_outlooklist { margin-left: -20px !important; } .embedImage { display:none !important; } </style><![endif]--><style> @font-face { font-family: 'Open Sans'; font-style: normal; font-weight: 700; font-display: swap; src: url('https://fonts.gstatic.com/s/opensans/v40/memSYaGs126MiZpBA-UvWbX2vVnXBbObj2OVZyOOSr4dVJWUgsg-1x4gaVIUwaEQbjA.woff2') format('woff2'); } @font-face { font-family: 'Open Sans'; font-style: italic; font-weight: 700; font-display: swap; src: url('https://fonts.googleapis.com/css2?family=Open+Sans:ital,wght@1,700&display=swap') format('woff2'); } </style></head><body class="a" style="margin:0px auto;padding:0px;word-wrap:normal;word-spacing:normal;background-color:#dedede;"><div role="article" aria-roledescription="email" aria-label="email_name" lang="en" style="font-size:1rem"><div style="display:none;max-height:0px;overflow:hidden;"> Plus more about Defeating Prompt Injections by Design and Reasoning to Learn from Latent Thoughts  ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ ‌ </div><table role="none" width="100%" border="0" cellspacing="0" align="center" cellpadding="0" class="gg"><tr><td align="center" valign="top"><table role="none" width="670" border="0" cellspacing="0" cellpadding="0" class="aa" style="width:670px;table-layout:fixed;"><tr><td class="bodyWrapper" align="center" valign="top" style="padding:7px 7px 7px 7px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td align="center" valign="top" style="border-width:0px 0px 0px 0px;border-style: solid; border-color: #2a2a2a;border-radius:10px 10px 0px 0px;background-color:#ffffff;" class="c"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr id="header"><td style="padding:20px 28px"><div style="padding-top:0px;padding-right:0px;padding-bottom:20px;padding-left:0px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td class="f" align="right" valign="top"><p> April 01, 2025 | <a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.c6q0w4g5sodbtO4I1B_pxSdB5RCIH6yy1Fm1CYma3EzM3iBS2UQ5NJ_ZTosRkjSkCEo3FGhTHhR_4c3uE-sUrL4ev_0pWgID_Tmi-JDneUZW9v8Ynfo6jmWwdCKOKI1UPSZvTxxqf882-cfmgH927wgm-qhQABQtQP6VVRg3pOHiNw-BTBW4E7y0MfMgfhDdLFMQAI1h6bc_usIxj8mHsfjvj38-objnQCBGtXKmKVKH3TJp8w8x0EO6h9my1YcozM3tm0BiAXnRIHCUpYPk6WhClb9Ci4Ay0pmEn-tZVYctNjzSGdHPOCS05JQTxdZUreeEIGAjGTY20gSvVu0APjKbWm5Yv1hwvAVDbSh1ZITO_CsH3q3A4dU-z4_BuX9ecqMgBIBlkPxiFe8NGQjcvwRdnD-mfJZNvQS7h9ozB3ykP91jPPkqqZgUfLCwp2OQ4vTePlV0x9-4_Mt0czugs5NjMJbA2aVexLEuxGI8lbRJBVzizmcrR3qeGSbCqpNU7MB-2cmIwgktgtDt9MOPGZQb91OFsTZTFj3pNMAkTOFntxB0BSlB0JhDqdc6za0kUWlysaf-OmqibcW2zy3GiVdoEr5ubSW-z9bNr0p0lZl9_58IUV73FyEKKUlFFajmPOmmKbiPldIJApa1Ids1XqHjgWY5jFy0bxIFUv-IjbkOlb4f3PCwXUlSvpZKZ-JT/4fa/NRovB4-NQV-yUQtq5HqcKA/h0/h001.dnB3Gka5AIzCi476XgSYrzv_WJ2302grpIJRlBzMszM"><span class="translation_missing" title="translation missing: en.templates.posts.email.v3.header.read_online">Read Online</span></a></p></td></tr><tr><td class="dd" align="center" valign="top" style="padding:15px 0 20;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td align="center" valign="top"><h1 style="text-align:left;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;font-weight:Bold;font-size:32px;color:#2A2A2A;padding:2px 0;line-height:38px;"> Anthropic's Research On The Biology of a LLM </h1><p style="text-align:left;font-family:'Helvetica',Arial,sans-serif;font-weight:normal;font-size:20px;color:#3E3E3E;padding:5px 0;line-height:24px;"> Plus more about Defeating Prompt Injections by Design and Reasoning to Learn from Latent Thoughts </p></td></tr></table></td></tr><tr><td style="height:0px;width:0px;"><div style="height:1px;" data-open-tracking="true"> <img src="https://elink4f7.mail.bycloud.ai/ss/o/u001.3wmUuY8gEWd4_869a_eXcg/4fa/NRovB4-NQV-yUQtq5HqcKA/ho.gif" alt="" width="1" height="1" border="0" style="height:1px !important;width:1px !important;border-width:0 !important;margin-top:0 !important;margin-bottom:0 !important;margin-right:0 !important;margin-left:0 !important;padding-top:0 !important;padding-bottom:0 !important;padding-right:0 !important;padding-left:0 !important;"/> </div></td></tr></table></div></td></tr><tr id="content-blocks"><td class="email-card-body" align="center" valign="top" style="padding-bottom:28px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td id="nov-18-th-nov-24-th-33-latest-ai-re" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h6 style="color:#2A2A2A;font-weight:normal;"><i>Mar 24th ~ Mar 30th</i><br><i>#49 Latest AI Research Explained Simply</i></h6></td></tr><tr><td><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" style=""><tr><td bgcolor="#222222" style="background-color:#222222;padding:0.0px 0.0px 0.0px 0.0px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0"><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"></p></td></tr></table></td></tr></table></td></tr><tr><td id="industry-news-in-1-line" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:Bold;padding:0px 28px;text-align:left;"><h2 style="color:#2A2A2A;font-weight:Bold;">🗞️ Industry News in 1 Line</h2></td></tr><tr><td style="padding-bottom:12px;padding-left:50px;padding-right:40px;padding-top:12px;" class="ee"><div style="margin-left:0px;" class="edm_outlooklist"><ol start="1" style="list-style-type:decimal;margin:0px 0px;padding:0px 0px 0px 0px;"><li class="listItem ultext"><p style="line-height:24px;padding:0px;text-align:left;word-break:break-word;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;">♥ 5.1k</span></span> Google has released <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.oB7zuO_W-X4Toa45C28ngyEFO5Gj_UpIFzWcxERDRtBmNn3kXjEwMzDmkE0qtKMgzDy7f0eg0dCy39vbw1ebVS1cUVe59FKgtA_qDFOayN3oSXeMU4hkVUUc071uc_dVjte-GgASPZPexvjDzcCJxAm-I-VURShRB-QZBQ9y5cDHPG2XU42dNxlf6mL6CquOOOWycfcbh9HQ0ghkIgUtSNfXTK5QDKpI7FOhn8L9USo/4fa/NRovB4-NQV-yUQtq5HqcKA/h1/h001.ORKMt18EGUzlVX2ALk2eRgk-qCOmmbEvDxTCZGlup6g" target="_blank" rel="noopener noreferrer nofollow"><span>Gemini 2.5 Pro</span></a>, a new state-of-the-art model with improved performance for reasoning, coding, and science tasks. An experimental version is now available <b>free</b> for all users, try it now on their <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.fUNb4GdFo9D3F8WuLArtoXy3TFQVRsN_o5nO7Vn3nW8lVxPNQUYuQwH5y7pqf7Y532fX_gNIL1ZqkGRRM64mMCOTzgN6SGDDbkH_yBVorHcV3OaX1EGXLSkiVSy0Is82/4fa/NRovB4-NQV-yUQtq5HqcKA/h2/h001.JU-s4IXFFVrQ3ctP29pS9YPX7h5T3_bel9B5Dn2ZEkM" target="_blank" rel="noopener noreferrer nofollow"><span>AI Studio</span></a> or <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.VomAAYwkCjux8i_FMc4kJVJebGQSxsAOZVK7PvOBHPVa8_uL3hfPic3tPZszdU57xTWLFwgn4hhwg6Cst69T5KZdwCyd6bmidnJ_jEU4nCyVnAbHlprfhpM9KmUyskwUeEtj8_4G_rl5gTvE-kPr2A/4fa/NRovB4-NQV-yUQtq5HqcKA/h3/h001.y0RNkAhGDk0fmUq8ag7vdJElsuvR-yNu4mv2sPwYM88" target="_blank" rel="noopener noreferrer nofollow"><span>Gemini app today</span></a>. </p><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:420px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/477fc442-a773-4e79-ab99-d9748e865c2a/gemini_benchmarks_cropped_light2x_1PPmDuP.gif?t=1743535475" alt="Gemini 2.5 Pro Benchmark" height="auto" width="420" style="display:block;width:100%;" border="0"/></td></tr><tr><td align="center" valign="top" class="t" style="width:420px;"><p>Gemini 2.5 Pro Benchmark</p></td></tr></table></li><li class="listItem ultext"><p style="line-height:24px;padding:0px;text-align:left;word-break:break-word;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;">♥ 1.5k</span></span> <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.gXpuuKq1N_-6FJq3Q1UVsQyNThfB2XpSx1B7s6Mw10ObTYbvMkbSV2Pwogv3C6mEMzX4glEiiEkpXeFIRromrRK6FYWeZMcXpZD1NjVNzEceJcQ4mN0_Etlaj7iY7T-sW1eRqLL8uNfQztNmIpbrsa31deh9cBYoa9V56fq3gmHq3ZHSwv7cQaRhkPHHYaxN/4fa/NRovB4-NQV-yUQtq5HqcKA/h4/h001.mK87t0A1c_STc8L8XZ63x9pph1V0vLF2RSmxpUrguWk" target="_blank" rel="noopener noreferrer nofollow"><span>Runway has introduced Gen-4</span></a>, an AI model that generates images and videos using visual references combined with text instructions for consistent results. The new model allows users to maintain continuity in style, subjects, and locations, as showcased in several short demonstration films. You can try the Image-to-Video functionality today by subscribing to paid or Enterprise plan. </p></li><li class="listItem ultext"><p style="line-height:24px;padding:0px;text-align:left;word-break:break-word;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;">♥ 1.9k</span></span> <span style="">The Qwen team has released </span><span style=""><a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.wcXdj6dB6nd1Cx4inzJNk_td6EHXs7wvRPySUkyhYFdrleYKASU0noDGzOke3yOVSb00CQo68Td1A9YOiigsL_ptAUNKK8_mB8M3oWQDh7YnBt7eXFMv1-L2k6ulx9acBBWyPRcATfSjy2GTl5GV7nHMWFIv5RGUyLwFCOa50tM/4fa/NRovB4-NQV-yUQtq5HqcKA/h5/h001.GXm4EddaQeyQHL9IlQ39u02zNhdDMxPP7C91Go0SESw" target="_blank" rel="noopener noreferrer nofollow"><span>Qwen2.5-VL-32B</span></a></span><span style="">, an updated vision-language AI model optimized for better mathematical reasoning, detailed image understanding, and providing responses more aligned with human preferences. This 32B parameter model has shown strong benchmark performance, surpassing comparable or even larger models on complex reasoning tasks. You can download its weights from </span><span style=""><a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.CxDkkVpJsBdVoe83c_tBWu5HHpecwStOph8cXvproQ8iXrsaVlyxiIKpZiT_UKna8bxiHd82oAuLy-cTt3vT_yUAMI6qJcarL-mXeXzWiuIDv3Kd3XuCsAwYoOx85nvQGfcEpvTRQ_arQBsCd5Fjd-eT3G4ljfeOjxC-olpSCM8/4fa/NRovB4-NQV-yUQtq5HqcKA/h6/h001.WawnQ6Y1dB3pRWcRubasBrKNmLxOU2PqamDDT_kA0YE" target="_blank" rel="noopener noreferrer nofollow"><span>Hugging Face</span></a></span><span style=""> or </span><span style=""><a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.c6q0w4g5sodbtO4I1B_pxbBqUjAzI8c4yNhiTrsgEvOBKQa4xIBg5KbqVM0Vcsbx86hAzv7Wdj6AWfxKyT4qTfERux35SIgA0ORi9TuIO3svMJwOBI9r9NXfG8nD6R1I5CCFc-XvppuQ1ponr3avKDLLpk4TQuuQFX2eWKN7z0PCkj8vil5LHXbCDuLSJTHZ/4fa/NRovB4-NQV-yUQtq5HqcKA/h7/h001.xln4frxlhji395H-v8Odbq6z5jxlZ7H9jSBwinrOqZo" target="_blank" rel="noopener noreferrer nofollow"><span>ModelScope</span></a></span><span style="">.</span></p><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:480px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/439c1b24-2923-4cab-a05f-a3935b7c1bea/qwen2.5vl-32b-vision.jpg?t=1743530893" alt="Qwen-2.5-VL-32B benchmark" height="auto" width="480" style="display:block;width:100%;" border="0"/></td></tr><tr><td align="center" valign="top" class="t" style="width:480px;"><p>Qwen-2.5-VL-32B benchmark</p></td></tr></table></li><li class="listItem ultext"><p style="line-height:24px;padding:0px;text-align:left;word-break:break-word;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;">♥ 21k</span></span> OpenAI has released <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.DIqyAo9xTeoWriogq2VlWeUmi9WmFR4pnC4wMSHAHOFgxp08QbDbo89acotot9UU8XS5w8Zv4Bpocxa5yfbULkh4ObqQyRJCSkeGuThVdnaVkej8XmSdsORM37zTuO3tFW8olQvQLPeC3DPxuJp3Hf4si-WGpU_MCindmPVAiX4/4fa/NRovB4-NQV-yUQtq5HqcKA/h8/h001.YcXvg4EHvBuo7y0OBj69S__M32Tjd9S0p2l4fXIQVDA" target="_blank" rel="noopener noreferrer nofollow"><span>GPT-4o’s native image generation</span></a>, along side improvements on its text generation capabilities. Its image generation is incredibly good at generating text within images and style transfer. GPT-4o is an autoregressive multimodal model, and you can read more about it on <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.ZsobsZmG6kUZ4LjqczYBVKgk36Ce7cLG9RJ0Va_3oE-INzBB49hLjmIAiWAI30ZMtDOSJL1ux7i3fSJ8Va2lSA71g7onFzoVoAT3rWrqXCACL1K1oNA5j5GYh7iCGzINsjDz8qt0pgCtJX2OTZCPChtV3DQIUX6Uq53icjq-5zj7bV-pn150y-JQGZ8yw5HdchLH_fpBB33fVvoPr5B59Q/4fa/NRovB4-NQV-yUQtq5HqcKA/h9/h001.MyVdPkZwnOTn3-Cnxdq1CHGIQL8vAFZYDFIC6dd0WFQ" target="_blank" rel="noopener noreferrer nofollow"><span>GPT-4o’s native image generation system card</span></a>. </p><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:510px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/4b57b6d5-7937-4d7f-a192-be4dc1ba8406/hero_image_1-whiteboard1.jpg?t=1743535966" alt="" height="auto" width="510" style="display:block;width:100%;" border="0"/></td></tr><tr><td align="center" valign="top" class="t" style="width:510px;"><p>an image generated with GPT-4o native image generation</p></td></tr></table><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:480px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/8847d6b7-e05f-4afc-accf-4c77a6072e0a/Screenshot_2025-04-01_153022.png?t=1743535852" alt="" height="auto" width="480" style="display:block;width:100%;" border="0"/></td></tr><tr><td align="center" valign="top" class="t" style="width:480px;"><p>1 million new users per hour</p></td></tr></table></li></ol></div></td></tr><tr><td><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" style=""><tr><td bgcolor="transparent" style="background-color:transparent;border-color:#2C81E5;border-style:solid;border-width:5px;padding:0.0px 0.0px 0.0px 0.0px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0"><tr><td class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:Bold;padding:0px 28px;text-align:left;"><h2 style="color:#2A2A2A;font-weight:Bold;"><span style="">Support My Newsletter</span></h2></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style="color:rgb(34, 34, 34);font-family:Georgia, "Times New Roman", serif;font-size:16px;">As I aim to keep this newsletter free forever, your support means a lot. If you like reading The AI Timeline, consider forwarding it to another research enthusiast, It helps us keep this up for free!</span></p></td></tr><tr><td align="center" valign="top"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td align="center" valign="top" style="font-size:0px;line-height:0px;padding:30px 0px 30px;" class="dd"><table class="j" role="none" width="50%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td> </td></tr></table></td></tr><tr><td class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:Bold;padding:0px 28px;text-align:left;"><h2 style="color:#2A2A2A;font-weight:Bold;">Share The AI Timeline</h2></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> You currently have <strong>0</strong> referrals. </p></td></tr><tr><td align="left" valign="top" style="padding: 20px 0px 20px 0px; display:none;width:0px;max-height:0px;overflow:hidden;mso-hide:all;height:0;font-size:0;max-height:0;line-height:0;margin:0 auto;" class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 0;"><tr><td align="center" valign="top" style="width:300px;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.c6q0w4g5sodbtO4I1B_pxWc4htTObwdorovK0nFHVH-4pUdVE0ELYH5DsNemk732SjNwhPNJ25r0O8B5vYifsGNUqyW5TiZkyMsF1yreu0byy2KW36J1wDdpoLuXg2TU1F1OW8OHoHaU4-ZmrZpPU4RN-crQCEimD190CSn9fPuxpIRojBJyu1VfV5KtQD3QMVdSg2JrjEj5-xm4r4E12Whf08itqPCb9Q5W0X4rt3ubYkqCmWnLeZpmb3_RZcbIk0UE5wZnFLCQJHLFs0qZ0OGpXp89o1HU4mWIBur5Or4tQGm5M_Y8m5PvTEfYfxLRyrcRv7GyVs5oLtFfiySZ2SqtZypLA-h50h61p0uPiA7iA_PiMqlVLtM-87XL33VZi05_O3UTpWE_0nAzFRJ4TW1ayz3_vn4Zlp9IERdbnnAd_1kPLD4lAQcR5PRXgtpCQek6Faby1-f5E96me8ItSRRtW5yxR6x6hOf_P5cMjFlyzkfwRsHM24hjEDcNfT5e14NsvJFZbNRy-G0VKeKZrSZ9yrk8QkyRE57DVDIWPQdf_GFzwteDfy97bbP_kVNAoveVczlOACPOpZDJWb0nPm2DRU2XA97fF_3sJKeIuuJrAnqDoDiV472SFJzYf2tL/4fa/NRovB4-NQV-yUQtq5HqcKA/h10/h001.sJMjBso_YJGyJGfAbV2hG2GVHEWVnE84yvvNRXtQwoQ" rel="noopener noreferrer nofollow" style="text-decoration:none;" target="_blank"><img src="" alt="" height="auto" width="300" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></a></td></tr></table></td></tr><tr><td align="left" valign="top" style="padding-bottom:14px;padding-left:28px;padding-right:28px;padding-top:14px;text-align:left;width:100%;word-break:break-word;" class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" align="left" style="margin:14px auto 14px auto;"><tr><td align="center" valign="middle" height="42" style="height:42px;background-color:#2C81E5;border-color:#DFD150;border-radius:px px px px;border-style:solid;border-width:px px px px;color:#FFFFFF;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.c6q0w4g5sodbtO4I1B_pxWc4htTObwdorovK0nFHVH-4pUdVE0ELYH5DsNemk732SjNwhPNJ25r0O8B5vYifsGNUqyW5TiZkyMsF1yreu0byy2KW36J1wDdpoLuXg2TU1F1OW8OHoHaU4-ZmrZpPU4RN-crQCEimD190CSn9fPuxpIRojBJyu1VfV5KtQD3QMVdSg2JrjEj5-xm4r4E12Whf08itqPCb9Q5W0X4rt3ubYkqCmWnLeZpmb3_RZcbIk0UE5wZnFLCQJHLFs0qZ0OGpXp89o1HU4mWIBur5Or4tQGm5M_Y8m5PvTEfYfxLRyrcRv7GyVs5oLtFfiySZ2SqtZypLA-h50h61p0uPiA7iA_PiMqlVLtM-87XL33VZi05_O3UTpWE_0nAzFRJ4TW1ayz3_vn4Zlp9IERdbnnAd_1kPLD4lAQcR5PRXgtpCQek6Faby1-f5E96me8ItSRRtW5yxR6x6hOf_P5cMjFlyzkfwRsHM24hjEDcNfT5e14NsvJFZbNRy-G0VKeKZrSZ9yrk8QkyRE57DVDIWPQdf_GFzwteDfy97bbP_kVNAoveVczlOACPOpZDJWb0nPm2DRU2XA97fF_3sJKeIuuJrAnqDoDiV472SFJzYf2tL/4fa/NRovB4-NQV-yUQtq5HqcKA/h11/h001.mrnEkf3kiPR7mD-scp2yXVxtPZIP_VjZ8QVuXxIL_fM" target="_blank" rel="noopener noreferrer nofollow" style="color:#FFFFFF;display:block;font-size:16px;font-size:16px;font-weight:normal;padding:0px 14px;padding:14px 14px 14px 14px;text-decoration:none;"> Click to Share </a></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Or copy and paste this link to others: <a class="link" href="https://mail.bycloud.ai/subscribe?ref=6SqUHb8KiF&_bhlid=7fecfad9eb7fd8bcdb529e945e11346b5897acdc" target="_blank" rel="noopener noreferrer nofollow" clicktracking="off"><span>https://mail.bycloud.ai/subscribe?ref=6SqUHb8KiF</span></a></p></td></tr><tr><td align="center" valign="top" style="font-size:0px;line-height:0px;padding:30px 0px 30px;" class="dd"><table class="j" role="none" width="50%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td> </td></tr></table></td></tr></table></td></tr><tr><td align="center" valign="top" style="padding-bottom:14px;padding-left:28px;padding-right:28px;padding-top:14px;text-align:center;width:100%;word-break:break-word;" class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" align="center" style="margin:14px auto 14px auto;"><tr><td align="center" valign="middle" height="42" style="height:42px;background-color:#2C81E5;border-color:#DFD150;border-radius:px px px px;border-style:solid;border-width:px px px px;color:#FFFFFF;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.zNfxTwpJFmrsCuJJphGRkKSrCVph9-fOYkcjx4VfJRwtQQsKrZC8pi-PiKai2fq4lAto9WepTJo69aQJ1T73b1BYaJHeCrLz1cWpFYfpKjdJ071BkzwRo9IrCS5YAIxy/4fa/NRovB4-NQV-yUQtq5HqcKA/h12/h001.cy2LyGIew7bNLfWtX4uk0cnBuosQ74ceJRxk5vi-iNc" target="_blank" rel="noopener noreferrer nofollow" style="color:#FFFFFF;display:block;font-size:16px;font-size:16px;font-weight:normal;padding:0px 14px;padding:14px 14px 14px 14px;text-decoration:none;"> Check Out My Patreon </a></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style=""><a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.tLfGW26lAwaS9gFg17HSoGymQ3NNPtd5dE5MV_8UgjIDFPVXngz8pvQBldSW42yhUe_Qiq6DgEPMEBuPL9yfRpXelTiuu2kS8pLFvsoem_XoZoy_n13sTKUhZIbl0VH6/4fa/NRovB4-NQV-yUQtq5HqcKA/h13/h001.hGp0_JdUUGttJCueL8a6roWEW7YsRe0PR5zqY32AD3M" target="_blank" rel="noopener noreferrer nofollow"><span>Advertise with The AI Timeline! </span></a></span></p></td></tr></table></td></tr></table></td></tr><tr><td><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" style=""><tr><td bgcolor="#222222" style="background-color:#222222;padding:0.0px 0.0px 0.0px 0.0px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0"><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"></p></td></tr></table></td></tr></table></td></tr><tr><td id="on-the-biology-of-a-large-language-" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:Bold;padding:0px 28px;text-align:left;"><h2 style="color:#2A2A2A;font-weight:Bold;">On the Biology of a Large Language Model</h2></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style=""><i>Lindsey et al. [Anthropic]</i></span></p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;"> ♥ 7.1k </span></span><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> </span><span style="background-color:#e0e0e0;"><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> LLM Interpretability </span></span></p></td></tr><tr><td id="breaking-down-claude-35-using-circu" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Breaking Down Claude 3.5 Using Circuit Tracing Methodology</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Large AI language models are incredibly capable, however, we don't fully understand <span style=""><i>how</i></span> they work on the inside. They operate like complex "black boxes," and as these models get smarter and are used in more important situations, simply knowing they <span style=""><i>work</i></span> isn't enough. We need to understand their internal reasoning to trust them, ensure they are safe, and judge if they're suitable for specific tasks. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> This technical paper tackles this "black box" problem by developing and applying new tools to peek inside these models, similar to how biologists use microscopes to understand cells. Building on previous work that identified basic concepts or "features" within models, this paper introduces methods, especially one called "<span style="font-weight:700;"><b>attribution graphs</b></span>," to map out how these features connect and interact. You can think of it like creating a wiring diagram for the AI's thought process. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> By tracing the steps the model takes internally from an input (like a question) to an output (its answer), the researchers aim to understand the specific mechanisms the AI uses for tasks like reasoning, planning, and even identifying harmful requests, using the <span style="font-weight:700;"><b>Claude 3.5 Haiku</b></span> model as a case study. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/e3d379dd-cc40-4348-92a2-3f9c4c47b3c8/image.png?t=1743529173" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td id="how-does-claude-35-work" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">How Does Claude 3.5 Work?</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> To understand how language models like Claude 3.5 Haiku arrive at their answers, the researchers developed a method focused on uncovering the hidden intermediate steps. They illustrate this using a simple example: completing the sentence "Fact: the capital of the state containing Dallas is..." with "Austin." While this seems intuitive (Dallas is in Texas, the capital of Texas is Austin), the question is whether the AI actually performs these two steps internally or uses some kind of memorized shortcut. This research provides evidence for genuine multi-step reasoning happening inside the model, coexisting with simpler pathways. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> The core technique involves generating an "attribution graph." This graph <span style="font-weight:700;"><b>visualizes the key internal "features"</b></span>, concepts the model represents internally, that become active during processing and how strongly they influence each other to produce the final output. The first step is to identify and <span style="font-weight:700;"><b>interpret these features</b></span>. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/4cbb9469-30f8-4cf3-9e63-48bc218aa938/image.png?t=1743529213" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> For instance, researchers found features specifically activated by the word "capital," but also more abstract features representing the concept of a capital, even activating across different languages (like "Hauptstadt" in German or "省会" in Chinese). Similarly, they identified features representing "Texas" (activated by "Dallas") and features that specifically push the model to output the word "Austin" ("say Austin" features) or any capital city ("say a capital" features). These individual features are then grouped into simplified categories called "supernodes" (e.g., a "Texas" supernode containing various Texas-related features) to make the interactions easier to analyze. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> The resulting attribution graph for the Dallas example revealed distinct pathways. Features in the "Dallas" supernode strongly activate features in the "Texas" supernode. Separately, the "capital" supernode activated the "say a capital" supernode. More importantly, the "Texas" and "say a capital" supernodes then jointly activated the "say Austin" supernode, leading to the final answer. This Dallas → Texas → Austin pathway provides clear evidence of the hypothesized two-step reasoning. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/1561b3b1-0f94-4c49-a445-2feed01e12e6/image.png?t=1743529245" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Interestingly, the graph also showed a direct "shortcut" link from "Dallas" to "say Austin," suggesting multiple mechanisms operate simultaneously. To confirm these aren't just artifacts of the analysis, the researchers performed "inhibition experiments." They artificially suppressed the activity of specific supernodes (like "Texas" or "capital") and observed the downstream effects. Inhibiting "Texas" reduced the activation of "say Austin" but not "say a capital," while inhibiting "capital" did the opposite, confirming the distinct roles of these pathways in the model's computation and altering the final predicted word in logical ways. </p></td></tr><tr><td id="results-and-real-world-implications" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Results and Real-World Implications of Transcendence</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> This experiment shows that tools like attribution graphs can successfully peek inside complex AI models like Claude 3.5 Haiku, and reveal surprisingly sophisticated internal mechanisms. The studies uncovered evidence of the model performing multi-step reasoning, using parallel computational pathways. This methodology is applicable for auditing specific model behaviors, understanding how capabilities emerge, identifying potential issues like hallucination triggers or ingrained biases, and assessing whether a model's explicit reasoning (like chain-of-thought) matches its internal processing. These insights confirm that modern <span style="font-weight:700;"><b>AI models develop intricate internal strategies far beyond simple pattern matching</b></span>. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> However, the researchers are clear about the limitations. These findings are based on specific examples and don't claim universality; the identified mechanisms might only apply in certain situations, and other mechanisms likely exist undiscovered. The methods currently <span style="font-weight:700;"><b>struggle with long prompts</b></span>, very complex reasoning chains spanning many steps, understanding why a model doesn't do something, and fully explaining the important role of attention mechanisms. </p></td></tr><tr><td align="center" valign="top" style="padding-bottom:14px;padding-left:28px;padding-right:28px;padding-top:14px;text-align:center;width:100%;word-break:break-word;" class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" align="center" style="margin:14px auto 14px auto;"><tr><td align="center" valign="middle" height="42" style="height:42px;background-color:#2C81E5;border-color:#DFD150;border-radius:px px px px;border-style:solid;border-width:px px px px;color:#FFFFFF;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.tLfGW26lAwaS9gFg17HSoCjz9klQtdBpBAnPAmkFOcZAKw7SwKO6lKoH2uiZnyjW1OHshDphzWK44e99IK1c6WvbAeY17_lF8MokWenzEtUIJx_AvvJAce1_yJ0P8E70RF2o_XGUWLXMjq5fF_YVdcm7MMP-X4aY2uiLg_3Jz9onBzZKUR7VNFeuoPOkFxlqBrNPzneQ-yyUHtPPXhd5dA/4fa/NRovB4-NQV-yUQtq5HqcKA/h14/h001.mFoEQk55MWXJ3mdq_yJX2XT7f8kzces3j1vU83940tk" target="_blank" rel="noopener noreferrer nofollow" style="color:#FFFFFF;display:block;font-size:16px;font-size:16px;font-weight:normal;padding:0px 14px;padding:14px 14px 14px 14px;text-decoration:none;"> Read Full Paper </a></td></tr></table></td></tr><tr><td><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" style=""><tr><td bgcolor="#222222" style="background-color:#222222;padding:0.0px 0.0px 0.0px 0.0px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0"><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"></p></td></tr></table></td></tr></table></td></tr><tr><td id="defeating-prompt-injections-by-desi" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:Bold;padding:0px 28px;text-align:left;"><h2 style="color:#2A2A2A;font-weight:Bold;">Defeating Prompt Injections by Design </h2></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style=""><i>Debenedetti et al. [Google, Google DeepMind, ETH Zurich]</i></span></p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;"> ♥ 549 </span></span><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> </span><span style="background-color:#e0e0e0;"><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> LLM Jailbreaking </span></span><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> </span></p></td></tr><tr><td id="securing-llm-agents-against-untrust" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Securing LLM Agents Against Untrusted Data</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> LLMs are increasingly used as the core component in agentic systems designed to interact with external environments, such as APIs, web pages, or user inputs. This interaction exposes a significant vulnerability: <span style="font-weight:700;"><b>prompt injection attacks</b></span>. When an LLM agent processes data from untrusted sources (e.g., content scraped from a website or output from an external tool), malicious instructions embedded within that data can potentially hijack the agent's behavior. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> These attacks can lead to undesirable outcomes, ranging from executing unintended actions to leaking sensitive user data. Current defense mechanisms often focus on training models to resist such manipulations or rely on system prompts defining security rules, but these approaches can be difficult, especially when the malicious instructions are cleverly disguised within legitimate-looking data. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/2d1f915c-c6d0-4ba7-8320-d688e7e48e6e/image.png?t=1743529401" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> This paper introduces <span style="font-weight:700;"><b>CaMeL (Capabilities for Machine Learning)</b></span>, a new defense mechanism that operates as a protective system layer around the LLM agent, and aims to mitigate prompt injection risks without altering the LLM itself. It achieves this by first explicitly extracting the intended control flow (the sequence of actions) and data flow (how information moves) directly from the initial, trusted user query. </p></td></tr><tr><td id="how-to-stop-prompt-injection-attack" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">How to Stop Prompt Injection Attacks via CaMeL</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> CaMeL (Capabilities for Machine Learning) introduces a system-level defense architecture designed to mitigate prompt injection vulnerabilities in LLM-based agentic systems without requiring modifications to the underlying language model. Its core principle draws heavily from established software security frameworks, namely Control Flow Integrity (CFI) and Information Flow Control (IFC). </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> It requires you to separate the determination of the agent's execution path (control flow) from the handling of potentially untrusted data retrieved during execution. CaMeL achieves this by treating the initial user query as the sole trusted source for <span style="font-weight:700;"><b>defining the intended program structure</b></span> and operational sequence. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/95763c8e-8021-41bc-b75d-a454a087e2e3/image.png?t=1743529350" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Operationally, CaMeL follows a distinct procedural flow. First, upon receiving a user query deemed trusted, the CaMeL system <span style="font-weight:700;"><b>performs static analysis</b></span> to extract an explicit representation of the intended control flow (e.g., a sequence of API calls or internal functions) and the corresponding data flow dependencies. This effectively defines the authorized execution graph before any interaction with external, untrusted sources. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Subsequently, when the LLM is invoked (for instance, to process data from a web page or an API response), its output, including any retrieved data, is <span style="font-weight:700;"><b>strictly compartmentalized</b></span>. This retrieved data, regardless of any embedded malicious instructions, is prohibited from influencing or altering the pre-determined control flow graph established from the original trusted query. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/93671888-9603-4da7-9f80-af4fc128ce34/image.png?t=1743529478" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> The enforcement of security guarantees within CaMeL relies on two key components: capabilities and a custom interpreter. Each data value within the system is associated with metadata, termed 'capabilities,' which encode fine-grained permissions dictating how that specific data can be used or propagated (implementing IFC). For instance, sensitive data might have capabilities restricting its flow to pre-approved, secure "sinks" (e.g., specific internal functions) and prohibiting transmission to external APIs. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> A custom Python interpreter then <span style="font-weight:700;"><b>executes the agent's program</b></span>, meticulously tracking data provenance and enforcing the constraints defined by these capabilities at runtime. This ensures that even if an LLM attempts to leak data due to injected prompts, the interpreter, guided by the capability system, will prevent the policy violation. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/f8f75c4e-36e0-4376-97e8-970de052b609/image.png?t=1743529442" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td id="results-and-evaluation" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Results and Evaluation</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> The researchers tested the CaMeL framework by using the comprehensive AgentDojo benchmark for testing its impact on both task completion utility and security against prompt injection across a diverse set of contemporary LLMs, including variants of Gemini, Claude, GPT, and the o1/o3 models. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Researchers concluded that CaMeL generally maintains high utility, with minimal performance degradation observed in most task suites; unexpectedly, utility even improved in certain model/task combinations (e.g., Gemini Pro 2.0 on Banking). The primary exception was the Travel suite, where reduced performance was linked to poorly documented APIs hindering the planning of LLM's ability to anticipate and parse tool output structures. This approach has some inherent limitations like "Data requires action" (where task logic depends on untrusted data inaccessible to the planning LLM) and instances of "Not enough context for Q-LLM". </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/3d94186a-ff8f-4b9c-9da2-52a9092cf988/image.png?t=1743529581" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td align="center" valign="top" style="padding-bottom:14px;padding-left:28px;padding-right:28px;padding-top:14px;text-align:center;width:100%;word-break:break-word;" class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" align="center" style="margin:14px auto 14px auto;"><tr><td align="center" valign="middle" height="42" style="height:42px;background-color:#2C81E5;border-color:#DFD150;border-radius:px px px px;border-style:solid;border-width:px px px px;color:#FFFFFF;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.DUiN96-Eq7pUHzwEhy5j28yjf9KIXZdsXoh1WlHvvKn9Mb2HOH4P1LD3-zFArpR0nZMtOUQzit9Uc4LkWI4zqFbql9KvRsdPK7P5p7TXVXWgZHPl6Ro45-xe4fPO0YFA/4fa/NRovB4-NQV-yUQtq5HqcKA/h15/h001.J5y6limUF1Eodc12oB9SbloMJDbwQeNeny5qef6S_II" target="_blank" rel="noopener noreferrer nofollow" style="color:#FFFFFF;display:block;font-size:16px;font-size:16px;font-weight:normal;padding:0px 14px;padding:14px 14px 14px 14px;text-decoration:none;"> Read Full Paper </a></td></tr></table></td></tr><tr><td><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" style=""><tr><td bgcolor="#222222" style="background-color:#222222;padding:0.0px 0.0px 0.0px 0.0px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0"><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"></p></td></tr></table></td></tr></table></td></tr><tr><td id="reasoning-to-learn-from-latent-thou" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:Bold;padding:0px 28px;text-align:left;"><h2 style="color:#2A2A2A;font-weight:Bold;">Reasoning to Learn from Latent Thoughts</h2></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style=""><i>Ruan et al. [Stanford University, University of Toronto, Vector Institute]</i></span></p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"><span style="background-color:#e0e0e0;"><span style="color:rgb(255, 58, 58);font-size:0.6rem;"> ♥ 640 </span></span><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> </span><span style="background-color:#e0e0e0;"><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> LLM Reasoning </span></span><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> </span><span style="background-color:#e0e0e0;"><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> bycloud’s pick </span></span><span style="color:rgb(44, 129, 229);font-size:0.6rem;"> </span></p></td></tr><tr><td id="introduction-to-bootstrapping-laten" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Introduction to Bootstrapping Latent Thoughts (BoLT)</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> LLMs are getting bigger at a very fast pace and we are running out of available human-written text to train them. This challenge is known as the data bottleneck and it can slow down the further progress in LLMs. To address this, the researchers introduced a new pretraining approach named "reasoning-to-learn." This approach says that standard web text is merely a compressed representation of a richer, underlying human thought process. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> By explicitly modeling and inferring these latent thoughts (Z) associated with the observed text (X), the researchers propose augmenting the pretraining data with this inferred reasoning, aiming to significantly enhance data efficiency. The core idea is that these latent thoughts contain contextual information and reasoning steps that enable more effective learning, much like how humans actively infer and decompress information when reading complex material. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/d8355397-7b0b-41aa-b304-ed9efbb99909/image.png?t=1743529743" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td id="inner-workings-of-bootstrapping-lat" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Inner-Workings of Bootstrapping Latent Thoughts (BoLT)</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> We want a language model (LM) to learn more efficiently, especially when high-quality training data is scarce. The core idea is that text we see online is often a summary of a longer thought process. If we could add those "missing thoughts" back into the training data, the LM might learn faster and better. The problem is, getting these thoughts usually requires a very powerful helper model, which limits how good our own LM can become. BoLT offers a clever workaround: it lets the LM teach itself to generate better thoughts over time. It does this using a two-step process called Expectation-Maximization (EM), repeated in cycles. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/fca5a3c4-9f2b-409a-be7a-ef59f210b6d7/image.png?t=1743529770" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> First is the <span style="font-weight:700;"><b>"Expectation" (E) step</b></span>: Here, the current LM takes a piece of text and tries to guess the underlying "thoughts." But instead of just taking its first guess, it generates multiple possible thought processes (this is the Monte Carlo part). Then, it evaluates these candidate thoughts; it gives higher scores to thoughts that are both logical on their own and do a good job explaining the original text, while penalizing obvious or unhelpful thoughts. It then picks one of the best-scoring thoughts to use. This whole process acts like a filter, selecting higher-quality reasoning than the model might produce on average. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Second is the <span style="font-weight:700;"><b>"Maximization" (M) step</b></span>: The LM is then trained using the original text paired with these carefully selected, higher-quality thoughts generated in the E-step. Because it's learning from better, more explicit reasoning, the LM itself becomes smarter and better at understanding context and logic. Instead of directly using latents sampled from the current model's posterior q(Z | X; Mt), BoLT samples multiple (K) candidate latents. It then calculates importance weights for each candidate based on the ratio p(Z(k), X; Mt) / q(Z(k) | X; Mt). </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/7ffc5557-e44e-4b23-b3b5-55b434c6886f/image.png?t=1743529839" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> The key is that this EM cycle repeats: the slightly smarter LM from the M-step then performs the next E-step, generating even better thoughts, which leads to more effective learning in the next M-step, creating a self-improvement loop where the model "bootstraps" its own reasoning capabilities. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/b3aa4fc4-317d-4abb-a0e6-cda317274717/image.png?t=1743529875" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td id="real-world-implications-of-bootstra" class="dd" align="left" valign="top" style="color:#2A2A2A;font-weight:normal;padding:0px 28px;text-align:left;"><h3 style="color:#2A2A2A;font-weight:normal;"><span style="color:rgb(67, 67, 67);">Real-World Implications of Bootstrapping Latent Thoughts (BoLT)</span></h3></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> The researchers tested their new "Bootstrapping Latent Thoughts" (BoLT) method to see if it really helps language models learn better with limited data, focusing on math problems. They found that <span style="font-weight:700;"><b>it works surprisingly well</b></span>! In their main experiment, they took a standard language model and had it go through several cycles of generating its own "thinking steps" for a fixed set of math text and then retraining itself using those thoughts. With each cycle, the model got consistently better, both at understanding the text (measured by standard model quality scores) and, more importantly, at <span style="font-weight:700;"><b>solving challenging math problems</b></span> from the MATH benchmark. This self-improvement continued for at least three cycles. </p></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> Additionally, the models trained using this BoLT method significantly outperformed models trained the old-fashioned way, i.e. just feeding them the raw text, even when those baseline models were trained for the same amount of computer time or saw the same amount of raw text. This shows that teaching the model to generate and learn from its own reasoning makes learning much more efficient. </p></td></tr><tr><td align="center" valign="top" style="padding: 20px 0px 20px 0px; " class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" style="margin:0 auto 0 auto;"><tr><td align="center" valign="top" style="width:600px;"><img src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/dc910ee7-fbb3-4019-aff0-35766f647099/image.png?t=1743529934" alt="" height="auto" width="600" style="margin: 0px 0 0px;display:block;width:100%;" border="0"/></td></tr></table></td></tr><tr><td class="dd" align="left" style="padding:0px 28px;text-align:left;word-break:break-word;"><p style="line-height:24px;"> They also discovered that letting the model generate <span style=""><i>more</i></span> candidate "thoughts" and picking the best one during each cycle further boosted performance. It suggests that <span style="font-weight:700;"><b>spending more computing power</b></span> on generating better reasoning during training <span style="font-weight:700;"><b>pays off</b></span>. While the method greatly improved complex math skills, they did note a slight dip in performance on simpler word problems in some tests (unless specifically fine-tuned), hinting that intense focus on one area might slightly affect others. </p></td></tr><tr><td align="center" valign="top" style="padding-bottom:14px;padding-left:28px;padding-right:28px;padding-top:14px;text-align:center;width:100%;word-break:break-word;" class="dd"><table role="none" border="0" cellspacing="0" cellpadding="0" align="center" style="margin:14px auto 14px auto;"><tr><td align="center" valign="middle" height="42" style="height:42px;background-color:#2C81E5;border-color:#DFD150;border-radius:px px px px;border-style:solid;border-width:px px px px;color:#FFFFFF;font-family:'Open Sans','Segoe UI','Apple SD Gothic Neo','Lucida Grande','Lucida Sans Unicode',sans-serif;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.DUiN96-Eq7pUHzwEhy5j28yjf9KIXZdsXoh1WlHvvKmcPEDXCsJPYvcBM-hywu2LP3-exkX1uUMcjVxgyXWJI9hsrM2adwO4e4MyMv1IjrcSY7atTZ3n8SEcVD16M_jS/4fa/NRovB4-NQV-yUQtq5HqcKA/h16/h001.G0mJVd1-0bVt8kMUxHhf00NtpvaayRDLMXgd6bBl0P0" target="_blank" rel="noopener noreferrer nofollow" style="color:#FFFFFF;display:block;font-size:16px;font-size:16px;font-weight:normal;padding:0px 14px;padding:14px 14px 14px 14px;text-decoration:none;"> Read Full Paper </a></td></tr></table></td></tr><tr><td class="dd" style="padding: 20px;"><table width="100%" cellpadding="0" cellspacing="0" role="none" style="max-width:520px;margin:0 auto;"><tr><td class="q" style="padding:16px 16px 6px 16px;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.tLfGW26lAwaS9gFg17HSoDDFT6eh5Nsg0xYVQj-h6I3o9m2k79_qw4izMYhmcI36rB-mWiKFHYGmdlqg4U94gGidag_UIRyb6D6zjDKQOi8hoM2AkcN0VT1TmohVBvDNdPMRZGv01xupvE9mIvZS-qCbCf1P8oWHc1p013bIXQ4/4fa/NRovB4-NQV-yUQtq5HqcKA/h17/h001.Uf619b7EPlqk0OMyMKwf_7E3TuYmoKZHgnEU88tp3ns" style="text-decoration:none !important;"><table width="100%" cellpadding="0" cellspacing="0" border="0" role="none"><tr><td width="100%" style="padding: 0 0 14px 0;text-decoration:none;width:100%;"><table width="100%" cellpadding="0" cellspacing="0" border="0" role="none"><tr><td width="36" style="width:36px;"><img src="https://pbs.twimg.com/profile_images/1698572487909400576/BvncwnrP_normal.jpg" alt="tw profile: The AI Timeline" style="display:block;width:36px;height:36px;border-radius:50%;border:0;"/></td><td width="400" style="padding:0 0 0 8px;text-decoration:none;"><span style="display:block;font-size:14px;color:#1c2022;font-weight:700;"> The AI Timeline </span><span style="display:block;color:#697882;font-size:14px;"> @TheAITimeline </span></td><td width="24" align="right" style="vertical-align:text-top;"><img width="24" height="24" loading="lazy" alt="tw" style="border:0;" src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/static_assets/x_logo.png"/></td></tr></table></td></tr><tr></tr><tr><td style="word-break:break-word;"><p>🚨This week's top AI/ML research papers:</p><p>- GPT-4o System Card: Native Image Generation <br>- Anthropic's On the Biology of a LLM <br>- Gemma 3 Technical Report <br>- Qwen2.5-Omni Technical Report <br>- Reasoning to Learn from Latent Thoughts <br>- Defeating Prompt Injections by Design <br>- Scaling</p></td></tr><tr><td style="padding:12px 0 0 0;"></td></tr><tr><td align="center" style="padding:8px 0 0 0;width:480px;"><img src="https://pbs.twimg.com/media/GnUk8utXcAAGs5I.jpg" width="480" height="auto" style="display:block;border:1px solid #E1E8ED;border-radius:5px;width:100%;max-width:480px;height:auto;"/></td></tr><tr><td height="8" style="line-height:1px;font-size:1px;height:8px;"> </td></tr><tr><td align="left" valign="top" class="s"><p>10:17 PM • Mar 30, 2025</p></td></tr><tr><td height="10" style="line-height: 1px; font-size: 1px; height: 10px;"> </td></tr><tr><td height="1" bgcolor="#e1e8ed" style="line-height:0px;font-size:0px;height:1px;"></td></tr><tr><td height="10" style="line-height:1px;font-size:1px;height:10px;"> </td></tr><tr><td align="left" valign="top" class="s"><p><b style="color:#1C2022">1.08K</b> Likes <b style="color:#1C2022">138</b> Retweets </p></td></tr><tr><td align="left" valign="top" class="s"><div align="center" style="text-align:center;margin-top:4px;margin-bottom:4px;padding:8px;border:1px solid #ccd6dd;border-radius:9999px;color:#1B95E0"><b>15 Replies</b></div></td></tr></table></a></td></tr></table></td></tr><tr><td class="dd" align="center" valign="top" style="padding:20px;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.amatuKKICSickUKplYJXmA5meQDBfUT78oHpH-wFd-DrtuKnPjf_uDYdc53dMIQKBri_aoSvEhy0iEFy3baKlQbMbp1T0aErJ3Ur-r-l8gAgrNiXcJ_TytfB0lY9ClNg/4fa/NRovB4-NQV-yUQtq5HqcKA/h18/h001.WBNDaOS71OLkCYp3ybfu5wBbFkynGBY-eWDriYapgjI" style="text-decoration:none;"><table align="center" width="100%" cellpadding="0" cellspacing="0" border="0" role="none" style="max-width:520px;margin:0 auto;"><tr><td class="p" width="100%" style="padding:2px;border:none;"><table width="100%" cellpadding="0" cellspacing="0" border="0" role="none"><tr><td align="center" valign="top" style="width:100%;"><div style="max-height:0;position:relative;opacity:0.999;width:100%;mso-hide:all;"><div style="display:inline-block;width:100%;padding-top:25%;"><img width="20%" height="auto" loading="lazy" alt="" style="border:0;" src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/static_assets/youtube_play_icon.png"/></div></div><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.amatuKKICSickUKplYJXmA5meQDBfUT78oHpH-wFd-CAfUfs-rxHxHVCBcWCJCBIE-ELA58CW_vORAMF9Ar-HkFaFnOWWpOdJ6_cNFwwbmj8sH83DA91fp2wbXRda8Yv/4fa/NRovB4-NQV-yUQtq5HqcKA/h19/h001.ymniBXaLROxlNJB8cTWkI7tG0EGLqQmRdDDqyOJUVis" style="text-decoration:none;"><img src="https://i.ytimg.com/vi/lkZTSUYfnTI/maxresdefault.jpg" width="480" height="auto" loading="lazy" alt="YouTube video by bycloud" style="display:block;height:auto;border:0;outline:none;text-decoration:none;background-color:#000000;width:100%;"/></a></td></tr><tr><td><p style="font-size:12px;font-weight:500;font-style:italic;font-family:Helvetica, Calibri, sans-serif;color: #686a6d; padding-top:0 !important;padding-bottom:6px !important; padding-left:4px !important;"> Anthropic found a "terrifying" consequence of adding reasoning to AI </p></td></tr></table></td></tr></table></a></td></tr></table></td></tr></table></td></tr><tr><td align="center" valign="top"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td><tr><td class="b" align="center" valign="top" bgcolor="#2a2a2a" style="padding:0px 0px 0px 0px;border-style:solid;border-width: 0px 0px 0px 0px;border-color: #2a2a2a;border-bottom-left-radius:10px;border-bottom-right-radius:10px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td align="center" valign="top" bgcolor="#73ddff" style="padding:12px"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td><span style="padding-left:1px;"></span></td><td align="center" valign="middle" width="75" style="width:75px;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.1muhFWIqieRYpaJ-FbWSCQqcWoV4NNHHr5SkP9THApWuHAAlWLQxI3Q_IqFmt_DcyAxeC8jDApCnHmMSBGpBb5sgtimvBYgxRX-Rp7s0F3LjCHoSwdhr83OBqRFhJ1y_/4fa/NRovB4-NQV-yUQtq5HqcKA/h20/h001.IS0l7bKVErodEKBw65H9x4MEAm_sQK10rUTbgqpb11g" style="text-decoration:none;"><img width="22" height="22" alt="tw" border="0" style="display:block;max-width:22px;color:Dark" src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/static_assets/x_dark.png"/></a></td><td align="center" valign="middle" width="75" style="width:75px;"><a href="https://elink4f7.mail.bycloud.ai/ss/c/u001.amatuKKICSickUKplYJXmBoQnQ9VXnB2zTxBG4HeHBgjMqVxpoXRdj01cjwyoVlHgiebEOgBvwHtevoVpsSvpn3Q1di2ml6sb3cBM-X6IStQbj_zQSVGWJ8AAmPw2en2/4fa/NRovB4-NQV-yUQtq5HqcKA/h21/h001.FGGWNH1obdQH8Yh2gM_yyYkBHLjw6yUKYZS4GGjPeGc" style="text-decoration:none;"><img width="22" height="22" alt="yt" border="0" style="display:block;max-width:22px;color:Dark" src="https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/static_assets/youtube_dark.png"/></a></td><td><span style="padding-left:1px;"></span></td></tr></table></td></tr><tr><td height="10" style="line-height:1px;font-size:1px;height:10px;"> </td></tr><tr><td class="w" align="center" valign="top" style="padding:15px 15px 15px 15px;"><table role="none" width="100%" border="0" cellspacing="0" cellpadding="0" align="center"><tr><td align="center" valign="top"><p style="font-family:'Verdana',Geneva,sans-serif;color:#FFFFFF!important;"> Update your email preferences or unsubscribe <a class="link" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.c6q0w4g5sodbtO4I1B_pxWc4htTObwdorovK0nFHVH-4pUdVE0ELYH5DsNemk732SjNwhPNJ25r0O8B5vYifsBhEpz-DJgyVFmavJPa0OyKRRnvw4o7XGyvIv7PRofnm-k7e_r64CW__v-Qo3gteJN7ld1e1T-pKKeuzTcQGNilUGXNQp2jAaG_I-ZCfPCzUdP9_oPsw87IijK5THeAfRImmmQCXTVvtPve429WqgWEvqEBULsMTa1ybH76iFm7YDj9rNpPQSyAVs7O63b2Lp8rzGWEga04Z15TVMZpg7gEG1Tavivrwm8Z1VAPfR-Kt2bUtDISFeeRAwgOoYqNzXOV369IU--_DSFbDk3tGJEK9cGj8A6DVrbkyTXRk8IH1XI7HDOD3t4C1Yz5ESmovuB4Q_99Zhr7WAy6OHFJIDLk_yOOb1Ci3mE4ZEmrW2WGOrfBrOXjHZnwJWv8_8JWJx57uAh80ZUHZr1yZcpDsevlqcbpezX3fz5hDy911etF1GrufZAU-2hDt1ijiYXqtxzzkYFpJqynTyDMIIiYpZ0YvUsTbNKSfMRr_vuYaa6XcBRJAdugOmgCPRXwReVlFmC2ucyVq3Yklas0Zpz9F9IcCN3cwpOND0CNcTvx37u_aXy3vHjB5SooZi5HJJX_CS28gBP6FrnR7E3vv82p63neoWnojb8K_F8MQOw4L0823FnrYM-_gWycUy5pZFSZLxEX5hH0_doXTIAHrGgrE9onQJC4wj8mKdZ4LiLYvRXYWrRL2sP3tY8R71k5ZkUTBZXj9t7AR0u1bc-fSi21fs2rULpcGzayjnfl0jbVtF4c5JZnP8PxW0-WP_3a34YKGhQ/4fa/NRovB4-NQV-yUQtq5HqcKA/h22/h001.T7xfYuaTEpJh-rg7iXG7cYSAcfDl9cksYA7ZR4PaRgw" style="text-decoration:underline;text-decoration-color:#FFFFFF!important;color:#FFFFFF!important;"> here</a></p><p class="copyright" style="font-family:'Verdana',Geneva,sans-serif;color:#FFFFFF!important;"> © 2025 bycloudai </p><p style="font-family:'Verdana',Geneva,sans-serif;color:#FFFFFF!important;"> 228 Park Ave S, #29976, New York, New York 10003, United States </p></td></tr><tr style="display: table-row !important;"><td align="center" valign="top" style="padding-top:20px;" style="display:table-cell !important;"><table role="none" border="0" cellspacing="0" cellpadding="0" align="center" style="display:table !important;"><tr style="display:table-row !important;"><td class="u" align="center" valign="middle" height="32" style="height:32px;display:table-cell !important; max-height: 32px !important;margin:0px !important; background-color: #ffffff !important;"><a style="line-height:32px !important;text-decoration:none;display:block !important;" href="https://elink4f7.mail.bycloud.ai/ss/c/u001.DUiN96-Eq7pUHzwEhy5j28olDWFpV5DDKfdk_OdOKOiR1tnBH_801gvjHMqOS108W5-Onr_QroJyEyQx8pM7mORX0EhycvJCnQ1OUDE06IlDmJaWzNFWqoplHRPGglb252fWTC9zEIy8rDd-RP3mAxyCWm8wD8UI8Rmr6Rf-tJt_oFayR9FEFGPIAMl0KyINHWu7TxY-J31mwYdu-y_FRJzzBL_h3wlQG30ydLVqSLNZvbxIXGrY-scgBJGciCND/4fa/NRovB4-NQV-yUQtq5HqcKA/h23/h001.rghRyOJEJyeeKoG7XnKmua-Tmij8Hx_8jH1DszZbjbM"><img src="https://media.beehiiv.com/output-onlinepngtools.png" width="16" alt="beehiiv logo" style="display:inline-block !important;max-width:16px !important; vertical-align:-3px !important;width: 16px !important;" border="0"/><span style="padding-left:11px !important;display: inline-block !important;">Powered by beehiiv</span></a></td></tr></table></td></tr><tr><td align="left" valign="top" height="2" style="height:2px;"><a href='https://elink4f7.mail.bycloud.ai/ss/c/u001.CxDkkVpJsBdVoe83c_tBWsHIaP4XNp0WgUYqLvHcKk_3uqk_KIkz4ddLinhFbud6JuxLFdSUhYnR7b1NSsmbtzXNGNblnEEMKUtkCAjkn8Y/4fa/NRovB4-NQV-yUQtq5HqcKA/h24/h001.MMhbmYQxJ4S03Ufe3Wzp1SGyHU_p5UAmSjh2F6ff_ic' style="color: #2a2a2a !important; cursor: default; font-size: 1px; text-decoration: none;"> Terms of Service </a></td></tr></table></td></tr></table></td></tr></td></tr></table></td></tr></table></td></tr></table></td></tr></table></div></body></html>