<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Agentic Systems on Sange Mehrab</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/</link><description>Recent content in Agentic Systems on Sange Mehrab</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Thu, 01 Jan 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/index.xml" rel="self" type="application/rss+xml"/><item><title>Section 7.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-01/</guid><description/></item><item><title>Section 7.2</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-02/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-02/</guid><description/></item><item><title>Karpathy autoresearch</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-03/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-03/</guid><description>&lt;h1 id="karpathy-autoresearch-explained"&gt;Karpathy Autoresearch Explained&lt;/h1&gt;
&lt;h2 id="introduction"&gt;Introduction&lt;/h2&gt;
&lt;p&gt;This lesson introduces &lt;strong&gt;autoresearch&lt;/strong&gt; as a practical workflow for letting an AI coding agent run experiments without waiting for a human to choose every next step. The basic pattern is simple: define the goal, freeze the evaluator, let the agent propose code changes, run the experiment, keep the change only if the metric improves, and repeat. The public examples make the idea concrete: single-GPU overnight runs improved &lt;code&gt;val_bpb&lt;/code&gt; from &lt;code&gt;0.997900&lt;/code&gt; to &lt;code&gt;0.969686&lt;/code&gt; in 126 experiments on an H100, and those smaller depth-12 findings later transferred to larger depth-24 &lt;code&gt;nanochat&lt;/code&gt; runs, reducing the &amp;ldquo;time to GPT-2&amp;rdquo; leaderboard entry from &lt;strong&gt;2.02 hours to 1.80 hours&lt;/strong&gt;, with a later entry at &lt;strong&gt;1.65 hours&lt;/strong&gt;. The rest of this section turns that workflow into a tutorial: first the naming and intuition, then the loop, comparisons, implementations, strengths, limitations, and a practical recipe for building a similar system.&lt;/p&gt;</description></item></channel></rss>