<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Sange Mehrab</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/</link><description>Recent content on Sange Mehrab</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Fri, 01 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://anwarshamim01.github.io/Sang_e_Mehrab/index.xml" rel="self" type="application/rss+xml"/><item><title>Exciton Diffusion as a Blind Search</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/research/exciton-diffusion-blind-search/</link><pubDate>Tue, 14 Apr 2026 10:30:00 +0300</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/research/exciton-diffusion-blind-search/</guid><description>Master&amp;#39;s thesis note on exciton diffusion, blind search, and first-hitting versus first-return statistics in disordered semiconductor environments.</description></item><item><title>1.1 From Scalars to Vectors: Data Points, Rows, Columns, and Transpose</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-01/</link><pubDate>Mon, 27 Apr 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-01/</guid><description>A careful introduction to how a single house measurement becomes a scalar, how many measurements become a vector, how rows become datasets, and how vectors are used in machine learning and neural networks.</description></item><item><title>1.2 From Matrices to Tensors: Tables, Images, Batches, and Multilinear Structure</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-02/</link><pubDate>Tue, 28 Apr 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-02/</guid><description>A deep introduction to matrices and tensors for AI mathematics: data matrices, images as matrices, color images and batches as tensors, matrix operations, linear maps, rank, norms, eigenvalues, SVD, tensor order, modes, unfolding, n-mode products, tensor contractions, CP decomposition, Tucker decomposition, and the formal tensor-product view.</description></item><item><title>1.3 Probability Distributions, Noise, Markov Chains, SDEs, and Stochastic PDEs</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-03/</link><pubDate>Wed, 29 Apr 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-03/</guid><description>A careful introduction to distributions and noise for AI mathematics, starting from real-world uncertainty and moving toward Gaussian random vectors, white and colored noise, Markov chains, Brownian motion, Itô SDEs, Fokker–Planck equations, stochastic PDEs, Euler–Maruyama, Milstein methods, and the mathematical foundation of diffusion generative models.</description></item><item><title>1.4 Mathematics of Large Language Models: Training, Inference, Attention, Scaling, and Alignment</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-04/</link><pubDate>Fri, 01 May 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-04/</guid><description>A beginner-to-advanced mathematical introduction to LLMs, covering autoregressive language modeling, tokenization, vector embeddings, positional encodings, transformer blocks, attention, softmax, cross-entropy, maximum likelihood, backpropagation, AdamW, scaling laws, compute-optimal training, MoE, efficient attention, KV caching, speculative decoding, quantization, LoRA, RLHF, DPO, PPO, and inference-time reasoning.</description></item><item><title>1.5 Mathematics of Machine Learning and Deep Learning for Time Series</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-05/</link><pubDate>Fri, 01 May 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-05/</guid><description>A beginner-to-advanced mathematical introduction to time-series machine learning and deep learning, covering classical forecasting, stochastic processes, spectral analysis, supervised time-series learning, recurrent neural networks, LSTMs, GRUs, temporal convolutional networks, transformers, PatchTST, iTransformer, TimesNet, TimeMixer, state-space models, S4, Mamba, DeepAR, N-BEATS, N-HiTS, Temporal Fusion Transformers, diffusion-based forecasting, neural ODEs, neural CDEs, foundation models, probabilistic losses, conformal prediction, anomaly detection, classification, and imputation.</description></item><item><title>Section 2.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-02/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-02/section-01/</guid><description/></item><item><title>Section 3.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-03/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-03/section-01/</guid><description/></item><item><title>Section 4.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-04/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-04/section-01/</guid><description/></item><item><title>Section 5.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-05/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-05/section-01/</guid><description/></item><item><title>Section 6.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-06/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-06/section-01/</guid><description/></item><item><title>Section 6.2</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-06/section-02/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-06/section-02/</guid><description/></item><item><title>Inference at scale for Large Language Models</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-06/section-03/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-06/section-03/</guid><description/></item><item><title>Section 7.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-01/</guid><description/></item><item><title>Section 7.2</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-02/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-02/</guid><description/></item><item><title>Karpathy autoresearch</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-03/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-07/section-03/</guid><description>&lt;h1 id="karpathy-autoresearch-explained"&gt;Karpathy Autoresearch Explained&lt;/h1&gt;
&lt;h2 id="introduction"&gt;Introduction&lt;/h2&gt;
&lt;p&gt;This lesson introduces &lt;strong&gt;autoresearch&lt;/strong&gt; as a practical workflow for letting an AI coding agent run experiments without waiting for a human to choose every next step. The basic pattern is simple: define the goal, freeze the evaluator, let the agent propose code changes, run the experiment, keep the change only if the metric improves, and repeat. The public examples make the idea concrete: single-GPU overnight runs improved &lt;code&gt;val_bpb&lt;/code&gt; from &lt;code&gt;0.997900&lt;/code&gt; to &lt;code&gt;0.969686&lt;/code&gt; in 126 experiments on an H100, and those smaller depth-12 findings later transferred to larger depth-24 &lt;code&gt;nanochat&lt;/code&gt; runs, reducing the &amp;ldquo;time to GPT-2&amp;rdquo; leaderboard entry from &lt;strong&gt;2.02 hours to 1.80 hours&lt;/strong&gt;, with a later entry at &lt;strong&gt;1.65 hours&lt;/strong&gt;. The rest of this section turns that workflow into a tutorial: first the naming and intuition, then the loop, comparisons, implementations, strengths, limitations, and a practical recipe for building a similar system.&lt;/p&gt;</description></item><item><title>Section 8.1</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-08/section-01/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-08/section-01/</guid><description/></item><item><title>Building a Systematic Technical Blog</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/posts/2026/04/building-a-systematic-technical-blog/</link><pubDate>Sun, 12 Apr 2026 22:00:00 +0300</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/posts/2026/04/building-a-systematic-technical-blog/</guid><description>A short note on the publishing system behind this blog: structured pages, mathematical typography, research notes, and reusable layouts for long-form technical writing.</description></item><item><title>Intro</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/posts/2026/04/intro/</link><pubDate>Wed, 08 Apr 2026 10:41:50 +0300</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/posts/2026/04/intro/</guid><description>Short introduction page used as a simple notebook-style entry in the archive.</description></item></channel></rss>