<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Course on Sange Mehrab</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/categories/course/</link><description>Recent content in Course on Sange Mehrab</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Fri, 01 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://anwarshamim01.github.io/Sang_e_Mehrab/categories/course/index.xml" rel="self" type="application/rss+xml"/><item><title>1.1 From Scalars to Vectors: Data Points, Rows, Columns, and Transpose</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-01/</link><pubDate>Mon, 27 Apr 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-01/</guid><description>A careful introduction to how a single house measurement becomes a scalar, how many measurements become a vector, how rows become datasets, and how vectors are used in machine learning and neural networks.</description></item><item><title>1.2 From Matrices to Tensors: Tables, Images, Batches, and Multilinear Structure</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-02/</link><pubDate>Tue, 28 Apr 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-02/</guid><description>A deep introduction to matrices and tensors for AI mathematics: data matrices, images as matrices, color images and batches as tensors, matrix operations, linear maps, rank, norms, eigenvalues, SVD, tensor order, modes, unfolding, n-mode products, tensor contractions, CP decomposition, Tucker decomposition, and the formal tensor-product view.</description></item><item><title>1.3 Probability Distributions, Noise, Markov Chains, SDEs, and Stochastic PDEs</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-03/</link><pubDate>Wed, 29 Apr 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-03/</guid><description>A careful introduction to distributions and noise for AI mathematics, starting from real-world uncertainty and moving toward Gaussian random vectors, white and colored noise, Markov chains, Brownian motion, Itô SDEs, Fokker–Planck equations, stochastic PDEs, Euler–Maruyama, Milstein methods, and the mathematical foundation of diffusion generative models.</description></item><item><title>1.4 Mathematics of Large Language Models: Training, Inference, Attention, Scaling, and Alignment</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-04/</link><pubDate>Fri, 01 May 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-04/</guid><description>A beginner-to-advanced mathematical introduction to LLMs, covering autoregressive language modeling, tokenization, vector embeddings, positional encodings, transformer blocks, attention, softmax, cross-entropy, maximum likelihood, backpropagation, AdamW, scaling laws, compute-optimal training, MoE, efficient attention, KV caching, speculative decoding, quantization, LoRA, RLHF, DPO, PPO, and inference-time reasoning.</description></item><item><title>1.5 Mathematics of Machine Learning and Deep Learning for Time Series</title><link>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-05/</link><pubDate>Fri, 01 May 2026 00:00:00 +0000</pubDate><guid>https://anwarshamim01.github.io/Sang_e_Mehrab/courses/course/chapter-01/section-05/</guid><description>A beginner-to-advanced mathematical introduction to time-series machine learning and deep learning, covering classical forecasting, stochastic processes, spectral analysis, supervised time-series learning, recurrent neural networks, LSTMs, GRUs, temporal convolutional networks, transformers, PatchTST, iTransformer, TimesNet, TimeMixer, state-space models, S4, Mamba, DeepAR, N-BEATS, N-HiTS, Temporal Fusion Transformers, diffusion-based forecasting, neural ODEs, neural CDEs, foundation models, probabilistic losses, conformal prediction, anomaly detection, classification, and imputation.</description></item></channel></rss>