<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Computer Vision on Blog</title><link>https://svfcode.github.io/en/tags/computer-vision/</link><description>Recent content in Computer Vision on Blog</description><generator>Hugo -- gohugo.io</generator><language>en</language><lastBuildDate>Thu, 12 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://svfcode.github.io/en/tags/computer-vision/index.xml" rel="self" type="application/rss+xml"/><item><title>Essential Mathematics — Overview</title><link>https://svfcode.github.io/en/p/math-essentials-overview/</link><pubDate>Thu, 12 Mar 2026 00:00:00 +0000</pubDate><guid>https://svfcode.github.io/en/p/math-essentials-overview/</guid><description>&lt;p>This is the intro to the «Essential Mathematics» series — an overview of which math areas are needed for AI and machine learning, and what they are used for.&lt;/p>
&lt;h2 id="linear-algebra">Linear Algebra&lt;/h2>
&lt;p>&lt;strong>What you need:&lt;/strong> vectors, matrices, matrix multiplication, linear transformations, dot product, vector norm, eigenvalues and eigenvectors (advanced).&lt;/p>
&lt;p>&lt;strong>Why:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>Vectors — the basic way to represent data: a word, an image, or a model&amp;rsquo;s «thought» is encoded as a vector of numbers&lt;/li>
&lt;li>Matrices — all neural network operations (convolutions, attention, linear layers) boil down to matrix multiplication&lt;/li>
&lt;li>Linear transformations — understanding how a layer «mixes» and combines features&lt;/li>
&lt;li>Without linear algebra you cannot read formulas in papers or understand what the code does (e.g. &lt;code>W @ x&lt;/code> is weight matrix times input vector)&lt;sup id="fnref:1">&lt;a href="#fn:1" class="footnote-ref" role="doc-noteref">1&lt;/a>&lt;/sup>&lt;/li>
&lt;/ul>
&lt;h2 id="derivatives">Derivatives&lt;/h2>
&lt;p>&lt;strong>What you need:&lt;/strong> derivatives, partial derivatives, gradient, chain rule, function optimization.&lt;/p>
&lt;p>&lt;strong>Why:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>Neural networks are trained via gradient descent: we move weights in the direction that reduces the error&lt;/li>
&lt;li>Gradient — a vector of partial derivatives showing which way to «tweak» each parameter&lt;/li>
&lt;li>Chain rule is the backbone of backpropagation — the algorithm that computes gradients for all layers&lt;/li>
&lt;li>Without calculus you cannot understand where weight updates come from or why training works at all&lt;/li>
&lt;/ul>
&lt;h2 id="probability-and-statistics">Probability and Statistics&lt;/h2>
&lt;p>&lt;strong>What you need:&lt;/strong> distributions (normal, Bernoulli, etc.), expectation, variance, conditional probability, Bayes&amp;rsquo; theorem, maximum likelihood estimation (MLE).&lt;/p>
&lt;p>&lt;strong>Why:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>Models are often formulated probabilistically: «probability of the next word», «how confident is the model»&lt;/li>
&lt;li>Loss functions often derive from likelihood (cross-entropy, MSE, etc.)&lt;/li>
&lt;li>Statistics for working with data: normalization, quality assessment, understanding metrics (precision, recall, AUC)&lt;/li>
&lt;li>Bayes — foundation of many classical algorithms and the way to «update beliefs» on new data&lt;/li>
&lt;/ul>
&lt;h2 id="computer-vision-cv">Computer Vision (CV)&lt;/h2>
&lt;p>&lt;strong>On top of the general minimum:&lt;/strong> discrete convolution (2D), affine and projective transformations, basic image geometry (camera calibration, epipolar geometry).&lt;/p>
&lt;p>&lt;strong>Why:&lt;/strong>&lt;/p>
&lt;ul>
&lt;li>Convolution — the core of CNNs: understanding how a filter «slides» over an image, and what kernel, padding, stride mean&lt;/li>
&lt;li>Affine transformations (rotation, scale, shift) — for augmentations and understanding invariance&lt;/li>
&lt;li>Projective geometry and homographies — for stereo vision, 3D reconstruction, AR&lt;/li>
&lt;li>For classical methods (SIFT, optical flow) — image derivatives and per-pixel gradients are useful&lt;/li>
&lt;/ul>
&lt;h2 id="optional-but-helpful">Optional but helpful&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>Information theory&lt;/strong> (entropy, cross-entropy) — deeper understanding of loss functions and data compression&lt;/li>
&lt;li>&lt;strong>Convex optimization&lt;/strong> — when the problem is «nice» and when SGD converges&lt;/li>
&lt;li>&lt;strong>Differential geometry&lt;/strong> — for advanced topics (diffusion models, representations)&lt;/li>
&lt;/ul>
&lt;h2 id="practical-minimum">Practical minimum&lt;/h2>
&lt;p>To comfortably read papers and code:&lt;/p>
&lt;ul>
&lt;li>Be able to multiply matrices and understand what a linear layer does&lt;/li>
&lt;li>Understand what a gradient is and why it matters for training&lt;/li>
&lt;li>Know basic distributions and quality metrics&lt;/li>
&lt;/ul>
&lt;p>You can pick up the rest as needed — when it appears in a specific problem or paper.&lt;/p>
&lt;div class="footnotes" role="doc-endnotes">
&lt;hr>
&lt;ol>
&lt;li id="fn:1">
&lt;p>&lt;code>W @ x&lt;/code> — in Python (NumPy, PyTorch) the matrix multiplication operator: the weight matrix W is multiplied by the input vector x. The result is the output vector. A linear layer essentially computes &lt;code>W·x + b&lt;/code>.&amp;#160;&lt;a href="#fnref:1" class="footnote-backref" role="doc-backlink">&amp;#x21a9;&amp;#xfe0e;&lt;/a>&lt;/p>
&lt;/li>
&lt;/ol>
&lt;/div></description></item></channel></rss>