Sign In

What are AI vision models REALLY!?

0

What are AI vision models REALLY!?

What Vision AI Models Actually Are

A grossly generalized analysis. I'm not going to even mention text encoders — they ARE my expertise — just know I'll explain that in the near future.

AAAAND I broke the longer article. Fantastic. I'm going to just leave it like this now.


Brief History

Vision models are a form of combined corruption. U-Nets were originally built for biomedical image segmentation — tracking cells under microscopy, delineating boundaries in tissue scans — a medical field that requires high precision, and the outcome showed deviance.

Perfect was imperfect.

Fast forward and you have the induction of DALL·E Mini and then shortly later Stable Diffusion appeared, like a magical box of images that made sense while DALL·E Mini made discolored abominations in comparison at the time.

What Are We Really Teaching?

Control of differential accumulation to deviate incorrectness over time.

Pause for a second. You'll need to understand that, because it's not something you can simply ingest.

It means we're teaching the model how to understand when it's okay to be wrong, and how to use that wrong to build a relational understanding towards something that WE use to decipher that wrong into something useful.

Think of it this way: we fill a leaking bucket to catch the leak's water, in order to slowly accumulate water in a glass, then discard the bucket to ensure the glass is only getting the water that passes through the leaky crack so it only accumulates in that one particular glass. Do this a billion times and you have a supervised masked classification gradient bucket. Then you use that to decide the model's loss based on that particular route, meant to modify the weights that declared those correct or incorrect values.

Boom. You have controlled chaotic insanity.

It's intentionally corrupting representations in order for those representations to provide a usefully deviant stepwise medium for extraction.

How Noise Drives Learning

Imagine a 3×3 grid of binary values. We count up 3 timesteps — simple: 0, 1, and 2.

During training we grab a random timestep between lowest and highest. Say we randomly land on timestep 2, and this is what it does to our grid:

Before:        After (t=2):
O O O          X O O
O O O          O X X
O O O          X O X

We have introduced a variance to the equation. The model must now account for this and predict the necessary solution based on the architecture, the losses, and the system's constraints. In diffusion, we accumulate noise and we predict the noise's reduction to reproduce images.

This commonly uses Euler-based scheduling with Gaussian noise. Newer models use a discrete schedule that is meant to be repeatable, and that repeatable schedule allows for direct linear trajectory noise prediction — flow matching — in a less chaotic fashion.

Next up, say we land on timestep 1 the next time the same image is seen:

Before:        After (t=1):
O O O          O O O
O O O          X O O
O O O          X O X

Less noise this time, since we're earlier on the corruption path from full data toward noise. Still reliable when seen enough times for it to become a trivial problem to solve.

Now timestep 3. Say our system ensures it never sees full noise — a reasonable design choice if you want to retain certain structural behavior:

After (t=3):
O X O
X X X
X O X

Run a few billion of these and you have a way to predict potential outcomes from any noise level the model has been trained on.

Why Imperfect?

When you grab a prediction from a certain set of sequences to train the model — say we teach a potato — a potato form is pretty specifically a potato. Given enough iterations, you'll see a potato. The complexity is enough to represent a similar pattern of deviance based on a potato.

Now say you want a blue potato. The model has likely seen a lot of blue, but the compositional binding between "blue" and "potato" may not be well-generalized. Did you teach <COLOR> <SUBJECT> as a composable relationship, or did you teach "red apple" and "taco salad" and "orange orange" as atomic concepts?

The model predicts the very best it can based on the current state, in relation to the need for whatever is requested. If you want a blue potato, the model may produce BLUE something — a wall, a tire, a chicken sandwich — while simultaneously having this drive, this pull, towards what is known as a potato. The patterns corrupt each other and accumulate together to form... incorrectness:

Step 1: Blue color is applied to the latent vector, but potato is pale orangeish-yellow.

Step 2: Blue color is applied to the latent vector, but potato is still orangeish-yellow, now sitting against a blue picture frame.

Step 3: Blue is applied, potato becomes a mixture of yellow and blue, background goes solid blue while the potato is incorrectly colored.

Step 4: Blue is applied, potato is now blue, background subsides revealing a more dominant average color cutting through.

So on and so forth, until the model generates a potato sitting on a blue plate. Perfect quality, perfect fidelity, fully relational — and yet entirely incorrect on multiple fronts, while being fully correct within the parameters of expectation.


Stars of the Show

These are the core components that make all of the above possible. Each one plays a distinct role in the pipeline, and understanding what they actually do matters more than memorizing their names.

The Linear Layer — Projection

The workhorse of modern AI. A linear layer performs a matrix multiplication plus bias — it projects input from one representational space into another. It predates most other structures and was the original idea behind many neural networks, yet it's quite new in its current optimized implementation. A series of hardware advances and mathematical reformulations took it from too slow to use to one of the fastest and most responsive operations available.

This is both the most commonly used layer in almost every system and the most commonly used representation of utility hidden under every other system that isn't using it directly.

The Conv — Local Pattern Extraction

A multi-tiered pattern scanner. Convolutional layers slide learned filters across spatial data, detecting local features — edges, textures, shapes — and building them into increasingly abstract representations. This allows rigid pixel values to be approximated along larger statistical curves, producing normalized feature maps that serve as tuning forks for backpropagation.

The Transformer — Contextual Weighting

This isn't a manifestation of AI — it's a statistical anomaly solver. It produces useful and informationally valid representations of information based on a series of complex weights and systematically organized lookups in an internalized ensemble behavior of differential equations.

These are some of the most powerful utilities that exist due to the very nature of causality and big number theorems. They simply fit most problems and can account for the majority of faults autonomously. With that, they are incredibly open-ended and take a long time to converge — so they aren't my favorite, but they get the job done.

The Attention — Learned Relevance

Commonly multi-headed attention (MHA) and cross-attention. If you think of each attention head as an opinion, you might assume you need to carefully teach each head what to focus on. You don't. The majority of the time, attention heads are treated like standalone opinion-givers that weight and pull based on the necessary task, trickling down those opinions from above to below through residual connections.

Most commonly these heads are left to self-organize, then queried for attention masks that eliminate less useful information downstream based on the current task request. Each head learns to attend to different relational patterns — some track spatial coherence, others track semantic relationships, others we may never fully understand.

The Normalization — The Fixer

This is the stopgap measure that ensures the numbers fit within hardware limits. Without normalizing values — commonly capping them between 0 and 1, or centering them around zero with unit variance — the hardware simply can't support the information. One of the first things you'll see without it is crashes. Hundreds of crashes as you try to figure out why the computer is not working how it's supposed to work.

Normalization makes sure everything can be usefully learnable, have reasonable peaks within reasonable spaces, fit certain mathematical rules, and above all fit within the hardware.


Bringing It Together

Every component described above exists to serve that single thesis: controlled differential accumulation to deviate incorrectness over time. The linear layers project, the convolutions extract, the transformers contextualize, the attention heads select, and the normalizations keep it all from exploding. Together they form a pipeline that takes intentional corruption and turns it into something we recognize as creation.

The model doesn't know what a potato is. It knows what patterns of deviance, accumulated over billions of noisy observations, converge toward something we would call a potato. And that distinction — between knowing and converging — is the entire game.

0