0.1.11 GPU β the artist and the parallel calculator
In one line: the GPU started as a chip for drawing triangles fast β and accidentally became the engine of the AI revolution.
A story β from Quake to ChatGPT
In 1993, three engineers in a Dennyβs restaurant in California sketched a chip that would do 3D graphics so fast that PC games could finally look like arcade games. They called the company NVIDIA. The name comes from βinvidiaβ β Latin for envy.
For a decade, GPUs (Graphics Processing Units) had one job: take a list of triangles, calculate where each pixel of each triangle should be drawn, and shade them. They were narrow but brutally good at it β hundreds of tiny processors all doing the same simple math in parallel.
In 2006 NVIDIA released CUDA β a way to write general-purpose code that ran on the GPU. Researchers started realising: βwait, neural networks are mostly just multiplying huge matrices, and thatβs the exact same shape as drawing triangles.β A 2012 paper called AlexNet trained a neural network on two GeForce GTX 580s β and obliterated every previous image-recognition record. The deep-learning era began.
By 2024, NVIDIA was the most valuable company in the world. Not because of games. Because every AI lab on Earth was queueing for H100 chips at $30,000 each. The GPU went from βthing that draws Doomβ to βthing that runs civilisationβ in 30 years.
Whatβs actually going on
A CPU has a few powerful cores (4-32) optimised for latency β finishing one complex task as fast as possible. A GPU has thousands of simple cores optimised for throughput β doing the same simple operation on millions of pieces of data simultaneously.
| CPU | GPU | |
|---|---|---|
| Cores | 4-64 | 1,000-20,000 |
| Per-core speed | Very fast | Slow |
| Best at | Branchy, sequential code | Parallel arithmetic on huge data |
| Memory | Shares RAM | Has its own dedicated VRAM (8-80 GB) |
Modern GPUs sit in a PCIe slot with their own fans. They have their own VRAM (Video RAM, very fast β typically GDDR6 or HBM3) and their own driver. Programs explicitly upload data to the GPU, ask the GPU to compute, and download the result.
If youβre not doing graphics, scientific simulation, crypto mining, or AI β your GPU is mostly idle. But for those workloads itβs 10-100Γ faster than a CPU.
Why a hacker cares
- GPU password cracking β a tool called Hashcat uses GPUs to try billions of password guesses per second. An 8-character password? Cracked in hours, not years. This is why everything modern uses slow key-derivation functions (bcrypt, scrypt, Argon2) β to make GPU brute-force impractical.
- Crypto-mining malware β quietly hijack the GPU for currency mining. Slow GPU, hot card, big power bill.
- AI/ML attack surface β the entire AI security niche (which weβre aiming for) is built on GPU-trained models. Attacks on ML models include adversarial examples, model extraction, training data poisoning, prompt injection β all happen at the GPU/model layer.
- Side channels β power and timing side-channels on GPUs are a research area. GPU-resident malware that hides from CPU-based detection has been demonstrated in academic papers.
- VRAM forensics β GPU memory often holds decrypted keys, frame buffers, and model weights long after the user thinks theyβre gone.
In one sketch
CPU GPU
ββββββ ββββββββββββββββββββββββ
β 4 β big cores β ββββββββββββββββββββ β
β to β good at one thing β β thousands of tiny β β
β 64 β at a time β β cores all doing β β
βcoresβ β β the same simple β β
ββββββ β β thing in parallel β β
β ββββββββββββββββββββ β
ββββββββββββββββββββββββ
sequential, low-latency parallel, high-throughput
Reference and image credit
- Nvidia_Geforce_6600GT_GPU_2009-01-27.jpg β Henry Muehlpfordt, CC BY-SA 3.0.
Memory peg
CPU is one chef cooking ten different dishes. GPU is a thousand line cooks all making the same omelette. AI is a lot of omelettes.