A hobbyist converted a Tesla V100 SMX GPU into a PCIe card for $200, achieving higher token‑per‑second throughput and better efficiency than a RTX 3060 and Radeon RX 7800 XT in LLM inference, highlighting the lingering value of Turing‑era HBM2 silicon in server‑grade AI workloads.
$200 Socketed Nvidia V100 Revived as a PCIe Card Beats Modern Mid‑Range GPUs in AI Inference
{{IMAGE:2}} Image credit: Nvidia
Announcement
A YouTube creator from the channel Hardware Haven has demonstrated that a 2017‑era Nvidia Tesla V100, originally sold for data‑center racks, can be repurposed into a consumer‑grade PCIe card for roughly $200. The modified board runs large language models (LLMs) faster than a contemporary RTX 3060 and even outperforms a Radeon RX 7800 XT in token‑generation speed, while delivering a modest edge in energy efficiency.
Technical specifications
- GPU core: Tesla V100, Turing architecture, SMX2 socket
- Memory: 16 GB HBM2, 900 GB/s bandwidth (32 GB version also exists)
- Base power envelope: 300 W (limited to 100 W in tests)
- Adapter: Custom SMX‑to‑PCIe x16 board, two 8‑pin PCIe power connectors, three 4‑pin PWM headers, no stock cooler
- Cooling solution: 3‑D‑printed duct with an 80 mm Notcua fan, directing airflow onto the GPU’s exposed heatsink
- Cost: ~US $100 for the V100 (eBay), ~US $100 for the adapter, total ≈ US $200
Benchmarks
| Test | GPU | Tokens / s | Power (W) | Tokens / W |
|---|---|---|---|---|
| Ollama – gpt‑oss‑20b | V100 | 130 | 293 | 0.44 |
| Ollama – gpt‑oss‑20b | RX 7800 XT | 90 | 300* | 0.30 |
| Google gemma‑4b‑e4b | V100 (300 W) | 108 | 293 | 0.37 |
| Google gemma‑4b‑e4b | RTX 3060 12 GB (full) | 76 | 235 | 0.32 |
| Same test, V100 limited to 100 W | V100 | 95 | 170 | 0.56 |
| Same test, RTX 3060 limited to 100 W | RTX 3060 | 68 | 171 | 0.40 |
*Power draw for the RX 7800 XT is an estimate based on typical board TDP.
Observations
- Throughput advantage – The V100’s HBM2 memory and 900 GB/s bus deliver roughly 30 % more tokens per second than the RX 7800 XT, despite both cards having 16 GB of VRAM.
- Energy efficiency – When throttled to 100 W, the V100 reaches 0.55 tokens / s / W, surpassing the RTX 3060’s 0.39 tokens / s / W under the same power ceiling.
- Idle consumption – The V100 idles at ~45 W, about 10 W higher than the RTX 3060’s 35 W, a factor to consider for always‑on servers.
- Thermal handling – The DIY duct and fan keep the GPU within safe limits (≈ 80 °C under load) without any proprietary cooling hardware.
Market implications
- Extended silicon life‑cycle – The experiment proves that server‑grade GPUs with HBM2 can remain competitive for inference workloads years after their nominal end‑of‑life. Enterprises looking to stretch AI budgets may source surplus V100s, retrofit them with inexpensive adapters, and achieve cost‑per‑token figures comparable to newer consumer silicon.
- Supply‑chain pressure – As the video gains traction, demand for used V100 units and SMX‑to‑PCIe adapters is likely to rise. Prices for the 16 GB variant could climb from $100 to $200‑$250, while the 32 GB version, currently around $500, may see a similar upward trend.
- Design considerations for OEMs – The success of a 3‑D‑printed airflow solution suggests that future low‑cost AI accelerators could adopt modular, user‑serviceable cooling rather than proprietary heat‑pipe assemblies, reducing BOM costs and simplifying aftermarket upgrades.
- Software ecosystem – Nvidia’s continued driver support for the V100 (CUDA 11.x, cuDNN 8) means that developers can run popular inference stacks (Ollama, TensorRT, PyTorch) without major compatibility hurdles, reinforcing the platform’s attractiveness despite its age.
What this means for small‑scale AI deployments
For hobbyists, startups, or edge‑compute sites that cannot justify a multi‑thousand‑dollar GPU, a refurbished V100 offers a sweet spot: high memory bandwidth, acceptable power draw, and software maturity at a fraction of the price of a brand‑new Ampere or Ada‑generation card. The trade‑off is higher idle power and the need for a custom cooling solution, but the performance‑per‑dollar ratio is hard to beat.
*For a visual walkthrough of the adapter build, watch the original YouTube video titled “This Ridiculous $200 AI GPU Shouldn’t Be This Good”.*
Comments
Please log in or register to join the discussion