
LLMs
Inside Nano-vLLM: How Modern Inference Engines Transform Prompts into Tokens
2/2/2026

LLMs
AutoBe's Hardcore Function Calling Benchmark Pushes LLM Backend Generation to New Limits
2/2/2026

LLMs
LLMs as Emacs Elisp Assistants: Reducing the Friction of Infinite Customization
2/2/2026

LLMs
Arcee AI launches Trinity Large, a 400B-parameter open-weight model challenging Meta's Llama 4 Maverick
1/29/2026

LLMs
GNU gettext Reaches Version 1.0 After 30+ Years In Development - Adds LLM Features
1/29/2026
LLMs
The Double-Edged Sword of AI-Assisted Code Porting: Lessons from Translating TinyEMU to Go
1/28/2026

LLMs
Turns out I was wrong about TDD
1/25/2026

LLMs
From 24 to 216: A Systematic Performance Engineering Journey for LLM-based TTS
1/25/2026

LLMs
Small Language Models Narrow the Gap: Efficiency Challenges Scale in Generative AI
1/25/2026

LLMs
Baidu's ERNIE 5.0: A 2.4 Trillion Parameter MoE Model, But What Does It Actually Do?
1/24/2026

LLMs
Why I Don't Have Fun With Claude Code
1/23/2026

LLMs
The best code is no code: composing APIs and CLIs in the era of LLMs
1/23/2026

LLMs
The End of One-Size-Fits-All LLM APIs: Why Workload-Specific Inference Is Taking Over
1/22/2026