
Hardware
Microsoft and NVIDIA Achieve Breakthrough DeepSeek-V3.2 Inference Performance with Blackwell Platform
2/28/2026

Hardware
SambaNova Unveils SN50 AI Accelerator, Partners with Intel for Xeon-Based Inference Infrastructure
2/25/2026

AI
AWS Brings Custom Nova Models to SageMaker Inference
2/16/2026

LLMs
Two Different Tricks for Fast LLM Inference: Anthropic vs OpenAI's Approaches
2/14/2026

Chips
Positron's Asimov AI Accelerator Takes on Nvidia with Laptop RAM and CXL Expansion
2/4/2026

LLMs
Inside Nano-vLLM: How Modern Inference Engines Transform Prompts into Tokens
2/2/2026
Hardware
OpenAI to serve ChatGPT on Cerebras' AI dinner plates • The Register
1/15/2026

Hardware
OpenAI's Strategic Compute Diversification: A 750MW Partnership with Cerebras for Low-Latency Inference
1/15/2026