vllm 4 Getting Started with running LLM models locally Nov 16, 2025 Speeding up local LLM inference 2x with Speculative Decoding Oct 26, 2025 Open Weights, Borrowed GPUs Oct 21, 2025 Harnessing GPT-OSS Built-in Tools Oct 11, 2025