<?xml version="1.0" encoding="UTF-8"?>
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9">
<url>
<loc>https://vllm.ai</loc>
<lastmod>2026-05-14T14:55:38.614Z</lastmod>
<changefreq>weekly</changefreq>
<priority>1</priority>
</url>
<url>
<loc>https://vllm.ai/events</loc>
<lastmod>2026-05-14T14:55:38.614Z</lastmod>
<changefreq>daily</changefreq>
<priority>0.8</priority>
</url>
<url>
<loc>https://vllm.ai/contact</loc>
<lastmod>2026-05-14T14:55:38.614Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/releases</loc>
<lastmod>2026-05-14T14:55:38.614Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>daily</changefreq>
<priority>0.9</priority>
</url>
<url>
<loc>https://vllm.ai/blog/archive</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>daily</changefreq>
<priority>0.8</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.6</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/performance</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/model-support</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/ecosystem</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/hardware</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/large-scale-serving</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/multimodal</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/quantization</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/speculative-decoding</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/community</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/developer</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/kv_cache</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/disaggregation</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/turboquant</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/benchmarking</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/kernel-fusion</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/agentic</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/fp8</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/mamba</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/engineering</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/triton</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/attention</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/tags/frontend</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>weekly</changefreq>
<priority>0.5</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-05-11-turboquant</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-05-11-vllm-tops-artificial-analysis</loc>
<lastmod>2026-05-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-05-06-mooncake-store</loc>
<lastmod>2026-05-06T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-28-nemotron-omni</loc>
<lastmod>2026-04-28T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-24-deepseek-v4</loc>
<lastmod>2026-04-24T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-22-fp8-kvcache</loc>
<lastmod>2026-04-22T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-21-hybrid-ssm-disagg</loc>
<lastmod>2026-04-21T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-14-vllm-korea-meetup-2026</loc>
<lastmod>2026-04-14T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-07-moriio-kv-connector</loc>
<lastmod>2026-04-07T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-04-02-gemma4</loc>
<lastmod>2026-04-02T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-03-30-extract-hidden-states</loc>
<lastmod>2026-03-30T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-03-24-mrv2</loc>
<lastmod>2026-03-24T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-03-13-p-eagle</loc>
<lastmod>2026-03-13T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-03-11-nemotron-3-super</loc>
<lastmod>2026-03-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-03-10-v0.2-vllm-sr-athena-release</loc>
<lastmod>2026-03-10T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-03-04-vllm-triton-backend-deep-dive</loc>
<lastmod>2026-03-04T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-02-27-rocm-attention-backend</loc>
<lastmod>2026-02-27T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-02-26-multi-lora</loc>
<lastmod>2026-02-26T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-02-13-gb300-deepseek</loc>
<lastmod>2026-02-13T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-02-03-dsr1-gb200-part1</loc>
<lastmod>2026-02-03T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-02-01-gpt-oss-optimizations</loc>
<lastmod>2026-02-01T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-01-31-streaming-realtime</loc>
<lastmod>2026-01-31T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-01-23-mom-on-amd-gpu</loc>
<lastmod>2026-01-23T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-01-08-kv-offloading-connector</loc>
<lastmod>2026-01-08T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-01-05-vllm-sr-iris</loc>
<lastmod>2026-01-05T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2026-01-02-introducing-vllm-playground</loc>
<lastmod>2026-01-02T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-27-vllm-ai-website</loc>
<lastmod>2025-12-27T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-19-vllm-omni-diffusion-cache-acceleration</loc>
<lastmod>2025-12-19T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-17-large-scale-serving</loc>
<lastmod>2025-12-17T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-16-vllm-sr-amd</loc>
<lastmod>2025-12-16T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-15-run-nvidia-nemotron-3-nano</loc>
<lastmod>2025-12-15T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-15-vllm-epd</loc>
<lastmod>2025-12-15T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-14-halugate</loc>
<lastmod>2025-12-14T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-13-speculators-v030</loc>
<lastmod>2025-12-13T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-13-vllm-router-release</loc>
<lastmod>2025-12-13T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-09-intel-autoround-llmc</loc>
<lastmod>2025-12-09T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-12-03-improved-cuda-debugging</loc>
<lastmod>2025-12-03T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-30-vllm-omni</loc>
<lastmod>2025-11-30T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-22-ray-symmetric-run</loc>
<lastmod>2025-11-22T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-20-vllm-plugin-system</loc>
<lastmod>2025-11-20T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-19-docker-model-runner-vllm</loc>
<lastmod>2025-11-19T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-19-signal-decision</loc>
<lastmod>2025-11-19T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-13-shm-ipc-cache</loc>
<lastmod>2025-11-13T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-11-intel-arc-pro-b</loc>
<lastmod>2025-11-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-11-10-bitwise-consistent-train-inference</loc>
<lastmod>2025-11-10T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-31-run-multimodal-reasoning-agents-nvidia-nemotron</loc>
<lastmod>2025-10-31T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-28-kimi-k2-accuracy</loc>
<lastmod>2025-10-28T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-27-semantic-router-modular</loc>
<lastmod>2025-10-27T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-26-sleep-mode</loc>
<lastmod>2025-10-26T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-23-now_serving_nvidia_nemotron_with_vllm</loc>
<lastmod>2025-10-23T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-22-agent-lightning</loc>
<lastmod>2025-10-22T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-16-vllm-tpu</loc>
<lastmod>2025-10-16T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-10-09-blackwell-inferencemax</loc>
<lastmod>2025-10-09T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-09-29-deepseek-v3-2</loc>
<lastmod>2025-09-29T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-09-16-vllm-meetup</loc>
<lastmod>2025-09-16T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-09-11-qwen3-next</loc>
<lastmod>2025-09-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-09-11-semantic-router</loc>
<lastmod>2025-09-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-09-05-anatomy-of-vllm</loc>
<lastmod>2025-09-05T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-09-05-beyond-text-generation</loc>
<lastmod>2025-09-05T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-08-20-torch-compile</loc>
<lastmod>2025-08-20T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-08-19-glm45-vllm</loc>
<lastmod>2025-08-19T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-08-11-cuda-debugging</loc>
<lastmod>2025-08-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-08-05-gpt-oss</loc>
<lastmod>2025-08-05T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-06-30-minimax-m1</loc>
<lastmod>2025-06-30T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-05-12-hardware-plugin</loc>
<lastmod>2025-05-12T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-04-23-openrlhf-vllm</loc>
<lastmod>2025-04-23T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-04-11-transformers-backend</loc>
<lastmod>2025-04-11T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-04-05-llama4</loc>
<lastmod>2025-04-05T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-02-24-ptpc-fp8-rocm</loc>
<lastmod>2025-02-24T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-02-21-aibrix-release</loc>
<lastmod>2025-02-21T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-02-17-distributed-inference</loc>
<lastmod>2025-02-17T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-01-27-intro-to-llama-stack-with-vllm</loc>
<lastmod>2025-01-27T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-01-27-v1-alpha-release</loc>
<lastmod>2025-01-27T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-01-21-stack-release</loc>
<lastmod>2025-01-21T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-01-14-struct-decode-intro</loc>
<lastmod>2025-01-14T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-01-10-dev-experience</loc>
<lastmod>2025-01-10T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2025-01-10-vllm-2024-wrapped-2025-vision</loc>
<lastmod>2025-01-10T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2024-10-23-vllm-serving-amd</loc>
<lastmod>2024-10-23T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2024-10-17-spec-decode</loc>
<lastmod>2024-10-17T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2024-09-05-perf-update</loc>
<lastmod>2024-09-05T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2024-07-25-lfai-perf</loc>
<lastmod>2024-07-25T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2024-07-23-llama31</loc>
<lastmod>2024-07-23T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2023-11-14-notes-vllm-vs-deepspeed</loc>
<lastmod>2023-11-14T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
<url>
<loc>https://vllm.ai/blog/2023-06-20-vllm</loc>
<lastmod>2023-06-20T00:00:00.000Z</lastmod>
<changefreq>monthly</changefreq>
<priority>0.7</priority>
</url>
</urlset>
