{"id":"4063394e-97ee-424c-9747-d19f2ce6e595","timestamp":1775407040548,"matcher":{"static":{},"wildcard":{},"dynamic":{}},"prerendered":["/","/cookies-policy","/blog/autodebug-telemetry-driven-inference-optimization-loop","/terms-of-service","/privacy-policy","/contact-us","/blog/traditional-observability-is-blind-to-inference","/blog/ai-debugging-and-optimization-for-production-inference","/pricing","/blog","/docs","/blog/vllm-production-observability-from-model-to-hardware","/blog/llm-api-latency-optimization-explained","/docs/guides/quick-start","/docs/guides/ai-debugging","/docs/guides/manual-tracing","/docs/security","/docs/concepts","/docs/guides/manual-profiling","/docs/guides/using-tags","/docs/integrations/fastapi","/docs/reference/python-api","/docs/integrations/nvidia","/docs/integrations/vllm","/docs/integrations/pytorch","/docs/integrations/dstack","/docs/integrations/sglang","/docs/integrations/hugging-face"]}