{"id":"8bb6e894-f99d-4e02-9b07-8273a3e91ebd","timestamp":1777196044098,"matcher":{"static":{},"wildcard":{},"dynamic":{}},"prerendered":["/","/contact-us","/blog/autodebug-telemetry-driven-inference-optimization-loop","/pricing","/cookies-policy","/blog/traditional-observability-is-blind-to-inference","/privacy-policy","/blog/ai-debugging-and-optimization-for-production-inference","/terms-of-service","/blog","/docs","/blog/vllm-production-observability-from-model-to-hardware","/blog/llm-api-latency-optimization-explained","/docs/guides/quick-start","/docs/guides/ai-optimization","/docs/concepts","/docs/security","/docs/guides/using-tags","/docs/integrations/hugging-face","/docs/integrations/pytorch","/docs/integrations/nvidia","/docs/guides/manual-profiling","/docs/integrations/dstack","/docs/guides/manual-tracing","/docs/integrations/vllm","/docs/integrations/sglang","/docs/integrations/fastapi","/docs/reference/rest-api","/docs/reference/python-api","/docs/reference/context-cli"]}