diff --git a/vllm/README.md b/vllm/README.md index 420aa6cfcf32e..b69854afb0e9a 100644 --- a/vllm/README.md +++ b/vllm/README.md @@ -48,6 +48,10 @@ In either case, make sure that the `source` value for your logs is `vllm`. This Need help? Contact [Datadog support][9]. +## Further Reading +Additional helpful documentation, links, and articles: +- [Optimize LLM application performance with Datadog's vLLM integration][13] + [1]: https://docs.vllm.ai/en/stable/ [2]: https://app.datadoghq.com/account/settings/agent/latest @@ -60,3 +64,4 @@ Need help? Contact [Datadog support][9]. [10]: https://docs.datadoghq.com/containers/docker/log/?tab=containerinstallation#installation [11]: https://docs.datadoghq.com/containers/docker/log/?tab=hostagent#installation [12]: https://docs.datadoghq.com/containers/docker/log/?tab=dockerfile#log-integrations +[13]: https://www.datadoghq.com/blog/vllm-integration/ \ No newline at end of file diff --git a/vllm/manifest.json b/vllm/manifest.json index c81c4669a8f8b..b5d9c61852e7a 100644 --- a/vllm/manifest.json +++ b/vllm/manifest.json @@ -19,6 +19,12 @@ "Category::AI/ML", "Submitted Data Type::Metrics", "Offering::Integration" + ], + "resources": [ + { + "resource_type": "blog", + "url": "https://www.datadoghq.com/blog/vllm-integration/" + } ] }, "assets": {