From d39c04f6f2d554bcc1c99b5974cd636f23adf9cb Mon Sep 17 00:00:00 2001 From: simon-mo Date: Wed, 7 Aug 2024 13:57:47 -0700 Subject: [PATCH] Add snowflake to llama3.1 post --- _posts/2024-07-23-llama31.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/_posts/2024-07-23-llama31.md b/_posts/2024-07-23-llama31.md index 378c102..e381fca 100644 --- a/_posts/2024-07-23-llama31.md +++ b/_posts/2024-07-23-llama31.md @@ -80,4 +80,4 @@ To learn more about distributed inference using vLLM please refer to [this doc]( ### Acknowledgements -We would like to thank Meta for the pre-release partnership and letting us test the model. Independently from the release, we thank the following vLLM contributors for the features mentioned in this blogpost: [Neural Magic](https://neuralmagic.com/) for FP8 quantization; [CentML](https://centml.ai/) for pipeline parallelism; [Anyscale](https://www.anyscale.com/) for the chunked prefill feature. The evaluation runs on [Lambda’s 1-Click Clusters](https://lambdalabs.com/service/gpu-cloud/1-click-clusters) with InfiniBand, and we thank Lambda for the resource and the smooth cluster setup experience. +We would like to thank Meta for the pre-release partnership and letting us test the model. Independently from the release, we thank the following vLLM contributors for the features mentioned in this blogpost: [Neural Magic](https://neuralmagic.com/) for FP8 quantization; [CentML](https://centml.ai/) and Snowflake for pipeline parallelism; [Anyscale](https://www.anyscale.com/) for the chunked prefill feature. The evaluation runs on [Lambda’s 1-Click Clusters](https://lambdalabs.com/service/gpu-cloud/1-click-clusters) with InfiniBand, and we thank Lambda for the resource and the smooth cluster setup experience.