InAWS in Plain EnglishbyRustem FeyzkhanovServerless compute for LLM — with a step-by-step guide for hosting Mistral 7B on AWS LambdaOne of the challenges with using LLM in production is finding the right way to host the models in the cloud. GPUs are expensive so hosting…Nov 16, 20237Nov 16, 20237
InAI MindbyJohn AdeojoDeploy Mixtral Quickly & Cheaply!Deploying Mixtral with Runpod and vLLM for LangChain, Haystack, LlamaIndex, and AutoGenJan 30, 2024Jan 30, 2024
InAWS in Plain EnglishbyRustem FeyzkhanovGuide for running Llama 2 using LLAMA.CPP on AWS FargateStep-by-step guide for deploying Llama 2 model to AWS using LLAMA.CPP as framework, Fargate for hardware and Copilot for deployment.Oct 17, 20232Oct 17, 20232