InAWS in Plain EnglishbyRustem FeyzkhanovServerless compute for LLM — with a step-by-step guide for hosting Mistral 7B on AWS LambdaOne of the challenges with using LLM in production is finding the right way to host the models in the cloud. GPUs are expensive so hosting…Nov 16, 20237Nov 16, 20237
Intel(R) Neural CompressorHighly-efficient LLM Inference on Intel PlatformsLeadership performance yet compatible with llama.cppOct 20, 20231Oct 20, 20231
InGovernment Digital Products, SingaporebyTerence Lucas YapFrom Conventional RAG to Graph RAGWhen Large Language Models Meet Knowledge GraphsMar 16, 20249Mar 16, 20249
InArtificial Intelligence in Plain EnglishbyAI TutorMasterHugging Face Models + Llama.cpp = Faster AI🚀How to Convert Models to GGUF FormatMar 17, 2024Mar 17, 2024
InTDS ArchivebyMaxime LabonneMerge Large Language Models with mergekitCreate your own models easily, no GPU required!Jan 8, 202417Jan 8, 202417
Peter StevensRun an LLM on Apple Silicon Mac using llama.cppExplore how to configure and experiment with large language models in your local environmentDec 27, 2023Dec 27, 2023
Ingrid StevensQuantization of LLMs with llama.cppUnderstanding and Implementing n-bit Quantization Techniques for Efficient Inference in LLMsMar 15, 20249Mar 15, 20249
InTDS ArchivebyIan HoAutomated Prompt EngineeringA mixture of reflections, lit reviews and an experiment (just for fun) on Automated Prompt Engineering for Large Language ModelsMar 10, 202412Mar 10, 202412
InSuperteams.aibyAkriti UpadhyayHow to Build an Advanced AI-Powered Enterprise Content Pipeline Using Mixtral 8x7B and QdrantBy Akriti UpadhyayFeb 19, 20241Feb 19, 20241
InWhyHow.AIbyChia Jeng YangWhy Gemini 1.5 (and other large context models) are bullish for RAGOptimization via RAG: How to overcome Accuracy, Cost, Latency and other performance limitations of large context models.Feb 18, 20248Feb 18, 20248
YaduvanshiharshBuild Your Own Chatbot Assistant in 2 Minutes with Hugging FaceHugging Face has introduced Assistant, built on top of Hugging Chat. With this platform, you can create your own chatbot assistant in under…Feb 17, 2024Feb 17, 2024
InLevel Up CodingbyYoussef Hosni14 Free Large Language Models Fine-Tuning NotebooksGetting Started with LLM Fune-Tunning through These Free Colab NotebooksFeb 6, 20242Feb 6, 20242
Dr. Ernesto LeeImplementing AI in Software Testing: Creating a Text Generation Model for Test AutomationIntroductionFeb 6, 20245Feb 6, 20245
InLevel Up CodingbyFareed KhanFree GenAI APIs You Can Use in 2024Exploring the Latest Free GenAI APIsFeb 4, 20242Feb 4, 20242
InTDS ArchivebyAnthony AlcarazEnhanced Large Language Models as Reasoning EnginesThe recent exponential advances in natural language processing capabilities from large language models (LLMs) have stirred tremendous…Dec 23, 20233Dec 23, 20233
Aaweg IImproving RAG: Self Querying RetrievalKEEP IN TOUCH | THE GEN AI SERIESFeb 11, 2024Feb 11, 2024
Sean RyanExtracting website CSS styling from a screenshot via the LLaVA LLM and prompt engineeringThe LLaVa LLM is a Large Language Model that can accept images as input. This opens up many possiblities…Feb 11, 2024Feb 11, 2024
InGenerative AIbyFabio MatricardiStop! Don’t read this until you get your LLM Under ControlLearn crucial “stop words” to avoid information overload and unlock focused conversations with your Large Language Model.Feb 10, 20241Feb 10, 20241
InLevel Up CodingbyGao Dalie (高達烈)LangGraph + Gemini Pro + Custom Tool + Streamlit = Multi-Agent Application DevelopmentIn this post, you are going to learn how we can create this chatbot in LangGraph, Gemini Pro or any model you like, Custom Function and…Feb 4, 20245Feb 4, 20245