Ertuğrul DemirReasoning Models at HomeHow to Train Your Own Specialized Reasoning Model at HomeFeb 81Feb 81
InTDS ArchivebyBenjamin MarieMulti-GPU Fine-tuning for Llama 3.1 70B with FSDP and QLoRAWhat you can do with only 2x24 GB GPUs and a lot of CPU RAMAug 8, 20242Aug 8, 20242
Abonia SojasingarayarRunning Ollama in Google Colab (Free Tier)A Step-by-Step TutorialSep 16, 20246Sep 16, 20246
Cobus GreylingEmergence of Large Action Models (LAMs) and Their Impact on AI AgentsWhile LLMs are great for understanding and producing unstructured content, LAMs are designed to bridge the gap by turning language into…Sep 13, 20242Sep 13, 20242
InTowards AIbyKaran Kaul | カランFine-Tune Llama3.1 8b in 2 Mins with only 12 training samples!Teach Llama3.1 your output format by creating a small dataset, then fine-tuning using Unsloth & Google Colab.Sep 8, 202413Sep 8, 202413
InTDS ArchivebyShaw TalebiCompressing Large Language Models (LLMs)Make LLMs 10X smaller without sacrificing performanceAug 30, 20245Aug 30, 20245
InTDS ArchivebyMarco PeixeiroTime-LLM: Reprogram an LLM for Time Series ForecastingDiscover the architecture of Time-LLM and apply it in a forecasting project with PythonMar 5, 20243Mar 5, 20243
Benjamin MarieRAG and Long-context LLMs, Which One to Choose?SummHay: A New BenchmarkJul 10, 2024Jul 10, 2024
InData Science at MicrosoftbyJane HuangEvaluating LLM systems: Metrics, challenges, and best practicesA detailed consideration of approaches to evaluation and selectionMar 5, 202421Mar 5, 202421
InTDS ArchivebyMarius StegerFine-Tune the Audio Spectrogram Transformer With 🤗 TransformersLearn how to fine-tune the Audio Spectrogram Transformer model for audio classification of your own data using the Hugging Face Ecosystem.Aug 21, 20242Aug 21, 20242
Mastering LLM (Large Language Model)How Much GPU Memory is Needed to Serve a Large Language Model (LLM)?In nearly all LLM interviews, there’s one question that consistently comes up: “How much GPU memory is needed to serve a Large Language…Aug 17, 202420Aug 17, 202420
InTDS ArchivebyMengliu ZhaoA Practical Guide to Contrastive LearningHow to build your very first SimSiam model with FashionMNISTJul 30, 2024Jul 30, 2024
InMad Chatter Tea PartybyNicola ProcopioMatryoshka Embedding Model for Ordinary PeopleA few months ago, after years of inaction, Open AI released its new embedding models text-embedding-3-*. Among the new features of these…May 9, 2024May 9, 2024
InTDS ArchivebyStefan KrawczykHow to Build a Streaming Agent with Burr, FastAPI, and ReactAn overview of how to leverage streaming using open source tools applied to building a simple agentic chat botJul 25, 2024Jul 25, 2024
InTDS ArchivebyMaxime LabonneFine-Tune Llama 3.1 Ultra-Efficiently with UnslothA beginner’s guide to state-of-the-art supervised fine-tuningJul 29, 20245Jul 29, 20245
Thiyagarajan PalaniyappanComprehensive Evaluation Metrics for Retrieval-Augmented Generation (RAG)Basic Performance MetricsJul 9, 2024Jul 9, 2024
InTDS ArchivebyMaarten GrootendorstA Visual Guide to QuantizationDemystifying the compression of large language modelsJul 24, 20243Jul 24, 20243
InAI AdvancesbyAnkush k SingalLangGraph and MapReduce: A Dynamic Duo for Parallel ProcessingAnkush k SingalJul 23, 2024Jul 23, 2024
Venkat Ram RaoTraining a Mini(114M Parameter) Llama 3 like Model from Scratchaka “How to Train you LLM”Jul 21, 20242Jul 21, 20242