Instruction Tuning — Topic Summaries
AI-powered summaries of 9 videos about Instruction Tuning.
9 summaries
LLM Foundations (LLM Bootcamp)
Large language models work because they turn text into numbers, then learn—via gradient-based training—to predict the next token using a Transformer...
Introducing Gemma - 2B 7B 6Trillion Tokens
Google’s new Gemma model suite brings open-weight, English text-only large language models in four sizes—2B and 7B, each available in base and...
Learn to Spell: Prompt Engineering (LLM Bootcamp)
Prompt engineering is the practical art of choosing the exact text you feed a language model so it behaves the way you need—often replacing what used...
Qwen 3 Embeddings & Rerankers
A new open suite of text embedding and reranking models from Qwen is aimed squarely at retrieval-augmented generation (RAG) use cases—especially...
Comparing LLMs with LangChain
Choosing a “good for production” large language model isn’t about picking the biggest name—it’s about matching model behavior to the task. A...
Mistral 7B - The New 7B LLaMA Killer?
Mistral AI’s newly released Mistral 7B is being positioned as a “7B LLaMA killer” because it delivers stronger benchmark performance than larger...
Investigating Alpaca 7B - Finetuned LLaMa LLM
Alpaca 7B is a newly released instruction-tuned 7-billion-parameter model built by Stanford that aims to match the quality of OpenAI’s...
Dolly 2.0: Free ChatGPT-like Model for Commercial Use
Dolly 2.0 is being released as a genuinely commercial-friendly, open instruction-tuned language model—complete with training code, dataset, and model...
Mixtral - Mixture of Experts (MoE) Free LLM that Rivals ChatGPT (3.5) by Mistral | Overview & Demo
Mistral AI’s Mixtral 8×7B (an open-weight sparse Mixture of Experts model) is positioned as a practical alternative to much larger LLMs by routing...