Get AI summaries of any video or article — Sign up free

Instruction Tuning — Topic Summaries

AI-powered summaries of 9 videos about Instruction Tuning.

9 summaries

No matches found.

LLM Foundations (LLM Bootcamp)

The Full Stack · 3 min read

Large language models work because they turn text into numbers, then learn—via gradient-based training—to predict the next token using a Transformer...

Transformer FoundationsAttention MechanismTokenization

Introducing Gemma - 2B 7B 6Trillion Tokens

Sam Witteveen · 2 min read

Google’s new Gemma model suite brings open-weight, English text-only large language models in four sizes—2B and 7B, each available in base and...

Gemma ModelsOpen-Weight LLMsTraining Tokens

Learn to Spell: Prompt Engineering (LLM Bootcamp)

The Full Stack · 3 min read

Prompt engineering is the practical art of choosing the exact text you feed a language model so it behaves the way you need—often replacing what used...

Prompt EngineeringConditioningInstruction Tuning

Qwen 3 Embeddings & Rerankers

Sam Witteveen · 2 min read

A new open suite of text embedding and reranking models from Qwen is aimed squarely at retrieval-augmented generation (RAG) use cases—especially...

Text EmbeddingsRerankingRAG

Comparing LLMs with LangChain

Sam Witteveen · 3 min read

Choosing a “good for production” large language model isn’t about picking the biggest name—it’s about matching model behavior to the task. A...

Model EvaluationLangChainInstruction Tuning

Mistral 7B - The New 7B LLaMA Killer?

Sam Witteveen · 3 min read

Mistral AI’s newly released Mistral 7B is being positioned as a “7B LLaMA killer” because it delivers stronger benchmark performance than larger...

Mistral 7BLLaMA BenchmarksInstruction Tuning

Investigating Alpaca 7B - Finetuned LLaMa LLM

Sam Witteveen · 2 min read

Alpaca 7B is a newly released instruction-tuned 7-billion-parameter model built by Stanford that aims to match the quality of OpenAI’s...

Instruction TuningLLaMA Fine-TuningModel Evaluation

Dolly 2.0: Free ChatGPT-like Model for Commercial Use

Venelin Valkov · 2 min read

Dolly 2.0 is being released as a genuinely commercial-friendly, open instruction-tuned language model—complete with training code, dataset, and model...

Dolly 2.0Instruction TuningDolly 15K

Mixtral - Mixture of Experts (MoE) Free LLM that Rivals ChatGPT (3.5) by Mistral | Overview & Demo

Venelin Valkov · 2 min read

Mistral AI’s Mixtral 8×7B (an open-weight sparse Mixture of Experts model) is positioned as a practical alternative to much larger LLMs by routing...

Mixture of ExpertsSparse RoutingInstruction Tuning