Get AI summaries of any video or article — Sign up free

Local LLMs — Topic Summaries

AI-powered summaries of 11 videos about Local LLMs.

11 summaries

No matches found.

Run your own AI (but private)

NetworkChuck · 3 min read

Local “private AI” is becoming practical: a person can run an LLM entirely on a laptop or workstation, keep data off third-party servers, and then...

Local LLMsOllama SetupRAG and Vector Databases

You've been using AI Wrong

NetworkChuck · 3 min read

Fabric is an open-source CLI framework built to reduce the friction of using AI by turning raw text (like YouTube transcripts or API data) into...

Fabric PatternsExtract WisdomCLI Workflows

Ollama - Local Models on your machine

Sam Witteveen · 2 min read

Ollama is a user-friendly way to run large language models locally on a Mac or Linux machine by downloading them and serving them through a local...

Local LLMsOllama SetupModel Downloads

LangChain + Retrieval Local LLMs for Retrieval QA - No OpenAI!!!

Sam Witteveen · 2 min read

Getting rid of OpenAI entirely for Retrieval QA with LangChain is feasible, but the quality hinges on the local LLM’s context limits, prompt format...

Retrieval QALangChainLocal LLMs

Smart Second Brain for Obsidian(Free & Offline)

Prakash Joshi Pax · 3 min read

A privacy-first “second brain” for Obsidian can run entirely offline by pairing the Obsidian plugin Smart Second Brain with locally hosted large...

Obsidian PluginOffline AILocal LLMs

Run LLMs Locally With Docker Model Runner

Krish Naik · 2 min read

Running open-source LLMs locally is now straightforward with Docker Model Runner, as long as Docker Desktop is updated and a few settings are...

Docker Model RunnerLocal LLMsDocker Desktop Setup

Microsoft's Phi 3.5 - The latest SLMs

Sam Witteveen · 2 min read

Microsoft has expanded its Phi 3 lineup with three new Phi 3.5 models—two instruction-tuned language models and an updated vision model—pushing...

Phi 3.5 ModelsLocal LLMsMixture of Experts

Open Source LLMs on GOD mode. Local LLMs MAXXED OUT on the RTX 5090!

MattVidPro · 2 min read

Running large language models entirely on a home PC is no longer a novelty—it’s practical, fast, and surprisingly capable when paired with a...

Local LLMsLM StudioDeepSeek R1

Build a Local AI App in 10 min with Docker (Zero Cloud Fees)

MattVidPro · 3 min read

Local AI apps can be built without paying per-request inference fees by running large language models entirely on a developer’s own machine—using...

Docker DesktopLocal LLMsQuantized Models

Ollama.ai: A Developer's Quick Start Guide!

AI Arcade · 3 min read

Local, on-device LLMs are moving from “cloud-only” APIs to a developer-friendly workflow where models download to a machine, run locally, and can be...

Local LLMsOllama SetupModel Variants

Run any LLMs locally: Ollama | LM Studio | GPT4All | WebUI | HuggingFace Transformers

AI Researcher · 3 min read

Running large language models locally boils down to one trade-off: keeping data on-device and gaining control over models and prompts, while paying...

Local LLMsGPU InferenceQuantization