Local LLMs — Topic Summaries
AI-powered summaries of 11 videos about Local LLMs.
11 summaries
Run your own AI (but private)
Local “private AI” is becoming practical: a person can run an LLM entirely on a laptop or workstation, keep data off third-party servers, and then...
You've been using AI Wrong
Fabric is an open-source CLI framework built to reduce the friction of using AI by turning raw text (like YouTube transcripts or API data) into...
Ollama - Local Models on your machine
Ollama is a user-friendly way to run large language models locally on a Mac or Linux machine by downloading them and serving them through a local...
LangChain + Retrieval Local LLMs for Retrieval QA - No OpenAI!!!
Getting rid of OpenAI entirely for Retrieval QA with LangChain is feasible, but the quality hinges on the local LLM’s context limits, prompt format...
Smart Second Brain for Obsidian(Free & Offline)
A privacy-first “second brain” for Obsidian can run entirely offline by pairing the Obsidian plugin Smart Second Brain with locally hosted large...
Run LLMs Locally With Docker Model Runner
Running open-source LLMs locally is now straightforward with Docker Model Runner, as long as Docker Desktop is updated and a few settings are...
Microsoft's Phi 3.5 - The latest SLMs
Microsoft has expanded its Phi 3 lineup with three new Phi 3.5 models—two instruction-tuned language models and an updated vision model—pushing...
Open Source LLMs on GOD mode. Local LLMs MAXXED OUT on the RTX 5090!
Running large language models entirely on a home PC is no longer a novelty—it’s practical, fast, and surprisingly capable when paired with a...
Build a Local AI App in 10 min with Docker (Zero Cloud Fees)
Local AI apps can be built without paying per-request inference fees by running large language models entirely on a developer’s own machine—using...
Ollama.ai: A Developer's Quick Start Guide!
Local, on-device LLMs are moving from “cloud-only” APIs to a developer-friendly workflow where models download to a machine, run locally, and can be...
Run any LLMs locally: Ollama | LM Studio | GPT4All | WebUI | HuggingFace Transformers
Running large language models locally boils down to one trade-off: keeping data on-device and gaining control over models and prompts, while paying...