Loading video player...
Want to run your own local LLM with LangChain using Ollama? In this tutorial, I show you how to set up and use Qwen2.5-Coder 7B locally for coding tasks — completely offline and private. Code link: https://github.com/Thought-Express/generative-ai If you’re a developer who wants to build AI agents without relying on OpenAI APIs, this step-by-step guide will help you integrate Ollama + LangChain + Qwen2.5-Coder into your Python workflow. We’ll cover everything from installing Ollama to running the qwen2.5-coder:7b model and connecting it to a LangChain agent. This setup works great on local machines (like M-series Macs or mid-range PCs) and gives you full control over your AI stack. What You’ll Learn How to install and configure Ollama How to download and run Qwen2.5-Coder 7B How to integrate Ollama with LangChain How to build a simple coding agent How to run a fully local AI development workflow Timestamps 00:00 – Introduction and revcap 01:05 – Installing Ollama 05:10 – Connecting LangChain to Ollama We at Thought Express are a team of enthusiastic educators who believe that learning is most fun when it’s most accessible and interesting 😁, and who likes boring lectures anyway. If you want to collaborate or have a coffee meet, email us at thoughtexe47@gmail.com #LocalLLM #Ollama #langchain #aiprogramming #genai