If you’ve ever wanted to run an LLM directly on your laptop without relying on the cloud, this new video is for you. I just released a hands‑on walkthrough of Foundry Local, where I show you exactly how to download an AI model to your machine and use it completely offline. In the video, I break down both methods developers use: - CLI workflow — install Foundry Local, pull a model, and run inference offline - Python SDK workflow — load the model in your code and build real offline AI features Whether you're a developer, an AI enthusiast, or someone who wants more privacy and zero token costs, this tutorial will help you get started in minutes.
This blog is all about my technical learnings pertaining to LLM, OpenAI, Azure OpenAI, C#, Azure, Python, AI, ML, Visual Studio Code and many more.