Quick and Dirty Guide: Getting started with open source AI

An image created by condensing the whole of human creative output into an image generation model

April 8, 2025

Here’s a quick and dirty guide to getting a local coding assistant up and running using the current state-of-the-art workflow for running large language models locally.

I’m doing this on a MacBook with an M3 chip and 24GB of RAM, but any M-series Mac (M1 through M4) will do. If you’re on a PC with a GPU and solid VRAM, the same steps apply — just adjusted for your setup.

Rule of thumb: Choose a model about 20% smaller in parameter size (in billions) than your available RAM/VRAM. With 24GB RAM, models in the 13B-18B parameter range typically run smoothly without straining your system.

For this example, I’ll be using the recently released DeepCoder-14B-Preview model from AgenticA — a compact, high-performance model built specifically for coding tasks. It’s a great fit for local inference, and it keeps up with the current frontier models while being lightweight enough to run on high-end consumer hardware.

Let’s get into it.

WHY RUN AI MODELS LOCALLY?

Privacy: Processing data on your local machine ensures that sensitive information remains secure and isn't transmitted to external servers.

Offline Access: Local models allow you to work without an internet connection, ensuring uninterrupted productivity.

Performance: Leveraging your device's hardware can lead to faster processing times, especially with optimized configurations.

PREREQUISITES

Before proceeding, ensure your system meets the following requirements:

Hardware:

Operating System:

METHOD 1: USING LM STUDIO

LM Studio is a user-friendly application that simplifies the process of discovering, downloading, and interacting with open-source AI models locally.

Download and Install LM Studio:

Optimize LM Studio for Apple Silicon (Mac Users):

This configuration ensures that LM Studio utilizes the Apple Metal GPU, enhancing the performance of large language models.

Download and Load a Model:

Interact with the Model:

METHOD 2: USING OLLAMA

Ollama is a command-line tool designed to facilitate the running of large language models locally.

Install Ollama:

For Mac users:

/bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)"

Install Ollama using Homebrew:

brew install ollama

For PC users:

Run a Model with Ollama:

ollama run deepseek-r1:14b

PERFORMANCE CONSIDERATIONS

Model Size and Hardware Capabilities:

Quantization:

GPU Acceleration:

GET HELP