Free, Unlimited, & Offline Al: Getting Started with Local Al

In this guide, we'll show you exactly how to get started with local AI using beginner-friendly apps like LM Studio and DiffusionBee. You'll learn the difference between cloud-based and local AI, the pros and cons of running models on your own machine, and how to set up your first local AI tools in minutes.

Written by
Matt Jasinski
and
Tom Nassr

November 19, 2025

Loading the Elevenlabs Text to Speech AudioNative Player...

Are you tired of AI usage limits? Are you concerned about your privacy when sending data to cloud-based tools like ChatGPT? Or perhaps you need to use AI while traveling, even without an internet connection? 

If any of these concerns sound familiar, it's time to start running AI locally.

In this post, we'll explain the difference between running AI locally versus using cloud-based apps like ChatGPT, cover the pros and cons, and show you how to get started with easy-to-use local AI apps.

Cloud vs. Local AI: How They Work

The most common ways to use AI are with online tools such as ChatGPT, Claude, and Gemini. Here is how these cloud-based tools operate:

• You send a request from your computer over the internet.

• The request goes to the AI company's servers, where it is processed, and a response is generated.

• The response is then sent back to you, and displayed in your browser. 

In this setup, your machine is only sending messages and receiving replies; it is not actually running the AI itself.

Running AI Locally

When you run locally, the actual AI model is stored on your device, and your requests never leave your machine

To achieve this, you need to download an AI model to your computer along with specialized software to run it.

Using local AI with LM Studio

Examples of this software include LM Studio, DiffusionBee, and Comfy. While these apps offer a limited selection of models you can download directly, a much larger repository of models can be found on platforms like Hugging Face.

Advantages of Running AI Locally

There are five key advantages to running AI locally:

1. No Usage Limits: Cloud apps like ChatGPT often limit your daily responses, especially during heavy usage. When running locally, you can send as many prompts as you want because you are using your own computer and resources.

2. It's Free: You don't need a subscription to run AI on your own machine. Apps like LM Studio, DiffusionBee, and Comfy are all free. 

3. Improved Privacy: Cloud-based AI companies like OpenAI, Anthropic, and Google can see every message you send to their tools and may include it in their training data. 

ChatGPT's privacy policy

By contrast, when running locally, your messages don't need to be sent over the internet, and those companies will never see them.

LM Studio's privacy policy

Note: be sure to check the privacy and data usage policies of any AI app before using it, particularly if you intend to use it for business. Above, you can see an excerpt of LM Studio’s data policies (screenshot taken Nov 17th, 2025).

4. Offline Access: Running AI locally is excellent for areas with spotty internet or for use while traveling, as it allows you to access AI tools without needing a strong, stable connection.

5. Choice of Model: You are not limited to models from a single company. You can choose from a wide range of models, including IBM Granite, open source versions of ChatGPT, Qwen, and others. 

Below, you can see a few of the models available to download in LM Studio. 

A selection of models available in LM Studio

Many of these models are fine-tuned to excel at specific tasks, industries, and niche use cases.

Disadvantages of Running AI Locally

While local AI offers many benefits, there are also some drawbacks:

Resource-Intensive: Good AI models are large and require significant resources. Processing the billions of parameters in these models demands a lot of RAM and VRAM. Working with images and video is especially resource-intensive.

Hardware Limitations: If you are using a standard laptop, you may be limited in the models you can choose and operate effectively. You'll typically need to stick with smaller, simpler models with fewer parameters. 

Less computing power also means slower answers from the AI. Average computers likely won’t offer the same lightning-fast speeds as ChatGPT, Claude, or Gemini. 

No Access to Top Models: You won't get access to the absolute latest and best proprietary models from companies like OpenAI, Anthropic, or Google, as they do not give these away for free. You will need to use open source alternatives instead.

We encourage you not to let these drawbacks dissuade you. Since downloading these tools and trying them out is completely free, getting first-hand experience is the best way to decide if local AI fits into your workflows.

Getting Started with Local AI: LM Studio and DiffusionBee

To see how running AI locally works in practice, we recommend starting with two free and easy-to-use apps:

LM Studio: Great for text-based work.

LM Studio

DiffusionBee: Good for generating images.

DiffusionBee

For generating AI video, we’d recommend trying out Comfy. However, getting started with Comfy is quite a bit more complicated, and we won’t be including a guide for it in this post. 

However, we may create a standalone tutorial for Comfy in the future.

LM Studio Setup

1. Download and Open: Download the app at lmstudio.ai and open it up. During initial setup, we recommend choosing "Power User Mode," as the "User" mode hides essential features and can be more confusing.

Choosing "Power user" mode in LM Studio

2. Get a Model: Once installed, you need an AI model to use. LM Studio will recommend a default model to use, such as gpt-oss-20b (an open-source model from OpenAI). 

The default model downloaded by LM Studio - gpt-oss-20b

However, you should be aware that the default model may be too resource-intensive for your computer, especially if you have older hardware. 

If your computer cannot run a model, LM studio will prevent you from using it and show you a warning message

3. Choose a Lighter Model (Optional):  If your computer is unable to run the model, you can download something lighter, like Qwen3 or IBM Granite 4 micro.

LM Studio lets you download models right from the app. Click on the Folder icon to access your installed models and search for new ones.  

Click on the Folder icon to access models in LM Studio

Models are tagged with "B" (e.g., 4B, 20B), which refers to billions of parameters. More parameters usually means better performance but requires higher resources. For lower-powered systems, go with a lower parameter count.

Some available models in LM Studio

Tip: You can visit livebench.ai to review performance benchmarks for any AI model. 

4. Chat: After downloading, go to the Chat tab and load your model using the dropdown. 

Loading a model in LM Studio

You can now send prompts just like with ChatGPT or Claude. 

Sending a prompt in LM Studio

The speed will depend on your hardware, but you can use the AI as much as you’d like with no subscription fees and no usage limits. 

Getting a response from Local AI

DiffusionBee Setup for Images

The DiffusionBee setup is just as simple as LM Studio:

1. Download and Open: Download the app at diffusionbee.com and open it up. 

2. Select a Model: DiffusionBee provides a default model you can use, but its output is very underwhelming. 

Downloading the default model in DiffusionBee

Click on “Models” (stacked cubes icon) in the lefthand menu, and download the model of your choice. 

Downloading additional models in DiffusionBee, like BluePencil XL

For example, you might choose Blue PencilXL for anime or illustration.

3. Generate: Select “Text to image” (document icon). Compose a prompt, adjust the settings as desired, and send your request. 

Sending a prompt in DiffusionBee

The generation may take a little while, but you should get your image within a minute.

An image generated locally by DiffusionBee

With DiffusionBee, you can regenerate as many images as you want with no cost or usage limits.

Ready to Transform Your Workflows?

Cloud-based AI is incredibly convenient, and it has become one of the biggest industries in the world. However, if you're looking for a free alternative with no usage limits and better privacy, you should give local AI tools a try. Download apps like LM Studio and DiffusionBee to get started.

At XRay, our first priority is cleaning up your most critical workflows to produce the results you need. Then, we use tools like automation and AI to make everything faster.

Click here to learn more and book a free call. 

We offer a long-term retainer for total workflow transformation, as well as flexible hourly support for smaller projects.

By clicking “Accept All Cookies”, you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts. View our Privacy Policy for more information.