Skip to content

Running LLMs locally

Published: at 09:30 AM
Author: Sebastian Talamoni

Why pay if you can also run it for free locally? This question clearly shows that i have finally finished my Dutch integration process. 😊 Here is will focus on the tools themselves, not in the models.

My 2 finalists

What do they have both in common?

LM Studio

To me LM Studio is the ‘fancy’ one , showing amazing UI/UX experience. If you love to customize via UI’s this is the one for you.

My personal short list:

alt text

Some screenshots to see the tool in action

Example: just asking for python API call (super simple,no challenge): alt text

the reaction was amazingly fast:

alt text

The UI/UX is just amazing: alt text

Ollama

Ollama is, in my humble opinion, the simplest and if you prefer cmd line (versus UI) this might be for you.

Ollama settings (via env vars)

alt text

changing model location: In my case, i have a pretty small c: drive so it’s wise to be sure that models can be stored on a separate drive. Was possible but not so easy as in the UI from LM Studio, but there are ENV variables for each topic.

Exposing OLLAMA to your network This is very useful if you use both windows and WSL. WSL v2 actually will not allow you to communicate to the host via localhost ( 127.0.0.1), so you need to make this change in your ENV variables in windows (if OLLAMA is installed there off course). So the OLLAMA_HOSTS and the OLLAMA_ORIGINS are needed for allowing this access.

Note: you need to restart the win app for this to take effect and you will probably get the windows firewall dialog alter after this change, so you will need to “allow access” if requested.

Using Ollama

If you want to check if something is running you can issue the ollama ps command to see model activity: Ollma PS command

References

Want to know more?

Book a meeting with me via Calendly or contact us!


Previous Post
Using ScrapeGraphAI for Modern Web Scraping
Next Post
About Us