BrainHoster

LLM Quickstart Guide

© 2025 Dev Advisory UK Ltd.

This guide provides step-by-step instructions for running Large Language Models (LLMs) on your BrainHoster VPS using the Ollama framework. Please note that some familiarity with executing commands in Linux and using the terminal is assumed.

Connect to your VPS using the ssh command from Windows, Linux, or macOS. Open a terminal and run:

ssh username@your_vps_ip
Tip: Use SSH keys for secure, password-free access. Generate a key pair with ssh-keygen and copy it to your VPS using `ssh-copy-id`.

Run your first model

Once connected, you can run a model using the following command:
ollama run <model_name>:<model_tag>

Example:
ollama run mistral:7b

Running another model

Execute the command: ollama run <model_name>:<model_tag> for each model you would like to run.

Examples:
ollama run qwen3:8b
ollama run deepseek-coder-v2:16b

To monitor the resources of you VPS, use the following commands:

Check RAM usage:

free -h

Check SSD space:
df -h

Monitor CPU/memory in real time:
top

To list the installed models and remove the unused ones, use the following commands:

List models:

ollama list

Remove a model:
ollama rm <model_name>:<model_tag>

DeepSeek-Coder-V2:16b


DeepSeek-R1:7b


Qwen3:8b


Llama 3:8b


Mistral:7b


For the full list of the models supported by Ollama, visit https://ollama.com/library.
Copyrights and credits for the models referenced in this quickstart guide are attributed to: