Loading...

Self-hosting OpenWebUI with Ollama the easy way

Jonas Scholz - Co-Founder von sliplane.ioJonas Scholz
4 min

OpenWebUI combined with Ollama gives you a powerful ChatGPT-like interface with local language models - completely private and under your control. No more sending sensitive data to OpenAI or paying per token! In this tutorial we're going to setup OpenWebUI with Ollama on sliplane, starting from just 9 euros per month!

Why self-host OpenWebUI?

  • Complete privacy: Your conversations never leave your server
  • No usage limits: Run as many queries as your server can handle
  • Choose your models: Use any Ollama-compatible model (Llama, Mistral, Phi, etc.)
  • Cost-effective: Fixed €9/month vs. pay-per-token pricing
  • Full control: Customize and extend as needed

Setup

  1. Signup at sliplane.io, the signup is free and you can use your Github Account to login.
  2. Create a server: if you just signed up you should already have a trial server that you can use for 48 hours for free. If not, go to > Servers and then click > Create Server
  3. Create a service: go to your projects (create a new one or use the default one) and click > Deploy Service (top right)
    • Select the OpenWebUI preset
  4. Click deploy, wait for the service to be up and then open the domain provided by sliplane. You can find that in the service settings, it should be service-name.sliplane.app
  5. Create your first account: When you first access OpenWebUI, you'll be prompted to create an admin account. This will be your main account with full admin privileges.
  6. Download a model:
    • Go to Admin Panel → Connections → Ollama
    • Click the download icon
    • Enter a model name (e.g., llama3.2 for the latest Llama model)
    • Browse available models at ollama.ai/library

setting

  1. Start chatting with your private AI!

Choosing the right server and model

The server you choose depends on whether you want to run models locally or use external APIs:

Using External APIs (€9/month)

The Base server (€9/month) is perfect if you want to use OpenWebUI with external API providers like OpenAI, Anthropic, or Google. While this doesn't give you the full privacy benefits of local models, it still provides a better interface and multi-user support at a fraction of the cost of team subscriptions.

Running Local Models

For true privacy with local models, you'll need more powerful servers:

Server TypePriceRAMRecommended ModelsUse Case
Base€92GBExternal APIs onlyOpenWebUI interface with cloud providers
Dedicated Base€498GBllama3.2:3b, mistral:7bSmall local models, basic chat
Dedicated Medium€9816GBllama3.2:7b, mixtral:8x7bMedium models, better quality
Dedicated Large€19632GBllama3.3:70b, qwen:32bLarge 20B+ models, best performance

Why self-hosting is the only option for many

For businesses and individuals who handle sensitive data, cloud AI services aren't just expensive - they're often not even an option:

  • Compliance Requirements: GDPR, HIPAA, and other regulations often prohibit sending data to US-based services
  • The Patriot Act Problem: US cloud services are forced to share data with authorities when requested
  • Corporate Secrets: Many companies simply cannot risk their intellectual property on external servers
  • Unlimited Team Access: One OpenWebUI instance can serve your entire team without per-seat pricing

Cost comparison for teams

Service5 Users10 Users20 UsersPrivacyCompliance
ChatGPT Team$150/month$300/month$600/month❌ US servers❌ Patriot Act
Claude Team$125/month$250/month$500/month❌ US servers❌ Patriot Act
Self-hosted OpenWebUI€9-196/month€9-196/month€9-196/month✅ Your servers✅ Full control

The real value isn't just cost savings - it's having an AI solution you can actually use for sensitive work.

FAQ

What models can I use?

Any model available in the Ollama library, including Llama, Mistral, Phi, Gemma, and many more. New models are added regularly.

How much RAM do I need for different models?

  • 3B parameter models: ~4GB RAM (Dedicated Base minimum)
  • 7B parameter models: ~8GB RAM (Dedicated Base minimum)
  • 13B parameter models: ~16GB RAM (Dedicated Medium recommended)
  • 20B+ parameter models: ~32GB RAM (Dedicated Large recommended)

Note: The €9 Base server cannot run local models. Use it with external API providers or upgrade to a Dedicated server for local model hosting.

Can I use multiple models?

Yes! You can download and switch between multiple models. OpenWebUI lets you select which model to use for each conversation.

Is it really private?

Yes! All processing happens on your server. No data is sent to external APIs unless you explicitly configure them.

How do I update models or OpenWebUI?

Simply redeploy the service in Sliplane to get the latest version. For models, you can download new versions through the OpenWebUI interface.

Welcome to the container cloud

Sliplane makes it simple to deploy containers in the cloud and scale up as you grow. Try it now and get started in minutes!