Take Control of Your Data: Run Your Own Private ChatGPT

AI is everywhere. Companies are racing to add it to everything they can, whether you want it there or not. So, this brings up a lot a questions about how secure our private data really is. Are they using your information to train their AI? Are they selling it? How secure it is? We have already seen case where peoples AI conversations are appearing in Google searches. You might be inclined to just not use or disable AI features, but you may start to fall behind the people who do use these features.

How can you utilize AI and keep your data private?

There is a way to use AI and keep your data private. You can set up your own local ChatGPT like service where your data never leaves your private network.

What can it do for you?

Much of what you can do with the “less” private AI services. Here are few ideas:

Getting Started

While getting the base up and running isn’t overly complex, it does require some technical knowledge. Please continue reading, and if you have any questions, I’m here to help!

Open WebUI

There is a relatively easy way to setup a service very much like ChatGPT. Enter open-webui. It will provide as user interface very similar to ChatGPT with the ability to use local AI models.

Ollama

Ollama is a service that you can run on your local network that allows Open WebUI to connect to AI models you have installed locally. Check out ollama.com/search to see a list of all the available models.

Setup

There are a number of ways to get this setup in your local environment, but the easiest is probably with Docker, as outlined in Open WebUI’s documentation. If you have docker-compose here is a simple docker-compose.yml to get you started:

services:
    webui:
        image: ghcr.io/open-webui/open-webui:main
        ports:
            - 8080:8080/tcp
        environment:
            - OLLAMA_BASE_URL=http://ollama:11434
        volumes:
            - openwebui:/app/backend/data
        depends_on:
            - ollama

    ollama:
        image: ollama/ollama:latest
        environment:
            - OLLAMA_ORIGINS="*"
        ports:
            - 11434:11434/tcp
        volumes:
            - ollama:/root/.ollama
        deploy:
            resources:
                reservations:
                    devices:
                        - driver: nvidia
                          count: all
                          capabilities: [gpu]
volumes:
    openwebui:
    ollama:

NOTE: This assumes you are using an Nvidia GPU. If not, you will need to modify it accordingly.

Once you have it up and running, you will need to install at least one model from the list I linked earlier. You can install the model right in Open WebUI in the admin settings.

Now when you click on “New Chat” the newly added model should already be selected, and you will see page that looks very similar to the ChatGPT.

image

Feeling Overwhelmed?

If this is all a bit overwhelming, I am here to help. Schedule a Free 30-Minute Consultation to see how I can improve your process.